diff --git a/.devcontainer/devcontainer.json b/.devcontainer/devcontainer.json index 025442f2ac..e4964e8909 100644 --- a/.devcontainer/devcontainer.json +++ b/.devcontainer/devcontainer.json @@ -7,98 +7,76 @@ "workspaceFolder": "/workspace", "shutdownAction": "stopCompose", "postCreateCommand": "/docker-init.sh", + "postStartCommand": "/docker-start.sh", "containerEnv": { "EDITOR_VSCODE": "true" }, "features": { - "docker-in-docker": { - "version": "latest" - } }, - - // Set *default* container specific settings.json values on container create. - "settings": { - "terminal.integrated.defaultProfile.linux": "zsh", - "python.pythonPath": "/usr/local/bin/python", - "python.languageServer": "Pylance", - "python.linting.enabled": true, - "python.linting.pylintEnabled": true, - "python.formatting.autopep8Path": "/usr/local/py-utils/bin/autopep8", - "python.formatting.blackPath": "/usr/local/py-utils/bin/black", - "python.formatting.yapfPath": "/usr/local/py-utils/bin/yapf", - "python.linting.banditPath": "/usr/local/py-utils/bin/bandit", - "python.linting.flake8Path": "/usr/local/py-utils/bin/flake8", - "python.linting.mypyPath": "/usr/local/py-utils/bin/mypy", - "python.linting.pycodestylePath": "/usr/local/py-utils/bin/pycodestyle", - "python.linting.pydocstylePath": "/usr/local/py-utils/bin/pydocstyle", - "python.linting.pylintPath": "/usr/local/py-utils/bin/pylint", - "python.testing.pytestArgs": [ - "ietf" - ], - "python.testing.unittestEnabled": true, - "python.testing.pytestEnabled": false, - "python.testing.unittestArgs": [ - "-v", - "-s", - "./ietf", - "-p", - "test*.py" - ], - "sqltools.connections": [ - // Default connection to dev DB container - { - "name": "Local Dev", - "server": "db", - "port": 3306, - "database": "ietf_utf8", - "username": "django", - "password": "RkTkDPFnKpko", - "driver": "MySQL", - "askForPassword": false, - "connectionTimeout": 60 + + "customizations": { + "vscode": { + "extensions": [ + "arcanis.vscode-zipfs", + "batisteo.vscode-django", + "dbaeumer.vscode-eslint", + "eamodio.gitlens", + "editorconfig.editorconfig", + "vue.volar@2.2.10", + "mrmlnc.vscode-duplicate", + "ms-azuretools.vscode-docker", + "ms-playwright.playwright", + "ms-python.python", + "ms-python.vscode-pylance", + "mutantdino.resourcemonitor", + "oderwat.indent-rainbow", + "redhat.vscode-yaml", + "ms-python.pylint", + "charliermarsh.ruff" + ], + "settings": { + "terminal.integrated.defaultProfile.linux": "zsh", + "python.pythonPath": "/usr/local/bin/python", + "python.languageServer": "Default", + "python.formatting.autopep8Path": "/usr/local/py-utils/bin/autopep8", + "python.formatting.blackPath": "/usr/local/py-utils/bin/black", + "python.formatting.yapfPath": "/usr/local/py-utils/bin/yapf", + "python.testing.pytestArgs": [ + "ietf" + ], + "python.testing.unittestEnabled": true, + "python.testing.pytestEnabled": false, + "python.testing.unittestArgs": [ + "-v", + "-s", + "./ietf", + "-p", + "test*.py" + ] } - ] - // "python.envFile": "${workspaceFolder}/.devcontainer/dev.env" + } }, - // Add the IDs of extensions you want installed when the container is created. - "extensions": [ - "arcanis.vscode-zipfs", - "batisteo.vscode-django", - "dbaeumer.vscode-eslint", - "eamodio.gitlens", - "editorconfig.editorconfig", - "vue.volar", - "mrmlnc.vscode-duplicate", - "ms-azuretools.vscode-docker", - "ms-playwright.playwright", - "ms-python.python", - "ms-python.vscode-pylance", - "mtxr.sqltools-driver-mysql", - "mtxr.sqltools", - "mutantdino.resourcemonitor", - "oderwat.indent-rainbow", - "redhat.vscode-yaml", - "spmeesseman.vscode-taskexplorer", - "visualstudioexptteam.vscodeintellicode" - ], - // Use 'forwardPorts' to make a list of ports inside the container available locally. - "forwardPorts": [8000, 3306], + "forwardPorts": [3000, 5432, 8000], "portsAttributes": { "3000": { "label": "Vite", "onAutoForward": "silent" }, - "3306": { - "label": "MariaDB", + "5432": { + "label": "PostgreSQL", "onAutoForward": "silent" }, "8000": { - "label": "Datatracker", + "label": "NGINX", "onAutoForward": "notify" + }, + "8001": { + "label": "Datatracker", + "onAutoForward": "ignore" } }, diff --git a/.devcontainer/docker-compose.extend.yml b/.devcontainer/docker-compose.extend.yml index fa9a412cf2..ce1ce259fd 100644 --- a/.devcontainer/docker-compose.extend.yml +++ b/.devcontainer/docker-compose.extend.yml @@ -1,5 +1,3 @@ -version: '3.8' - services: app: environment: @@ -14,6 +12,10 @@ services: # - datatracker-vscode-ext:/root/.vscode-server/extensions # Runs app on the same network as the database container, allows "forwardPorts" in devcontainer.json function. network_mode: service:db + blobstore: + ports: + - '9000:9000' + - '9001:9001' volumes: datatracker-vscode-ext: diff --git a/.editorconfig b/.editorconfig index d6eafe8d8f..7e5ce6236a 100644 --- a/.editorconfig +++ b/.editorconfig @@ -50,3 +50,9 @@ indent_size = 2 [ietf/**.html] insert_final_newline = false + +# Settings for Kubernetes yaml +# --------------------------------------------------------- +# Use 2-space indents +[k8s/**.yaml] +indent_size = 2 diff --git a/.gitattributes b/.gitattributes index 937c0eb379..62f4aae432 100644 --- a/.gitattributes +++ b/.gitattributes @@ -1,2 +1,280 @@ -/.yarn/releases/** binary -/.yarn/plugins/** binary +# Auto detect text files and perform LF normalization +* text=auto + +# --------------------------------------------------- +# Python Projects +# --------------------------------------------------- + +# Source files +*.pxd text diff=python +*.py text diff=python +*.py3 text diff=python +*.pyw text diff=python +*.pyx text diff=python +*.pyz text diff=python +*.pyi text diff=python + +# Binary files +*.db binary +*.p binary +*.pkl binary +*.pickle binary +*.pyc binary export-ignore +*.pyo binary export-ignore +*.pyd binary + +# Jupyter notebook +*.ipynb text eol=lf + +# --------------------------------------------------- +# Web Projects +# --------------------------------------------------- + +# Source code +*.bash text eol=lf +*.bat text eol=crlf +*.cmd text eol=crlf +*.coffee text +*.css text diff=css +*.htm text diff=html +*.html text diff=html +*.inc text +*.ini text +*.js text +*.mjs text +*.cjs text +*.json text +*.jsx text +*.less text +*.ls text +*.map text -diff +*.od text +*.onlydata text +*.php text diff=php +*.pl text +*.ps1 text eol=crlf +*.py text diff=python +*.rb text diff=ruby +*.sass text +*.scm text +*.scss text diff=css +*.sh text eol=lf +.husky/* text eol=lf +*.sql text +*.styl text +*.tag text +*.ts text +*.tsx text +*.xml text +*.xhtml text diff=html + +# Docker +Dockerfile text + +# Documentation +*.ipynb text eol=lf +*.markdown text diff=markdown +*.md text diff=markdown +*.mdwn text diff=markdown +*.mdown text diff=markdown +*.mkd text diff=markdown +*.mkdn text diff=markdown +*.mdtxt text +*.mdtext text +*.txt text +AUTHORS text +CHANGELOG text +CHANGES text +CONTRIBUTING text +COPYING text +copyright text +*COPYRIGHT* text +INSTALL text +license text +LICENSE text +NEWS text +readme text +*README* text +TODO text + +# Templates +*.dot text +*.ejs text +*.erb text +*.haml text +*.handlebars text +*.hbs text +*.hbt text +*.jade text +*.latte text +*.mustache text +*.njk text +*.phtml text +*.pug text +*.svelte text +*.tmpl text +*.tpl text +*.twig text +*.vue text + +# Configs +*.cnf text +*.conf text +*.config text +.editorconfig text +.env text +.gitattributes text +.gitconfig text +.htaccess text +*.lock text -diff +package.json text eol=lf +package-lock.json text eol=lf -diff +pnpm-lock.yaml text eol=lf -diff +.prettierrc text +yarn.lock text -diff +*.toml text +*.yaml text +*.yml text +browserslist text +Makefile text +makefile text +# Fixes syntax highlighting on GitHub to allow comments +tsconfig.json linguist-language=JSON-with-Comments + +# Heroku +Procfile text + +# Graphics +*.ai binary +*.bmp binary +*.eps binary +*.gif binary +*.gifv binary +*.ico binary +*.jng binary +*.jp2 binary +*.jpg binary +*.jpeg binary +*.jpx binary +*.jxr binary +*.pdf binary +*.png binary +*.psb binary +*.psd binary +*.svg text +*.svgz binary +*.tif binary +*.tiff binary +*.wbmp binary +*.webp binary + +# Audio +*.kar binary +*.m4a binary +*.mid binary +*.midi binary +*.mp3 binary +*.ogg binary +*.ra binary + +# Video +*.3gpp binary +*.3gp binary +*.as binary +*.asf binary +*.asx binary +*.avi binary +*.fla binary +*.flv binary +*.m4v binary +*.mng binary +*.mov binary +*.mp4 binary +*.mpeg binary +*.mpg binary +*.ogv binary +*.swc binary +*.swf binary +*.webm binary + +# Archives +*.7z binary +*.gz binary +*.jar binary +*.rar binary +*.tar binary +*.zip binary + +# Fonts +*.ttf binary +*.eot binary +*.otf binary +*.woff binary +*.woff2 binary + +# Executables +*.exe binary +*.pyc binary +# Prevents massive diffs caused by vendored, minified files +**/.yarn/releases/** binary +**/.yarn/plugins/** binary + +# RC files (like .babelrc or .eslintrc) +*.*rc text + +# Ignore files (like .npmignore or .gitignore) +*.*ignore text + +# Prevents massive diffs from built files +dist/* binary + +# --------------------------------------------------- +# Common +# --------------------------------------------------- + +# Documents +*.bibtex text diff=bibtex +*.doc diff=astextplain +*.DOC diff=astextplain +*.docx diff=astextplain +*.DOCX diff=astextplain +*.dot diff=astextplain +*.DOT diff=astextplain +*.pdf diff=astextplain +*.PDF diff=astextplain +*.rtf diff=astextplain +*.RTF diff=astextplain +*.md text diff=markdown +*.mdx text diff=markdown +*.tex text diff=tex +*.adoc text +*.textile text +*.mustache text +*.csv text eol=crlf +*.tab text +*.tsv text +*.txt text +*.sql text +*.epub diff=astextplain + +# Text files where line endings should be preserved +*.patch -text + +# --------------------------------------------------- +# Vzic specific +# --------------------------------------------------- + +*.pl text diff=perl +*.pm text diff=perl + +# C/C++ +*.c text diff=cpp +*.cc text diff=cpp +*.cxx text diff=cpp +*.cpp text diff=cpp +*.cpi text diff=cpp +*.c++ text diff=cpp +*.hpp text diff=cpp +*.h text diff=cpp +*.h++ text diff=cpp +*.hh text diff=cpp \ No newline at end of file diff --git a/.github/ISSUE_TEMPLATE/config.yml b/.github/ISSUE_TEMPLATE/config.yml index ef8822f5b6..320614b17e 100644 --- a/.github/ISSUE_TEMPLATE/config.yml +++ b/.github/ISSUE_TEMPLATE/config.yml @@ -1,8 +1,8 @@ -blank_issues_enabled: false -contact_links: - - name: Help / Questions - url: https://github.com/ietf-tools/datatracker/discussions/categories/help-questions - about: Need help? Have a question on setting up the project or its usage? - - name: Discuss New Ideas - url: https://github.com/ietf-tools/datatracker/discussions/categories/ideas - about: Submit ideas for new features or improvements to be discussed. +blank_issues_enabled: false +contact_links: + - name: Help and questions + url: https://github.com/ietf-tools/datatracker/discussions/categories/help-questions + about: Need help? Have a question on setting up the project or its usage? + - name: Discuss new ideas + url: https://github.com/ietf-tools/datatracker/discussions/categories/ideas + about: Submit ideas for new features or improvements to be discussed. diff --git a/.github/ISSUE_TEMPLATE/new-feature.yml b/.github/ISSUE_TEMPLATE/new-feature.yml index cf67176892..285081e1c8 100644 --- a/.github/ISSUE_TEMPLATE/new-feature.yml +++ b/.github/ISSUE_TEMPLATE/new-feature.yml @@ -1,16 +1,17 @@ -name: New Feature / Enhancement -description: Propose a new idea to be implemented +name: Suggest new feature or enhancement +description: Propose a new idea to be implemented. labels: ["enhancement"] +type: Feature body: - type: markdown attributes: value: | - Thanks for taking the time to propose a new feature / enhancement idea. + Thanks for taking the time to propose a new feature or enhancement idea. - type: textarea id: description attributes: label: Description - description: Include as much info as possible, including mockups / screenshots if available. + description: Include as much info as possible, including mockups or screenshots if available. placeholder: Description validations: required: true diff --git a/.github/ISSUE_TEMPLATE/report-a-bug.yml b/.github/ISSUE_TEMPLATE/report-a-bug.yml index 0749f78956..47fa1185b4 100644 --- a/.github/ISSUE_TEMPLATE/report-a-bug.yml +++ b/.github/ISSUE_TEMPLATE/report-a-bug.yml @@ -1,6 +1,7 @@ -name: Report a Bug -description: Something isn't right? File a bug report +name: Report a Datatracker bug +description: Something in the datatracker's behavior isn't right? File a bug report. Don't use this to report RFC errata or issues with the content of Internet-Drafts. labels: ["bug"] +type: Bug body: - type: markdown attributes: @@ -10,7 +11,7 @@ body: id: description attributes: label: Describe the issue - description: Include as much info as possible, including the current behavior, expected behavior, screenshots, etc. If this is a display / UX issue, make sure to list the browser(s) you're experiencing the issue on. + description: Include as much info as possible, including the current behavior, expected behavior, screenshots, etc. If this is a display or user interface issue, make sure to list the browser(s) you're experiencing the issue on. placeholder: Description validations: required: true @@ -22,3 +23,11 @@ body: options: - label: I agree to follow the [IETF's Code of Conduct](https://github.com/ietf-tools/.github/blob/main/CODE_OF_CONDUCT.md) required: true + - type: markdown + attributes: + value: | + If you are having trouble logging into the datatracker, please do not open an issue here. Instead, please send email to support@ietf.org providing your name and username. + - type: markdown + attributes: + value: | + **Please do not report issues with the content of Internet-Drafts or RFCs using this repository. Send email to the relevant group instead. Some Internet-Drafts have their own github repositories where issues can be reported. See the datatracker's page for the I-D for links and email addresses. Errata for published RFCs are submitted at https://www.rfc-editor.org/errata.php#reportnew** diff --git a/.github/dependabot.yml b/.github/dependabot.yml new file mode 100644 index 0000000000..17d89f1aab --- /dev/null +++ b/.github/dependabot.yml @@ -0,0 +1,61 @@ +# To get started with Dependabot version updates, you'll need to specify which +# package ecosystems to update and where the package manifests are located. +# Please see the documentation for all configuration options: +# https://docs.github.com/github/administering-a-repository/configuration-options-for-dependency-updates + +version: 2 +updates: + - package-ecosystem: "github-actions" + directory: "/" + schedule: + interval: "weekly" + - package-ecosystem: "docker" + directory: "/docker" + schedule: + interval: "weekly" + - package-ecosystem: "pip" + directory: "/" + schedule: + interval: "weekly" + reviewers: + - "rjsparks" + - package-ecosystem: "npm" + directory: "/" + schedule: + interval: "weekly" + groups: + yarn: + patterns: + - "*" + - package-ecosystem: "npm" + directory: "/playwright" + schedule: + interval: "weekly" + groups: + npm: + patterns: + - "*" + - package-ecosystem: "npm" + directory: "/dev/coverage-action" + schedule: + interval: "weekly" + groups: + npm: + patterns: + - "*" + - package-ecosystem: "npm" + directory: "/dev/deploy-to-container" + schedule: + interval: "weekly" + groups: + npm: + patterns: + - "*" + - package-ecosystem: "npm" + directory: "/dev/diff" + schedule: + interval: "weekly" + groups: + npm: + patterns: + - "*" diff --git a/.github/semantic.yml b/.github/semantic.yml new file mode 100644 index 0000000000..be3439f6b9 --- /dev/null +++ b/.github/semantic.yml @@ -0,0 +1,3 @@ +enabled: true +titleOnly: true +targetUrl: "https://www.conventionalcommits.org/en/v1.0.0/#summary" diff --git a/.github/workflows/build-base-app.yml b/.github/workflows/build-base-app.yml index 6bc771def0..35172aa299 100644 --- a/.github/workflows/build-base-app.yml +++ b/.github/workflows/build-base-app.yml @@ -1,42 +1,67 @@ -name: Build Base App Docker Image - -on: - push: - branches: - - 'main' - paths: - - 'docker/base.Dockerfile' - - workflow_dispatch: - -jobs: - publish: - runs-on: ubuntu-latest - permissions: - contents: read - packages: write - - steps: - - uses: actions/checkout@v2 - - - name: Set up QEMU - uses: docker/setup-qemu-action@v2 - - - name: Set up Docker Buildx - uses: docker/setup-buildx-action@v2 - - - name: Login to GitHub Container Registry - uses: docker/login-action@v2 - with: - registry: ghcr.io - username: ${{ github.actor }} - password: ${{ secrets.GITHUB_TOKEN }} - - - name: Docker Build & Push - uses: docker/build-push-action@v3 - with: - context: . - file: docker/base.Dockerfile - platforms: linux/amd64,linux/arm64 - push: true - tags: ghcr.io/ietf-tools/datatracker-app-base:latest +name: Build Base App Docker Image + +on: + push: + branches: + - 'main' + paths: + - 'docker/base.Dockerfile' + - 'requirements.txt' + + workflow_dispatch: + +jobs: + publish: + runs-on: ubuntu-latest + permissions: + contents: write + packages: write + + steps: + - uses: actions/checkout@v6 + with: + token: ${{ secrets.GH_COMMON_TOKEN }} + + - name: Set Version + run: | + printf -v CURDATE '%(%Y%m%dT%H%M)T' -1 + echo "IMGVERSION=$CURDATE" >> $GITHUB_ENV + + - name: Set up QEMU + uses: docker/setup-qemu-action@v4 + + - name: Set up Docker Buildx + uses: docker/setup-buildx-action@v4 + + - name: Login to GitHub Container Registry + uses: docker/login-action@v4 + with: + registry: ghcr.io + username: ${{ github.actor }} + password: ${{ secrets.GITHUB_TOKEN }} + + - name: Docker Build & Push + uses: docker/build-push-action@v7 + env: + DOCKER_BUILD_SUMMARY: false + with: + context: . + file: docker/base.Dockerfile + platforms: linux/amd64,linux/arm64 + push: true + tags: | + ghcr.io/ietf-tools/datatracker-app-base:${{ env.IMGVERSION }} + ghcr.io/ietf-tools/datatracker-app-base:py312 + ${{ github.ref == 'refs/heads/main' && 'ghcr.io/ietf-tools/datatracker-app-base:latest' || '' }} + + - name: Update version references + run: | + sed -i "1s/.*/FROM ghcr.io\/ietf-tools\/datatracker-app-base:${{ env.IMGVERSION }}/" dev/build/Dockerfile + echo "${{ env.IMGVERSION }}" > dev/build/TARGET_BASE + + - name: Commit CHANGELOG.md + uses: stefanzweifel/git-auto-commit-action@v7 + with: + branch: ${{ github.ref_name }} + commit_message: 'ci: update base image target version to ${{ env.IMGVERSION }}' + file_pattern: dev/build/Dockerfile dev/build/TARGET_BASE diff --git a/.github/workflows/build-celery-worker.yml b/.github/workflows/build-celery-worker.yml deleted file mode 100644 index 87a7d18517..0000000000 --- a/.github/workflows/build-celery-worker.yml +++ /dev/null @@ -1,45 +0,0 @@ -name: Build Celery Worker Docker Image - -on: - push: - branches: - - 'main' - paths: - - 'requirements.txt' - - 'dev/celery/**' - - '.github/workflows/build-celery-worker.yml' - - workflow_dispatch: - -jobs: - publish: - runs-on: ubuntu-latest - permissions: - contents: read - packages: write - - steps: - - uses: actions/checkout@v2 - - - name: Set up QEMU - uses: docker/setup-qemu-action@v2 - - - name: Set up Docker Buildx - uses: docker/setup-buildx-action@v2 - - - name: Login to GitHub Container Registry - uses: docker/login-action@v2 - with: - registry: ghcr.io - username: ${{ github.actor }} - password: ${{ secrets.GITHUB_TOKEN }} - - - name: Docker Build & Push - uses: docker/build-push-action@v3 - with: - context: . - file: dev/celery/Dockerfile - platforms: linux/amd64,linux/arm64 - push: true - tags: ghcr.io/ietf-tools/datatracker-celery:latest - diff --git a/.github/workflows/build-devblobstore.yml b/.github/workflows/build-devblobstore.yml new file mode 100644 index 0000000000..14c4b1a135 --- /dev/null +++ b/.github/workflows/build-devblobstore.yml @@ -0,0 +1,47 @@ +name: Build Dev/Test Blobstore Docker Image + +on: + push: + branches: + - 'main' + paths: + - '.github/workflows/build-devblobstore.yml' + + workflow_dispatch: + +env: + MINIO_VERSION: latest + +jobs: + publish: + runs-on: ubuntu-latest + permissions: + contents: read + packages: write + + steps: + - uses: actions/checkout@v6 + + - name: Set up Docker Buildx + uses: docker/setup-buildx-action@v4 + + - name: Login to GitHub Container Registry + uses: docker/login-action@v4 + with: + registry: ghcr.io + username: ${{ github.actor }} + password: ${{ secrets.GITHUB_TOKEN }} + + - name: Docker Build & Push + uses: docker/build-push-action@v7 + env: + DOCKER_BUILD_SUMMARY: false + with: + context: . + file: docker/devblobstore.Dockerfile + platforms: linux/amd64,linux/arm64 + push: true + build-args: MINIO_VERSION=${{ env.MINIO_VERSION }} + tags: | + ghcr.io/ietf-tools/datatracker-devblobstore:${{ env.MINIO_VERSION }} + ghcr.io/ietf-tools/datatracker-devblobstore:latest diff --git a/.github/workflows/build-mq-broker.yml b/.github/workflows/build-mq-broker.yml index 2ba41bb9c0..b297e34b47 100644 --- a/.github/workflows/build-mq-broker.yml +++ b/.github/workflows/build-mq-broker.yml @@ -8,7 +8,13 @@ on: - 'dev/mq/**' - '.github/workflows/build-mq-broker.yml' - workflow_dispatch: + workflow_dispatch: + inputs: + rabbitmq_version: + description: 'RabbitMQ Version' + default: '3.13-alpine' + required: true + type: string jobs: publish: @@ -18,27 +24,40 @@ jobs: packages: write steps: - - uses: actions/checkout@v2 + - uses: actions/checkout@v6 - name: Set up QEMU - uses: docker/setup-qemu-action@v2 + uses: docker/setup-qemu-action@v4 - name: Set up Docker Buildx - uses: docker/setup-buildx-action@v2 + uses: docker/setup-buildx-action@v4 - name: Login to GitHub Container Registry - uses: docker/login-action@v2 + uses: docker/login-action@v4 with: registry: ghcr.io username: ${{ github.actor }} password: ${{ secrets.GITHUB_TOKEN }} + - name: Set rabbitmq version + id: rabbitmq-version + run: | + if [[ "${{ inputs.rabbitmq_version }}" == "" ]]; then + echo "RABBITMQ_VERSION=3.13-alpine" >> $GITHUB_OUTPUT + else + echo "RABBITMQ_VERSION=${{ inputs.rabbitmq_version }}" >> $GITHUB_OUTPUT + fi + - name: Docker Build & Push - uses: docker/build-push-action@v3 + uses: docker/build-push-action@v7 + env: + DOCKER_BUILD_SUMMARY: false with: context: . file: dev/mq/Dockerfile platforms: linux/amd64,linux/arm64 push: true - tags: ghcr.io/ietf-tools/datatracker-mq:latest - + build-args: RABBITMQ_VERSION=${{ steps.rabbitmq-version.outputs.RABBITMQ_VERSION }} + tags: | + ghcr.io/ietf-tools/datatracker-mq:${{ steps.rabbitmq-version.outputs.RABBITMQ_VERSION }} + ghcr.io/ietf-tools/datatracker-mq:latest diff --git a/.github/workflows/build.yml b/.github/workflows/build.yml index 2c581e70ab..49a0e5b53b 100644 --- a/.github/workflows/build.yml +++ b/.github/workflows/build.yml @@ -1,33 +1,41 @@ name: Build and Release -run-name: ${{ github.event.inputs.publish == 'true' && '[Prod]' || '[Dev]' }} Build ${{ github.run_number }} of branch ${{ github.ref_name }} by @${{ github.actor }} +run-name: ${{ github.ref_name == 'release' && '[Prod]' || '[Dev]' }} Build ${{ github.run_number }} of branch ${{ github.ref_name }} by @${{ github.actor }} on: push: - tags: - - '*' - + branches: [release] + workflow_dispatch: inputs: - summary: - description: 'Release Summary' - required: false - type: string - default: '' - publish: - description: 'Create Production Release' - default: false + deploy: + description: 'Deploy to K8S' + default: 'Skip' required: true - type: boolean - sandbox: - description: 'Deploy to Sandbox' + type: choice + options: + - Skip + - Staging Only + - Staging + Prod + dev: + description: 'Deploy to Dev' default: true required: true type: boolean + devNoDbRefresh: + description: 'Dev Disable Daily DB Refresh' + default: false + required: true + type: boolean skiptests: description: 'Skip Tests' default: false required: true type: boolean + skiparm: + description: 'Skip ARM64 Build' + default: false + required: true + type: boolean ignoreLowerCoverage: description: 'Ignore Lower Coverage' default: false @@ -38,11 +46,10 @@ on: default: false required: true type: boolean - dryrun: - description: 'Dry Run' - default: false - required: true - type: boolean + +concurrency: + group: ${{ github.workflow }}-${{ github.ref }} + cancel-in-progress: true jobs: # ----------------------------------------------------------------- @@ -56,283 +63,147 @@ jobs: pkg_version: ${{ steps.buildvars.outputs.pkg_version }} from_tag: ${{ steps.semver.outputs.nextStrict }} to_tag: ${{ steps.semver.outputs.current }} + base_image_version: ${{ steps.baseimgversion.outputs.base_image_version }} - steps: - - uses: actions/checkout@v3 + steps: + - uses: actions/checkout@v6 with: - fetch-depth: 0 - - - name: Dry Run Notify - if: ${{ github.event.inputs.dryrun == 'true' }} - run: | - echo "::notice::This is a DRY RUN of a production release. No release will be created." + fetch-depth: 1 + fetch-tags: false - - name: Get Next Version - if: ${{ github.event.inputs.publish == 'true' || github.event.inputs.dryrun == 'true' }} + - name: Get Next Version (Prod) + if: ${{ github.ref_name == 'release' }} id: semver uses: ietf-tools/semver-action@v1 with: token: ${{ github.token }} - branch: main + branch: release skipInvalidTags: true - - - name: Set Next Version Env Var - if: ${{ github.event.inputs.publish == 'true' || github.event.inputs.dryrun == 'true' }} + patchList: fix, bugfix, perf, refactor, test, tests, chore + + - name: Get Dev Version + if: ${{ github.ref_name != 'release' }} + id: semverdev + uses: ietf-tools/semver-action@v1 + with: + token: ${{ github.token }} + branch: release + skipInvalidTags: true + noVersionBumpBehavior: 'current' + noNewCommitBehavior: 'current' + + - name: Set Release Flag + if: ${{ github.ref_name == 'release' }} run: | - echo "NEXT_VERSION=$nextStrict" >> $GITHUB_ENV + echo "IS_RELEASE=true" >> $GITHUB_ENV - name: Create Draft Release - uses: ncipollo/release-action@v1.12.0 - if: ${{ github.event.inputs.publish == 'true' && github.event.inputs.dryrun == 'false' }} + uses: ncipollo/release-action@v1.21.0 + if: ${{ github.ref_name == 'release' }} with: prerelease: true draft: false commit: ${{ github.sha }} - tag: ${{ env.NEXT_VERSION }} - name: ${{ env.NEXT_VERSION }} + tag: ${{ steps.semver.outputs.nextStrict }} + name: ${{ steps.semver.outputs.nextStrict }} body: '*pending*' token: ${{ secrets.GITHUB_TOKEN }} - name: Set Build Variables id: buildvars run: | - if [[ $NEXT_VERSION ]]; then - echo "Using AUTO SEMVER mode: $NEXT_VERSION" + if [[ $IS_RELEASE ]]; then + echo "Using AUTO SEMVER mode: ${{ steps.semver.outputs.nextStrict }}" echo "should_deploy=true" >> $GITHUB_OUTPUT - echo "pkg_version=$NEXT_VERSION" >> $GITHUB_OUTPUT - echo "::notice::Release $NEXT_VERSION created using branch $GITHUB_REF_NAME" - elif [[ "$GITHUB_REF" =~ ^refs/tags/* ]]; then - echo "Using TAG mode: $GITHUB_REF_NAME" - echo "should_deploy=true" >> $GITHUB_OUTPUT - echo "pkg_version=$GITHUB_REF_NAME" >> $GITHUB_OUTPUT - echo "::notice::Release $GITHUB_REF_NAME created using tag $GITHUB_REF_NAME" + echo "pkg_version=${{ steps.semver.outputs.nextStrict }}" >> $GITHUB_OUTPUT + echo "::notice::Release ${{ steps.semver.outputs.nextStrict }} created using branch $GITHUB_REF_NAME" else - echo "Using TEST mode: 9.0.0-dev.$GITHUB_RUN_NUMBER" + echo "Using TEST mode: ${{ steps.semverdev.outputs.nextMajorStrict }}.0.0-dev.$GITHUB_RUN_NUMBER" echo "should_deploy=false" >> $GITHUB_OUTPUT - echo "pkg_version=9.0.0-dev.$GITHUB_RUN_NUMBER" >> $GITHUB_OUTPUT - echo "::notice::Non-production build 9.0.0-dev.$GITHUB_RUN_NUMBER created using branch $GITHUB_REF_NAME" + echo "pkg_version=${{ steps.semverdev.outputs.nextMajorStrict }}.0.0-dev.$GITHUB_RUN_NUMBER" >> $GITHUB_OUTPUT + echo "::notice::Non-production build ${{ steps.semverdev.outputs.nextMajorStrict }}.0.0-dev.$GITHUB_RUN_NUMBER created using branch $GITHUB_REF_NAME" fi + - name: Get Base Image Target Version + id: baseimgversion + run: | + echo "base_image_version=$(sed -n '1p' dev/build/TARGET_BASE)" >> $GITHUB_OUTPUT + # ----------------------------------------------------------------- # TESTS # ----------------------------------------------------------------- - tests-python: - name: Run Tests (Python) - if: ${{ github.event.inputs.skiptests == 'false' }} + + tests: + name: Run Tests + uses: ./.github/workflows/tests.yml + if: ${{ github.event.inputs.skiptests == 'false' || github.ref_name == 'release' }} needs: [prepare] - runs-on: ubuntu-latest - container: ghcr.io/ietf-tools/datatracker-app-base:latest - - services: - db: - image: ghcr.io/ietf-tools/datatracker-db:latest - volumes: - - mariadb-data:/var/lib/mysql - env: - MYSQL_ROOT_PASSWORD: ietf - MYSQL_DATABASE: ietf_utf8 - MYSQL_USER: django - MYSQL_PASSWORD: RkTkDPFnKpko - - steps: - - uses: actions/checkout@v3 - - - name: Prepare for tests - run: | - chmod +x ./dev/tests/prepare.sh - sh ./dev/tests/prepare.sh - - - name: Ensure DB is ready - run: | - /usr/local/bin/wait-for db:3306 -- echo "DB ready" - - - name: Run all tests - shell: bash - run: | - echo "Running checks..." - ./ietf/manage.py check - ./ietf/manage.py migrate - echo "Validating migrations..." - if ! ( ietf/manage.py makemigrations --dry-run --check --verbosity 3 ) ; then - echo "Model changes without migrations found." - exit 1 - fi - echo "Running tests..." - if [[ "x${{ github.event.inputs.ignoreLowerCoverage }}" == "xtrue" ]]; then - echo "Lower coverage failures will be ignored." - ./ietf/manage.py test --settings=settings_sqlitetest --ignore-lower-coverage - else - ./ietf/manage.py test --settings=settings_sqlitetest - fi - coverage xml - - - name: Upload Coverage Results to Codecov - uses: codecov/codecov-action@v2.1.0 - with: - files: coverage.xml - - - name: Convert Coverage Results - if: ${{ always() }} - run: | - mv latest-coverage.json coverage.json - - - name: Upload Coverage Results as Build Artifact - uses: actions/upload-artifact@v3 - if: ${{ always() }} - with: - name: coverage - path: coverage.json + secrets: inherit + with: + ignoreLowerCoverage: ${{ github.event.inputs.ignoreLowerCoverage == 'true' }} + skipSelenium: true + targetBaseVersion: ${{ needs.prepare.outputs.base_image_version }} - tests-playwright: - name: Run Tests (Playwright) - if: ${{ github.event.inputs.skiptests == 'false' }} - needs: [prepare] - runs-on: macos-latest - strategy: - fail-fast: false - matrix: - project: [chromium, firefox] - - steps: - - uses: actions/checkout@v3 - - - uses: actions/setup-node@v3 - with: - node-version: '18' - - - name: Run all tests - run: | - echo "Installing dependencies..." - yarn - echo "Installing Playwright..." - cd playwright - mkdir test-results - npm ci - npx playwright install --with-deps ${{ matrix.project }} - echo "Running tests..." - npx playwright test --project=${{ matrix.project }} - - - name: Upload Report - uses: actions/upload-artifact@v3 - if: ${{ always() }} - continue-on-error: true - with: - name: playwright-results-${{ matrix.project }} - path: playwright/test-results/ - if-no-files-found: ignore - - tests-playwright-legacy: - name: Run Tests (Playwright Legacy) - if: ${{ github.event.inputs.skiptests == 'false' }} - needs: [prepare] - runs-on: ubuntu-latest - container: ghcr.io/ietf-tools/datatracker-app-base:latest - strategy: - fail-fast: false - matrix: - project: [chromium, firefox] - - services: - db: - image: ghcr.io/ietf-tools/datatracker-db:latest - volumes: - - mariadb-data:/var/lib/mysql - env: - MYSQL_ROOT_PASSWORD: ietf - MYSQL_DATABASE: ietf_utf8 - MYSQL_USER: django - MYSQL_PASSWORD: RkTkDPFnKpko - - steps: - - uses: actions/checkout@v3 - - - name: Prepare for tests - run: | - chmod +x ./dev/tests/prepare.sh - sh ./dev/tests/prepare.sh - - - name: Ensure DB is ready - run: | - /usr/local/bin/wait-for db:3306 -- echo "DB ready" - - - name: Start Datatracker - run: | - echo "Running checks..." - ./ietf/manage.py check - echo "Starting datatracker..." - ./ietf/manage.py runserver 0.0.0.0:8000 --settings=settings_local & - echo "Waiting for datatracker to be ready..." - /usr/local/bin/wait-for localhost:8000 -- echo "Datatracker ready" - - - name: Run all tests - env: - # Required to get firefox to run as root: - HOME: "" - run: | - echo "Installing dependencies..." - yarn - echo "Installing Playwright..." - cd playwright - mkdir test-results - npm ci - npx playwright install --with-deps ${{ matrix.project }} - echo "Running tests..." - npx playwright test --project=${{ matrix.project }} -c playwright-legacy.config.js - - - name: Upload Report - uses: actions/upload-artifact@v3 - if: ${{ always() }} - continue-on-error: true - with: - name: playwright-legacy-results-${{ matrix.project }} - path: playwright/test-results/ - if-no-files-found: ignore - # ----------------------------------------------------------------- # RELEASE # ----------------------------------------------------------------- release: name: Make Release - if: ${{ always() }} - needs: [tests-python, tests-playwright, tests-playwright-legacy, prepare] - runs-on: ubuntu-latest + if: ${{ !failure() && !cancelled() }} + needs: [tests, prepare] + runs-on: + group: hperf-8c32r + permissions: + contents: write + packages: write env: SHOULD_DEPLOY: ${{needs.prepare.outputs.should_deploy}} PKG_VERSION: ${{needs.prepare.outputs.pkg_version}} FROM_TAG: ${{needs.prepare.outputs.from_tag}} TO_TAG: ${{needs.prepare.outputs.to_tag}} + TARGET_BASE: ${{needs.prepare.outputs.base_image_version}} steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v6 with: - fetch-depth: 0 - - - name: Setup Node.js - uses: actions/setup-node@v3 + fetch-depth: 1 + fetch-tags: false + + - name: Setup Node.js environment + uses: actions/setup-node@v6 with: - node-version: '16' - + node-version: 18.x + - name: Setup Python - uses: actions/setup-python@v4 + uses: actions/setup-python@v6 with: - python-version: '3.x' - + python-version: "3.x" + + - name: Setup AWS CLI + uses: unfor19/install-aws-cli-action@v1 + with: + version: 2.22.35 + - name: Download a Coverage Results - if: ${{ github.event.inputs.skiptests == 'false' }} - uses: actions/download-artifact@v3.0.0 + if: ${{ github.event.inputs.skiptests == 'false' || github.ref_name == 'release' }} + uses: actions/download-artifact@v8.0.1 with: name: coverage - name: Make Release Build env: DEBIAN_FRONTEND: noninteractive + BROWSERSLIST_IGNORE_OLD_DATA: 1 run: | echo "PKG_VERSION: $PKG_VERSION" echo "GITHUB_SHA: $GITHUB_SHA" echo "GITHUB_REF_NAME: $GITHUB_REF_NAME" - echo "Running build script..." - chmod +x ./dev/deploy/build.sh - sh ./dev/deploy/build.sh + echo "Running frontend build script..." + echo "Compiling native node packages..." + yarn rebuild + echo "Packaging static assets..." + yarn build --base=https://static.ietf.org/dt/$PKG_VERSION/ + yarn legacy:build echo "Setting version $PKG_VERSION..." sed -i -r -e "s|^__version__ += '.*'$|__version__ = '$PKG_VERSION'|" ietf/__init__.py sed -i -r -e "s|^__release_hash__ += '.*'$|__release_hash__ = '$GITHUB_SHA'|" ietf/__init__.py @@ -350,40 +221,102 @@ jobs: run: | echo "Build release tarball..." mkdir -p /home/runner/work/release - tar -czf /home/runner/work/release/release.tar.gz -X dev/deploy/exclude-patterns.txt . - + tar -czf /home/runner/work/release/release.tar.gz -X dev/build/exclude-patterns.txt . + + - name: Collect + Push Statics + env: + DEBIAN_FRONTEND: noninteractive + AWS_ACCESS_KEY_ID: ${{ secrets.CF_R2_STATIC_KEY_ID }} + AWS_SECRET_ACCESS_KEY: ${{ secrets.CF_R2_STATIC_KEY_SECRET }} + AWS_DEFAULT_REGION: auto + AWS_ENDPOINT_URL: ${{ secrets.CF_R2_ENDPOINT }} + run: | + echo "Collecting statics..." + echo "Using ghcr.io/ietf-tools/datatracker-app-base:${{ env.TARGET_BASE }}" + docker run --rm --name collectstatics -v $(pwd):/workspace ghcr.io/ietf-tools/datatracker-app-base:${{ env.TARGET_BASE }} sh dev/build/collectstatics.sh + echo "Pushing statics..." + cd static + aws s3 sync . s3://static/dt/$PKG_VERSION --only-show-errors + + - name: Augment dockerignore for docker image build + env: + DEBIAN_FRONTEND: noninteractive + run: | + cat >> .dockerignore <> $GITHUB_ENV + + - name: Build Images + uses: docker/build-push-action@v7 + env: + DOCKER_BUILD_SUMMARY: false + with: + context: . + file: dev/build/Dockerfile + platforms: ${{ github.event.inputs.skiparm == 'true' && 'linux/amd64' || 'linux/amd64,linux/arm64' }} + push: true + tags: | + ghcr.io/ietf-tools/datatracker:${{ env.PKG_VERSION }} + ${{ env.FEATURE_LATEST_TAG && format('ghcr.io/ietf-tools/datatracker:{0}-latest', env.FEATURE_LATEST_TAG) || null }} + - name: Update CHANGELOG id: changelog uses: Requarks/changelog-action@v1 - if: ${{ env.SHOULD_DEPLOY == 'true' && github.event.inputs.dryrun == 'false' }} + if: ${{ env.SHOULD_DEPLOY == 'true' }} with: token: ${{ github.token }} fromTag: ${{ env.FROM_TAG }} toTag: ${{ env.TO_TAG }} writeToFile: false + - name: Download Coverage Results + if: ${{ github.event.inputs.skiptests == 'false' || github.ref_name == 'release' }} + uses: actions/download-artifact@v8.0.1 + with: + name: coverage + - name: Prepare Coverage Action - if: ${{ github.event.inputs.skiptests == 'false' }} + if: ${{ github.event.inputs.skiptests == 'false' || github.ref_name == 'release' }} working-directory: ./dev/coverage-action run: npm install - name: Process Coverage Stats + Chart id: covprocess uses: ./dev/coverage-action/ - if: ${{ github.event.inputs.skiptests == 'false' }} + if: ${{ github.event.inputs.skiptests == 'false' || github.ref_name == 'release' }} with: token: ${{ github.token }} tokenCommon: ${{ secrets.GH_COMMON_TOKEN }} repoCommon: common version: ${{needs.prepare.outputs.pkg_version}} changelog: ${{ steps.changelog.outputs.changes }} - summary: ${{ github.event.inputs.summary }} + summary: '' coverageResultsPath: coverage.json histCoveragePath: historical-coverage.json - name: Create Release - uses: ncipollo/release-action@v1.12.0 - if: ${{ env.SHOULD_DEPLOY == 'true' && github.event.inputs.dryrun == 'false' }} + uses: ncipollo/release-action@v1.21.0 + if: ${{ env.SHOULD_DEPLOY == 'true' }} with: allowUpdates: true makeLatest: true @@ -395,8 +328,8 @@ jobs: token: ${{ secrets.GITHUB_TOKEN }} - name: Update Baseline Coverage - uses: ncipollo/release-action@v1.12.0 - if: ${{ github.event.inputs.updateCoverage == 'true' && github.event.inputs.dryrun == 'false' }} + uses: ncipollo/release-action@v1.21.0 + if: ${{ github.event.inputs.updateCoverage == 'true' || github.ref_name == 'release' }} with: allowUpdates: true tag: baseline @@ -408,57 +341,158 @@ jobs: token: ${{ secrets.GITHUB_TOKEN }} - name: Upload Build Artifacts - uses: actions/upload-artifact@v3 + uses: actions/upload-artifact@v7 with: name: release-${{ env.PKG_VERSION }} path: /home/runner/work/release/release.tar.gz - - name: Notify on Slack - if: ${{ always() }} - uses: slackapi/slack-github-action@v1.23.0 + # ----------------------------------------------------------------- + # NOTIFY + # ----------------------------------------------------------------- + notify: + name: Notify + if: ${{ always() }} + needs: [prepare, tests, release] + runs-on: ubuntu-latest + env: + PKG_VERSION: ${{needs.prepare.outputs.pkg_version}} + + steps: + - name: Notify on Slack (Success) + if: ${{ !contains(join(needs.*.result, ','), 'failure') }} + uses: slackapi/slack-github-action@v3 with: - channel-id: ${{ secrets.SLACK_GH_BUILDS_CHANNEL_ID }} + token: ${{ secrets.SLACK_GH_BOT }} + method: chat.postMessage payload: | - { - "text": "Datatracker - Build by ${{ github.triggering_actor }} completed - <@${{ secrets.SLACK_UID_RJSPARKS }}>" - } - env: - SLACK_BOT_TOKEN: ${{ secrets.SLACK_GH_BOT }} + channel: ${{ secrets.SLACK_GH_BUILDS_CHANNEL_ID }} + text: "Datatracker Build by ${{ github.triggering_actor }}" + attachments: + - color: "28a745" + fields: + - title: "Status" + short: true + value: "Completed" + - name: Notify on Slack (Failure) + if: ${{ contains(join(needs.*.result, ','), 'failure') }} + uses: slackapi/slack-github-action@v3 + with: + token: ${{ secrets.SLACK_GH_BOT }} + method: chat.postMessage + payload: | + channel: ${{ secrets.SLACK_GH_BUILDS_CHANNEL_ID }} + text: "Datatracker Build by ${{ github.triggering_actor }}" + attachments: + - color: "a82929" + fields: + - title: "Status" + short: true + value: "Failed" # ----------------------------------------------------------------- - # SANDBOX + # DEV # ----------------------------------------------------------------- - sandbox: - name: Deploy to Sandbox - if: ${{ always() && github.event.inputs.sandbox == 'true' }} + dev: + name: Deploy to Dev + if: ${{ !failure() && !cancelled() && github.event.inputs.dev == 'true' }} needs: [prepare, release] - runs-on: [self-hosted, dev-server] + runs-on: ubuntu-latest + environment: + name: dev env: PKG_VERSION: ${{needs.prepare.outputs.pkg_version}} steps: - - uses: actions/checkout@v3 - - - name: Download a Release Artifact - uses: actions/download-artifact@v3.0.0 - with: - name: release-${{ env.PKG_VERSION }} - - - name: Deploy to containers - env: - DEBIAN_FRONTEND: noninteractive - run: | - echo "Reset production flags in settings.py..." - sed -i -r -e 's/^DEBUG *= *.*$/DEBUG = True/' -e "s/^SERVER_MODE *= *.*\$/SERVER_MODE = 'development'/" ietf/settings.py - echo "Install Deploy to Container CLI dependencies..." - cd dev/deploy-to-container - npm ci - cd ../.. - echo "Start Deploy..." - node ./dev/deploy-to-container/cli.js --branch ${{ github.ref_name }} --domain dev.ietf.org --appversion ${{ env.PKG_VERSION }} --commit ${{ github.sha }} --ghrunid ${{ github.run_id }} - - - name: Cleanup old docker resources - env: - DEBIAN_FRONTEND: noninteractive - run: | - docker image prune -a -f + - uses: actions/checkout@v6 + with: + ref: main + + - name: Get Deploy Name + env: + DEBIAN_FRONTEND: noninteractive + run: | + echo "Install Get Deploy Name CLI dependencies..." + cd dev/k8s-get-deploy-name + npm ci + echo "Get Deploy Name..." + echo "DEPLOY_NAMESPACE=$(node cli.js --branch ${{ github.ref_name }})" >> "$GITHUB_ENV" + + - name: Deploy to dev + uses: the-actions-org/workflow-dispatch@v4 + with: + workflow: deploy-dev.yml + repo: ietf-tools/infra-k8s + ref: main + token: ${{ secrets.GH_INFRA_K8S_TOKEN }} + inputs: '{ "app":"datatracker", "appVersion":"${{ env.PKG_VERSION }}", "remoteRef":"${{ github.sha }}", "namespace":"${{ env.DEPLOY_NAMESPACE }}", "disableDailyDbRefresh":${{ inputs.devNoDbRefresh }} }' + wait-for-completion: true + wait-for-completion-timeout: 60m + wait-for-completion-interval: 30s + display-workflow-run-url: false + + # ----------------------------------------------------------------- + # STAGING + # ----------------------------------------------------------------- + staging: + name: Deploy to Staging + if: ${{ !failure() && !cancelled() && (github.event.inputs.deploy == 'Staging Only' || github.event.inputs.deploy == 'Staging + Prod' || github.ref_name == 'release') }} + needs: [prepare, release] + runs-on: ubuntu-latest + environment: + name: staging + env: + PKG_VERSION: ${{needs.prepare.outputs.pkg_version}} + + steps: + - name: Refresh Staging DB + uses: the-actions-org/workflow-dispatch@v4 + with: + workflow: deploy-db.yml + repo: ietf-tools/infra-k8s + ref: main + token: ${{ secrets.GH_INFRA_K8S_TOKEN }} + inputs: '{ "environment":"${{ secrets.GHA_K8S_CLUSTER }}", "app":"datatracker", "manifest":"postgres", "forceRecreate":true, "restoreToLastFullSnapshot":true, "waitClusterReady":true }' + wait-for-completion: true + wait-for-completion-timeout: 120m + wait-for-completion-interval: 20s + display-workflow-run-url: false + + - name: Deploy to staging + uses: the-actions-org/workflow-dispatch@v4 + with: + workflow: deploy.yml + repo: ietf-tools/infra-k8s + ref: main + token: ${{ secrets.GH_INFRA_K8S_TOKEN }} + inputs: '{ "environment":"${{ secrets.GHA_K8S_CLUSTER }}", "app":"datatracker", "appVersion":"${{ env.PKG_VERSION }}", "remoteRef":"${{ github.sha }}" }' + wait-for-completion: true + wait-for-completion-timeout: 10m + wait-for-completion-interval: 30s + display-workflow-run-url: false + + # ----------------------------------------------------------------- + # PROD + # ----------------------------------------------------------------- + prod: + name: Deploy to Production + if: ${{ !failure() && !cancelled() && (github.event.inputs.deploy == 'Staging + Prod' || github.ref_name == 'release') }} + needs: [prepare, staging] + runs-on: ubuntu-latest + environment: + name: production + env: + PKG_VERSION: ${{needs.prepare.outputs.pkg_version}} + + steps: + - name: Deploy to production + uses: the-actions-org/workflow-dispatch@v4 + with: + workflow: deploy.yml + repo: ietf-tools/infra-k8s + ref: main + token: ${{ secrets.GH_INFRA_K8S_TOKEN }} + inputs: '{ "environment":"${{ secrets.GHA_K8S_CLUSTER }}", "app":"datatracker", "appVersion":"${{ env.PKG_VERSION }}", "remoteRef":"${{ github.sha }}" }' + wait-for-completion: true + wait-for-completion-timeout: 10m + wait-for-completion-interval: 30s + display-workflow-run-url: false diff --git a/.github/workflows/ci-run-tests.yml b/.github/workflows/ci-run-tests.yml index e2924ceaa3..5349f1ac7a 100644 --- a/.github/workflows/ci-run-tests.yml +++ b/.github/workflows/ci-run-tests.yml @@ -1,10 +1,10 @@ -name: Run All Tests +name: PR - Run All Tests on: pull_request: branches: - 'main' - - 'feat/tzaware' + - 'feat/rfc' paths: - 'client/**' - 'ietf/**' @@ -13,163 +13,34 @@ on: - 'package.json' jobs: - tests-python: - name: Run Tests (Python) + # ----------------------------------------------------------------- + # PREPARE + # ----------------------------------------------------------------- + prepare: + name: Prepare runs-on: ubuntu-latest - container: ghcr.io/ietf-tools/datatracker-app-base:latest - - services: - db: - image: ghcr.io/ietf-tools/datatracker-db:latest - volumes: - - mariadb-data:/var/lib/mysql - env: - MYSQL_ROOT_PASSWORD: ietf - MYSQL_DATABASE: ietf_utf8 - MYSQL_USER: django - MYSQL_PASSWORD: RkTkDPFnKpko - - steps: - - uses: actions/checkout@v3 - - - name: Prepare for tests - run: | - chmod +x ./dev/tests/prepare.sh - sh ./dev/tests/prepare.sh - - - name: Ensure DB is ready - run: | - /usr/local/bin/wait-for db:3306 -- echo "DB ready" - - - name: Run all tests - run: | - echo "Running checks..." - ./ietf/manage.py check - ./ietf/manage.py migrate - echo "Validating migrations..." - if ! ( ietf/manage.py makemigrations --dry-run --check --verbosity 3 ) ; then - echo "Model changes without migrations found." - echo ${MSG} - exit 1 - fi - echo "Running tests..." - ./ietf/manage.py test --settings=settings_sqlitetest - coverage xml - - - name: Upload Coverage Results to Codecov - uses: codecov/codecov-action@v2.1.0 - with: - files: coverage.xml - - - name: Convert Coverage Results - if: ${{ always() }} - run: | - mv latest-coverage.json coverage.json - - - name: Upload Coverage Results as Build Artifact - uses: actions/upload-artifact@v3.0.0 - if: ${{ always() }} - with: - name: coverage - path: coverage.json - - tests-playwright: - name: Run Tests (Playwright) - runs-on: macos-latest - strategy: - fail-fast: false - matrix: - project: [chromium, firefox] - - steps: - - uses: actions/checkout@v3 + outputs: + base_image_version: ${{ steps.baseimgversion.outputs.base_image_version }} - - uses: actions/setup-node@v3 + steps: + - uses: actions/checkout@v6 with: - node-version: '18' + fetch-depth: 1 + fetch-tags: false - - name: Run all tests + - name: Get Base Image Target Version + id: baseimgversion run: | - echo "Installing dependencies..." - yarn - echo "Installing Playwright..." - cd playwright - mkdir test-results - npm ci - npx playwright install --with-deps ${{ matrix.project }} - echo "Running tests..." - npx playwright test --project=${{ matrix.project }} - - - name: Upload Report - uses: actions/upload-artifact@v3.0.0 - if: ${{ always() }} - continue-on-error: true - with: - name: playwright-results-${{ matrix.project }} - path: playwright/test-results/ - if-no-files-found: ignore + echo "base_image_version=$(sed -n '1p' dev/build/TARGET_BASE)" >> $GITHUB_OUTPUT - tests-playwright-legacy: - name: Run Tests (Playwright Legacy) - runs-on: ubuntu-latest - container: ghcr.io/ietf-tools/datatracker-app-base:latest - strategy: - fail-fast: false - matrix: - project: [chromium, firefox] - - services: - db: - image: ghcr.io/ietf-tools/datatracker-db:latest - volumes: - - mariadb-data:/var/lib/mysql - env: - MYSQL_ROOT_PASSWORD: ietf - MYSQL_DATABASE: ietf_utf8 - MYSQL_USER: django - MYSQL_PASSWORD: RkTkDPFnKpko - - steps: - - uses: actions/checkout@v3 - - - name: Prepare for tests - run: | - chmod +x ./dev/tests/prepare.sh - sh ./dev/tests/prepare.sh - - - name: Ensure DB is ready - run: | - /usr/local/bin/wait-for db:3306 -- echo "DB ready" - - - name: Start Datatracker - run: | - echo "Running checks..." - ./ietf/manage.py check - echo "Starting datatracker..." - ./ietf/manage.py runserver 0.0.0.0:8000 --settings=settings_local & - echo "Waiting for datatracker to be ready..." - /usr/local/bin/wait-for localhost:8000 -- echo "Datatracker ready" - - - name: Run all tests - env: - # Required to get firefox to run as root: - HOME: "" - run: | - echo "Installing dependencies..." - yarn - echo "Installing Playwright..." - cd playwright - mkdir test-results - npm ci - npx playwright install --with-deps ${{ matrix.project }} - echo "Running tests..." - npx playwright test --project=${{ matrix.project }} -c playwright-legacy.config.js - - - name: Upload Report - uses: actions/upload-artifact@v3 - if: ${{ always() }} - continue-on-error: true - with: - name: playwright-legacy-results-${{ matrix.project }} - path: playwright/test-results/ - if-no-files-found: ignore + # ----------------------------------------------------------------- + # TESTS + # ----------------------------------------------------------------- + tests: + name: Run Tests + uses: ./.github/workflows/tests.yml + needs: [prepare] + with: + ignoreLowerCoverage: false + skipSelenium: true + targetBaseVersion: ${{ needs.prepare.outputs.base_image_version }} diff --git a/.github/workflows/codeql-analysis.yml b/.github/workflows/codeql-analysis.yml index 2ed7034d6c..bc20779ae6 100644 --- a/.github/workflows/codeql-analysis.yml +++ b/.github/workflows/codeql-analysis.yml @@ -26,12 +26,12 @@ jobs: steps: - name: Checkout repository - uses: actions/checkout@v2 + uses: actions/checkout@v6 - name: Initialize CodeQL - uses: github/codeql-action/init@v1 + uses: github/codeql-action/init@v4 with: languages: ${{ matrix.language }} - name: Perform CodeQL Analysis - uses: github/codeql-action/analyze@v1 + uses: github/codeql-action/analyze@v4 diff --git a/.github/workflows/dependency-review.yml b/.github/workflows/dependency-review.yml index fe461b4243..e255b270ff 100644 --- a/.github/workflows/dependency-review.yml +++ b/.github/workflows/dependency-review.yml @@ -15,6 +15,8 @@ jobs: runs-on: ubuntu-latest steps: - name: 'Checkout Repository' - uses: actions/checkout@v3 + uses: actions/checkout@v6 - name: 'Dependency Review' - uses: actions/dependency-review-action@v2 + uses: actions/dependency-review-action@v4 + with: + vulnerability-check: false diff --git a/.github/workflows/dev-assets-sync-nightly.yml b/.github/workflows/dev-assets-sync-nightly.yml index 9a9f6b2977..cd986f06f3 100644 --- a/.github/workflows/dev-assets-sync-nightly.yml +++ b/.github/workflows/dev-assets-sync-nightly.yml @@ -29,33 +29,21 @@ jobs: contents: read packages: write steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v6 - name: Login to GitHub Container Registry - uses: docker/login-action@v2 + uses: docker/login-action@v4 with: registry: ghcr.io username: ${{ github.actor }} password: ${{ secrets.GITHUB_TOKEN }} - name: Docker Build & Push - uses: docker/build-push-action@v3 + uses: docker/build-push-action@v7 + env: + DOCKER_BUILD_SUMMARY: false with: context: . file: dev/shared-assets-sync/Dockerfile push: true tags: ghcr.io/ietf-tools/datatracker-rsync-assets:latest - - sync: - name: Run assets rsync - if: ${{ always() }} - runs-on: [self-hosted, dev-server] - needs: [build] - steps: - - name: Run rsync - env: - DEBIAN_FRONTEND: noninteractive - run: | - docker pull ghcr.io/ietf-tools/datatracker-rsync-assets:latest - docker run --rm -v dt-assets:/assets ghcr.io/ietf-tools/datatracker-rsync-assets:latest - docker image prune -a -f diff --git a/.github/workflows/dev-db-nightly.yml b/.github/workflows/dev-db-nightly.yml deleted file mode 100644 index 1287805440..0000000000 --- a/.github/workflows/dev-db-nightly.yml +++ /dev/null @@ -1,186 +0,0 @@ -# GITHUB ACTIONS - WORKFLOW - -# Build the database dev docker image with the latest database dump every night -# so that developers don't have to manually build it themselves. - -# DB dump becomes available at around 0700 UTC, so schedule is set to 0800 UTC -# to account for variations. - -name: Nightly Dev DB Image - -# Controls when the workflow will run -on: - # Run every night - schedule: - - cron: '0 8 * * *' - - # Run on db.Dockerfile changes - push: - branches: - - main - paths: - - 'docker/db.Dockerfile' - - # Allows you to run this workflow manually from the Actions tab - workflow_dispatch: - -jobs: - build-mariadb: - name: Build MariaDB Docker Images - runs-on: ubuntu-latest - if: ${{ github.ref == 'refs/heads/main' }} - permissions: - contents: read - packages: write - strategy: - matrix: - include: - - platform: "linux/arm64" - docker: "arm64" - - platform: "linux/amd64" - docker: "x64" - steps: - - uses: actions/checkout@v3 - - - name: Set up QEMU - uses: docker/setup-qemu-action@v2 - - - name: Set up Docker Buildx - uses: docker/setup-buildx-action@v2 - - - name: Login to GitHub Container Registry - uses: docker/login-action@v2 - with: - registry: ghcr.io - username: ${{ github.actor }} - password: ${{ secrets.GITHUB_TOKEN }} - - - name: Docker Build & Push - uses: docker/build-push-action@v3 - with: - context: . - file: docker/db.Dockerfile - platforms: ${{ matrix.platform }} - push: true - tags: ghcr.io/ietf-tools/datatracker-db:latest-${{ matrix.docker }} - - combine-mariadb: - name: Create MariaDB Docker Manifests - runs-on: ubuntu-latest - needs: [build-mariadb] - permissions: - packages: write - steps: - - name: Get Current Date as Tag - id: date - run: echo "date=$(date +'%Y%m%d')" >> $GITHUB_OUTPUT - - - name: Login to GitHub Container Registry - uses: docker/login-action@v2 - with: - registry: ghcr.io - username: ${{ github.actor }} - password: ${{ secrets.GITHUB_TOKEN }} - - - name: Create and Push Manifests - run: | - echo "Creating the manifests..." - docker manifest create ghcr.io/ietf-tools/datatracker-db:nightly-${{ steps.date.outputs.date }} ghcr.io/ietf-tools/datatracker-db:latest-x64 ghcr.io/ietf-tools/datatracker-db:latest-arm64 - docker manifest create ghcr.io/ietf-tools/datatracker-db:latest ghcr.io/ietf-tools/datatracker-db:latest-x64 ghcr.io/ietf-tools/datatracker-db:latest-arm64 - echo "Pushing the manifests..." - docker manifest push -p ghcr.io/ietf-tools/datatracker-db:nightly-${{ steps.date.outputs.date }} - docker manifest push -p ghcr.io/ietf-tools/datatracker-db:latest - - migrate: - name: Migrate MySQL to PostgreSQL DB - runs-on: ubuntu-latest - container: ghcr.io/ietf-tools/datatracker-app-base:latest - needs: [combine-mariadb] - permissions: - contents: read - packages: write - services: - db: - image: ghcr.io/ietf-tools/datatracker-db:latest - volumes: - - mariadb-data:/var/lib/mysql - env: - MYSQL_ROOT_PASSWORD: ietf - MYSQL_DATABASE: ietf_utf8 - MYSQL_USER: django - MYSQL_PASSWORD: RkTkDPFnKpko - pgdb: - image: postgres:14.5 - volumes: - - /pgdata:/var/lib/postgresql/data - env: - POSTGRES_PASSWORD: RkTkDPFnKpko - POSTGRES_USER: django - POSTGRES_DB: ietf - POSTGRES_HOST_AUTH_METHOD: trust - steps: - - uses: actions/checkout@v3 - with: - ref: 'feat/postgres' - - - name: Migrate - uses: nick-fields/retry@v2 - with: - timeout_minutes: 30 - max_attempts: 3 - command: | - chmod +x ./docker/scripts/db-pg-migrate.sh - sh ./docker/scripts/db-pg-migrate.sh - on_retry_command: | - psql -U django -h pgdb -d ietf -v ON_ERROR_STOP=1 -c '\x' -c 'DROP SCHEMA ietf_utf8 CASCADE;' - rm -f cast.load - - - name: Upload DB Dump - uses: actions/upload-artifact@v3 - with: - name: dump - path: ietf.dump - - build: - name: Build PostgreSQL Docker Images - runs-on: ubuntu-latest - needs: [migrate] - permissions: - contents: read - packages: write - - steps: - - uses: actions/checkout@v3 - with: - ref: 'feat/postgres' - - - name: Download DB Dump - uses: actions/download-artifact@v3 - with: - name: dump - - - name: Get Current Date as Tag - id: date - run: echo "date=$(date +'%Y%m%d')" >> $GITHUB_OUTPUT - - - name: Set up QEMU - uses: docker/setup-qemu-action@v2 - - - name: Set up Docker Buildx - uses: docker/setup-buildx-action@v2 - - - name: Login to GitHub Container Registry - uses: docker/login-action@v2 - with: - registry: ghcr.io - username: ${{ github.actor }} - password: ${{ secrets.GITHUB_TOKEN }} - - - name: Docker Build & Push - uses: docker/build-push-action@v3 - with: - context: . - file: docker/db-pg.Dockerfile - platforms: linux/amd64,linux/arm64 - push: true - tags: ghcr.io/ietf-tools/datatracker-db-pg:latest,ghcr.io/ietf-tools/datatracker-db-pg:nightly-${{ steps.date.outputs.date }} diff --git a/.github/workflows/tests-az.yml b/.github/workflows/tests-az.yml new file mode 100644 index 0000000000..833ca89bef --- /dev/null +++ b/.github/workflows/tests-az.yml @@ -0,0 +1,109 @@ +name: Tests (Azure Test) + +on: + workflow_dispatch: + +jobs: + main: + name: Run Tests on Azure temp VM + runs-on: ubuntu-latest + + permissions: + contents: read + + steps: + - name: Launch VM on Azure + id: azlaunch + run: | + echo "Authenticating to Azure..." + az login --service-principal -u ${{ secrets.AZ_TESTS_APP_ID }} -p ${{ secrets.AZ_TESTS_PWD }} --tenant ${{ secrets.AZ_TESTS_TENANT_ID }} + echo "Creating VM..." + vminfo=$(az vm create \ + --resource-group ghaDatatrackerTests \ + --name tmpGhaVM2 \ + --image Ubuntu2204 \ + --admin-username azureuser \ + --generate-ssh-keys \ + --priority Spot \ + --size Standard_D4as_v5 \ + --max-price -1 \ + --os-disk-size-gb 30 \ + --eviction-policy Delete \ + --nic-delete-option Delete \ + --output tsv \ + --query "publicIpAddress") + echo "ipaddr=$vminfo" >> "$GITHUB_OUTPUT" + echo "VM Public IP: $vminfo" + cat ~/.ssh/id_rsa > ${{ github.workspace }}/prvkey.key + ssh-keyscan -t rsa $vminfo >> ~/.ssh/known_hosts + + - name: Remote SSH into VM + uses: appleboy/ssh-action@0ff4204d59e8e51228ff73bce53f80d53301dee2 + env: + GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }} + with: + host: ${{ steps.azlaunch.outputs.ipaddr }} + port: 22 + username: azureuser + command_timeout: 60m + key_path: ${{ github.workspace }}/prvkey.key + envs: GITHUB_TOKEN + script_stop: true + script: | + export DEBIAN_FRONTEND=noninteractive + lsb_release -a + sudo apt-get update + sudo apt-get upgrade -y + + echo "Installing Docker..." + curl -fsSL https://get.docker.com -o get-docker.sh + sudo sh get-docker.sh + + echo "Starting Containers..." + sudo docker network create dtnet + sudo docker run -d --name db --network=dtnet ghcr.io/ietf-tools/datatracker-db:latest & + sudo docker run -d --name app --network=dtnet ghcr.io/ietf-tools/datatracker-app-base:latest sleep infinity & + wait + + echo "Cloning datatracker repo..." + sudo docker exec app git clone --depth=1 https://github.com/ietf-tools/datatracker.git . + echo "Prepare tests..." + sudo docker exec app chmod +x ./dev/tests/prepare.sh + sudo docker exec app sh ./dev/tests/prepare.sh + echo "Running checks..." + sudo docker exec app ietf/manage.py check + sudo docker exec app ietf/manage.py migrate --fake-initial + echo "Running tests..." + sudo docker exec app ietf/manage.py test -v2 --validate-html-harder --settings=settings_test + + - name: Destroy VM + resources + if: always() + shell: pwsh + run: | + echo "Destroying VM..." + az vm delete -g ghaDatatrackerTests -n tmpGhaVM2 --yes --force-deletion true + + $resourceOrderRemovalOrder = [ordered]@{ + "Microsoft.Compute/virtualMachines" = 0 + "Microsoft.Compute/disks" = 1 + "Microsoft.Network/networkInterfaces" = 2 + "Microsoft.Network/publicIpAddresses" = 3 + "Microsoft.Network/networkSecurityGroups" = 4 + "Microsoft.Network/virtualNetworks" = 5 + } + echo "Fetching remaining resources..." + $resources = az resource list --resource-group ghaDatatrackerTests | ConvertFrom-Json + + $orderedResources = $resources + | Sort-Object @{ + Expression = {$resourceOrderRemovalOrder[$_.type]} + Descending = $False + } + + echo "Deleting remaining resources..." + $orderedResources | ForEach-Object { + az resource delete --resource-group ghaDatatrackerTests --ids $_.id --verbose + } + + echo "Logout from Azure..." + az logout diff --git a/.github/workflows/tests.yml b/.github/workflows/tests.yml new file mode 100644 index 0000000000..ad2e35408d --- /dev/null +++ b/.github/workflows/tests.yml @@ -0,0 +1,190 @@ +name: Reusable Tests Workflow + +on: + workflow_call: + inputs: + ignoreLowerCoverage: + description: 'Ignore Lower Coverage' + default: false + required: true + type: boolean + skipSelenium: + description: 'Skip Selenium Tests' + default: false + required: false + type: boolean + targetBaseVersion: + description: 'Target Base Image Version' + default: latest + required: false + type: string + +jobs: + tests-python: + name: Python Tests + runs-on: ubuntu-latest + container: ghcr.io/ietf-tools/datatracker-app-base:${{ inputs.targetBaseVersion }} + + services: + db: + image: ghcr.io/ietf-tools/datatracker-db:latest + blobstore: + image: ghcr.io/ietf-tools/datatracker-devblobstore:latest + + steps: + - uses: actions/checkout@v6 + + - name: Prepare for tests + run: | + chmod +x ./dev/tests/prepare.sh + sh ./dev/tests/prepare.sh + + - name: Ensure DB is ready + run: | + /usr/local/bin/wait-for db:5432 -- echo "DB ready" + + - name: Run all tests + shell: bash + run: | + echo "Running checks..." + ./ietf/manage.py check + ./ietf/manage.py migrate --fake-initial + echo "Validating migrations..." + if ! ( ietf/manage.py makemigrations --dry-run --check --verbosity 3 ) ; then + echo "Model changes without migrations found." + exit 1 + fi + if [[ "x${{ inputs.skipSelenium }}" == "xtrue" ]]; then + echo "Disable selenium tests..." + rm /usr/bin/geckodriver + fi + echo "Running tests..." + if [[ "x${{ inputs.ignoreLowerCoverage }}" == "xtrue" ]]; then + echo "Lower coverage failures will be ignored." + HOME=/root ./ietf/manage.py test -v2 --validate-html-harder --settings=settings_test --ignore-lower-coverage + else + HOME=/root ./ietf/manage.py test -v2 --validate-html-harder --settings=settings_test + fi + coverage xml + + - name: Upload geckodriver.log + uses: actions/upload-artifact@v7 + if: ${{ failure() }} + with: + name: geckodriverlog + path: geckodriver.log + + - name: Upload Coverage Results to Codecov + uses: codecov/codecov-action@v6 + with: + disable_search: true + files: coverage.xml + token: ${{ secrets.CODECOV_TOKEN }} + + - name: Convert Coverage Results + if: ${{ always() }} + run: | + mv latest-coverage.json coverage.json + + - name: Upload Coverage Results as Build Artifact + uses: actions/upload-artifact@v7 + if: ${{ always() }} + with: + name: coverage + path: coverage.json + + tests-playwright: + name: Playwright Tests + runs-on: macos-latest + strategy: + fail-fast: false + matrix: + project: [chromium, firefox] + + steps: + - uses: actions/checkout@v6 + + - uses: actions/setup-node@v6 + with: + node-version: '18' + + - name: Run all tests + run: | + echo "Installing dependencies..." + yarn + echo "Installing Playwright..." + cd playwright + mkdir test-results + npm ci + npx playwright install --with-deps ${{ matrix.project }} + echo "Running tests..." + npx playwright test --project=${{ matrix.project }} + + - name: Upload Report + uses: actions/upload-artifact@v7 + if: ${{ always() }} + continue-on-error: true + with: + name: playwright-results-${{ matrix.project }} + path: playwright/test-results/ + if-no-files-found: ignore + + tests-playwright-legacy: + if: ${{ false }} # disable until we sort out suspected test runner issue + name: Playwright Legacy Tests + runs-on: ubuntu-latest + container: ghcr.io/ietf-tools/datatracker-app-base:${{ inputs.targetBaseVersion }} + strategy: + fail-fast: false + matrix: + project: [chromium, firefox] + + services: + db: + image: ghcr.io/ietf-tools/datatracker-db:latest + + steps: + - uses: actions/checkout@v6 + + - name: Prepare for tests + run: | + chmod +x ./dev/tests/prepare.sh + sh ./dev/tests/prepare.sh + + - name: Ensure DB is ready + run: | + /usr/local/bin/wait-for db:5432 -- echo "DB ready" + + - name: Start Datatracker + run: | + echo "Running checks..." + ./ietf/manage.py check + ./ietf/manage.py migrate --fake-initial + echo "Starting datatracker..." + ./ietf/manage.py runserver 0.0.0.0:8000 --settings=settings_local & + echo "Waiting for datatracker to be ready..." + /usr/local/bin/wait-for localhost:8000 -- echo "Datatracker ready" + + - name: Run all tests + env: + # Required to get firefox to run as root: + HOME: "" + run: | + echo "Installing dependencies..." + yarn + echo "Installing Playwright..." + cd playwright + mkdir test-results + npm ci + npx playwright install --with-deps ${{ matrix.project }} + echo "Running tests..." + npx playwright test --project=${{ matrix.project }} -c playwright-legacy.config.js + + - name: Upload Report + uses: actions/upload-artifact@v7 + if: ${{ always() }} + continue-on-error: true + with: + name: playwright-legacy-results-${{ matrix.project }} + path: playwright/test-results/ + if-no-files-found: ignore diff --git a/.gitignore b/.gitignore index 80e5f0228b..ccc7a46b08 100644 --- a/.gitignore +++ b/.gitignore @@ -1,6 +1,7 @@ .DS_store datatracker.sublime-project datatracker.sublime-workspace +/.claude /.coverage /.factoryboy_random_state /.mypy_cache @@ -17,14 +18,17 @@ datatracker.sublime-workspace /docker/docker-compose.extend-custom.yml /env /ghostdriver.log +/geckodriver.log /htmlcov /ietf/static/dist-neue /latest-coverage.json /media /node_modules /release-coverage.json +/static /tmp-* /.testresult +*.swp *.pyc __pycache__ .yarn/* diff --git a/.pnp.cjs b/.pnp.cjs index 46892da0d3..5fcce34d2f 100644 --- a/.pnp.cjs +++ b/.pnp.cjs @@ -33,73 +33,84 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { [null, {\ "packageLocation": "./",\ "packageDependencies": [\ - ["@faker-js/faker", "npm:7.6.0"],\ - ["@fullcalendar/bootstrap5", "npm:5.11.3"],\ - ["@fullcalendar/core", "npm:5.11.3"],\ - ["@fullcalendar/daygrid", "npm:5.11.3"],\ - ["@fullcalendar/interaction", "npm:5.11.3"],\ - ["@fullcalendar/list", "npm:5.11.3"],\ - ["@fullcalendar/luxon2", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.11.3"],\ - ["@fullcalendar/timegrid", "npm:5.11.3"],\ - ["@fullcalendar/vue3", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.11.3"],\ - ["@parcel/optimizer-data-url", "npm:2.8.2"],\ - ["@parcel/transformer-inline-string", "npm:2.8.2"],\ - ["@parcel/transformer-sass", "npm:2.8.2"],\ - ["@popperjs/core", "npm:2.11.6"],\ - ["@rollup/pluginutils", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.0.2"],\ + ["@fullcalendar/bootstrap5", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/core", "npm:6.1.11"],\ + ["@fullcalendar/daygrid", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/icalendar", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/interaction", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/list", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/luxon3", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/timegrid", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/vue3", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@parcel/optimizer-data-url", "npm:2.12.0"],\ + ["@parcel/transformer-inline-string", "npm:2.12.0"],\ + ["@parcel/transformer-sass", "npm:2.12.0"],\ + ["@popperjs/core", "npm:2.11.8"],\ + ["@rollup/pluginutils", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.1.0"],\ ["@twuni/emojify", "npm:1.0.2"],\ - ["@vitejs/plugin-vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.2.0"],\ - ["bootstrap", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.2.3"],\ - ["bootstrap-icons", "npm:1.10.3"],\ - ["browser-fs-access", "npm:0.31.1"],\ + ["@vitejs/plugin-vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:4.6.2"],\ + ["@vue/language-plugin-pug", "npm:2.0.7"],\ + ["bootstrap", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.3.3"],\ + ["bootstrap-icons", "npm:1.11.3"],\ + ["browser-fs-access", "npm:0.35.0"],\ ["browserlist", "npm:1.0.1"],\ - ["c8", "npm:7.12.0"],\ - ["caniuse-lite", "npm:1.0.30001442"],\ - ["d3", "npm:7.8.0"],\ - ["eslint", "npm:8.31.0"],\ - ["eslint-config-standard", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:17.0.0"],\ - ["eslint-plugin-cypress", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.12.1"],\ - ["eslint-plugin-import", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.26.0"],\ - ["eslint-plugin-n", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:15.6.1"],\ + ["c8", "npm:9.1.0"],\ + ["caniuse-lite", "npm:1.0.30001603"],\ + ["d3", "npm:7.9.0"],\ + ["eslint", "npm:8.57.0"],\ + ["eslint-config-standard", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:17.1.0"],\ + ["eslint-plugin-cypress", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.15.1"],\ + ["eslint-plugin-import", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.29.1"],\ + ["eslint-plugin-n", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:16.6.2"],\ ["eslint-plugin-node", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:11.1.0"],\ ["eslint-plugin-promise", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.1"],\ - ["eslint-plugin-vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:9.8.0"],\ + ["eslint-plugin-vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:9.24.0"],\ ["file-saver", "npm:2.0.5"],\ - ["highcharts", "npm:10.3.2"],\ - ["html-validate", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:7.12.2"],\ - ["jquery", "npm:3.6.3"],\ - ["jquery-migrate", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.0"],\ - ["jquery-ui-dist", "npm:1.13.2"],\ - ["js-cookie", "npm:3.0.1"],\ + ["highcharts", "npm:11.4.0"],\ + ["html-validate", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:8.18.1"],\ + ["ical.js", "npm:1.5.0"],\ + ["jquery", "npm:3.7.1"],\ + ["jquery-migrate", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.1"],\ + ["js-cookie", "npm:3.0.5"],\ ["list.js", "npm:2.3.1"],\ ["lodash", "npm:4.17.21"],\ ["lodash-es", "npm:4.17.21"],\ - ["luxon", "npm:3.2.1"],\ - ["moment", "npm:2.29.4"],\ - ["moment-timezone", "npm:0.5.40"],\ + ["luxon", "npm:3.4.4"],\ + ["moment", "npm:2.30.1"],\ + ["moment-timezone", "npm:0.5.45"],\ ["ms", "npm:2.1.3"],\ ["murmurhash-js", "npm:1.0.0"],\ - ["naive-ui", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.34.3"],\ - ["parcel", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.8.2"],\ - ["pinia", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.0.28"],\ + ["naive-ui", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.38.1"],\ + ["parcel", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.12.0"],\ + ["pinia", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.1.7"],\ ["pinia-plugin-persist", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:1.0.0"],\ ["pug", "npm:3.0.2"],\ - ["sass", "npm:1.57.1"],\ + ["sass", "npm:1.72.0"],\ ["seedrandom", "npm:3.0.5"],\ ["select2", "npm:4.1.0-rc.0"],\ ["select2-bootstrap-5-theme", "npm:1.3.0"],\ ["send", "npm:0.18.0"],\ - ["shepherd.js", "npm:10.0.1"],\ - ["slugify", "npm:1.6.5"],\ - ["sortablejs", "npm:1.15.0"],\ - ["vite", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.2.5"],\ - ["vue", "npm:3.2.45"],\ - ["vue-router", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:4.1.6"],\ + ["shepherd.js", "npm:11.2.0"],\ + ["slugify", "npm:1.6.6"],\ + ["sortablejs", "npm:1.15.2"],\ + ["vanillajs-datepicker", "npm:1.3.4"],\ + ["vite", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:4.5.3"],\ + ["vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.21"],\ + ["vue-router", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:4.3.0"],\ ["zxcvbn", "npm:4.4.2"]\ ],\ "linkType": "SOFT"\ }]\ ]],\ + ["@aashutoshrathi/word-wrap", [\ + ["npm:1.2.6", {\ + "packageLocation": "./.yarn/cache/@aashutoshrathi-word-wrap-npm-1.2.6-5b1d95e487-ada901b9e7.zip/node_modules/@aashutoshrathi/word-wrap/",\ + "packageDependencies": [\ + ["@aashutoshrathi/word-wrap", "npm:1.2.6"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["@babel/code-frame", [\ ["npm:7.16.7", {\ "packageLocation": "./.yarn/cache/@babel-code-frame-npm-7.16.7-093eb9e124-db2f7faa31.zip/node_modules/@babel/code-frame/",\ @@ -139,6 +150,24 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@babel/types", "npm:7.18.4"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:7.23.9", {\ + "packageLocation": "./.yarn/cache/@babel-parser-npm-7.23.9-720a0b56cb-e7cd4960ac.zip/node_modules/@babel/parser/",\ + "packageDependencies": [\ + ["@babel/parser", "npm:7.23.9"],\ + ["@babel/types", "npm:7.18.4"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@babel/runtime", [\ + ["npm:7.23.2", {\ + "packageLocation": "./.yarn/cache/@babel-runtime-npm-7.23.2-d013d6cf7e-6c4df4839e.zip/node_modules/@babel/runtime/",\ + "packageDependencies": [\ + ["@babel/runtime", "npm:7.23.2"],\ + ["regenerator-runtime", "npm:0.14.0"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@babel/types", [\ @@ -162,19 +191,19 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@css-render/plugin-bem", [\ - ["npm:0.15.10", {\ - "packageLocation": "./.yarn/cache/@css-render-plugin-bem-npm-0.15.10-41ccecaa2f-cbab72a7b5.zip/node_modules/@css-render/plugin-bem/",\ + ["npm:0.15.12", {\ + "packageLocation": "./.yarn/cache/@css-render-plugin-bem-npm-0.15.12-bf8b43dc1f-9fa7ddd62b.zip/node_modules/@css-render/plugin-bem/",\ "packageDependencies": [\ - ["@css-render/plugin-bem", "npm:0.15.10"]\ + ["@css-render/plugin-bem", "npm:0.15.12"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:0.15.10", {\ - "packageLocation": "./.yarn/__virtual__/@css-render-plugin-bem-virtual-1bbf78173b/0/cache/@css-render-plugin-bem-npm-0.15.10-41ccecaa2f-cbab72a7b5.zip/node_modules/@css-render/plugin-bem/",\ + ["virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:0.15.12", {\ + "packageLocation": "./.yarn/__virtual__/@css-render-plugin-bem-virtual-105b1b654b/0/cache/@css-render-plugin-bem-npm-0.15.12-bf8b43dc1f-9fa7ddd62b.zip/node_modules/@css-render/plugin-bem/",\ "packageDependencies": [\ - ["@css-render/plugin-bem", "virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:0.15.10"],\ + ["@css-render/plugin-bem", "virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:0.15.12"],\ ["@types/css-render", null],\ - ["css-render", "npm:0.15.10"]\ + ["css-render", "npm:0.15.12"]\ ],\ "packagePeers": [\ "@types/css-render",\ @@ -191,12 +220,32 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "SOFT"\ }],\ - ["virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:0.15.10", {\ - "packageLocation": "./.yarn/__virtual__/@css-render-vue3-ssr-virtual-75d553c878/0/cache/@css-render-vue3-ssr-npm-0.15.10-b8526cc313-7977e0c440.zip/node_modules/@css-render/vue3-ssr/",\ + ["npm:0.15.12", {\ + "packageLocation": "./.yarn/cache/@css-render-vue3-ssr-npm-0.15.12-a130f4db3a-a5505ae161.zip/node_modules/@css-render/vue3-ssr/",\ "packageDependencies": [\ - ["@css-render/vue3-ssr", "virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:0.15.10"],\ + ["@css-render/vue3-ssr", "npm:0.15.12"]\ + ],\ + "linkType": "SOFT"\ + }],\ + ["virtual:2366be83ef58a728ebb5a5e9ed4600f4465f98b2a844262fcfbe89415361d5d5f9e964ec3b9a72d6a5004f37c1024d017c65e67473dd9cc39cd61f51768c65e6#npm:0.15.10", {\ + "packageLocation": "./.yarn/__virtual__/@css-render-vue3-ssr-virtual-8cb63dbe2e/0/cache/@css-render-vue3-ssr-npm-0.15.10-b8526cc313-7977e0c440.zip/node_modules/@css-render/vue3-ssr/",\ + "packageDependencies": [\ + ["@css-render/vue3-ssr", "virtual:2366be83ef58a728ebb5a5e9ed4600f4465f98b2a844262fcfbe89415361d5d5f9e964ec3b9a72d6a5004f37c1024d017c65e67473dd9cc39cd61f51768c65e6#npm:0.15.10"],\ ["@types/vue", null],\ - ["vue", "npm:3.2.45"]\ + ["vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.21"]\ + ],\ + "packagePeers": [\ + "@types/vue",\ + "vue"\ + ],\ + "linkType": "HARD"\ + }],\ + ["virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:0.15.12", {\ + "packageLocation": "./.yarn/__virtual__/@css-render-vue3-ssr-virtual-18db73fb22/0/cache/@css-render-vue3-ssr-npm-0.15.12-a130f4db3a-a5505ae161.zip/node_modules/@css-render/vue3-ssr/",\ + "packageDependencies": [\ + ["@css-render/vue3-ssr", "virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:0.15.12"],\ + ["@types/vue", null],\ + ["vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.21"]\ ],\ "packagePeers": [\ "@types/vue",\ @@ -215,31 +264,250 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@esbuild/android-arm", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/@esbuild-android-arm-npm-0.15.11-21c0733fef/node_modules/@esbuild/android-arm/",\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-android-arm-npm-0.18.20-a30c33e9ed/node_modules/@esbuild/android-arm/",\ + "packageDependencies": [\ + ["@esbuild/android-arm", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/android-arm64", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-android-arm64-npm-0.18.20-fd4fb45ae7/node_modules/@esbuild/android-arm64/",\ + "packageDependencies": [\ + ["@esbuild/android-arm64", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/android-x64", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-android-x64-npm-0.18.20-22b610e3f4/node_modules/@esbuild/android-x64/",\ + "packageDependencies": [\ + ["@esbuild/android-x64", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/darwin-arm64", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-darwin-arm64-npm-0.18.20-00b3504077/node_modules/@esbuild/darwin-arm64/",\ "packageDependencies": [\ - ["@esbuild/android-arm", "npm:0.15.11"]\ + ["@esbuild/darwin-arm64", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/darwin-x64", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-darwin-x64-npm-0.18.20-767fe27d1b/node_modules/@esbuild/darwin-x64/",\ + "packageDependencies": [\ + ["@esbuild/darwin-x64", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/freebsd-arm64", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-freebsd-arm64-npm-0.18.20-797e8c8987/node_modules/@esbuild/freebsd-arm64/",\ + "packageDependencies": [\ + ["@esbuild/freebsd-arm64", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/freebsd-x64", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-freebsd-x64-npm-0.18.20-f7563ff3dd/node_modules/@esbuild/freebsd-x64/",\ + "packageDependencies": [\ + ["@esbuild/freebsd-x64", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/linux-arm", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-linux-arm-npm-0.18.20-06b400b09e/node_modules/@esbuild/linux-arm/",\ + "packageDependencies": [\ + ["@esbuild/linux-arm", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/linux-arm64", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-linux-arm64-npm-0.18.20-7b48b328fe/node_modules/@esbuild/linux-arm64/",\ + "packageDependencies": [\ + ["@esbuild/linux-arm64", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/linux-ia32", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-linux-ia32-npm-0.18.20-2f5a035f9e/node_modules/@esbuild/linux-ia32/",\ + "packageDependencies": [\ + ["@esbuild/linux-ia32", "npm:0.18.20"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@esbuild/linux-loong64", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/@esbuild-linux-loong64-npm-0.15.11-b380a37e51/node_modules/@esbuild/linux-loong64/",\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-linux-loong64-npm-0.18.20-e91b93ee90/node_modules/@esbuild/linux-loong64/",\ + "packageDependencies": [\ + ["@esbuild/linux-loong64", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/linux-mips64el", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-linux-mips64el-npm-0.18.20-a5e9429f2a/node_modules/@esbuild/linux-mips64el/",\ + "packageDependencies": [\ + ["@esbuild/linux-mips64el", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/linux-ppc64", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-linux-ppc64-npm-0.18.20-218f398134/node_modules/@esbuild/linux-ppc64/",\ + "packageDependencies": [\ + ["@esbuild/linux-ppc64", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/linux-riscv64", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-linux-riscv64-npm-0.18.20-6a2972f753/node_modules/@esbuild/linux-riscv64/",\ + "packageDependencies": [\ + ["@esbuild/linux-riscv64", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/linux-s390x", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-linux-s390x-npm-0.18.20-ff9d596142/node_modules/@esbuild/linux-s390x/",\ + "packageDependencies": [\ + ["@esbuild/linux-s390x", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/linux-x64", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-linux-x64-npm-0.18.20-de8e99b449/node_modules/@esbuild/linux-x64/",\ + "packageDependencies": [\ + ["@esbuild/linux-x64", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/netbsd-x64", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-netbsd-x64-npm-0.18.20-39b460150f/node_modules/@esbuild/netbsd-x64/",\ + "packageDependencies": [\ + ["@esbuild/netbsd-x64", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/openbsd-x64", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-openbsd-x64-npm-0.18.20-90ab921595/node_modules/@esbuild/openbsd-x64/",\ + "packageDependencies": [\ + ["@esbuild/openbsd-x64", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/sunos-x64", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-sunos-x64-npm-0.18.20-d18b46b343/node_modules/@esbuild/sunos-x64/",\ + "packageDependencies": [\ + ["@esbuild/sunos-x64", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/win32-arm64", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-win32-arm64-npm-0.18.20-a58fe6c6a3/node_modules/@esbuild/win32-arm64/",\ + "packageDependencies": [\ + ["@esbuild/win32-arm64", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/win32-ia32", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-win32-ia32-npm-0.18.20-d7ee926338/node_modules/@esbuild/win32-ia32/",\ + "packageDependencies": [\ + ["@esbuild/win32-ia32", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@esbuild/win32-x64", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/@esbuild-win32-x64-npm-0.18.20-37a9ab2bda/node_modules/@esbuild/win32-x64/",\ "packageDependencies": [\ - ["@esbuild/linux-loong64", "npm:0.15.11"]\ + ["@esbuild/win32-x64", "npm:0.18.20"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@eslint-community/eslint-utils", [\ + ["npm:4.4.0", {\ + "packageLocation": "./.yarn/cache/@eslint-community-eslint-utils-npm-4.4.0-d1791bd5a3-cdfe3ae42b.zip/node_modules/@eslint-community/eslint-utils/",\ + "packageDependencies": [\ + ["@eslint-community/eslint-utils", "npm:4.4.0"]\ + ],\ + "linkType": "SOFT"\ + }],\ + ["virtual:4286e12a3a0f74af013bc8f16c6d8fdde823cfbf6389660266b171e551f576c805b0a7a8eb2a7087a5cee7dfe6ebb6e1ea3808d93daf915edc95656907a381bb#npm:4.4.0", {\ + "packageLocation": "./.yarn/__virtual__/@eslint-community-eslint-utils-virtual-1c7da85a1a/0/cache/@eslint-community-eslint-utils-npm-4.4.0-d1791bd5a3-cdfe3ae42b.zip/node_modules/@eslint-community/eslint-utils/",\ + "packageDependencies": [\ + ["@eslint-community/eslint-utils", "virtual:4286e12a3a0f74af013bc8f16c6d8fdde823cfbf6389660266b171e551f576c805b0a7a8eb2a7087a5cee7dfe6ebb6e1ea3808d93daf915edc95656907a381bb#npm:4.4.0"],\ + ["@types/eslint", null],\ + ["eslint", "npm:8.57.0"],\ + ["eslint-visitor-keys", "npm:3.3.0"]\ + ],\ + "packagePeers": [\ + "@types/eslint",\ + "eslint"\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@eslint-community/regexpp", [\ + ["npm:4.10.0", {\ + "packageLocation": "./.yarn/cache/@eslint-community-regexpp-npm-4.10.0-6bfb984c81-2a6e345429.zip/node_modules/@eslint-community/regexpp/",\ + "packageDependencies": [\ + ["@eslint-community/regexpp", "npm:4.10.0"]\ + ],\ + "linkType": "HARD"\ + }],\ + ["npm:4.8.0", {\ + "packageLocation": "./.yarn/cache/@eslint-community-regexpp-npm-4.8.0-92ece47e3d-601e6d033d.zip/node_modules/@eslint-community/regexpp/",\ + "packageDependencies": [\ + ["@eslint-community/regexpp", "npm:4.8.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@eslint/eslintrc", [\ - ["npm:1.4.1", {\ - "packageLocation": "./.yarn/cache/@eslint-eslintrc-npm-1.4.1-007f670de2-cd3e5a8683.zip/node_modules/@eslint/eslintrc/",\ + ["npm:2.1.4", {\ + "packageLocation": "./.yarn/cache/@eslint-eslintrc-npm-2.1.4-1ff4b5f908-10957c7592.zip/node_modules/@eslint/eslintrc/",\ "packageDependencies": [\ - ["@eslint/eslintrc", "npm:1.4.1"],\ + ["@eslint/eslintrc", "npm:2.1.4"],\ ["ajv", "npm:6.12.6"],\ ["debug", "virtual:b86a9fb34323a98c6519528ed55faa0d9b44ca8879307c0b29aa384bde47ff59a7d0c9051b31246f14521dfb71ba3c5d6d0b35c29fffc17bf875aa6ad977d9e8#npm:4.3.4"],\ - ["espree", "npm:9.4.0"],\ + ["espree", "npm:9.6.1"],\ ["globals", "npm:13.19.0"],\ ["ignore", "npm:5.2.0"],\ ["import-fresh", "npm:3.3.0"],\ @@ -250,99 +518,189 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ - ["@faker-js/faker", [\ - ["npm:7.6.0", {\ - "packageLocation": "./.yarn/cache/@faker-js-faker-npm-7.6.0-fa135883e9-942af62217.zip/node_modules/@faker-js/faker/",\ + ["@eslint/js", [\ + ["npm:8.57.0", {\ + "packageLocation": "./.yarn/cache/@eslint-js-npm-8.57.0-00ead3710a-315dc65b0e.zip/node_modules/@eslint/js/",\ "packageDependencies": [\ - ["@faker-js/faker", "npm:7.6.0"]\ + ["@eslint/js", "npm:8.57.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ - ["@fullcalendar/bootstrap5", [\ - ["npm:5.11.3", {\ - "packageLocation": "./.yarn/cache/@fullcalendar-bootstrap5-npm-5.11.3-3e86f39d7d-a63a500d72.zip/node_modules/@fullcalendar/bootstrap5/",\ + ["@floating-ui/core", [\ + ["npm:1.4.1", {\ + "packageLocation": "./.yarn/cache/@floating-ui-core-npm-1.4.1-fe89c45d92-be4ab864fe.zip/node_modules/@floating-ui/core/",\ "packageDependencies": [\ - ["@fullcalendar/bootstrap5", "npm:5.11.3"],\ - ["@fullcalendar/common", "npm:5.11.3"],\ - ["tslib", "npm:2.4.0"]\ + ["@floating-ui/core", "npm:1.4.1"],\ + ["@floating-ui/utils", "npm:0.1.2"]\ ],\ "linkType": "HARD"\ }]\ ]],\ - ["@fullcalendar/common", [\ - ["npm:5.11.3", {\ - "packageLocation": "./.yarn/cache/@fullcalendar-common-npm-5.11.3-6268994b76-be4b365dca.zip/node_modules/@fullcalendar/common/",\ + ["@floating-ui/dom", [\ + ["npm:1.5.2", {\ + "packageLocation": "./.yarn/cache/@floating-ui-dom-npm-1.5.2-f1b8ca0c30-3c71eed50b.zip/node_modules/@floating-ui/dom/",\ "packageDependencies": [\ - ["@fullcalendar/common", "npm:5.11.3"],\ - ["tslib", "npm:2.4.0"]\ + ["@floating-ui/dom", "npm:1.5.2"],\ + ["@floating-ui/core", "npm:1.4.1"],\ + ["@floating-ui/utils", "npm:0.1.2"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@floating-ui/utils", [\ + ["npm:0.1.2", {\ + "packageLocation": "./.yarn/cache/@floating-ui-utils-npm-0.1.2-22eefe56f0-3e29fd3c69.zip/node_modules/@floating-ui/utils/",\ + "packageDependencies": [\ + ["@floating-ui/utils", "npm:0.1.2"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@fullcalendar/bootstrap5", [\ + ["npm:6.1.11", {\ + "packageLocation": "./.yarn/cache/@fullcalendar-bootstrap5-npm-6.1.11-6e0fbf281a-a0c3b94346.zip/node_modules/@fullcalendar/bootstrap5/",\ + "packageDependencies": [\ + ["@fullcalendar/bootstrap5", "npm:6.1.11"]\ + ],\ + "linkType": "SOFT"\ + }],\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11", {\ + "packageLocation": "./.yarn/__virtual__/@fullcalendar-bootstrap5-virtual-50942c1c6f/0/cache/@fullcalendar-bootstrap5-npm-6.1.11-6e0fbf281a-a0c3b94346.zip/node_modules/@fullcalendar/bootstrap5/",\ + "packageDependencies": [\ + ["@fullcalendar/bootstrap5", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/core", "npm:6.1.11"],\ + ["@types/fullcalendar__core", null]\ + ],\ + "packagePeers": [\ + "@fullcalendar/core",\ + "@types/fullcalendar__core"\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@fullcalendar/core", [\ - ["npm:5.11.3", {\ - "packageLocation": "./.yarn/cache/@fullcalendar-core-npm-5.11.3-ed98a1ea9f-2774d0fa18.zip/node_modules/@fullcalendar/core/",\ + ["npm:6.1.11", {\ + "packageLocation": "./.yarn/cache/@fullcalendar-core-npm-6.1.11-ae049c8ace-0078a6f96b.zip/node_modules/@fullcalendar/core/",\ "packageDependencies": [\ - ["@fullcalendar/core", "npm:5.11.3"],\ - ["@fullcalendar/common", "npm:5.11.3"],\ - ["preact", "npm:10.7.2"],\ - ["tslib", "npm:2.4.0"]\ + ["@fullcalendar/core", "npm:6.1.11"],\ + ["preact", "npm:10.12.1"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@fullcalendar/daygrid", [\ - ["npm:5.11.3", {\ - "packageLocation": "./.yarn/cache/@fullcalendar-daygrid-npm-5.11.3-b387dff934-426b53c5bb.zip/node_modules/@fullcalendar/daygrid/",\ + ["npm:6.1.11", {\ + "packageLocation": "./.yarn/cache/@fullcalendar-daygrid-npm-6.1.11-2187ca1b8f-6eb5606de5.zip/node_modules/@fullcalendar/daygrid/",\ "packageDependencies": [\ - ["@fullcalendar/daygrid", "npm:5.11.3"],\ - ["@fullcalendar/common", "npm:5.11.3"],\ - ["tslib", "npm:2.4.0"]\ + ["@fullcalendar/daygrid", "npm:6.1.11"]\ + ],\ + "linkType": "SOFT"\ + }],\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11", {\ + "packageLocation": "./.yarn/__virtual__/@fullcalendar-daygrid-virtual-b91d1ffe14/0/cache/@fullcalendar-daygrid-npm-6.1.11-2187ca1b8f-6eb5606de5.zip/node_modules/@fullcalendar/daygrid/",\ + "packageDependencies": [\ + ["@fullcalendar/daygrid", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/core", "npm:6.1.11"],\ + ["@types/fullcalendar__core", null]\ + ],\ + "packagePeers": [\ + "@fullcalendar/core",\ + "@types/fullcalendar__core"\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@fullcalendar/icalendar", [\ + ["npm:6.1.11", {\ + "packageLocation": "./.yarn/cache/@fullcalendar-icalendar-npm-6.1.11-73807e790d-4e6eff15a8.zip/node_modules/@fullcalendar/icalendar/",\ + "packageDependencies": [\ + ["@fullcalendar/icalendar", "npm:6.1.11"]\ + ],\ + "linkType": "SOFT"\ + }],\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11", {\ + "packageLocation": "./.yarn/__virtual__/@fullcalendar-icalendar-virtual-636a290006/0/cache/@fullcalendar-icalendar-npm-6.1.11-73807e790d-4e6eff15a8.zip/node_modules/@fullcalendar/icalendar/",\ + "packageDependencies": [\ + ["@fullcalendar/icalendar", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/core", "npm:6.1.11"],\ + ["@types/fullcalendar__core", null],\ + ["@types/ical.js", null],\ + ["ical.js", "npm:1.5.0"]\ + ],\ + "packagePeers": [\ + "@fullcalendar/core",\ + "@types/fullcalendar__core",\ + "@types/ical.js",\ + "ical.js"\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@fullcalendar/interaction", [\ - ["npm:5.11.3", {\ - "packageLocation": "./.yarn/cache/@fullcalendar-interaction-npm-5.11.3-15335cb10a-e8a1b49f2f.zip/node_modules/@fullcalendar/interaction/",\ + ["npm:6.1.11", {\ + "packageLocation": "./.yarn/cache/@fullcalendar-interaction-npm-6.1.11-39630596c7-c67d4cfa0b.zip/node_modules/@fullcalendar/interaction/",\ "packageDependencies": [\ - ["@fullcalendar/interaction", "npm:5.11.3"],\ - ["@fullcalendar/common", "npm:5.11.3"],\ - ["tslib", "npm:2.4.0"]\ + ["@fullcalendar/interaction", "npm:6.1.11"]\ + ],\ + "linkType": "SOFT"\ + }],\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11", {\ + "packageLocation": "./.yarn/__virtual__/@fullcalendar-interaction-virtual-3ebf8b0646/0/cache/@fullcalendar-interaction-npm-6.1.11-39630596c7-c67d4cfa0b.zip/node_modules/@fullcalendar/interaction/",\ + "packageDependencies": [\ + ["@fullcalendar/interaction", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/core", "npm:6.1.11"],\ + ["@types/fullcalendar__core", null]\ + ],\ + "packagePeers": [\ + "@fullcalendar/core",\ + "@types/fullcalendar__core"\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@fullcalendar/list", [\ - ["npm:5.11.3", {\ - "packageLocation": "./.yarn/cache/@fullcalendar-list-npm-5.11.3-6174d0e1da-976da49b12.zip/node_modules/@fullcalendar/list/",\ + ["npm:6.1.11", {\ + "packageLocation": "./.yarn/cache/@fullcalendar-list-npm-6.1.11-8f1846f302-84a8cd6e63.zip/node_modules/@fullcalendar/list/",\ "packageDependencies": [\ - ["@fullcalendar/list", "npm:5.11.3"],\ - ["@fullcalendar/common", "npm:5.11.3"],\ - ["tslib", "npm:2.4.0"]\ + ["@fullcalendar/list", "npm:6.1.11"]\ + ],\ + "linkType": "SOFT"\ + }],\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11", {\ + "packageLocation": "./.yarn/__virtual__/@fullcalendar-list-virtual-1c555df506/0/cache/@fullcalendar-list-npm-6.1.11-8f1846f302-84a8cd6e63.zip/node_modules/@fullcalendar/list/",\ + "packageDependencies": [\ + ["@fullcalendar/list", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/core", "npm:6.1.11"],\ + ["@types/fullcalendar__core", null]\ + ],\ + "packagePeers": [\ + "@fullcalendar/core",\ + "@types/fullcalendar__core"\ ],\ "linkType": "HARD"\ }]\ ]],\ - ["@fullcalendar/luxon2", [\ - ["npm:5.11.3", {\ - "packageLocation": "./.yarn/cache/@fullcalendar-luxon2-npm-5.11.3-ccde7500a8-7533018590.zip/node_modules/@fullcalendar/luxon2/",\ + ["@fullcalendar/luxon3", [\ + ["npm:6.1.11", {\ + "packageLocation": "./.yarn/cache/@fullcalendar-luxon3-npm-6.1.11-3e90656a71-8e7f45aab2.zip/node_modules/@fullcalendar/luxon3/",\ "packageDependencies": [\ - ["@fullcalendar/luxon2", "npm:5.11.3"]\ + ["@fullcalendar/luxon3", "npm:6.1.11"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.11.3", {\ - "packageLocation": "./.yarn/__virtual__/@fullcalendar-luxon2-virtual-efa9fdf749/0/cache/@fullcalendar-luxon2-npm-5.11.3-ccde7500a8-7533018590.zip/node_modules/@fullcalendar/luxon2/",\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11", {\ + "packageLocation": "./.yarn/__virtual__/@fullcalendar-luxon3-virtual-38643019c2/0/cache/@fullcalendar-luxon3-npm-6.1.11-3e90656a71-8e7f45aab2.zip/node_modules/@fullcalendar/luxon3/",\ "packageDependencies": [\ - ["@fullcalendar/luxon2", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.11.3"],\ - ["@fullcalendar/common", "npm:5.11.3"],\ + ["@fullcalendar/luxon3", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/core", "npm:6.1.11"],\ + ["@types/fullcalendar__core", null],\ ["@types/luxon", null],\ - ["luxon", "npm:3.2.1"],\ - ["tslib", "npm:2.4.0"]\ + ["luxon", "npm:3.4.4"]\ ],\ "packagePeers": [\ + "@fullcalendar/core",\ + "@types/fullcalendar__core",\ "@types/luxon",\ "luxon"\ ],\ @@ -350,35 +708,48 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@fullcalendar/timegrid", [\ - ["npm:5.11.3", {\ - "packageLocation": "./.yarn/cache/@fullcalendar-timegrid-npm-5.11.3-4075b09051-ce675eca7d.zip/node_modules/@fullcalendar/timegrid/",\ + ["npm:6.1.11", {\ + "packageLocation": "./.yarn/cache/@fullcalendar-timegrid-npm-6.1.11-1d43455bfd-4a11e6dd90.zip/node_modules/@fullcalendar/timegrid/",\ "packageDependencies": [\ - ["@fullcalendar/timegrid", "npm:5.11.3"],\ - ["@fullcalendar/common", "npm:5.11.3"],\ - ["@fullcalendar/daygrid", "npm:5.11.3"],\ - ["tslib", "npm:2.4.0"]\ + ["@fullcalendar/timegrid", "npm:6.1.11"]\ + ],\ + "linkType": "SOFT"\ + }],\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11", {\ + "packageLocation": "./.yarn/__virtual__/@fullcalendar-timegrid-virtual-5e951d78a6/0/cache/@fullcalendar-timegrid-npm-6.1.11-1d43455bfd-4a11e6dd90.zip/node_modules/@fullcalendar/timegrid/",\ + "packageDependencies": [\ + ["@fullcalendar/timegrid", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/core", "npm:6.1.11"],\ + ["@fullcalendar/daygrid", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@types/fullcalendar__core", null]\ + ],\ + "packagePeers": [\ + "@fullcalendar/core",\ + "@types/fullcalendar__core"\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@fullcalendar/vue3", [\ - ["npm:5.11.3", {\ - "packageLocation": "./.yarn/cache/@fullcalendar-vue3-npm-5.11.3-047b9981f6-13a648a0c5.zip/node_modules/@fullcalendar/vue3/",\ + ["npm:6.1.11", {\ + "packageLocation": "./.yarn/cache/@fullcalendar-vue3-npm-6.1.11-f6b8b48da4-5891a596e9.zip/node_modules/@fullcalendar/vue3/",\ "packageDependencies": [\ - ["@fullcalendar/vue3", "npm:5.11.3"]\ + ["@fullcalendar/vue3", "npm:6.1.11"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.11.3", {\ - "packageLocation": "./.yarn/__virtual__/@fullcalendar-vue3-virtual-623c0672ef/0/cache/@fullcalendar-vue3-npm-5.11.3-047b9981f6-13a648a0c5.zip/node_modules/@fullcalendar/vue3/",\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11", {\ + "packageLocation": "./.yarn/__virtual__/@fullcalendar-vue3-virtual-cb317bc2d1/0/cache/@fullcalendar-vue3-npm-6.1.11-f6b8b48da4-5891a596e9.zip/node_modules/@fullcalendar/vue3/",\ "packageDependencies": [\ - ["@fullcalendar/vue3", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.11.3"],\ - ["@fullcalendar/core", "npm:5.11.3"],\ + ["@fullcalendar/vue3", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/core", "npm:6.1.11"],\ + ["@types/fullcalendar__core", null],\ ["@types/vue", null],\ - ["tslib", "npm:2.4.0"],\ - ["vue", "npm:3.2.45"]\ + ["vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.21"]\ ],\ "packagePeers": [\ + "@fullcalendar/core",\ + "@types/fullcalendar__core",\ "@types/vue",\ "vue"\ ],\ @@ -395,21 +766,21 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@html-validate/stylish", [\ - ["npm:3.0.0", {\ - "packageLocation": "./.yarn/cache/@html-validate-stylish-npm-3.0.0-6d9dccafda-818efd25ac.zip/node_modules/@html-validate/stylish/",\ + ["npm:4.1.0", {\ + "packageLocation": "./.yarn/cache/@html-validate-stylish-npm-4.1.0-aba0cf2d6c-4af90db4f9.zip/node_modules/@html-validate/stylish/",\ "packageDependencies": [\ - ["@html-validate/stylish", "npm:3.0.0"],\ + ["@html-validate/stylish", "npm:4.1.0"],\ ["kleur", "npm:4.1.4"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@humanwhocodes/config-array", [\ - ["npm:0.11.8", {\ - "packageLocation": "./.yarn/cache/@humanwhocodes-config-array-npm-0.11.8-7955bfecc2-0fd6b3c54f.zip/node_modules/@humanwhocodes/config-array/",\ + ["npm:0.11.14", {\ + "packageLocation": "./.yarn/cache/@humanwhocodes-config-array-npm-0.11.14-94a02fcc87-861ccce9ea.zip/node_modules/@humanwhocodes/config-array/",\ "packageDependencies": [\ - ["@humanwhocodes/config-array", "npm:0.11.8"],\ - ["@humanwhocodes/object-schema", "npm:1.2.1"],\ + ["@humanwhocodes/config-array", "npm:0.11.14"],\ + ["@humanwhocodes/object-schema", "npm:2.0.2"],\ ["debug", "virtual:b86a9fb34323a98c6519528ed55faa0d9b44ca8879307c0b29aa384bde47ff59a7d0c9051b31246f14521dfb71ba3c5d6d0b35c29fffc17bf875aa6ad977d9e8#npm:4.3.4"],\ ["minimatch", "npm:3.1.2"]\ ],\ @@ -426,10 +797,34 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@humanwhocodes/object-schema", [\ - ["npm:1.2.1", {\ - "packageLocation": "./.yarn/cache/@humanwhocodes-object-schema-npm-1.2.1-eb622b5d0e-a824a1ec31.zip/node_modules/@humanwhocodes/object-schema/",\ + ["npm:2.0.2", {\ + "packageLocation": "./.yarn/cache/@humanwhocodes-object-schema-npm-2.0.2-77b42018f9-2fc1150336.zip/node_modules/@humanwhocodes/object-schema/",\ "packageDependencies": [\ - ["@humanwhocodes/object-schema", "npm:1.2.1"]\ + ["@humanwhocodes/object-schema", "npm:2.0.2"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@isaacs/cliui", [\ + ["npm:8.0.2", {\ + "packageLocation": "./.yarn/cache/@isaacs-cliui-npm-8.0.2-f4364666d5-4a473b9b32.zip/node_modules/@isaacs/cliui/",\ + "packageDependencies": [\ + ["@isaacs/cliui", "npm:8.0.2"],\ + ["string-width", "npm:5.1.2"],\ + ["string-width-cjs", [\ + "string-width",\ + "npm:4.2.3"\ + ]],\ + ["strip-ansi", "npm:7.0.1"],\ + ["strip-ansi-cjs", [\ + "strip-ansi",\ + "npm:6.0.1"\ + ]],\ + ["wrap-ansi", "npm:8.1.0"],\ + ["wrap-ansi-cjs", [\ + "wrap-ansi",\ + "npm:7.0.0"\ + ]]\ ],\ "linkType": "HARD"\ }]\ @@ -459,6 +854,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@jridgewell/sourcemap-codec", "npm:1.4.14"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:1.4.15", {\ + "packageLocation": "./.yarn/cache/@jridgewell-sourcemap-codec-npm-1.4.15-a055fb62cf-b881c7e503.zip/node_modules/@jridgewell/sourcemap-codec/",\ + "packageDependencies": [\ + ["@jridgewell/sourcemap-codec", "npm:1.4.15"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@jridgewell/trace-mapping", [\ @@ -507,6 +909,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@lmdb/lmdb-darwin-arm64", "npm:2.5.2"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:2.8.5", {\ + "packageLocation": "./.yarn/unplugged/@lmdb-lmdb-darwin-arm64-npm-2.8.5-a9ab00615c/node_modules/@lmdb/lmdb-darwin-arm64/",\ + "packageDependencies": [\ + ["@lmdb/lmdb-darwin-arm64", "npm:2.8.5"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@lmdb/lmdb-darwin-x64", [\ @@ -516,6 +925,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@lmdb/lmdb-darwin-x64", "npm:2.5.2"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:2.8.5", {\ + "packageLocation": "./.yarn/unplugged/@lmdb-lmdb-darwin-x64-npm-2.8.5-080b8c9329/node_modules/@lmdb/lmdb-darwin-x64/",\ + "packageDependencies": [\ + ["@lmdb/lmdb-darwin-x64", "npm:2.8.5"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@lmdb/lmdb-linux-arm", [\ @@ -525,6 +941,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@lmdb/lmdb-linux-arm", "npm:2.5.2"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:2.8.5", {\ + "packageLocation": "./.yarn/unplugged/@lmdb-lmdb-linux-arm-npm-2.8.5-081004004c/node_modules/@lmdb/lmdb-linux-arm/",\ + "packageDependencies": [\ + ["@lmdb/lmdb-linux-arm", "npm:2.8.5"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@lmdb/lmdb-linux-arm64", [\ @@ -534,6 +957,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@lmdb/lmdb-linux-arm64", "npm:2.5.2"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:2.8.5", {\ + "packageLocation": "./.yarn/unplugged/@lmdb-lmdb-linux-arm64-npm-2.8.5-9dfda9f24f/node_modules/@lmdb/lmdb-linux-arm64/",\ + "packageDependencies": [\ + ["@lmdb/lmdb-linux-arm64", "npm:2.8.5"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@lmdb/lmdb-linux-x64", [\ @@ -543,6 +973,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@lmdb/lmdb-linux-x64", "npm:2.5.2"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:2.8.5", {\ + "packageLocation": "./.yarn/unplugged/@lmdb-lmdb-linux-x64-npm-2.8.5-0f668ba9a7/node_modules/@lmdb/lmdb-linux-x64/",\ + "packageDependencies": [\ + ["@lmdb/lmdb-linux-x64", "npm:2.8.5"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@lmdb/lmdb-win32-x64", [\ @@ -552,6 +989,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@lmdb/lmdb-win32-x64", "npm:2.5.2"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:2.8.5", {\ + "packageLocation": "./.yarn/unplugged/@lmdb-lmdb-win32-x64-npm-2.8.5-3702de4edb/node_modules/@lmdb/lmdb-win32-x64/",\ + "packageDependencies": [\ + ["@lmdb/lmdb-win32-x64", "npm:2.8.5"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@mischnic/json-sourcemap", [\ @@ -573,6 +1017,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@msgpackr-extract/msgpackr-extract-darwin-arm64", "npm:2.0.2"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:3.0.2", {\ + "packageLocation": "./.yarn/unplugged/@msgpackr-extract-msgpackr-extract-darwin-arm64-npm-3.0.2-18ac236cc4/node_modules/@msgpackr-extract/msgpackr-extract-darwin-arm64/",\ + "packageDependencies": [\ + ["@msgpackr-extract/msgpackr-extract-darwin-arm64", "npm:3.0.2"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@msgpackr-extract/msgpackr-extract-darwin-x64", [\ @@ -582,6 +1033,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@msgpackr-extract/msgpackr-extract-darwin-x64", "npm:2.0.2"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:3.0.2", {\ + "packageLocation": "./.yarn/unplugged/@msgpackr-extract-msgpackr-extract-darwin-x64-npm-3.0.2-39dd07082a/node_modules/@msgpackr-extract/msgpackr-extract-darwin-x64/",\ + "packageDependencies": [\ + ["@msgpackr-extract/msgpackr-extract-darwin-x64", "npm:3.0.2"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@msgpackr-extract/msgpackr-extract-linux-arm", [\ @@ -591,6 +1049,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@msgpackr-extract/msgpackr-extract-linux-arm", "npm:2.0.2"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:3.0.2", {\ + "packageLocation": "./.yarn/unplugged/@msgpackr-extract-msgpackr-extract-linux-arm-npm-3.0.2-808a652e0b/node_modules/@msgpackr-extract/msgpackr-extract-linux-arm/",\ + "packageDependencies": [\ + ["@msgpackr-extract/msgpackr-extract-linux-arm", "npm:3.0.2"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@msgpackr-extract/msgpackr-extract-linux-arm64", [\ @@ -600,6 +1065,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@msgpackr-extract/msgpackr-extract-linux-arm64", "npm:2.0.2"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:3.0.2", {\ + "packageLocation": "./.yarn/unplugged/@msgpackr-extract-msgpackr-extract-linux-arm64-npm-3.0.2-cfbf50d4c6/node_modules/@msgpackr-extract/msgpackr-extract-linux-arm64/",\ + "packageDependencies": [\ + ["@msgpackr-extract/msgpackr-extract-linux-arm64", "npm:3.0.2"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@msgpackr-extract/msgpackr-extract-linux-x64", [\ @@ -609,6 +1081,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@msgpackr-extract/msgpackr-extract-linux-x64", "npm:2.0.2"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:3.0.2", {\ + "packageLocation": "./.yarn/unplugged/@msgpackr-extract-msgpackr-extract-linux-x64-npm-3.0.2-262fca760d/node_modules/@msgpackr-extract/msgpackr-extract-linux-x64/",\ + "packageDependencies": [\ + ["@msgpackr-extract/msgpackr-extract-linux-x64", "npm:3.0.2"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@msgpackr-extract/msgpackr-extract-win32-x64", [\ @@ -618,6 +1097,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@msgpackr-extract/msgpackr-extract-win32-x64", "npm:2.0.2"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:3.0.2", {\ + "packageLocation": "./.yarn/unplugged/@msgpackr-extract-msgpackr-extract-win32-x64-npm-3.0.2-c627beab89/node_modules/@msgpackr-extract/msgpackr-extract-win32-x64/",\ + "packageDependencies": [\ + ["@msgpackr-extract/msgpackr-extract-win32-x64", "npm:3.0.2"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@nodelib/fs.scandir", [\ @@ -674,61 +1160,45 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@parcel/bundler-default", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-bundler-default-npm-2.8.2-497641ec3a-8330a76248.zip/node_modules/@parcel/bundler-default/",\ - "packageDependencies": [\ - ["@parcel/bundler-default", "npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/graph", "npm:2.8.2"],\ - ["@parcel/hash", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-bundler-default-npm-2.12.0-9ba57d919c-f211a76f55.zip/node_modules/@parcel/bundler-default/",\ + "packageDependencies": [\ + ["@parcel/bundler-default", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/graph", "npm:3.2.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/rust", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["nullthrows", "npm:1.1.1"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/cache", [\ - ["npm:2.6.2", {\ - "packageLocation": "./.yarn/cache/@parcel-cache-npm-2.6.2-7c97030a45-e7b540fe10.zip/node_modules/@parcel/cache/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-cache-npm-2.12.0-3389909f2c-a45e799809.zip/node_modules/@parcel/cache/",\ "packageDependencies": [\ - ["@parcel/cache", "npm:2.6.2"]\ + ["@parcel/cache", "npm:2.12.0"]\ ],\ "linkType": "SOFT"\ }],\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-cache-npm-2.8.2-4957caf228-7d1c951e3f.zip/node_modules/@parcel/cache/",\ + ["npm:2.6.2", {\ + "packageLocation": "./.yarn/cache/@parcel-cache-npm-2.6.2-7c97030a45-e7b540fe10.zip/node_modules/@parcel/cache/",\ "packageDependencies": [\ - ["@parcel/cache", "npm:2.8.2"]\ + ["@parcel/cache", "npm:2.6.2"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:4a1952be093bd8b413333d6452f7e82b253ed9257f221c7341bb9d244e18522a0f05e6b82b53db04b831a339d732c092e9b930055d598f0acbd03dc62701ab32#npm:2.8.2", {\ - "packageLocation": "./.yarn/__virtual__/@parcel-cache-virtual-3c8bad9556/0/cache/@parcel-cache-npm-2.8.2-4957caf228-7d1c951e3f.zip/node_modules/@parcel/cache/",\ - "packageDependencies": [\ - ["@parcel/cache", "virtual:4a1952be093bd8b413333d6452f7e82b253ed9257f221c7341bb9d244e18522a0f05e6b82b53db04b831a339d732c092e9b930055d598f0acbd03dc62701ab32#npm:2.8.2"],\ - ["@parcel/core", "npm:2.6.2"],\ - ["@parcel/fs", "virtual:4a1952be093bd8b413333d6452f7e82b253ed9257f221c7341bb9d244e18522a0f05e6b82b53db04b831a339d732c092e9b930055d598f0acbd03dc62701ab32#npm:2.8.2"],\ - ["@parcel/logger", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0", {\ + "packageLocation": "./.yarn/__virtual__/@parcel-cache-virtual-a2e9499dbb/0/cache/@parcel-cache-npm-2.12.0-3389909f2c-a45e799809.zip/node_modules/@parcel/cache/",\ + "packageDependencies": [\ + ["@parcel/cache", "virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0"],\ + ["@parcel/core", "npm:2.12.0"],\ + ["@parcel/fs", "virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0"],\ + ["@parcel/logger", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["@types/parcel__core", null],\ - ["lmdb", "npm:2.5.2"]\ - ],\ - "packagePeers": [\ - "@types/parcel__core"\ - ],\ - "linkType": "HARD"\ - }],\ - ["virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2", {\ - "packageLocation": "./.yarn/__virtual__/@parcel-cache-virtual-860bd7931d/0/cache/@parcel-cache-npm-2.8.2-4957caf228-7d1c951e3f.zip/node_modules/@parcel/cache/",\ - "packageDependencies": [\ - ["@parcel/cache", "virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2"],\ - ["@parcel/core", "npm:2.8.2"],\ - ["@parcel/fs", "virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2"],\ - ["@parcel/logger", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ - ["@types/parcel__core", null],\ - ["lmdb", "npm:2.5.2"]\ + ["lmdb", "npm:2.8.5"]\ ],\ "packagePeers": [\ "@parcel/core",\ @@ -751,79 +1221,96 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "@types/parcel__core"\ ],\ "linkType": "HARD"\ + }],\ + ["virtual:ffe47febbf7847f9b64454e506be514f3cbd8bbd1821ba64e8e762685b5100c3f7867a926c2aa7f5349f2a1370184e7d2f8f70428bcab9b21701f56d9632c378#npm:2.12.0", {\ + "packageLocation": "./.yarn/__virtual__/@parcel-cache-virtual-6f5cc88243/0/cache/@parcel-cache-npm-2.12.0-3389909f2c-a45e799809.zip/node_modules/@parcel/cache/",\ + "packageDependencies": [\ + ["@parcel/cache", "virtual:ffe47febbf7847f9b64454e506be514f3cbd8bbd1821ba64e8e762685b5100c3f7867a926c2aa7f5349f2a1370184e7d2f8f70428bcab9b21701f56d9632c378#npm:2.12.0"],\ + ["@parcel/core", "npm:2.6.2"],\ + ["@parcel/fs", "virtual:ffe47febbf7847f9b64454e506be514f3cbd8bbd1821ba64e8e762685b5100c3f7867a926c2aa7f5349f2a1370184e7d2f8f70428bcab9b21701f56d9632c378#npm:2.12.0"],\ + ["@parcel/logger", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ + ["@types/parcel__core", null],\ + ["lmdb", "npm:2.8.5"]\ + ],\ + "packagePeers": [\ + "@types/parcel__core"\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@parcel/codeframe", [\ - ["npm:2.6.2", {\ - "packageLocation": "./.yarn/cache/@parcel-codeframe-npm-2.6.2-39f0ef1504-3253f42b90.zip/node_modules/@parcel/codeframe/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-codeframe-npm-2.12.0-aa8027940e-265c4d7ebe.zip/node_modules/@parcel/codeframe/",\ "packageDependencies": [\ - ["@parcel/codeframe", "npm:2.6.2"],\ + ["@parcel/codeframe", "npm:2.12.0"],\ ["chalk", "npm:4.1.2"]\ ],\ "linkType": "HARD"\ }],\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-codeframe-npm-2.8.2-77f4dce4ad-a2638353c6.zip/node_modules/@parcel/codeframe/",\ + ["npm:2.6.2", {\ + "packageLocation": "./.yarn/cache/@parcel-codeframe-npm-2.6.2-39f0ef1504-3253f42b90.zip/node_modules/@parcel/codeframe/",\ "packageDependencies": [\ - ["@parcel/codeframe", "npm:2.8.2"],\ + ["@parcel/codeframe", "npm:2.6.2"],\ ["chalk", "npm:4.1.2"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/compressor-raw", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-compressor-raw-npm-2.8.2-0d385dde76-61a1299615.zip/node_modules/@parcel/compressor-raw/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-compressor-raw-npm-2.12.0-19f313c172-16c56704f3.zip/node_modules/@parcel/compressor-raw/",\ "packageDependencies": [\ - ["@parcel/compressor-raw", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"]\ + ["@parcel/compressor-raw", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/config-default", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-config-default-npm-2.8.2-89026bc258-035db3ab37.zip/node_modules/@parcel/config-default/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-config-default-npm-2.12.0-aefd3c699e-72877c5dc4.zip/node_modules/@parcel/config-default/",\ "packageDependencies": [\ - ["@parcel/config-default", "npm:2.8.2"]\ + ["@parcel/config-default", "npm:2.12.0"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:7ff1d17261e888ed45770ab8af325407870f62c23cd6264c3a2830a9a45cf064c4196c0c92d06cfbc9b69b3bbe2e4c7776dce9f9d39af95141f0808c9e3cc9ec#npm:2.8.2", {\ - "packageLocation": "./.yarn/__virtual__/@parcel-config-default-virtual-a8d3e47300/0/cache/@parcel-config-default-npm-2.8.2-89026bc258-035db3ab37.zip/node_modules/@parcel/config-default/",\ - "packageDependencies": [\ - ["@parcel/config-default", "virtual:7ff1d17261e888ed45770ab8af325407870f62c23cd6264c3a2830a9a45cf064c4196c0c92d06cfbc9b69b3bbe2e4c7776dce9f9d39af95141f0808c9e3cc9ec#npm:2.8.2"],\ - ["@parcel/bundler-default", "npm:2.8.2"],\ - ["@parcel/compressor-raw", "npm:2.8.2"],\ - ["@parcel/core", "npm:2.8.2"],\ - ["@parcel/namer-default", "npm:2.8.2"],\ - ["@parcel/optimizer-css", "npm:2.8.2"],\ - ["@parcel/optimizer-htmlnano", "npm:2.8.2"],\ - ["@parcel/optimizer-image", "npm:2.8.2"],\ - ["@parcel/optimizer-svgo", "npm:2.8.2"],\ - ["@parcel/optimizer-terser", "npm:2.8.2"],\ - ["@parcel/packager-css", "npm:2.8.2"],\ - ["@parcel/packager-html", "npm:2.8.2"],\ - ["@parcel/packager-js", "npm:2.8.2"],\ - ["@parcel/packager-raw", "npm:2.8.2"],\ - ["@parcel/packager-svg", "npm:2.8.2"],\ - ["@parcel/reporter-dev-server", "npm:2.8.2"],\ - ["@parcel/resolver-default", "npm:2.8.2"],\ - ["@parcel/runtime-browser-hmr", "npm:2.8.2"],\ - ["@parcel/runtime-js", "npm:2.8.2"],\ - ["@parcel/runtime-react-refresh", "npm:2.8.2"],\ - ["@parcel/runtime-service-worker", "npm:2.8.2"],\ - ["@parcel/transformer-babel", "npm:2.8.2"],\ - ["@parcel/transformer-css", "npm:2.8.2"],\ - ["@parcel/transformer-html", "npm:2.8.2"],\ - ["@parcel/transformer-image", "virtual:a8d3e47300ee51a368ec8290994b964e28af6a1cbbe38a4f7a9755ee0929a36efd1109fa5eb1b35945f18ecd707c76db2ec79a9b4ff78c272d3ff1debe2b54e9#npm:2.8.2"],\ - ["@parcel/transformer-js", "virtual:a8d3e47300ee51a368ec8290994b964e28af6a1cbbe38a4f7a9755ee0929a36efd1109fa5eb1b35945f18ecd707c76db2ec79a9b4ff78c272d3ff1debe2b54e9#npm:2.8.2"],\ - ["@parcel/transformer-json", "npm:2.8.2"],\ - ["@parcel/transformer-postcss", "npm:2.8.2"],\ - ["@parcel/transformer-posthtml", "npm:2.8.2"],\ - ["@parcel/transformer-raw", "npm:2.8.2"],\ - ["@parcel/transformer-react-refresh-wrap", "npm:2.8.2"],\ - ["@parcel/transformer-svg", "npm:2.8.2"],\ + ["virtual:fdd74b573cf769bcde15fb47c39fbe0d73f59838182900fd59d3d43b2214ea01b1d45084fb49d0c192fc3e8a49adea5782afcb7fe14e09c63bedaf09f4939e35#npm:2.12.0", {\ + "packageLocation": "./.yarn/__virtual__/@parcel-config-default-virtual-284acdc258/0/cache/@parcel-config-default-npm-2.12.0-aefd3c699e-72877c5dc4.zip/node_modules/@parcel/config-default/",\ + "packageDependencies": [\ + ["@parcel/config-default", "virtual:fdd74b573cf769bcde15fb47c39fbe0d73f59838182900fd59d3d43b2214ea01b1d45084fb49d0c192fc3e8a49adea5782afcb7fe14e09c63bedaf09f4939e35#npm:2.12.0"],\ + ["@parcel/bundler-default", "npm:2.12.0"],\ + ["@parcel/compressor-raw", "npm:2.12.0"],\ + ["@parcel/core", "npm:2.12.0"],\ + ["@parcel/namer-default", "npm:2.12.0"],\ + ["@parcel/optimizer-css", "npm:2.12.0"],\ + ["@parcel/optimizer-htmlnano", "npm:2.12.0"],\ + ["@parcel/optimizer-image", "virtual:284acdc258f2328e304855ff98dec9e5e8952a2bd7797a2e11c082f6cad2e0d3068e07fb498d46b810d8efae36becee510ac53186a75e438e809dc472f832ab2#npm:2.12.0"],\ + ["@parcel/optimizer-svgo", "npm:2.12.0"],\ + ["@parcel/optimizer-swc", "npm:2.12.0"],\ + ["@parcel/packager-css", "npm:2.12.0"],\ + ["@parcel/packager-html", "npm:2.12.0"],\ + ["@parcel/packager-js", "npm:2.12.0"],\ + ["@parcel/packager-raw", "npm:2.12.0"],\ + ["@parcel/packager-svg", "npm:2.12.0"],\ + ["@parcel/packager-wasm", "npm:2.12.0"],\ + ["@parcel/reporter-dev-server", "npm:2.12.0"],\ + ["@parcel/resolver-default", "npm:2.12.0"],\ + ["@parcel/runtime-browser-hmr", "npm:2.12.0"],\ + ["@parcel/runtime-js", "npm:2.12.0"],\ + ["@parcel/runtime-react-refresh", "npm:2.12.0"],\ + ["@parcel/runtime-service-worker", "npm:2.12.0"],\ + ["@parcel/transformer-babel", "npm:2.12.0"],\ + ["@parcel/transformer-css", "npm:2.12.0"],\ + ["@parcel/transformer-html", "npm:2.12.0"],\ + ["@parcel/transformer-image", "virtual:284acdc258f2328e304855ff98dec9e5e8952a2bd7797a2e11c082f6cad2e0d3068e07fb498d46b810d8efae36becee510ac53186a75e438e809dc472f832ab2#npm:2.12.0"],\ + ["@parcel/transformer-js", "virtual:284acdc258f2328e304855ff98dec9e5e8952a2bd7797a2e11c082f6cad2e0d3068e07fb498d46b810d8efae36becee510ac53186a75e438e809dc472f832ab2#npm:2.12.0"],\ + ["@parcel/transformer-json", "npm:2.12.0"],\ + ["@parcel/transformer-postcss", "npm:2.12.0"],\ + ["@parcel/transformer-posthtml", "npm:2.12.0"],\ + ["@parcel/transformer-raw", "npm:2.12.0"],\ + ["@parcel/transformer-react-refresh-wrap", "npm:2.12.0"],\ + ["@parcel/transformer-svg", "npm:2.12.0"],\ ["@types/parcel__core", null]\ ],\ "packagePeers": [\ @@ -834,6 +1321,38 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@parcel/core", [\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-core-npm-2.12.0-8f08b883d4-5bf6746308.zip/node_modules/@parcel/core/",\ + "packageDependencies": [\ + ["@parcel/core", "npm:2.12.0"],\ + ["@mischnic/json-sourcemap", "npm:0.1.0"],\ + ["@parcel/cache", "virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/events", "npm:2.12.0"],\ + ["@parcel/fs", "virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0"],\ + ["@parcel/graph", "npm:3.2.0"],\ + ["@parcel/logger", "npm:2.12.0"],\ + ["@parcel/package-manager", "virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/profiler", "npm:2.12.0"],\ + ["@parcel/rust", "npm:2.12.0"],\ + ["@parcel/source-map", "npm:2.1.1"],\ + ["@parcel/types", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ + ["@parcel/workers", "virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0"],\ + ["abortcontroller-polyfill", "npm:1.7.3"],\ + ["base-x", "npm:3.0.9"],\ + ["browserslist", "npm:4.20.3"],\ + ["clone", "npm:2.1.2"],\ + ["dotenv", "npm:7.0.0"],\ + ["dotenv-expand", "npm:5.1.0"],\ + ["json5", "npm:2.2.1"],\ + ["msgpackr", "npm:1.10.1"],\ + ["nullthrows", "npm:1.1.1"],\ + ["semver", "npm:7.5.4"]\ + ],\ + "linkType": "HARD"\ + }],\ ["npm:2.6.2", {\ "packageLocation": "./.yarn/cache/@parcel-core-npm-2.6.2-f04091cfa7-f550cbbd5e.zip/node_modules/@parcel/core/",\ "packageDependencies": [\ @@ -864,53 +1383,22 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["semver", "npm:5.7.1"]\ ],\ "linkType": "HARD"\ - }],\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-core-npm-2.8.2-7ac9ecd9f9-0c989ef087.zip/node_modules/@parcel/core/",\ - "packageDependencies": [\ - ["@parcel/core", "npm:2.8.2"],\ - ["@mischnic/json-sourcemap", "npm:0.1.0"],\ - ["@parcel/cache", "virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/events", "npm:2.8.2"],\ - ["@parcel/fs", "virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2"],\ - ["@parcel/graph", "npm:2.8.2"],\ - ["@parcel/hash", "npm:2.8.2"],\ - ["@parcel/logger", "npm:2.8.2"],\ - ["@parcel/package-manager", "virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["@parcel/source-map", "npm:2.1.1"],\ - ["@parcel/types", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ - ["@parcel/workers", "virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2"],\ - ["abortcontroller-polyfill", "npm:1.7.3"],\ - ["base-x", "npm:3.0.9"],\ - ["browserslist", "npm:4.20.3"],\ - ["clone", "npm:2.1.2"],\ - ["dotenv", "npm:7.0.0"],\ - ["dotenv-expand", "npm:5.1.0"],\ - ["json5", "npm:2.2.1"],\ - ["msgpackr", "npm:1.6.0"],\ - ["nullthrows", "npm:1.1.1"],\ - ["semver", "npm:5.7.1"]\ - ],\ - "linkType": "HARD"\ }]\ ]],\ ["@parcel/diagnostic", [\ - ["npm:2.6.2", {\ - "packageLocation": "./.yarn/cache/@parcel-diagnostic-npm-2.6.2-ad66c9d460-c20c7b12c4.zip/node_modules/@parcel/diagnostic/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-diagnostic-npm-2.12.0-6e89ddad28-a4b918c1a0.zip/node_modules/@parcel/diagnostic/",\ "packageDependencies": [\ - ["@parcel/diagnostic", "npm:2.6.2"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ ["@mischnic/json-sourcemap", "npm:0.1.0"],\ ["nullthrows", "npm:1.1.1"]\ ],\ "linkType": "HARD"\ }],\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-diagnostic-npm-2.8.2-7f2dfb035e-91ca29cce4.zip/node_modules/@parcel/diagnostic/",\ + ["npm:2.6.2", {\ + "packageLocation": "./.yarn/cache/@parcel-diagnostic-npm-2.6.2-ad66c9d460-c20c7b12c4.zip/node_modules/@parcel/diagnostic/",\ "packageDependencies": [\ - ["@parcel/diagnostic", "npm:2.8.2"],\ + ["@parcel/diagnostic", "npm:2.6.2"],\ ["@mischnic/json-sourcemap", "npm:0.1.0"],\ ["nullthrows", "npm:1.1.1"]\ ],\ @@ -918,63 +1406,46 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@parcel/events", [\ - ["npm:2.6.2", {\ - "packageLocation": "./.yarn/cache/@parcel-events-npm-2.6.2-c1dc15633e-272898db0c.zip/node_modules/@parcel/events/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-events-npm-2.12.0-e6eff18c8c-136a8a2921.zip/node_modules/@parcel/events/",\ "packageDependencies": [\ - ["@parcel/events", "npm:2.6.2"]\ + ["@parcel/events", "npm:2.12.0"]\ ],\ "linkType": "HARD"\ }],\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-events-npm-2.8.2-ddf12da1ba-99aad2e735.zip/node_modules/@parcel/events/",\ + ["npm:2.6.2", {\ + "packageLocation": "./.yarn/cache/@parcel-events-npm-2.6.2-c1dc15633e-272898db0c.zip/node_modules/@parcel/events/",\ "packageDependencies": [\ - ["@parcel/events", "npm:2.8.2"]\ + ["@parcel/events", "npm:2.6.2"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/fs", [\ - ["npm:2.6.2", {\ - "packageLocation": "./.yarn/cache/@parcel-fs-npm-2.6.2-1670f601e3-b5e324d93b.zip/node_modules/@parcel/fs/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-fs-npm-2.12.0-3c46842e62-43d454d55d.zip/node_modules/@parcel/fs/",\ "packageDependencies": [\ - ["@parcel/fs", "npm:2.6.2"]\ + ["@parcel/fs", "npm:2.12.0"]\ ],\ "linkType": "SOFT"\ }],\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-fs-npm-2.8.2-97422ca16d-c25408fe2d.zip/node_modules/@parcel/fs/",\ + ["npm:2.6.2", {\ + "packageLocation": "./.yarn/cache/@parcel-fs-npm-2.6.2-1670f601e3-b5e324d93b.zip/node_modules/@parcel/fs/",\ "packageDependencies": [\ - ["@parcel/fs", "npm:2.8.2"]\ + ["@parcel/fs", "npm:2.6.2"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:4a1952be093bd8b413333d6452f7e82b253ed9257f221c7341bb9d244e18522a0f05e6b82b53db04b831a339d732c092e9b930055d598f0acbd03dc62701ab32#npm:2.8.2", {\ - "packageLocation": "./.yarn/__virtual__/@parcel-fs-virtual-c518c9b1bf/0/cache/@parcel-fs-npm-2.8.2-97422ca16d-c25408fe2d.zip/node_modules/@parcel/fs/",\ - "packageDependencies": [\ - ["@parcel/fs", "virtual:4a1952be093bd8b413333d6452f7e82b253ed9257f221c7341bb9d244e18522a0f05e6b82b53db04b831a339d732c092e9b930055d598f0acbd03dc62701ab32#npm:2.8.2"],\ - ["@parcel/core", "npm:2.6.2"],\ - ["@parcel/fs-search", "npm:2.8.2"],\ - ["@parcel/types", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ - ["@parcel/watcher", "npm:2.0.7"],\ - ["@parcel/workers", "virtual:4a1952be093bd8b413333d6452f7e82b253ed9257f221c7341bb9d244e18522a0f05e6b82b53db04b831a339d732c092e9b930055d598f0acbd03dc62701ab32#npm:2.8.2"],\ - ["@types/parcel__core", null]\ - ],\ - "packagePeers": [\ - "@types/parcel__core"\ - ],\ - "linkType": "HARD"\ - }],\ - ["virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2", {\ - "packageLocation": "./.yarn/__virtual__/@parcel-fs-virtual-4dbb1e3886/0/cache/@parcel-fs-npm-2.8.2-97422ca16d-c25408fe2d.zip/node_modules/@parcel/fs/",\ - "packageDependencies": [\ - ["@parcel/fs", "virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2"],\ - ["@parcel/core", "npm:2.8.2"],\ - ["@parcel/fs-search", "npm:2.8.2"],\ - ["@parcel/types", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0", {\ + "packageLocation": "./.yarn/__virtual__/@parcel-fs-virtual-762e5c5add/0/cache/@parcel-fs-npm-2.12.0-3c46842e62-43d454d55d.zip/node_modules/@parcel/fs/",\ + "packageDependencies": [\ + ["@parcel/fs", "virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0"],\ + ["@parcel/core", "npm:2.12.0"],\ + ["@parcel/rust", "npm:2.12.0"],\ + ["@parcel/types", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["@parcel/watcher", "npm:2.0.7"],\ - ["@parcel/workers", "virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2"],\ + ["@parcel/workers", "virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0"],\ ["@types/parcel__core", null]\ ],\ "packagePeers": [\ @@ -999,6 +1470,23 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "@types/parcel__core"\ ],\ "linkType": "HARD"\ + }],\ + ["virtual:ffe47febbf7847f9b64454e506be514f3cbd8bbd1821ba64e8e762685b5100c3f7867a926c2aa7f5349f2a1370184e7d2f8f70428bcab9b21701f56d9632c378#npm:2.12.0", {\ + "packageLocation": "./.yarn/__virtual__/@parcel-fs-virtual-ae7dde1116/0/cache/@parcel-fs-npm-2.12.0-3c46842e62-43d454d55d.zip/node_modules/@parcel/fs/",\ + "packageDependencies": [\ + ["@parcel/fs", "virtual:ffe47febbf7847f9b64454e506be514f3cbd8bbd1821ba64e8e762685b5100c3f7867a926c2aa7f5349f2a1370184e7d2f8f70428bcab9b21701f56d9632c378#npm:2.12.0"],\ + ["@parcel/core", "npm:2.6.2"],\ + ["@parcel/rust", "npm:2.12.0"],\ + ["@parcel/types", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ + ["@parcel/watcher", "npm:2.0.7"],\ + ["@parcel/workers", "virtual:ffe47febbf7847f9b64454e506be514f3cbd8bbd1821ba64e8e762685b5100c3f7867a926c2aa7f5349f2a1370184e7d2f8f70428bcab9b21701f56d9632c378#npm:2.12.0"],\ + ["@types/parcel__core", null]\ + ],\ + "packagePeers": [\ + "@types/parcel__core"\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@parcel/fs-search", [\ @@ -1009,14 +1497,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["detect-libc", "npm:1.0.3"]\ ],\ "linkType": "HARD"\ - }],\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/unplugged/@parcel-fs-search-npm-2.8.2-a3c70b64fe/node_modules/@parcel/fs-search/",\ - "packageDependencies": [\ - ["@parcel/fs-search", "npm:2.8.2"],\ - ["detect-libc", "npm:1.0.3"]\ - ],\ - "linkType": "HARD"\ }]\ ]],\ ["@parcel/graph", [\ @@ -1029,10 +1509,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "HARD"\ }],\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-graph-npm-2.8.2-039d19c5f3-d503597911.zip/node_modules/@parcel/graph/",\ + ["npm:3.2.0", {\ + "packageLocation": "./.yarn/cache/@parcel-graph-npm-3.2.0-92821d4289-b4d31624fc.zip/node_modules/@parcel/graph/",\ "packageDependencies": [\ - ["@parcel/graph", "npm:2.8.2"],\ + ["@parcel/graph", "npm:3.2.0"],\ ["nullthrows", "npm:1.1.1"]\ ],\ "linkType": "HARD"\ @@ -1047,18 +1527,18 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["xxhash-wasm", "npm:0.4.2"]\ ],\ "linkType": "HARD"\ - }],\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/unplugged/@parcel-hash-npm-2.8.2-4189a2e2e3/node_modules/@parcel/hash/",\ - "packageDependencies": [\ - ["@parcel/hash", "npm:2.8.2"],\ - ["detect-libc", "npm:1.0.3"],\ - ["xxhash-wasm", "npm:0.4.2"]\ - ],\ - "linkType": "HARD"\ }]\ ]],\ ["@parcel/logger", [\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-logger-npm-2.12.0-7d2f85a906-be3fe9d9ea.zip/node_modules/@parcel/logger/",\ + "packageDependencies": [\ + ["@parcel/logger", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/events", "npm:2.12.0"]\ + ],\ + "linkType": "HARD"\ + }],\ ["npm:2.6.2", {\ "packageLocation": "./.yarn/cache/@parcel-logger-npm-2.6.2-d7fe563ebb-d3536408da.zip/node_modules/@parcel/logger/",\ "packageDependencies": [\ @@ -1067,69 +1547,63 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@parcel/events", "npm:2.6.2"]\ ],\ "linkType": "HARD"\ - }],\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-logger-npm-2.8.2-0b40fa2df8-8d9b4264cb.zip/node_modules/@parcel/logger/",\ - "packageDependencies": [\ - ["@parcel/logger", "npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/events", "npm:2.8.2"]\ - ],\ - "linkType": "HARD"\ }]\ ]],\ - ["@parcel/markdown-ansi", [\ - ["npm:2.6.2", {\ - "packageLocation": "./.yarn/cache/@parcel-markdown-ansi-npm-2.6.2-16ce118d53-742c64c5db.zip/node_modules/@parcel/markdown-ansi/",\ + ["@parcel/markdown-ansi", [\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-markdown-ansi-npm-2.12.0-6b0fe453df-850ee665d9.zip/node_modules/@parcel/markdown-ansi/",\ "packageDependencies": [\ - ["@parcel/markdown-ansi", "npm:2.6.2"],\ + ["@parcel/markdown-ansi", "npm:2.12.0"],\ ["chalk", "npm:4.1.2"]\ ],\ "linkType": "HARD"\ }],\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-markdown-ansi-npm-2.8.2-3a4b50f123-aaff302f12.zip/node_modules/@parcel/markdown-ansi/",\ + ["npm:2.6.2", {\ + "packageLocation": "./.yarn/cache/@parcel-markdown-ansi-npm-2.6.2-16ce118d53-742c64c5db.zip/node_modules/@parcel/markdown-ansi/",\ "packageDependencies": [\ - ["@parcel/markdown-ansi", "npm:2.8.2"],\ + ["@parcel/markdown-ansi", "npm:2.6.2"],\ ["chalk", "npm:4.1.2"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/namer-default", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-namer-default-npm-2.8.2-d3e74161c0-c9592f4022.zip/node_modules/@parcel/namer-default/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-namer-default-npm-2.12.0-28980cfd47-dc92ec0945.zip/node_modules/@parcel/namer-default/",\ "packageDependencies": [\ - ["@parcel/namer-default", "npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ + ["@parcel/namer-default", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ ["nullthrows", "npm:1.1.1"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/node-resolver-core", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-node-resolver-core-npm-2.8.2-5629a9b021-92f0e2bf4b.zip/node_modules/@parcel/node-resolver-core/",\ + ["npm:3.3.0", {\ + "packageLocation": "./.yarn/cache/@parcel-node-resolver-core-npm-3.3.0-53804df663-acc3721678.zip/node_modules/@parcel/node-resolver-core/",\ "packageDependencies": [\ - ["@parcel/node-resolver-core", "npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["@parcel/node-resolver-core", "npm:3.3.0"],\ + ["@mischnic/json-sourcemap", "npm:0.1.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/fs", "virtual:ffe47febbf7847f9b64454e506be514f3cbd8bbd1821ba64e8e762685b5100c3f7867a926c2aa7f5349f2a1370184e7d2f8f70428bcab9b21701f56d9632c378#npm:2.12.0"],\ + ["@parcel/rust", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["nullthrows", "npm:1.1.1"],\ - ["semver", "npm:5.7.1"]\ + ["semver", "npm:7.5.4"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/optimizer-css", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-optimizer-css-npm-2.8.2-6de222af5e-8298155bac.zip/node_modules/@parcel/optimizer-css/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-optimizer-css-npm-2.12.0-f95bd4d060-abcdf58c29.zip/node_modules/@parcel/optimizer-css/",\ "packageDependencies": [\ - ["@parcel/optimizer-css", "npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ + ["@parcel/optimizer-css", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ ["@parcel/source-map", "npm:2.1.1"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["browserslist", "npm:4.20.3"],\ ["lightningcss", "npm:1.17.1"],\ ["nullthrows", "npm:1.1.1"]\ @@ -1138,12 +1612,12 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@parcel/optimizer-data-url", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-optimizer-data-url-npm-2.8.2-2b95b0c045-e0966a5e18.zip/node_modules/@parcel/optimizer-data-url/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-optimizer-data-url-npm-2.12.0-dad3731170-0397293961.zip/node_modules/@parcel/optimizer-data-url/",\ "packageDependencies": [\ - ["@parcel/optimizer-data-url", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["@parcel/optimizer-data-url", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["isbinaryfile", "npm:4.0.10"],\ ["mime", "npm:2.6.0"]\ ],\ @@ -1151,12 +1625,12 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@parcel/optimizer-htmlnano", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-optimizer-htmlnano-npm-2.8.2-989bccf2aa-3913b51ccd.zip/node_modules/@parcel/optimizer-htmlnano/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-optimizer-htmlnano-npm-2.12.0-cdd2835c12-64e571f56f.zip/node_modules/@parcel/optimizer-htmlnano/",\ "packageDependencies": [\ - ["@parcel/optimizer-htmlnano", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["htmlnano", "virtual:989bccf2aa3cb6d741c7ce5643ed4e817accfff58f84aaf6345d18a48f60699e8de9c81c8278d6bdf8deed7d9e463b2cc20d78724ce4bd72d1bbf84cb8c02220#npm:2.0.2"],\ + ["@parcel/optimizer-htmlnano", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["htmlnano", "virtual:cdd2835c1202e86fad55b2266578ff3755267672440481af37bdfff670fd205f561469a10385c20d1ff403af7fad49006bc71ffff21d12592a8ebd0c8be79c0c#npm:2.0.2"],\ ["nullthrows", "npm:1.1.1"],\ ["posthtml", "npm:0.16.6"],\ ["svgo", "npm:2.8.0"]\ @@ -1165,95 +1639,90 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@parcel/optimizer-image", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/unplugged/@parcel-optimizer-image-npm-2.8.2-eb7453ba87/node_modules/@parcel/optimizer-image/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-optimizer-image-npm-2.12.0-4cbc56f72d-7d28379bf1.zip/node_modules/@parcel/optimizer-image/",\ "packageDependencies": [\ - ["@parcel/optimizer-image", "npm:2.8.2"],\ - ["@parcel/core", "npm:2.6.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ - ["@parcel/workers", "virtual:4a1952be093bd8b413333d6452f7e82b253ed9257f221c7341bb9d244e18522a0f05e6b82b53db04b831a339d732c092e9b930055d598f0acbd03dc62701ab32#npm:2.8.2"],\ - ["detect-libc", "npm:1.0.3"]\ + ["@parcel/optimizer-image", "npm:2.12.0"]\ + ],\ + "linkType": "SOFT"\ + }],\ + ["virtual:284acdc258f2328e304855ff98dec9e5e8952a2bd7797a2e11c082f6cad2e0d3068e07fb498d46b810d8efae36becee510ac53186a75e438e809dc472f832ab2#npm:2.12.0", {\ + "packageLocation": "./.yarn/__virtual__/@parcel-optimizer-image-virtual-8c3b1760b5/0/cache/@parcel-optimizer-image-npm-2.12.0-4cbc56f72d-7d28379bf1.zip/node_modules/@parcel/optimizer-image/",\ + "packageDependencies": [\ + ["@parcel/optimizer-image", "virtual:284acdc258f2328e304855ff98dec9e5e8952a2bd7797a2e11c082f6cad2e0d3068e07fb498d46b810d8efae36becee510ac53186a75e438e809dc472f832ab2#npm:2.12.0"],\ + ["@parcel/core", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/rust", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ + ["@parcel/workers", "virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0"],\ + ["@types/parcel__core", null]\ + ],\ + "packagePeers": [\ + "@parcel/core",\ + "@types/parcel__core"\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/optimizer-svgo", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-optimizer-svgo-npm-2.8.2-d86f49e88e-608179fb18.zip/node_modules/@parcel/optimizer-svgo/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-optimizer-svgo-npm-2.12.0-08c0f1b17f-d3a4d2de9f.zip/node_modules/@parcel/optimizer-svgo/",\ "packageDependencies": [\ - ["@parcel/optimizer-svgo", "npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["@parcel/optimizer-svgo", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["svgo", "npm:2.8.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ - ["@parcel/optimizer-terser", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-optimizer-terser-npm-2.8.2-8af8c43b6e-e5cc9ef648.zip/node_modules/@parcel/optimizer-terser/",\ + ["@parcel/optimizer-swc", [\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-optimizer-swc-npm-2.12.0-fb535e4283-0b7fdf3df1.zip/node_modules/@parcel/optimizer-swc/",\ "packageDependencies": [\ - ["@parcel/optimizer-terser", "npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ + ["@parcel/optimizer-swc", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ ["@parcel/source-map", "npm:2.1.1"],\ - ["@parcel/utils", "npm:2.8.2"],\ - ["nullthrows", "npm:1.1.1"],\ - ["terser", "npm:5.13.1"]\ + ["@parcel/utils", "npm:2.12.0"],\ + ["@swc/core", "virtual:5f8211ac5fe0096c8679c8fc747f0917af84ce168460ce1b592cb42613ababf55139691f5b329cd10e1e2b99af39861401c7b9633ed396447c506b02a80144b0#npm:1.3.62"],\ + ["nullthrows", "npm:1.1.1"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/package-manager", [\ - ["npm:2.6.2", {\ - "packageLocation": "./.yarn/cache/@parcel-package-manager-npm-2.6.2-41edbfb7da-0c7dfce953.zip/node_modules/@parcel/package-manager/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-package-manager-npm-2.12.0-fc90aacf70-a517e9efe1.zip/node_modules/@parcel/package-manager/",\ "packageDependencies": [\ - ["@parcel/package-manager", "npm:2.6.2"]\ + ["@parcel/package-manager", "npm:2.12.0"]\ ],\ "linkType": "SOFT"\ }],\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-package-manager-npm-2.8.2-40215edd8a-99d022d3fa.zip/node_modules/@parcel/package-manager/",\ + ["npm:2.6.2", {\ + "packageLocation": "./.yarn/cache/@parcel-package-manager-npm-2.6.2-41edbfb7da-0c7dfce953.zip/node_modules/@parcel/package-manager/",\ "packageDependencies": [\ - ["@parcel/package-manager", "npm:2.8.2"]\ + ["@parcel/package-manager", "npm:2.6.2"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:4a1952be093bd8b413333d6452f7e82b253ed9257f221c7341bb9d244e18522a0f05e6b82b53db04b831a339d732c092e9b930055d598f0acbd03dc62701ab32#npm:2.8.2", {\ - "packageLocation": "./.yarn/__virtual__/@parcel-package-manager-virtual-50f7d94dbb/0/cache/@parcel-package-manager-npm-2.8.2-40215edd8a-99d022d3fa.zip/node_modules/@parcel/package-manager/",\ - "packageDependencies": [\ - ["@parcel/package-manager", "virtual:4a1952be093bd8b413333d6452f7e82b253ed9257f221c7341bb9d244e18522a0f05e6b82b53db04b831a339d732c092e9b930055d598f0acbd03dc62701ab32#npm:2.8.2"],\ - ["@parcel/core", "npm:2.6.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/fs", "virtual:4a1952be093bd8b413333d6452f7e82b253ed9257f221c7341bb9d244e18522a0f05e6b82b53db04b831a339d732c092e9b930055d598f0acbd03dc62701ab32#npm:2.8.2"],\ - ["@parcel/logger", "npm:2.8.2"],\ - ["@parcel/types", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ - ["@parcel/workers", "virtual:4a1952be093bd8b413333d6452f7e82b253ed9257f221c7341bb9d244e18522a0f05e6b82b53db04b831a339d732c092e9b930055d598f0acbd03dc62701ab32#npm:2.8.2"],\ - ["@types/parcel__core", null],\ - ["semver", "npm:5.7.1"]\ - ],\ - "packagePeers": [\ - "@types/parcel__core"\ - ],\ - "linkType": "HARD"\ - }],\ - ["virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2", {\ - "packageLocation": "./.yarn/__virtual__/@parcel-package-manager-virtual-1690fe77c3/0/cache/@parcel-package-manager-npm-2.8.2-40215edd8a-99d022d3fa.zip/node_modules/@parcel/package-manager/",\ - "packageDependencies": [\ - ["@parcel/package-manager", "virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2"],\ - ["@parcel/core", "npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/fs", "virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2"],\ - ["@parcel/logger", "npm:2.8.2"],\ - ["@parcel/types", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ - ["@parcel/workers", "virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2"],\ + ["virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0", {\ + "packageLocation": "./.yarn/__virtual__/@parcel-package-manager-virtual-8612c9adea/0/cache/@parcel-package-manager-npm-2.12.0-fc90aacf70-a517e9efe1.zip/node_modules/@parcel/package-manager/",\ + "packageDependencies": [\ + ["@parcel/package-manager", "virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0"],\ + ["@parcel/core", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/fs", "virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0"],\ + ["@parcel/logger", "npm:2.12.0"],\ + ["@parcel/node-resolver-core", "npm:3.3.0"],\ + ["@parcel/types", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ + ["@parcel/workers", "virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0"],\ + ["@swc/core", "virtual:5f8211ac5fe0096c8679c8fc747f0917af84ce168460ce1b592cb42613ababf55139691f5b329cd10e1e2b99af39861401c7b9633ed396447c506b02a80144b0#npm:1.3.62"],\ ["@types/parcel__core", null],\ - ["semver", "npm:5.7.1"]\ + ["semver", "npm:7.5.4"]\ ],\ "packagePeers": [\ "@parcel/core",\ @@ -1279,29 +1748,52 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "@types/parcel__core"\ ],\ "linkType": "HARD"\ + }],\ + ["virtual:ffe47febbf7847f9b64454e506be514f3cbd8bbd1821ba64e8e762685b5100c3f7867a926c2aa7f5349f2a1370184e7d2f8f70428bcab9b21701f56d9632c378#npm:2.12.0", {\ + "packageLocation": "./.yarn/__virtual__/@parcel-package-manager-virtual-5f8211ac5f/0/cache/@parcel-package-manager-npm-2.12.0-fc90aacf70-a517e9efe1.zip/node_modules/@parcel/package-manager/",\ + "packageDependencies": [\ + ["@parcel/package-manager", "virtual:ffe47febbf7847f9b64454e506be514f3cbd8bbd1821ba64e8e762685b5100c3f7867a926c2aa7f5349f2a1370184e7d2f8f70428bcab9b21701f56d9632c378#npm:2.12.0"],\ + ["@parcel/core", "npm:2.6.2"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/fs", "virtual:ffe47febbf7847f9b64454e506be514f3cbd8bbd1821ba64e8e762685b5100c3f7867a926c2aa7f5349f2a1370184e7d2f8f70428bcab9b21701f56d9632c378#npm:2.12.0"],\ + ["@parcel/logger", "npm:2.12.0"],\ + ["@parcel/node-resolver-core", "npm:3.3.0"],\ + ["@parcel/types", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ + ["@parcel/workers", "virtual:ffe47febbf7847f9b64454e506be514f3cbd8bbd1821ba64e8e762685b5100c3f7867a926c2aa7f5349f2a1370184e7d2f8f70428bcab9b21701f56d9632c378#npm:2.12.0"],\ + ["@swc/core", "virtual:5f8211ac5fe0096c8679c8fc747f0917af84ce168460ce1b592cb42613ababf55139691f5b329cd10e1e2b99af39861401c7b9633ed396447c506b02a80144b0#npm:1.3.62"],\ + ["@types/parcel__core", null],\ + ["semver", "npm:7.5.4"]\ + ],\ + "packagePeers": [\ + "@types/parcel__core"\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@parcel/packager-css", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-packager-css-npm-2.8.2-63302c1b3b-18ba8e43b3.zip/node_modules/@parcel/packager-css/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-packager-css-npm-2.12.0-b1c27a8323-684aaa1d85.zip/node_modules/@parcel/packager-css/",\ "packageDependencies": [\ - ["@parcel/packager-css", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ + ["@parcel/packager-css", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ ["@parcel/source-map", "npm:2.1.1"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["@parcel/utils", "npm:2.12.0"],\ + ["lightningcss", "npm:1.17.1"],\ ["nullthrows", "npm:1.1.1"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/packager-html", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-packager-html-npm-2.8.2-b901dd589c-e4975a4869.zip/node_modules/@parcel/packager-html/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-packager-html-npm-2.12.0-ad361b1265-ee558ad616.zip/node_modules/@parcel/packager-html/",\ "packageDependencies": [\ - ["@parcel/packager-html", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["@parcel/types", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["@parcel/packager-html", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/types", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["nullthrows", "npm:1.1.1"],\ ["posthtml", "npm:0.16.6"]\ ],\ @@ -1309,15 +1801,16 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@parcel/packager-js", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-packager-js-npm-2.8.2-9730c3d7a1-5c4a74e9b2.zip/node_modules/@parcel/packager-js/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-packager-js-npm-2.12.0-093e3200cd-2189b7ff15.zip/node_modules/@parcel/packager-js/",\ "packageDependencies": [\ - ["@parcel/packager-js", "npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/hash", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ + ["@parcel/packager-js", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/rust", "npm:2.12.0"],\ ["@parcel/source-map", "npm:2.1.1"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["@parcel/types", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["globals", "npm:13.15.0"],\ ["nullthrows", "npm:1.1.1"]\ ],\ @@ -1325,29 +1818,47 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@parcel/packager-raw", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-packager-raw-npm-2.8.2-e7b417ac32-198984e93e.zip/node_modules/@parcel/packager-raw/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-packager-raw-npm-2.12.0-b7f15635f8-39ce2fc7ae.zip/node_modules/@parcel/packager-raw/",\ "packageDependencies": [\ - ["@parcel/packager-raw", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"]\ + ["@parcel/packager-raw", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/packager-svg", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-packager-svg-npm-2.8.2-a7884bf9a1-7e10546425.zip/node_modules/@parcel/packager-svg/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-packager-svg-npm-2.12.0-fa921ce522-436ac9ea39.zip/node_modules/@parcel/packager-svg/",\ "packageDependencies": [\ - ["@parcel/packager-svg", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["@parcel/types", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["@parcel/packager-svg", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/types", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["posthtml", "npm:0.16.6"]\ ],\ "linkType": "HARD"\ }]\ ]],\ + ["@parcel/packager-wasm", [\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-packager-wasm-npm-2.12.0-ec551a9e29-a10e1cd988.zip/node_modules/@parcel/packager-wasm/",\ + "packageDependencies": [\ + ["@parcel/packager-wasm", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["@parcel/plugin", [\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-plugin-npm-2.12.0-947dec85d3-0b52f1dd06.zip/node_modules/@parcel/plugin/",\ + "packageDependencies": [\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/types", "npm:2.12.0"]\ + ],\ + "linkType": "HARD"\ + }],\ ["npm:2.6.2", {\ "packageLocation": "./.yarn/cache/@parcel-plugin-npm-2.6.2-d1ea2dda44-23da0fa372.zip/node_modules/@parcel/plugin/",\ "packageDependencies": [\ @@ -1355,24 +1866,28 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@parcel/types", "npm:2.6.2"]\ ],\ "linkType": "HARD"\ - }],\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-plugin-npm-2.8.2-1747a062e1-5c9f0ec6ff.zip/node_modules/@parcel/plugin/",\ + }]\ + ]],\ + ["@parcel/profiler", [\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-profiler-npm-2.12.0-69720a23ab-b683b74e10.zip/node_modules/@parcel/profiler/",\ "packageDependencies": [\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["@parcel/types", "npm:2.8.2"]\ + ["@parcel/profiler", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/events", "npm:2.12.0"],\ + ["chrome-trace-event", "npm:1.0.3"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/reporter-cli", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-reporter-cli-npm-2.8.2-57fd49365f-5ac5cbb7c3.zip/node_modules/@parcel/reporter-cli/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-reporter-cli-npm-2.12.0-b3e4c5fe19-8cc524fa15.zip/node_modules/@parcel/reporter-cli/",\ "packageDependencies": [\ - ["@parcel/reporter-cli", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["@parcel/types", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["@parcel/reporter-cli", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/types", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["chalk", "npm:4.1.2"],\ ["term-size", "npm:2.2.1"]\ ],\ @@ -1380,57 +1895,71 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@parcel/reporter-dev-server", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-reporter-dev-server-npm-2.8.2-55972e618f-1efff76ed9.zip/node_modules/@parcel/reporter-dev-server/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-reporter-dev-server-npm-2.12.0-aed1d2c68c-43957b4656.zip/node_modules/@parcel/reporter-dev-server/",\ + "packageDependencies": [\ + ["@parcel/reporter-dev-server", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@parcel/reporter-tracer", [\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-reporter-tracer-npm-2.12.0-5cec9ab2d5-24cddacd19.zip/node_modules/@parcel/reporter-tracer/",\ "packageDependencies": [\ - ["@parcel/reporter-dev-server", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"]\ + ["@parcel/reporter-tracer", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ + ["chrome-trace-event", "npm:1.0.3"],\ + ["nullthrows", "npm:1.1.1"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/resolver-default", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-resolver-default-npm-2.8.2-f0fe8ef74c-66e0233ed6.zip/node_modules/@parcel/resolver-default/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-resolver-default-npm-2.12.0-8da790891c-f3652eea09.zip/node_modules/@parcel/resolver-default/",\ "packageDependencies": [\ - ["@parcel/resolver-default", "npm:2.8.2"],\ - ["@parcel/node-resolver-core", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"]\ + ["@parcel/resolver-default", "npm:2.12.0"],\ + ["@parcel/node-resolver-core", "npm:3.3.0"],\ + ["@parcel/plugin", "npm:2.12.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/runtime-browser-hmr", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-runtime-browser-hmr-npm-2.8.2-bfd277b18f-64543de8cf.zip/node_modules/@parcel/runtime-browser-hmr/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-runtime-browser-hmr-npm-2.12.0-6f0da66673-bbba57ecee.zip/node_modules/@parcel/runtime-browser-hmr/",\ "packageDependencies": [\ - ["@parcel/runtime-browser-hmr", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"]\ + ["@parcel/runtime-browser-hmr", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/runtime-js", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-runtime-js-npm-2.8.2-171208460f-a5c0c7d2ad.zip/node_modules/@parcel/runtime-js/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-runtime-js-npm-2.12.0-e21acc0f42-6afa3e7eb2.zip/node_modules/@parcel/runtime-js/",\ "packageDependencies": [\ - ["@parcel/runtime-js", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["@parcel/runtime-js", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["nullthrows", "npm:1.1.1"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/runtime-react-refresh", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-runtime-react-refresh-npm-2.8.2-2b20ac8c6d-6483b8ed55.zip/node_modules/@parcel/runtime-react-refresh/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-runtime-react-refresh-npm-2.12.0-2b09615691-41aee9a874.zip/node_modules/@parcel/runtime-react-refresh/",\ "packageDependencies": [\ - ["@parcel/runtime-react-refresh", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["@parcel/runtime-react-refresh", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["react-error-overlay", "npm:6.0.9"],\ ["react-refresh", "npm:0.9.0"]\ ],\ @@ -1438,17 +1967,26 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@parcel/runtime-service-worker", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-runtime-service-worker-npm-2.8.2-1ec24cff9d-4b52703d3b.zip/node_modules/@parcel/runtime-service-worker/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-runtime-service-worker-npm-2.12.0-7d227ff0bf-c71246428e.zip/node_modules/@parcel/runtime-service-worker/",\ "packageDependencies": [\ - ["@parcel/runtime-service-worker", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["@parcel/runtime-service-worker", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["nullthrows", "npm:1.1.1"]\ ],\ "linkType": "HARD"\ }]\ ]],\ + ["@parcel/rust", [\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/unplugged/@parcel-rust-npm-2.12.0-0cf943f3e5/node_modules/@parcel/rust/",\ + "packageDependencies": [\ + ["@parcel/rust", "npm:2.12.0"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["@parcel/source-map", [\ ["npm:2.0.5", {\ "packageLocation": "./.yarn/unplugged/@parcel-source-map-npm-2.0.5-2444d2c092/node_modules/@parcel/source-map/",\ @@ -1468,31 +2006,31 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@parcel/transformer-babel", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-transformer-babel-npm-2.8.2-94dae9d0e8-4b2064aaba.zip/node_modules/@parcel/transformer-babel/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-transformer-babel-npm-2.12.0-953de52432-b8c457c0be.zip/node_modules/@parcel/transformer-babel/",\ "packageDependencies": [\ - ["@parcel/transformer-babel", "npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ + ["@parcel/transformer-babel", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ ["@parcel/source-map", "npm:2.1.1"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["browserslist", "npm:4.20.3"],\ ["json5", "npm:2.2.1"],\ ["nullthrows", "npm:1.1.1"],\ - ["semver", "npm:5.7.1"]\ + ["semver", "npm:7.5.4"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/transformer-css", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-transformer-css-npm-2.8.2-283cfa7f07-d0d3121d2b.zip/node_modules/@parcel/transformer-css/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-transformer-css-npm-2.12.0-24ddc31ae3-3a6f16321d.zip/node_modules/@parcel/transformer-css/",\ "packageDependencies": [\ - ["@parcel/transformer-css", "npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ + ["@parcel/transformer-css", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ ["@parcel/source-map", "npm:2.1.1"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["browserslist", "npm:4.20.3"],\ ["lightningcss", "npm:1.17.1"],\ ["nullthrows", "npm:1.1.1"]\ @@ -1501,38 +2039,39 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@parcel/transformer-html", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-transformer-html-npm-2.8.2-998bc39b95-e3bead4866.zip/node_modules/@parcel/transformer-html/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-transformer-html-npm-2.12.0-be2b9ee40c-7fcfac62ca.zip/node_modules/@parcel/transformer-html/",\ "packageDependencies": [\ - ["@parcel/transformer-html", "npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/hash", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ + ["@parcel/transformer-html", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/rust", "npm:2.12.0"],\ ["nullthrows", "npm:1.1.1"],\ ["posthtml", "npm:0.16.6"],\ ["posthtml-parser", "npm:0.10.2"],\ ["posthtml-render", "npm:3.0.0"],\ - ["semver", "npm:5.7.1"]\ + ["semver", "npm:7.5.4"],\ + ["srcset", "npm:4.0.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/transformer-image", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-transformer-image-npm-2.8.2-c8f5d0643b-acfe6e06f3.zip/node_modules/@parcel/transformer-image/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-transformer-image-npm-2.12.0-53f04e21c0-0a1581eacc.zip/node_modules/@parcel/transformer-image/",\ "packageDependencies": [\ - ["@parcel/transformer-image", "npm:2.8.2"]\ + ["@parcel/transformer-image", "npm:2.12.0"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:a8d3e47300ee51a368ec8290994b964e28af6a1cbbe38a4f7a9755ee0929a36efd1109fa5eb1b35945f18ecd707c76db2ec79a9b4ff78c272d3ff1debe2b54e9#npm:2.8.2", {\ - "packageLocation": "./.yarn/__virtual__/@parcel-transformer-image-virtual-b4b3db3f89/0/cache/@parcel-transformer-image-npm-2.8.2-c8f5d0643b-acfe6e06f3.zip/node_modules/@parcel/transformer-image/",\ - "packageDependencies": [\ - ["@parcel/transformer-image", "virtual:a8d3e47300ee51a368ec8290994b964e28af6a1cbbe38a4f7a9755ee0929a36efd1109fa5eb1b35945f18ecd707c76db2ec79a9b4ff78c272d3ff1debe2b54e9#npm:2.8.2"],\ - ["@parcel/core", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ - ["@parcel/workers", "virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2"],\ + ["virtual:284acdc258f2328e304855ff98dec9e5e8952a2bd7797a2e11c082f6cad2e0d3068e07fb498d46b810d8efae36becee510ac53186a75e438e809dc472f832ab2#npm:2.12.0", {\ + "packageLocation": "./.yarn/__virtual__/@parcel-transformer-image-virtual-acc9c20c9c/0/cache/@parcel-transformer-image-npm-2.12.0-53f04e21c0-0a1581eacc.zip/node_modules/@parcel/transformer-image/",\ + "packageDependencies": [\ + ["@parcel/transformer-image", "virtual:284acdc258f2328e304855ff98dec9e5e8952a2bd7797a2e11c082f6cad2e0d3068e07fb498d46b810d8efae36becee510ac53186a75e438e809dc472f832ab2#npm:2.12.0"],\ + ["@parcel/core", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ + ["@parcel/workers", "virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0"],\ ["@types/parcel__core", null],\ ["nullthrows", "npm:1.1.1"]\ ],\ @@ -1544,40 +2083,40 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@parcel/transformer-inline-string", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-transformer-inline-string-npm-2.8.2-3a03397064-5f6f4be447.zip/node_modules/@parcel/transformer-inline-string/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-transformer-inline-string-npm-2.12.0-a33f10bafa-5f63c08695.zip/node_modules/@parcel/transformer-inline-string/",\ "packageDependencies": [\ - ["@parcel/transformer-inline-string", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"]\ + ["@parcel/transformer-inline-string", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/transformer-js", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/unplugged/@parcel-transformer-js-virtual-e9a5222b61/node_modules/@parcel/transformer-js/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-transformer-js-npm-2.12.0-404d54db18-b9fe4c887b.zip/node_modules/@parcel/transformer-js/",\ "packageDependencies": [\ - ["@parcel/transformer-js", "npm:2.8.2"]\ + ["@parcel/transformer-js", "npm:2.12.0"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:a8d3e47300ee51a368ec8290994b964e28af6a1cbbe38a4f7a9755ee0929a36efd1109fa5eb1b35945f18ecd707c76db2ec79a9b4ff78c272d3ff1debe2b54e9#npm:2.8.2", {\ - "packageLocation": "./.yarn/unplugged/@parcel-transformer-js-virtual-e9a5222b61/node_modules/@parcel/transformer-js/",\ - "packageDependencies": [\ - ["@parcel/transformer-js", "virtual:a8d3e47300ee51a368ec8290994b964e28af6a1cbbe38a4f7a9755ee0929a36efd1109fa5eb1b35945f18ecd707c76db2ec79a9b4ff78c272d3ff1debe2b54e9#npm:2.8.2"],\ - ["@parcel/core", "npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ + ["virtual:284acdc258f2328e304855ff98dec9e5e8952a2bd7797a2e11c082f6cad2e0d3068e07fb498d46b810d8efae36becee510ac53186a75e438e809dc472f832ab2#npm:2.12.0", {\ + "packageLocation": "./.yarn/__virtual__/@parcel-transformer-js-virtual-567f83ac24/0/cache/@parcel-transformer-js-npm-2.12.0-404d54db18-b9fe4c887b.zip/node_modules/@parcel/transformer-js/",\ + "packageDependencies": [\ + ["@parcel/transformer-js", "virtual:284acdc258f2328e304855ff98dec9e5e8952a2bd7797a2e11c082f6cad2e0d3068e07fb498d46b810d8efae36becee510ac53186a75e438e809dc472f832ab2#npm:2.12.0"],\ + ["@parcel/core", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/rust", "npm:2.12.0"],\ ["@parcel/source-map", "npm:2.1.1"],\ - ["@parcel/utils", "npm:2.8.2"],\ - ["@parcel/workers", "virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2"],\ - ["@swc/helpers", "npm:0.4.14"],\ + ["@parcel/utils", "npm:2.12.0"],\ + ["@parcel/workers", "virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0"],\ + ["@swc/helpers", "npm:0.5.1"],\ ["@types/parcel__core", null],\ ["browserslist", "npm:4.20.3"],\ - ["detect-libc", "npm:1.0.3"],\ ["nullthrows", "npm:1.1.1"],\ ["regenerator-runtime", "npm:0.13.9"],\ - ["semver", "npm:5.7.1"]\ + ["semver", "npm:7.5.4"]\ ],\ "packagePeers": [\ "@parcel/core",\ @@ -1587,77 +2126,77 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@parcel/transformer-json", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-transformer-json-npm-2.8.2-98e2e0cf80-b22a609ae9.zip/node_modules/@parcel/transformer-json/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-transformer-json-npm-2.12.0-652d8d99d2-a711cb65a8.zip/node_modules/@parcel/transformer-json/",\ "packageDependencies": [\ - ["@parcel/transformer-json", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ + ["@parcel/transformer-json", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ ["json5", "npm:2.2.1"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/transformer-postcss", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-transformer-postcss-npm-2.8.2-547cd470da-ee152a91fb.zip/node_modules/@parcel/transformer-postcss/",\ - "packageDependencies": [\ - ["@parcel/transformer-postcss", "npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/hash", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-transformer-postcss-npm-2.12.0-f0cfb95fac-b210044a7f.zip/node_modules/@parcel/transformer-postcss/",\ + "packageDependencies": [\ + ["@parcel/transformer-postcss", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/rust", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["clone", "npm:2.1.2"],\ ["nullthrows", "npm:1.1.1"],\ ["postcss-value-parser", "npm:4.2.0"],\ - ["semver", "npm:5.7.1"]\ + ["semver", "npm:7.5.4"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/transformer-posthtml", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-transformer-posthtml-npm-2.8.2-76f67e31b6-4865968546.zip/node_modules/@parcel/transformer-posthtml/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-transformer-posthtml-npm-2.12.0-41c570db12-b62582ae7e.zip/node_modules/@parcel/transformer-posthtml/",\ "packageDependencies": [\ - ["@parcel/transformer-posthtml", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["@parcel/transformer-posthtml", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["nullthrows", "npm:1.1.1"],\ ["posthtml", "npm:0.16.6"],\ ["posthtml-parser", "npm:0.10.2"],\ ["posthtml-render", "npm:3.0.0"],\ - ["semver", "npm:5.7.1"]\ + ["semver", "npm:7.5.4"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/transformer-raw", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-transformer-raw-npm-2.8.2-a43c4fa2f7-386f64445a.zip/node_modules/@parcel/transformer-raw/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-transformer-raw-npm-2.12.0-bd2cb66ddf-de6681e2e7.zip/node_modules/@parcel/transformer-raw/",\ "packageDependencies": [\ - ["@parcel/transformer-raw", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"]\ + ["@parcel/transformer-raw", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/transformer-react-refresh-wrap", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-transformer-react-refresh-wrap-npm-2.8.2-62e6dd04c2-d091ab4a25.zip/node_modules/@parcel/transformer-react-refresh-wrap/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-transformer-react-refresh-wrap-npm-2.12.0-59ed68910f-9aba8c1ab0.zip/node_modules/@parcel/transformer-react-refresh-wrap/",\ "packageDependencies": [\ - ["@parcel/transformer-react-refresh-wrap", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["@parcel/transformer-react-refresh-wrap", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["react-refresh", "npm:0.9.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/transformer-sass", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-transformer-sass-npm-2.8.2-4e0c2f2900-42bbfa9401.zip/node_modules/@parcel/transformer-sass/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-transformer-sass-npm-2.12.0-ef787eef35-ce6b4d329b.zip/node_modules/@parcel/transformer-sass/",\ "packageDependencies": [\ - ["@parcel/transformer-sass", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ + ["@parcel/transformer-sass", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ ["@parcel/source-map", "npm:2.1.1"],\ ["sass", "npm:1.52.1"]\ ],\ @@ -1665,23 +2204,37 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@parcel/transformer-svg", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-transformer-svg-npm-2.8.2-c8870e67e5-e4522b69e3.zip/node_modules/@parcel/transformer-svg/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-transformer-svg-npm-2.12.0-f41b181676-92b7c65894.zip/node_modules/@parcel/transformer-svg/",\ "packageDependencies": [\ - ["@parcel/transformer-svg", "npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/hash", "npm:2.8.2"],\ - ["@parcel/plugin", "npm:2.8.2"],\ + ["@parcel/transformer-svg", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/plugin", "npm:2.12.0"],\ + ["@parcel/rust", "npm:2.12.0"],\ ["nullthrows", "npm:1.1.1"],\ ["posthtml", "npm:0.16.6"],\ ["posthtml-parser", "npm:0.10.2"],\ ["posthtml-render", "npm:3.0.0"],\ - ["semver", "npm:5.7.1"]\ + ["semver", "npm:7.5.4"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@parcel/types", [\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-types-npm-2.12.0-ffe47febbf-250f95580c.zip/node_modules/@parcel/types/",\ + "packageDependencies": [\ + ["@parcel/types", "npm:2.12.0"],\ + ["@parcel/cache", "virtual:ffe47febbf7847f9b64454e506be514f3cbd8bbd1821ba64e8e762685b5100c3f7867a926c2aa7f5349f2a1370184e7d2f8f70428bcab9b21701f56d9632c378#npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/fs", "virtual:ffe47febbf7847f9b64454e506be514f3cbd8bbd1821ba64e8e762685b5100c3f7867a926c2aa7f5349f2a1370184e7d2f8f70428bcab9b21701f56d9632c378#npm:2.12.0"],\ + ["@parcel/package-manager", "virtual:ffe47febbf7847f9b64454e506be514f3cbd8bbd1821ba64e8e762685b5100c3f7867a926c2aa7f5349f2a1370184e7d2f8f70428bcab9b21701f56d9632c378#npm:2.12.0"],\ + ["@parcel/source-map", "npm:2.1.1"],\ + ["@parcel/workers", "virtual:ffe47febbf7847f9b64454e506be514f3cbd8bbd1821ba64e8e762685b5100c3f7867a926c2aa7f5349f2a1370184e7d2f8f70428bcab9b21701f56d9632c378#npm:2.12.0"],\ + ["utility-types", "npm:3.10.0"]\ + ],\ + "linkType": "HARD"\ + }],\ ["npm:2.6.2", {\ "packageLocation": "./.yarn/cache/@parcel-types-npm-2.6.2-aa1797faca-16f3c3ac36.zip/node_modules/@parcel/types/",\ "packageDependencies": [\ @@ -1695,23 +2248,24 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["utility-types", "npm:3.10.0"]\ ],\ "linkType": "HARD"\ - }],\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-types-npm-2.8.2-4a1952be09-04b3d5f199.zip/node_modules/@parcel/types/",\ - "packageDependencies": [\ - ["@parcel/types", "npm:2.8.2"],\ - ["@parcel/cache", "virtual:4a1952be093bd8b413333d6452f7e82b253ed9257f221c7341bb9d244e18522a0f05e6b82b53db04b831a339d732c092e9b930055d598f0acbd03dc62701ab32#npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/fs", "virtual:4a1952be093bd8b413333d6452f7e82b253ed9257f221c7341bb9d244e18522a0f05e6b82b53db04b831a339d732c092e9b930055d598f0acbd03dc62701ab32#npm:2.8.2"],\ - ["@parcel/package-manager", "virtual:4a1952be093bd8b413333d6452f7e82b253ed9257f221c7341bb9d244e18522a0f05e6b82b53db04b831a339d732c092e9b930055d598f0acbd03dc62701ab32#npm:2.8.2"],\ - ["@parcel/source-map", "npm:2.1.1"],\ - ["@parcel/workers", "virtual:4a1952be093bd8b413333d6452f7e82b253ed9257f221c7341bb9d244e18522a0f05e6b82b53db04b831a339d732c092e9b930055d598f0acbd03dc62701ab32#npm:2.8.2"],\ - ["utility-types", "npm:3.10.0"]\ - ],\ - "linkType": "HARD"\ }]\ ]],\ ["@parcel/utils", [\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-utils-npm-2.12.0-d8a9a48a66-ba80a60fed.zip/node_modules/@parcel/utils/",\ + "packageDependencies": [\ + ["@parcel/utils", "npm:2.12.0"],\ + ["@parcel/codeframe", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/logger", "npm:2.12.0"],\ + ["@parcel/markdown-ansi", "npm:2.12.0"],\ + ["@parcel/rust", "npm:2.12.0"],\ + ["@parcel/source-map", "npm:2.1.1"],\ + ["chalk", "npm:4.1.2"],\ + ["nullthrows", "npm:1.1.1"]\ + ],\ + "linkType": "HARD"\ + }],\ ["npm:2.6.2", {\ "packageLocation": "./.yarn/cache/@parcel-utils-npm-2.6.2-cab87aed21-a74fdca966.zip/node_modules/@parcel/utils/",\ "packageDependencies": [\ @@ -1725,20 +2279,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["chalk", "npm:4.1.2"]\ ],\ "linkType": "HARD"\ - }],\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-utils-npm-2.8.2-8c378b4d3a-fcbc70426e.zip/node_modules/@parcel/utils/",\ - "packageDependencies": [\ - ["@parcel/utils", "npm:2.8.2"],\ - ["@parcel/codeframe", "npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/hash", "npm:2.8.2"],\ - ["@parcel/logger", "npm:2.8.2"],\ - ["@parcel/markdown-ansi", "npm:2.8.2"],\ - ["@parcel/source-map", "npm:2.1.1"],\ - ["chalk", "npm:4.1.2"]\ - ],\ - "linkType": "HARD"\ }]\ ]],\ ["@parcel/watcher", [\ @@ -1764,49 +2304,31 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@parcel/workers", [\ - ["npm:2.6.2", {\ - "packageLocation": "./.yarn/cache/@parcel-workers-npm-2.6.2-a30e38db52-92b65cd3fd.zip/node_modules/@parcel/workers/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/@parcel-workers-npm-2.12.0-3ddd4664bc-e19c3c0a66.zip/node_modules/@parcel/workers/",\ "packageDependencies": [\ - ["@parcel/workers", "npm:2.6.2"]\ + ["@parcel/workers", "npm:2.12.0"]\ ],\ "linkType": "SOFT"\ }],\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/@parcel-workers-npm-2.8.2-48e612dc38-df3f793301.zip/node_modules/@parcel/workers/",\ + ["npm:2.6.2", {\ + "packageLocation": "./.yarn/cache/@parcel-workers-npm-2.6.2-a30e38db52-92b65cd3fd.zip/node_modules/@parcel/workers/",\ "packageDependencies": [\ - ["@parcel/workers", "npm:2.8.2"]\ + ["@parcel/workers", "npm:2.6.2"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:4a1952be093bd8b413333d6452f7e82b253ed9257f221c7341bb9d244e18522a0f05e6b82b53db04b831a339d732c092e9b930055d598f0acbd03dc62701ab32#npm:2.8.2", {\ - "packageLocation": "./.yarn/__virtual__/@parcel-workers-virtual-851062ddba/0/cache/@parcel-workers-npm-2.8.2-48e612dc38-df3f793301.zip/node_modules/@parcel/workers/",\ - "packageDependencies": [\ - ["@parcel/workers", "virtual:4a1952be093bd8b413333d6452f7e82b253ed9257f221c7341bb9d244e18522a0f05e6b82b53db04b831a339d732c092e9b930055d598f0acbd03dc62701ab32#npm:2.8.2"],\ - ["@parcel/core", "npm:2.6.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/logger", "npm:2.8.2"],\ - ["@parcel/types", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ - ["@types/parcel__core", null],\ - ["chrome-trace-event", "npm:1.0.3"],\ - ["nullthrows", "npm:1.1.1"]\ - ],\ - "packagePeers": [\ - "@types/parcel__core"\ - ],\ - "linkType": "HARD"\ - }],\ - ["virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2", {\ - "packageLocation": "./.yarn/__virtual__/@parcel-workers-virtual-e52c4ea4f5/0/cache/@parcel-workers-npm-2.8.2-48e612dc38-df3f793301.zip/node_modules/@parcel/workers/",\ - "packageDependencies": [\ - ["@parcel/workers", "virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2"],\ - ["@parcel/core", "npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/logger", "npm:2.8.2"],\ - ["@parcel/types", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0", {\ + "packageLocation": "./.yarn/__virtual__/@parcel-workers-virtual-fbd6240557/0/cache/@parcel-workers-npm-2.12.0-3ddd4664bc-e19c3c0a66.zip/node_modules/@parcel/workers/",\ + "packageDependencies": [\ + ["@parcel/workers", "virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0"],\ + ["@parcel/core", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/logger", "npm:2.12.0"],\ + ["@parcel/profiler", "npm:2.12.0"],\ + ["@parcel/types", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["@types/parcel__core", null],\ - ["chrome-trace-event", "npm:1.0.3"],\ ["nullthrows", "npm:1.1.1"]\ ],\ "packagePeers": [\ @@ -1832,6 +2354,33 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "@types/parcel__core"\ ],\ "linkType": "HARD"\ + }],\ + ["virtual:ffe47febbf7847f9b64454e506be514f3cbd8bbd1821ba64e8e762685b5100c3f7867a926c2aa7f5349f2a1370184e7d2f8f70428bcab9b21701f56d9632c378#npm:2.12.0", {\ + "packageLocation": "./.yarn/__virtual__/@parcel-workers-virtual-0f6ac1cb6e/0/cache/@parcel-workers-npm-2.12.0-3ddd4664bc-e19c3c0a66.zip/node_modules/@parcel/workers/",\ + "packageDependencies": [\ + ["@parcel/workers", "virtual:ffe47febbf7847f9b64454e506be514f3cbd8bbd1821ba64e8e762685b5100c3f7867a926c2aa7f5349f2a1370184e7d2f8f70428bcab9b21701f56d9632c378#npm:2.12.0"],\ + ["@parcel/core", "npm:2.6.2"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/logger", "npm:2.12.0"],\ + ["@parcel/profiler", "npm:2.12.0"],\ + ["@parcel/types", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ + ["@types/parcel__core", null],\ + ["nullthrows", "npm:1.1.1"]\ + ],\ + "packagePeers": [\ + "@types/parcel__core"\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@pkgjs/parseargs", [\ + ["npm:0.11.0", {\ + "packageLocation": "./.yarn/cache/@pkgjs-parseargs-npm-0.11.0-cd2a3fe948-6ad6a00fc4.zip/node_modules/@pkgjs/parseargs/",\ + "packageDependencies": [\ + ["@pkgjs/parseargs", "npm:0.11.0"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@popperjs/core", [\ @@ -1842,26 +2391,26 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "HARD"\ }],\ - ["npm:2.11.6", {\ - "packageLocation": "./.yarn/cache/@popperjs-core-npm-2.11.6-5bcdc104bd-47fb328cec.zip/node_modules/@popperjs/core/",\ + ["npm:2.11.8", {\ + "packageLocation": "./.yarn/cache/@popperjs-core-npm-2.11.8-f1692e11a0-e5c69fdebf.zip/node_modules/@popperjs/core/",\ "packageDependencies": [\ - ["@popperjs/core", "npm:2.11.6"]\ + ["@popperjs/core", "npm:2.11.8"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@rollup/pluginutils", [\ - ["npm:5.0.2", {\ - "packageLocation": "./.yarn/cache/@rollup-pluginutils-npm-5.0.2-6aa9d0ddd4-edea15e543.zip/node_modules/@rollup/pluginutils/",\ + ["npm:5.1.0", {\ + "packageLocation": "./.yarn/cache/@rollup-pluginutils-npm-5.1.0-6939820ef8-3cc5a6d914.zip/node_modules/@rollup/pluginutils/",\ "packageDependencies": [\ - ["@rollup/pluginutils", "npm:5.0.2"]\ + ["@rollup/pluginutils", "npm:5.1.0"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.0.2", {\ - "packageLocation": "./.yarn/__virtual__/@rollup-pluginutils-virtual-ca58d3a074/0/cache/@rollup-pluginutils-npm-5.0.2-6aa9d0ddd4-edea15e543.zip/node_modules/@rollup/pluginutils/",\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.1.0", {\ + "packageLocation": "./.yarn/__virtual__/@rollup-pluginutils-virtual-e968017249/0/cache/@rollup-pluginutils-npm-5.1.0-6939820ef8-3cc5a6d914.zip/node_modules/@rollup/pluginutils/",\ "packageDependencies": [\ - ["@rollup/pluginutils", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.0.2"],\ + ["@rollup/pluginutils", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.1.0"],\ ["@types/estree", "npm:1.0.0"],\ ["@types/rollup", null],\ ["estree-walker", "npm:2.0.2"],\ @@ -1876,17 +2425,17 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@sidvind/better-ajv-errors", [\ - ["npm:2.0.0", {\ - "packageLocation": "./.yarn/cache/@sidvind-better-ajv-errors-npm-2.0.0-3531bddef9-12b0d87855.zip/node_modules/@sidvind/better-ajv-errors/",\ + ["npm:2.1.3", {\ + "packageLocation": "./.yarn/cache/@sidvind-better-ajv-errors-npm-2.1.3-e3d1c524a8-949cb805a1.zip/node_modules/@sidvind/better-ajv-errors/",\ "packageDependencies": [\ - ["@sidvind/better-ajv-errors", "npm:2.0.0"]\ + ["@sidvind/better-ajv-errors", "npm:2.1.3"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:ef051354894d83cc0cda7d7a4a6eca604681f35eb8f2a25fe305435ac599a4d71af04fcaa26daa3f5aa064e73442d202ba9dfbda5c3f8c7b1daf472f3a17da86#npm:2.0.0", {\ - "packageLocation": "./.yarn/__virtual__/@sidvind-better-ajv-errors-virtual-65bece6091/0/cache/@sidvind-better-ajv-errors-npm-2.0.0-3531bddef9-12b0d87855.zip/node_modules/@sidvind/better-ajv-errors/",\ + ["virtual:640261ed3b7a9880a388cc504caacf8ea790dd52f1cb31fbc3be445cb2adc6e73fc87097de620863105eb917510145ef2457d30000c7361456ab67ec0b895136#npm:2.1.3", {\ + "packageLocation": "./.yarn/__virtual__/@sidvind-better-ajv-errors-virtual-ff98ba00e3/0/cache/@sidvind-better-ajv-errors-npm-2.1.3-e3d1c524a8-949cb805a1.zip/node_modules/@sidvind/better-ajv-errors/",\ "packageDependencies": [\ - ["@sidvind/better-ajv-errors", "virtual:ef051354894d83cc0cda7d7a4a6eca604681f35eb8f2a25fe305435ac599a4d71af04fcaa26daa3f5aa064e73442d202ba9dfbda5c3f8c7b1daf472f3a17da86#npm:2.0.0"],\ + ["@sidvind/better-ajv-errors", "virtual:640261ed3b7a9880a388cc504caacf8ea790dd52f1cb31fbc3be445cb2adc6e73fc87097de620863105eb917510145ef2457d30000c7361456ab67ec0b895136#npm:2.1.3"],\ ["@babel/code-frame", "npm:7.16.7"],\ ["@types/ajv", null],\ ["ajv", "npm:8.11.0"],\ @@ -1899,11 +2448,133 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["@swc/core", [\ + ["npm:1.3.62", {\ + "packageLocation": "./.yarn/unplugged/@swc-core-virtual-8fda1c3f9b/node_modules/@swc/core/",\ + "packageDependencies": [\ + ["@swc/core", "npm:1.3.62"]\ + ],\ + "linkType": "SOFT"\ + }],\ + ["virtual:5f8211ac5fe0096c8679c8fc747f0917af84ce168460ce1b592cb42613ababf55139691f5b329cd10e1e2b99af39861401c7b9633ed396447c506b02a80144b0#npm:1.3.62", {\ + "packageLocation": "./.yarn/unplugged/@swc-core-virtual-8fda1c3f9b/node_modules/@swc/core/",\ + "packageDependencies": [\ + ["@swc/core", "virtual:5f8211ac5fe0096c8679c8fc747f0917af84ce168460ce1b592cb42613ababf55139691f5b329cd10e1e2b99af39861401c7b9633ed396447c506b02a80144b0#npm:1.3.62"],\ + ["@swc/core-darwin-arm64", "npm:1.3.62"],\ + ["@swc/core-darwin-x64", "npm:1.3.62"],\ + ["@swc/core-linux-arm-gnueabihf", "npm:1.3.62"],\ + ["@swc/core-linux-arm64-gnu", "npm:1.3.62"],\ + ["@swc/core-linux-arm64-musl", "npm:1.3.62"],\ + ["@swc/core-linux-x64-gnu", "npm:1.3.62"],\ + ["@swc/core-linux-x64-musl", "npm:1.3.62"],\ + ["@swc/core-win32-arm64-msvc", "npm:1.3.62"],\ + ["@swc/core-win32-ia32-msvc", "npm:1.3.62"],\ + ["@swc/core-win32-x64-msvc", "npm:1.3.62"],\ + ["@swc/helpers", null],\ + ["@types/swc__helpers", null]\ + ],\ + "packagePeers": [\ + "@swc/helpers",\ + "@types/swc__helpers"\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@swc/core-darwin-arm64", [\ + ["npm:1.3.62", {\ + "packageLocation": "./.yarn/unplugged/@swc-core-darwin-arm64-npm-1.3.62-b4af5d9b32/node_modules/@swc/core-darwin-arm64/",\ + "packageDependencies": [\ + ["@swc/core-darwin-arm64", "npm:1.3.62"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@swc/core-darwin-x64", [\ + ["npm:1.3.62", {\ + "packageLocation": "./.yarn/unplugged/@swc-core-darwin-x64-npm-1.3.62-7d7bc99502/node_modules/@swc/core-darwin-x64/",\ + "packageDependencies": [\ + ["@swc/core-darwin-x64", "npm:1.3.62"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@swc/core-linux-arm-gnueabihf", [\ + ["npm:1.3.62", {\ + "packageLocation": "./.yarn/unplugged/@swc-core-linux-arm-gnueabihf-npm-1.3.62-2528581a9c/node_modules/@swc/core-linux-arm-gnueabihf/",\ + "packageDependencies": [\ + ["@swc/core-linux-arm-gnueabihf", "npm:1.3.62"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@swc/core-linux-arm64-gnu", [\ + ["npm:1.3.62", {\ + "packageLocation": "./.yarn/unplugged/@swc-core-linux-arm64-gnu-npm-1.3.62-7b527a3356/node_modules/@swc/core-linux-arm64-gnu/",\ + "packageDependencies": [\ + ["@swc/core-linux-arm64-gnu", "npm:1.3.62"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@swc/core-linux-arm64-musl", [\ + ["npm:1.3.62", {\ + "packageLocation": "./.yarn/unplugged/@swc-core-linux-arm64-musl-npm-1.3.62-5faf35783f/node_modules/@swc/core-linux-arm64-musl/",\ + "packageDependencies": [\ + ["@swc/core-linux-arm64-musl", "npm:1.3.62"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@swc/core-linux-x64-gnu", [\ + ["npm:1.3.62", {\ + "packageLocation": "./.yarn/unplugged/@swc-core-linux-x64-gnu-npm-1.3.62-1fc43a8907/node_modules/@swc/core-linux-x64-gnu/",\ + "packageDependencies": [\ + ["@swc/core-linux-x64-gnu", "npm:1.3.62"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@swc/core-linux-x64-musl", [\ + ["npm:1.3.62", {\ + "packageLocation": "./.yarn/unplugged/@swc-core-linux-x64-musl-npm-1.3.62-ffabf9bf27/node_modules/@swc/core-linux-x64-musl/",\ + "packageDependencies": [\ + ["@swc/core-linux-x64-musl", "npm:1.3.62"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@swc/core-win32-arm64-msvc", [\ + ["npm:1.3.62", {\ + "packageLocation": "./.yarn/unplugged/@swc-core-win32-arm64-msvc-npm-1.3.62-f4199145ca/node_modules/@swc/core-win32-arm64-msvc/",\ + "packageDependencies": [\ + ["@swc/core-win32-arm64-msvc", "npm:1.3.62"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@swc/core-win32-ia32-msvc", [\ + ["npm:1.3.62", {\ + "packageLocation": "./.yarn/unplugged/@swc-core-win32-ia32-msvc-npm-1.3.62-56dc98262c/node_modules/@swc/core-win32-ia32-msvc/",\ + "packageDependencies": [\ + ["@swc/core-win32-ia32-msvc", "npm:1.3.62"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@swc/core-win32-x64-msvc", [\ + ["npm:1.3.62", {\ + "packageLocation": "./.yarn/unplugged/@swc-core-win32-x64-msvc-npm-1.3.62-200450bac0/node_modules/@swc/core-win32-x64-msvc/",\ + "packageDependencies": [\ + ["@swc/core-win32-x64-msvc", "npm:1.3.62"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["@swc/helpers", [\ - ["npm:0.4.14", {\ - "packageLocation": "./.yarn/cache/@swc-helpers-npm-0.4.14-f806c3fb16-273fd3f3fc.zip/node_modules/@swc/helpers/",\ + ["npm:0.5.1", {\ + "packageLocation": "./.yarn/cache/@swc-helpers-npm-0.5.1-424376f311-71e0e27234.zip/node_modules/@swc/helpers/",\ "packageDependencies": [\ - ["@swc/helpers", "npm:0.4.14"],\ + ["@swc/helpers", "npm:0.5.1"],\ ["tslib", "npm:2.4.0"]\ ],\ "linkType": "HARD"\ @@ -1954,17 +2625,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ - ["@types/jest", [\ - ["npm:27.4.1", {\ - "packageLocation": "./.yarn/cache/@types-jest-npm-27.4.1-31d07cd0d8-5184f3eef4.zip/node_modules/@types/jest/",\ - "packageDependencies": [\ - ["@types/jest", "npm:27.4.1"],\ - ["jest-matcher-utils", "npm:27.5.1"],\ - ["pretty-format", "npm:27.5.1"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ ["@types/json5", [\ ["npm:0.0.29", {\ "packageLocation": "./.yarn/cache/@types-json5-npm-0.0.29-f63a7916bd-e60b153664.zip/node_modules/@types/json5/",\ @@ -1975,10 +2635,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@types/katex", [\ - ["npm:0.14.0", {\ - "packageLocation": "./.yarn/cache/@types-katex-npm-0.14.0-acd5bc3e87-330e0d0337.zip/node_modules/@types/katex/",\ + ["npm:0.16.5", {\ + "packageLocation": "./.yarn/cache/@types-katex-npm-0.16.5-ff9336f176-a1ce22cd87.zip/node_modules/@types/katex/",\ "packageDependencies": [\ - ["@types/katex", "npm:0.14.0"]\ + ["@types/katex", "npm:0.16.5"]\ ],\ "linkType": "HARD"\ }]\ @@ -1990,13 +2650,20 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@types/lodash", "npm:4.14.182"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:4.14.200", {\ + "packageLocation": "./.yarn/cache/@types-lodash-npm-4.14.200-8559f51fce-6471f8bb5d.zip/node_modules/@types/lodash/",\ + "packageDependencies": [\ + ["@types/lodash", "npm:4.14.200"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["@types/lodash-es", [\ - ["npm:4.17.6", {\ - "packageLocation": "./.yarn/cache/@types-lodash-es-npm-4.17.6-fd5abbdc74-9bd239dd52.zip/node_modules/@types/lodash-es/",\ + ["npm:4.17.10", {\ + "packageLocation": "./.yarn/cache/@types-lodash-es-npm-4.17.10-a7dae21818-129e9dde83.zip/node_modules/@types/lodash-es/",\ "packageDependencies": [\ - ["@types/lodash-es", "npm:4.17.6"],\ + ["@types/lodash-es", "npm:4.17.10"],\ ["@types/lodash", "npm:4.14.182"]\ ],\ "linkType": "HARD"\ @@ -2020,22 +2687,31 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["@ungap/structured-clone", [\ + ["npm:1.2.0", {\ + "packageLocation": "./.yarn/cache/@ungap-structured-clone-npm-1.2.0-648f0b82e0-4f656b7b46.zip/node_modules/@ungap/structured-clone/",\ + "packageDependencies": [\ + ["@ungap/structured-clone", "npm:1.2.0"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["@vitejs/plugin-vue", [\ - ["npm:3.2.0", {\ - "packageLocation": "./.yarn/cache/@vitejs-plugin-vue-npm-3.2.0-d467fde943-64774f770e.zip/node_modules/@vitejs/plugin-vue/",\ + ["npm:4.6.2", {\ + "packageLocation": "./.yarn/cache/@vitejs-plugin-vue-npm-4.6.2-d7ace53203-01bc4ed643.zip/node_modules/@vitejs/plugin-vue/",\ "packageDependencies": [\ - ["@vitejs/plugin-vue", "npm:3.2.0"]\ + ["@vitejs/plugin-vue", "npm:4.6.2"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.2.0", {\ - "packageLocation": "./.yarn/__virtual__/@vitejs-plugin-vue-virtual-154906c651/0/cache/@vitejs-plugin-vue-npm-3.2.0-d467fde943-64774f770e.zip/node_modules/@vitejs/plugin-vue/",\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:4.6.2", {\ + "packageLocation": "./.yarn/__virtual__/@vitejs-plugin-vue-virtual-090b584a9c/0/cache/@vitejs-plugin-vue-npm-4.6.2-d7ace53203-01bc4ed643.zip/node_modules/@vitejs/plugin-vue/",\ "packageDependencies": [\ - ["@vitejs/plugin-vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.2.0"],\ + ["@vitejs/plugin-vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:4.6.2"],\ ["@types/vite", null],\ ["@types/vue", null],\ - ["vite", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.2.5"],\ - ["vue", "npm:3.2.45"]\ + ["vite", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:4.5.3"],\ + ["vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.21"]\ ],\ "packagePeers": [\ "@types/vite",\ @@ -2046,132 +2722,178 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["@volar/language-core", [\ + ["npm:2.1.4", {\ + "packageLocation": "./.yarn/cache/@volar-language-core-npm-2.1.4-18ee1a037d-7430f65143.zip/node_modules/@volar/language-core/",\ + "packageDependencies": [\ + ["@volar/language-core", "npm:2.1.4"],\ + ["@volar/source-map", "npm:2.1.4"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@volar/language-service", [\ + ["npm:2.1.4", {\ + "packageLocation": "./.yarn/cache/@volar-language-service-npm-2.1.4-2d34cb628f-06cdcfacf0.zip/node_modules/@volar/language-service/",\ + "packageDependencies": [\ + ["@volar/language-service", "npm:2.1.4"],\ + ["@volar/language-core", "npm:2.1.4"],\ + ["vscode-languageserver-protocol", "npm:3.17.5"],\ + ["vscode-languageserver-textdocument", "npm:1.0.11"],\ + ["vscode-uri", "npm:3.0.8"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@volar/source-map", [\ + ["npm:2.1.4", {\ + "packageLocation": "./.yarn/cache/@volar-source-map-npm-2.1.4-5963b1701f-e2f65bcfd6.zip/node_modules/@volar/source-map/",\ + "packageDependencies": [\ + ["@volar/source-map", "npm:2.1.4"],\ + ["muggle-string", "npm:0.4.1"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["@vscode/l10n", [\ + ["npm:0.0.18", {\ + "packageLocation": "./.yarn/cache/@vscode-l10n-npm-0.0.18-8a12efe4b5-c33876cebd.zip/node_modules/@vscode/l10n/",\ + "packageDependencies": [\ + ["@vscode/l10n", "npm:0.0.18"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["@vue/compiler-core", [\ - ["npm:3.2.45", {\ - "packageLocation": "./.yarn/cache/@vue-compiler-core-npm-3.2.45-2a68bebbd0-e3c687b24c.zip/node_modules/@vue/compiler-core/",\ + ["npm:3.4.21", {\ + "packageLocation": "./.yarn/cache/@vue-compiler-core-npm-3.4.21-ec7f24d7f5-0d6b7732bc.zip/node_modules/@vue/compiler-core/",\ "packageDependencies": [\ - ["@vue/compiler-core", "npm:3.2.45"],\ - ["@babel/parser", "npm:7.18.4"],\ - ["@vue/shared", "npm:3.2.45"],\ + ["@vue/compiler-core", "npm:3.4.21"],\ + ["@babel/parser", "npm:7.23.9"],\ + ["@vue/shared", "npm:3.4.21"],\ + ["entities", "npm:4.5.0"],\ ["estree-walker", "npm:2.0.2"],\ - ["source-map", "npm:0.6.1"]\ + ["source-map-js", "npm:1.0.2"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@vue/compiler-dom", [\ - ["npm:3.2.45", {\ - "packageLocation": "./.yarn/cache/@vue-compiler-dom-npm-3.2.45-e742186d0b-8911553863.zip/node_modules/@vue/compiler-dom/",\ + ["npm:3.4.21", {\ + "packageLocation": "./.yarn/cache/@vue-compiler-dom-npm-3.4.21-3d49f99020-f53e4f4e0a.zip/node_modules/@vue/compiler-dom/",\ "packageDependencies": [\ - ["@vue/compiler-dom", "npm:3.2.45"],\ - ["@vue/compiler-core", "npm:3.2.45"],\ - ["@vue/shared", "npm:3.2.45"]\ + ["@vue/compiler-dom", "npm:3.4.21"],\ + ["@vue/compiler-core", "npm:3.4.21"],\ + ["@vue/shared", "npm:3.4.21"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@vue/compiler-sfc", [\ - ["npm:3.2.45", {\ - "packageLocation": "./.yarn/cache/@vue-compiler-sfc-npm-3.2.45-f1fe8426df-bec375faa0.zip/node_modules/@vue/compiler-sfc/",\ - "packageDependencies": [\ - ["@vue/compiler-sfc", "npm:3.2.45"],\ - ["@babel/parser", "npm:7.18.4"],\ - ["@vue/compiler-core", "npm:3.2.45"],\ - ["@vue/compiler-dom", "npm:3.2.45"],\ - ["@vue/compiler-ssr", "npm:3.2.45"],\ - ["@vue/reactivity-transform", "npm:3.2.45"],\ - ["@vue/shared", "npm:3.2.45"],\ + ["npm:3.4.21", {\ + "packageLocation": "./.yarn/cache/@vue-compiler-sfc-npm-3.4.21-c2b76ee1ff-226dc404be.zip/node_modules/@vue/compiler-sfc/",\ + "packageDependencies": [\ + ["@vue/compiler-sfc", "npm:3.4.21"],\ + ["@babel/parser", "npm:7.23.9"],\ + ["@vue/compiler-core", "npm:3.4.21"],\ + ["@vue/compiler-dom", "npm:3.4.21"],\ + ["@vue/compiler-ssr", "npm:3.4.21"],\ + ["@vue/shared", "npm:3.4.21"],\ ["estree-walker", "npm:2.0.2"],\ - ["magic-string", "npm:0.25.9"],\ - ["postcss", "npm:8.4.12"],\ - ["source-map", "npm:0.6.1"]\ + ["magic-string", "npm:0.30.7"],\ + ["postcss", "npm:8.4.35"],\ + ["source-map-js", "npm:1.0.2"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@vue/compiler-ssr", [\ - ["npm:3.2.45", {\ - "packageLocation": "./.yarn/cache/@vue-compiler-ssr-npm-3.2.45-0b951e5028-830c475506.zip/node_modules/@vue/compiler-ssr/",\ + ["npm:3.4.21", {\ + "packageLocation": "./.yarn/cache/@vue-compiler-ssr-npm-3.4.21-e6f043341e-c510bee68b.zip/node_modules/@vue/compiler-ssr/",\ "packageDependencies": [\ - ["@vue/compiler-ssr", "npm:3.2.45"],\ - ["@vue/compiler-dom", "npm:3.2.45"],\ - ["@vue/shared", "npm:3.2.45"]\ + ["@vue/compiler-ssr", "npm:3.4.21"],\ + ["@vue/compiler-dom", "npm:3.4.21"],\ + ["@vue/shared", "npm:3.4.21"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@vue/devtools-api", [\ - ["npm:6.4.5", {\ - "packageLocation": "./.yarn/cache/@vue-devtools-api-npm-6.4.5-bcd56e5fec-40c5adc878.zip/node_modules/@vue/devtools-api/",\ + ["npm:6.5.0", {\ + "packageLocation": "./.yarn/cache/@vue-devtools-api-npm-6.5.0-0dc0468299-ec819ef3a4.zip/node_modules/@vue/devtools-api/",\ + "packageDependencies": [\ + ["@vue/devtools-api", "npm:6.5.0"]\ + ],\ + "linkType": "HARD"\ + }],\ + ["npm:6.6.1", {\ + "packageLocation": "./.yarn/cache/@vue-devtools-api-npm-6.6.1-ef3c82703e-cf12b5ebcc.zip/node_modules/@vue/devtools-api/",\ "packageDependencies": [\ - ["@vue/devtools-api", "npm:6.4.5"]\ + ["@vue/devtools-api", "npm:6.6.1"]\ ],\ "linkType": "HARD"\ }]\ ]],\ - ["@vue/reactivity", [\ - ["npm:3.2.45", {\ - "packageLocation": "./.yarn/cache/@vue-reactivity-npm-3.2.45-bc3378a52c-4ba609744a.zip/node_modules/@vue/reactivity/",\ + ["@vue/language-plugin-pug", [\ + ["npm:2.0.7", {\ + "packageLocation": "./.yarn/cache/@vue-language-plugin-pug-npm-2.0.7-547300c7e0-11cc96eb5f.zip/node_modules/@vue/language-plugin-pug/",\ "packageDependencies": [\ - ["@vue/reactivity", "npm:3.2.45"],\ - ["@vue/shared", "npm:3.2.45"]\ + ["@vue/language-plugin-pug", "npm:2.0.7"],\ + ["@volar/source-map", "npm:2.1.4"],\ + ["volar-service-pug", "npm:0.0.34"]\ ],\ "linkType": "HARD"\ }]\ ]],\ - ["@vue/reactivity-transform", [\ - ["npm:3.2.45", {\ - "packageLocation": "./.yarn/cache/@vue-reactivity-transform-npm-3.2.45-05914b9134-4010408189.zip/node_modules/@vue/reactivity-transform/",\ + ["@vue/reactivity", [\ + ["npm:3.4.21", {\ + "packageLocation": "./.yarn/cache/@vue-reactivity-npm-3.4.21-fd3e254d08-79c7ebe3ec.zip/node_modules/@vue/reactivity/",\ "packageDependencies": [\ - ["@vue/reactivity-transform", "npm:3.2.45"],\ - ["@babel/parser", "npm:7.18.4"],\ - ["@vue/compiler-core", "npm:3.2.45"],\ - ["@vue/shared", "npm:3.2.45"],\ - ["estree-walker", "npm:2.0.2"],\ - ["magic-string", "npm:0.25.9"]\ + ["@vue/reactivity", "npm:3.4.21"],\ + ["@vue/shared", "npm:3.4.21"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@vue/runtime-core", [\ - ["npm:3.2.45", {\ - "packageLocation": "./.yarn/cache/@vue-runtime-core-npm-3.2.45-084482e779-0ac376a760.zip/node_modules/@vue/runtime-core/",\ + ["npm:3.4.21", {\ + "packageLocation": "./.yarn/cache/@vue-runtime-core-npm-3.4.21-7bf985040b-4eb9b5d91f.zip/node_modules/@vue/runtime-core/",\ "packageDependencies": [\ - ["@vue/runtime-core", "npm:3.2.45"],\ - ["@vue/reactivity", "npm:3.2.45"],\ - ["@vue/shared", "npm:3.2.45"]\ + ["@vue/runtime-core", "npm:3.4.21"],\ + ["@vue/reactivity", "npm:3.4.21"],\ + ["@vue/shared", "npm:3.4.21"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@vue/runtime-dom", [\ - ["npm:3.2.45", {\ - "packageLocation": "./.yarn/cache/@vue-runtime-dom-npm-3.2.45-6ab018299f-c66c71a2fc.zip/node_modules/@vue/runtime-dom/",\ + ["npm:3.4.21", {\ + "packageLocation": "./.yarn/cache/@vue-runtime-dom-npm-3.4.21-40f99cf9a2-ebfdaa081f.zip/node_modules/@vue/runtime-dom/",\ "packageDependencies": [\ - ["@vue/runtime-dom", "npm:3.2.45"],\ - ["@vue/runtime-core", "npm:3.2.45"],\ - ["@vue/shared", "npm:3.2.45"],\ - ["csstype", "npm:2.6.20"]\ + ["@vue/runtime-dom", "npm:3.4.21"],\ + ["@vue/runtime-core", "npm:3.4.21"],\ + ["@vue/shared", "npm:3.4.21"],\ + ["csstype", "npm:3.1.3"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["@vue/server-renderer", [\ - ["npm:3.2.45", {\ - "packageLocation": "./.yarn/cache/@vue-server-renderer-npm-3.2.45-dbee798520-062812235c.zip/node_modules/@vue/server-renderer/",\ + ["npm:3.4.21", {\ + "packageLocation": "./.yarn/cache/@vue-server-renderer-npm-3.4.21-bf6b2daebb-faa3dc4876.zip/node_modules/@vue/server-renderer/",\ "packageDependencies": [\ - ["@vue/server-renderer", "npm:3.2.45"]\ + ["@vue/server-renderer", "npm:3.4.21"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:06b4b60efe017e0fdf8405fef0e04a64d7dc86cf13e04f05fb681889d996c75efd151033b2e012c72dfb9a0b8d1a647b6d3a8115078891aebe2fa1a4e81f50bf#npm:3.2.45", {\ - "packageLocation": "./.yarn/__virtual__/@vue-server-renderer-virtual-8298a8cec7/0/cache/@vue-server-renderer-npm-3.2.45-dbee798520-062812235c.zip/node_modules/@vue/server-renderer/",\ + ["virtual:b79af6274dddda2b283f42be2b827e30c3e5389bce2938ee73bdb74ee9781811fc079c6836719e57940708d59b3beeb14d9e3c12f37f2d22582a53e6c32e4c97#npm:3.4.21", {\ + "packageLocation": "./.yarn/__virtual__/@vue-server-renderer-virtual-4c61378d94/0/cache/@vue-server-renderer-npm-3.4.21-bf6b2daebb-faa3dc4876.zip/node_modules/@vue/server-renderer/",\ "packageDependencies": [\ - ["@vue/server-renderer", "virtual:06b4b60efe017e0fdf8405fef0e04a64d7dc86cf13e04f05fb681889d996c75efd151033b2e012c72dfb9a0b8d1a647b6d3a8115078891aebe2fa1a4e81f50bf#npm:3.2.45"],\ + ["@vue/server-renderer", "virtual:b79af6274dddda2b283f42be2b827e30c3e5389bce2938ee73bdb74ee9781811fc079c6836719e57940708d59b3beeb14d9e3c12f37f2d22582a53e6c32e4c97#npm:3.4.21"],\ ["@types/vue", null],\ - ["@vue/compiler-ssr", "npm:3.2.45"],\ - ["@vue/shared", "npm:3.2.45"],\ - ["vue", "npm:3.2.45"]\ + ["@vue/compiler-ssr", "npm:3.4.21"],\ + ["@vue/shared", "npm:3.4.21"],\ + ["vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.21"]\ ],\ "packagePeers": [\ "@types/vue",\ @@ -2181,10 +2903,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["@vue/shared", [\ - ["npm:3.2.45", {\ - "packageLocation": "./.yarn/cache/@vue-shared-npm-3.2.45-1855c9c551-ff3205056c.zip/node_modules/@vue/shared/",\ + ["npm:3.4.21", {\ + "packageLocation": "./.yarn/cache/@vue-shared-npm-3.4.21-2aee4ae0bc-5f30a40891.zip/node_modules/@vue/shared/",\ "packageDependencies": [\ - ["@vue/shared", "npm:3.2.45"]\ + ["@vue/shared", "npm:3.4.21"]\ ],\ "linkType": "HARD"\ }]\ @@ -2215,17 +2937,17 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "HARD"\ }],\ - ["npm:8.7.1", {\ - "packageLocation": "./.yarn/cache/acorn-npm-8.7.1-7c7a019990-aca0aabf98.zip/node_modules/acorn/",\ + ["npm:8.10.0", {\ + "packageLocation": "./.yarn/cache/acorn-npm-8.10.0-2230c9e83e-538ba38af0.zip/node_modules/acorn/",\ "packageDependencies": [\ - ["acorn", "npm:8.7.1"]\ + ["acorn", "npm:8.10.0"]\ ],\ "linkType": "HARD"\ }],\ - ["npm:8.8.0", {\ - "packageLocation": "./.yarn/cache/acorn-npm-8.8.0-9ef399ab45-7270ca82b2.zip/node_modules/acorn/",\ + ["npm:8.7.1", {\ + "packageLocation": "./.yarn/cache/acorn-npm-8.7.1-7c7a019990-aca0aabf98.zip/node_modules/acorn/",\ "packageDependencies": [\ - ["acorn", "npm:8.8.0"]\ + ["acorn", "npm:8.7.1"]\ ],\ "linkType": "HARD"\ }]\ @@ -2238,12 +2960,12 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "SOFT"\ }],\ - ["virtual:0371ef3614c1182c8fcb05e5954a5fd9c124be4b821bd43f5ef7bdb1bf9603ab5ec16aa2a2a1de93fa397b424774f98f00c5e66d99eec56be0bd9f2a1ab2c75f#npm:5.3.2", {\ - "packageLocation": "./.yarn/__virtual__/acorn-jsx-virtual-c72fe04f43/0/cache/acorn-jsx-npm-5.3.2-d7594599ea-c3d3b2a89c.zip/node_modules/acorn-jsx/",\ + ["virtual:a50722a5a9326b6a5f12350c494c4db3aa0f4caeac45e3e9e5fe071da20014ecfe738fe2ebe2c9c98abae81a4ea86b42f56d776b3bd5ec37f9ad3670c242b242#npm:5.3.2", {\ + "packageLocation": "./.yarn/__virtual__/acorn-jsx-virtual-834321b202/0/cache/acorn-jsx-npm-5.3.2-d7594599ea-c3d3b2a89c.zip/node_modules/acorn-jsx/",\ "packageDependencies": [\ - ["acorn-jsx", "virtual:0371ef3614c1182c8fcb05e5954a5fd9c124be4b821bd43f5ef7bdb1bf9603ab5ec16aa2a2a1de93fa397b424774f98f00c5e66d99eec56be0bd9f2a1ab2c75f#npm:5.3.2"],\ + ["acorn-jsx", "virtual:a50722a5a9326b6a5f12350c494c4db3aa0f4caeac45e3e9e5fe071da20014ecfe738fe2ebe2c9c98abae81a4ea86b42f56d776b3bd5ec37f9ad3670c242b242#npm:5.3.2"],\ ["@types/acorn", null],\ - ["acorn", "npm:8.8.0"]\ + ["acorn", "npm:8.10.0"]\ ],\ "packagePeers": [\ "@types/acorn",\ @@ -2265,15 +2987,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ - ["acorn-walk", [\ - ["npm:8.2.0", {\ - "packageLocation": "./.yarn/cache/acorn-walk-npm-8.2.0-2f2cac3177-1715e76c01.zip/node_modules/acorn-walk/",\ - "packageDependencies": [\ - ["acorn-walk", "npm:8.2.0"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ ["agent-base", [\ ["npm:6.0.2", {\ "packageLocation": "./.yarn/cache/agent-base-npm-6.0.2-428f325a93-f52b6872cc.zip/node_modules/agent-base/",\ @@ -2338,6 +3051,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["ansi-regex", "npm:5.0.1"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:6.0.1", {\ + "packageLocation": "./.yarn/cache/ansi-regex-npm-6.0.1-8d663a607d-1ff8b7667c.zip/node_modules/ansi-regex/",\ + "packageDependencies": [\ + ["ansi-regex", "npm:6.0.1"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["ansi-styles", [\ @@ -2357,10 +3077,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "HARD"\ }],\ - ["npm:5.2.0", {\ - "packageLocation": "./.yarn/cache/ansi-styles-npm-5.2.0-72fc7003e3-d7f4e97ce0.zip/node_modules/ansi-styles/",\ + ["npm:6.2.1", {\ + "packageLocation": "./.yarn/cache/ansi-styles-npm-6.2.1-d43647018c-ef940f2f0c.zip/node_modules/ansi-styles/",\ "packageDependencies": [\ - ["ansi-styles", "npm:5.2.0"]\ + ["ansi-styles", "npm:6.2.1"]\ ],\ "linkType": "HARD"\ }]\ @@ -2405,33 +3125,87 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["array-buffer-byte-length", [\ + ["npm:1.0.0", {\ + "packageLocation": "./.yarn/cache/array-buffer-byte-length-npm-1.0.0-331671f28a-044e101ce1.zip/node_modules/array-buffer-byte-length/",\ + "packageDependencies": [\ + ["array-buffer-byte-length", "npm:1.0.0"],\ + ["call-bind", "npm:1.0.2"],\ + ["is-array-buffer", "npm:3.0.1"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["array-includes", [\ - ["npm:3.1.4", {\ - "packageLocation": "./.yarn/cache/array-includes-npm-3.1.4-79bb883109-69967c38c5.zip/node_modules/array-includes/",\ + ["npm:3.1.7", {\ + "packageLocation": "./.yarn/cache/array-includes-npm-3.1.7-d32a5ee179-06f9e4598f.zip/node_modules/array-includes/",\ "packageDependencies": [\ - ["array-includes", "npm:3.1.4"],\ + ["array-includes", "npm:3.1.7"],\ ["call-bind", "npm:1.0.2"],\ - ["define-properties", "npm:1.1.4"],\ - ["es-abstract", "npm:1.19.5"],\ - ["get-intrinsic", "npm:1.1.1"],\ + ["define-properties", "npm:1.2.0"],\ + ["es-abstract", "npm:1.22.3"],\ + ["get-intrinsic", "npm:1.2.1"],\ ["is-string", "npm:1.0.7"]\ ],\ "linkType": "HARD"\ }]\ ]],\ + ["array.prototype.findlastindex", [\ + ["npm:1.2.3", {\ + "packageLocation": "./.yarn/cache/array.prototype.findlastindex-npm-1.2.3-2a36f4417b-31f35d7b37.zip/node_modules/array.prototype.findlastindex/",\ + "packageDependencies": [\ + ["array.prototype.findlastindex", "npm:1.2.3"],\ + ["call-bind", "npm:1.0.2"],\ + ["define-properties", "npm:1.2.0"],\ + ["es-abstract", "npm:1.22.3"],\ + ["es-shim-unscopables", "npm:1.0.0"],\ + ["get-intrinsic", "npm:1.2.1"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["array.prototype.flat", [\ - ["npm:1.3.0", {\ - "packageLocation": "./.yarn/cache/array.prototype.flat-npm-1.3.0-6c5c4292bd-2a652b3e8d.zip/node_modules/array.prototype.flat/",\ + ["npm:1.3.2", {\ + "packageLocation": "./.yarn/cache/array.prototype.flat-npm-1.3.2-350729f7f4-5d6b4bf102.zip/node_modules/array.prototype.flat/",\ "packageDependencies": [\ - ["array.prototype.flat", "npm:1.3.0"],\ + ["array.prototype.flat", "npm:1.3.2"],\ ["call-bind", "npm:1.0.2"],\ - ["define-properties", "npm:1.1.4"],\ - ["es-abstract", "npm:1.19.5"],\ + ["define-properties", "npm:1.2.0"],\ + ["es-abstract", "npm:1.22.3"],\ + ["es-shim-unscopables", "npm:1.0.0"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["array.prototype.flatmap", [\ + ["npm:1.3.2", {\ + "packageLocation": "./.yarn/cache/array.prototype.flatmap-npm-1.3.2-5c6a4af226-ce09fe21dc.zip/node_modules/array.prototype.flatmap/",\ + "packageDependencies": [\ + ["array.prototype.flatmap", "npm:1.3.2"],\ + ["call-bind", "npm:1.0.2"],\ + ["define-properties", "npm:1.2.0"],\ + ["es-abstract", "npm:1.22.3"],\ ["es-shim-unscopables", "npm:1.0.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ + ["arraybuffer.prototype.slice", [\ + ["npm:1.0.2", {\ + "packageLocation": "./.yarn/cache/arraybuffer.prototype.slice-npm-1.0.2-4eda52ad8c-c200faf437.zip/node_modules/arraybuffer.prototype.slice/",\ + "packageDependencies": [\ + ["arraybuffer.prototype.slice", "npm:1.0.2"],\ + ["array-buffer-byte-length", "npm:1.0.0"],\ + ["call-bind", "npm:1.0.2"],\ + ["define-properties", "npm:1.2.0"],\ + ["es-abstract", "npm:1.22.3"],\ + ["get-intrinsic", "npm:1.2.1"],\ + ["is-array-buffer", "npm:3.0.2"],\ + ["is-shared-array-buffer", "npm:1.0.2"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["asap", [\ ["npm:2.0.6", {\ "packageLocation": "./.yarn/cache/asap-npm-2.0.6-36714d439d-b296c92c4b.zip/node_modules/asap/",\ @@ -2451,10 +3225,19 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["async-validator", [\ - ["npm:4.1.1", {\ - "packageLocation": "./.yarn/cache/async-validator-npm-4.1.1-470b8d5b59-88590ab8ad.zip/node_modules/async-validator/",\ + ["npm:4.2.5", {\ + "packageLocation": "./.yarn/cache/async-validator-npm-4.2.5-4d61110c66-3e3d891a2e.zip/node_modules/async-validator/",\ + "packageDependencies": [\ + ["async-validator", "npm:4.2.5"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["available-typed-arrays", [\ + ["npm:1.0.5", {\ + "packageLocation": "./.yarn/cache/available-typed-arrays-npm-1.0.5-88f321e4d3-20eb47b3ce.zip/node_modules/available-typed-arrays/",\ "packageDependencies": [\ - ["async-validator", "npm:4.1.1"]\ + ["available-typed-arrays", "npm:1.0.5"]\ ],\ "linkType": "HARD"\ }]\ @@ -2514,10 +3297,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "SOFT"\ }],\ - ["npm:5.2.3", {\ - "packageLocation": "./.yarn/cache/bootstrap-npm-5.2.3-7458283a23-0211805dec.zip/node_modules/bootstrap/",\ + ["npm:5.3.3", {\ + "packageLocation": "./.yarn/cache/bootstrap-npm-5.3.3-da08e2f0fe-537b68db30.zip/node_modules/bootstrap/",\ "packageDependencies": [\ - ["bootstrap", "npm:5.2.3"]\ + ["bootstrap", "npm:5.3.3"]\ ],\ "linkType": "SOFT"\ }],\ @@ -2534,11 +3317,11 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "HARD"\ }],\ - ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.2.3", {\ - "packageLocation": "./.yarn/__virtual__/bootstrap-virtual-c4952ffff0/0/cache/bootstrap-npm-5.2.3-7458283a23-0211805dec.zip/node_modules/bootstrap/",\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.3.3", {\ + "packageLocation": "./.yarn/__virtual__/bootstrap-virtual-2c24090b13/0/cache/bootstrap-npm-5.3.3-da08e2f0fe-537b68db30.zip/node_modules/bootstrap/",\ "packageDependencies": [\ - ["bootstrap", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.2.3"],\ - ["@popperjs/core", "npm:2.11.6"],\ + ["bootstrap", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.3.3"],\ + ["@popperjs/core", "npm:2.11.8"],\ ["@types/popperjs__core", null]\ ],\ "packagePeers": [\ @@ -2549,10 +3332,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["bootstrap-icons", [\ - ["npm:1.10.3", {\ - "packageLocation": "./.yarn/cache/bootstrap-icons-npm-1.10.3-d595a95ca4-bef7f83ce0.zip/node_modules/bootstrap-icons/",\ + ["npm:1.11.3", {\ + "packageLocation": "./.yarn/cache/bootstrap-icons-npm-1.11.3-8d5387bef2-d5cdb90fe3.zip/node_modules/bootstrap-icons/",\ "packageDependencies": [\ - ["bootstrap-icons", "npm:1.10.3"]\ + ["bootstrap-icons", "npm:1.11.3"]\ ],\ "linkType": "HARD"\ }]\ @@ -2587,10 +3370,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["browser-fs-access", [\ - ["npm:0.31.1", {\ - "packageLocation": "./.yarn/cache/browser-fs-access-npm-0.31.1-c276b62f78-4a5c88839e.zip/node_modules/browser-fs-access/",\ + ["npm:0.35.0", {\ + "packageLocation": "./.yarn/cache/browser-fs-access-npm-0.35.0-1577b5a7ba-5f3bf1ec17.zip/node_modules/browser-fs-access/",\ "packageDependencies": [\ - ["browser-fs-access", "npm:0.31.1"]\ + ["browser-fs-access", "npm:0.35.0"]\ ],\ "linkType": "HARD"\ }]\ @@ -2619,11 +3402,11 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ - ["buffer-from", [\ - ["npm:1.1.2", {\ - "packageLocation": "./.yarn/cache/buffer-from-npm-1.1.2-03d2f20d7e-0448524a56.zip/node_modules/buffer-from/",\ + ["builtin-modules", [\ + ["npm:3.3.0", {\ + "packageLocation": "./.yarn/cache/builtin-modules-npm-3.3.0-db4f3d32de-db021755d7.zip/node_modules/builtin-modules/",\ "packageDependencies": [\ - ["buffer-from", "npm:1.1.2"]\ + ["builtin-modules", "npm:3.3.0"]\ ],\ "linkType": "HARD"\ }]\ @@ -2639,22 +3422,21 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["c8", [\ - ["npm:7.12.0", {\ - "packageLocation": "./.yarn/cache/c8-npm-7.12.0-c808cac509-3b7fa9ad7c.zip/node_modules/c8/",\ + ["npm:9.1.0", {\ + "packageLocation": "./.yarn/cache/c8-npm-9.1.0-92c3d37f46-c5249bf9c3.zip/node_modules/c8/",\ "packageDependencies": [\ - ["c8", "npm:7.12.0"],\ + ["c8", "npm:9.1.0"],\ ["@bcoe/v8-coverage", "npm:0.2.3"],\ ["@istanbuljs/schema", "npm:0.1.3"],\ ["find-up", "npm:5.0.0"],\ - ["foreground-child", "npm:2.0.0"],\ + ["foreground-child", "npm:3.1.1"],\ ["istanbul-lib-coverage", "npm:3.2.0"],\ - ["istanbul-lib-report", "npm:3.0.0"],\ - ["istanbul-reports", "npm:3.1.4"],\ - ["rimraf", "npm:3.0.2"],\ + ["istanbul-lib-report", "npm:3.0.1"],\ + ["istanbul-reports", "npm:3.1.6"],\ ["test-exclude", "npm:6.0.0"],\ ["v8-to-istanbul", "npm:9.0.1"],\ - ["yargs", "npm:16.2.0"],\ - ["yargs-parser", "npm:20.2.9"]\ + ["yargs", "npm:17.7.2"],\ + ["yargs-parser", "npm:21.1.1"]\ ],\ "linkType": "HARD"\ }]\ @@ -2695,6 +3477,16 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["get-intrinsic", "npm:1.1.1"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:1.0.5", {\ + "packageLocation": "./.yarn/cache/call-bind-npm-1.0.5-65600fae47-449e83ecbd.zip/node_modules/call-bind/",\ + "packageDependencies": [\ + ["call-bind", "npm:1.0.5"],\ + ["function-bind", "npm:1.1.2"],\ + ["get-intrinsic", "npm:1.2.1"],\ + ["set-function-length", "npm:1.1.1"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["callsites", [\ @@ -2714,10 +3506,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "HARD"\ }],\ - ["npm:1.0.30001442", {\ - "packageLocation": "./.yarn/cache/caniuse-lite-npm-1.0.30001442-4206643829-c1bff65bd4.zip/node_modules/caniuse-lite/",\ + ["npm:1.0.30001603", {\ + "packageLocation": "./.yarn/cache/caniuse-lite-npm-1.0.30001603-77af81f60b-e66e0d24b8.zip/node_modules/caniuse-lite/",\ "packageDependencies": [\ - ["caniuse-lite", "npm:1.0.30001442"]\ + ["caniuse-lite", "npm:1.0.30001603"]\ ],\ "linkType": "HARD"\ }]\ @@ -2798,10 +3590,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["cliui", [\ - ["npm:7.0.4", {\ - "packageLocation": "./.yarn/cache/cliui-npm-7.0.4-d6b8a9edb6-ce2e8f578a.zip/node_modules/cliui/",\ + ["npm:8.0.1", {\ + "packageLocation": "./.yarn/cache/cliui-npm-8.0.1-3b029092cf-79648b3b00.zip/node_modules/cliui/",\ "packageDependencies": [\ - ["cliui", "npm:7.0.4"],\ + ["cliui", "npm:8.0.1"],\ ["string-width", "npm:4.2.3"],\ ["strip-ansi", "npm:6.0.1"],\ ["wrap-ansi", "npm:7.0.0"]\ @@ -2862,13 +3654,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["commander", [\ - ["npm:2.20.3", {\ - "packageLocation": "./.yarn/cache/commander-npm-2.20.3-d8dcbaa39b-ab8c07884e.zip/node_modules/commander/",\ - "packageDependencies": [\ - ["commander", "npm:2.20.3"]\ - ],\ - "linkType": "HARD"\ - }],\ ["npm:7.2.0", {\ "packageLocation": "./.yarn/cache/commander-npm-7.2.0-19178180f8-53501cbeee.zip/node_modules/commander/",\ "packageDependencies": [\ @@ -2952,6 +3737,15 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["csstype", "npm:3.0.11"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:0.15.12", {\ + "packageLocation": "./.yarn/cache/css-render-npm-0.15.12-ff93ab2bdd-80265c5055.zip/node_modules/css-render/",\ + "packageDependencies": [\ + ["css-render", "npm:0.15.12"],\ + ["@emotion/hash", "npm:0.8.0"],\ + ["csstype", "npm:3.0.11"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["css-select", [\ @@ -3008,26 +3802,26 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["csstype", [\ - ["npm:2.6.20", {\ - "packageLocation": "./.yarn/cache/csstype-npm-2.6.20-7c929732a1-cb5d5ded49.zip/node_modules/csstype/",\ + ["npm:3.0.11", {\ + "packageLocation": "./.yarn/cache/csstype-npm-3.0.11-b49897178d-95e56abfe9.zip/node_modules/csstype/",\ "packageDependencies": [\ - ["csstype", "npm:2.6.20"]\ + ["csstype", "npm:3.0.11"]\ ],\ "linkType": "HARD"\ }],\ - ["npm:3.0.11", {\ - "packageLocation": "./.yarn/cache/csstype-npm-3.0.11-b49897178d-95e56abfe9.zip/node_modules/csstype/",\ + ["npm:3.1.3", {\ + "packageLocation": "./.yarn/cache/csstype-npm-3.1.3-e9a1c85013-8db785cc92.zip/node_modules/csstype/",\ "packageDependencies": [\ - ["csstype", "npm:3.0.11"]\ + ["csstype", "npm:3.1.3"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["d3", [\ - ["npm:7.8.0", {\ - "packageLocation": "./.yarn/cache/d3-npm-7.8.0-43f19bbccd-383d2c8aa6.zip/node_modules/d3/",\ + ["npm:7.9.0", {\ + "packageLocation": "./.yarn/cache/d3-npm-7.9.0-d293821ce6-1c0e9135f1.zip/node_modules/d3/",\ "packageDependencies": [\ - ["d3", "npm:7.8.0"],\ + ["d3", "npm:7.9.0"],\ ["d3-array", "npm:3.1.6"],\ ["d3-axis", "npm:3.0.0"],\ ["d3-brush", "npm:3.0.0"],\ @@ -3394,28 +4188,29 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["date-fns", [\ - ["npm:2.28.0", {\ - "packageLocation": "./.yarn/cache/date-fns-npm-2.28.0-c19c5add1b-a0516b2e4f.zip/node_modules/date-fns/",\ + ["npm:2.30.0", {\ + "packageLocation": "./.yarn/cache/date-fns-npm-2.30.0-895c790e0f-f7be015232.zip/node_modules/date-fns/",\ "packageDependencies": [\ - ["date-fns", "npm:2.28.0"]\ + ["date-fns", "npm:2.30.0"],\ + ["@babel/runtime", "npm:7.23.2"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["date-fns-tz", [\ - ["npm:1.3.3", {\ - "packageLocation": "./.yarn/cache/date-fns-tz-npm-1.3.3-4b42de3dcf-52111dffb4.zip/node_modules/date-fns-tz/",\ + ["npm:2.0.0", {\ + "packageLocation": "./.yarn/cache/date-fns-tz-npm-2.0.0-9b7996f292-a6553603a9.zip/node_modules/date-fns-tz/",\ "packageDependencies": [\ - ["date-fns-tz", "npm:1.3.3"]\ + ["date-fns-tz", "npm:2.0.0"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:1.3.3", {\ - "packageLocation": "./.yarn/__virtual__/date-fns-tz-virtual-ec6b2c4cf2/0/cache/date-fns-tz-npm-1.3.3-4b42de3dcf-52111dffb4.zip/node_modules/date-fns-tz/",\ + ["virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:2.0.0", {\ + "packageLocation": "./.yarn/__virtual__/date-fns-tz-virtual-6610d5adee/0/cache/date-fns-tz-npm-2.0.0-9b7996f292-a6553603a9.zip/node_modules/date-fns-tz/",\ "packageDependencies": [\ - ["date-fns-tz", "virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:1.3.3"],\ + ["date-fns-tz", "virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:2.0.0"],\ ["@types/date-fns", null],\ - ["date-fns", "npm:2.28.0"]\ + ["date-fns", "npm:2.30.0"]\ ],\ "packagePeers": [\ "@types/date-fns",\ @@ -3446,10 +4241,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "SOFT"\ }],\ - ["virtual:b86a9fb34323a98c6519528ed55faa0d9b44ca8879307c0b29aa384bde47ff59a7d0c9051b31246f14521dfb71ba3c5d6d0b35c29fffc17bf875aa6ad977d9e8#npm:4.3.4", {\ - "packageLocation": "./.yarn/__virtual__/debug-virtual-4488998e89/0/cache/debug-npm-4.3.4-4513954577-3dbad3f94e.zip/node_modules/debug/",\ + ["virtual:2a426afc4b2eef43db12a540d29c2b5476640459bfcd5c24f86bb401cf8cce97e63bd81794d206a5643057e7f662643afd5ce3dfc4d4bfd8e706006c6309c5fa#npm:3.2.7", {\ + "packageLocation": "./.yarn/__virtual__/debug-virtual-d2345003b7/0/cache/debug-npm-3.2.7-754e818c7a-b3d8c59407.zip/node_modules/debug/",\ "packageDependencies": [\ - ["debug", "virtual:b86a9fb34323a98c6519528ed55faa0d9b44ca8879307c0b29aa384bde47ff59a7d0c9051b31246f14521dfb71ba3c5d6d0b35c29fffc17bf875aa6ad977d9e8#npm:4.3.4"],\ + ["debug", "virtual:2a426afc4b2eef43db12a540d29c2b5476640459bfcd5c24f86bb401cf8cce97e63bd81794d206a5643057e7f662643afd5ce3dfc4d4bfd8e706006c6309c5fa#npm:3.2.7"],\ ["@types/supports-color", null],\ ["ms", "npm:2.1.2"],\ ["supports-color", null]\ @@ -3460,10 +4255,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "HARD"\ }],\ - ["virtual:d9426786c635bc4b52511d6cc4b56156f50d780a698c0e20fc6caf10d3be51cbf176e79cff882f4d42a23ff4d0f89fe94222849578214e7fbae0f2754c82af02#npm:3.2.7", {\ - "packageLocation": "./.yarn/__virtual__/debug-virtual-b810fb6338/0/cache/debug-npm-3.2.7-754e818c7a-b3d8c59407.zip/node_modules/debug/",\ + ["virtual:b86a9fb34323a98c6519528ed55faa0d9b44ca8879307c0b29aa384bde47ff59a7d0c9051b31246f14521dfb71ba3c5d6d0b35c29fffc17bf875aa6ad977d9e8#npm:4.3.4", {\ + "packageLocation": "./.yarn/__virtual__/debug-virtual-4488998e89/0/cache/debug-npm-4.3.4-4513954577-3dbad3f94e.zip/node_modules/debug/",\ "packageDependencies": [\ - ["debug", "virtual:d9426786c635bc4b52511d6cc4b56156f50d780a698c0e20fc6caf10d3be51cbf176e79cff882f4d42a23ff4d0f89fe94222849578214e7fbae0f2754c82af02#npm:3.2.7"],\ + ["debug", "virtual:b86a9fb34323a98c6519528ed55faa0d9b44ca8879307c0b29aa384bde47ff59a7d0c9051b31246f14521dfb71ba3c5d6d0b35c29fffc17bf875aa6ad977d9e8#npm:4.3.4"],\ ["@types/supports-color", null],\ ["ms", "npm:2.1.2"],\ ["supports-color", null]\ @@ -3499,10 +4294,22 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["deepmerge", [\ - ["npm:4.2.2", {\ - "packageLocation": "./.yarn/cache/deepmerge-npm-4.2.2-112165ced2-a8c43a1ed8.zip/node_modules/deepmerge/",\ + ["npm:4.3.1", {\ + "packageLocation": "./.yarn/cache/deepmerge-npm-4.3.1-4f751a0844-2024c6a980.zip/node_modules/deepmerge/",\ + "packageDependencies": [\ + ["deepmerge", "npm:4.3.1"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["define-data-property", [\ + ["npm:1.1.1", {\ + "packageLocation": "./.yarn/cache/define-data-property-npm-1.1.1-2b5156d112-a29855ad3f.zip/node_modules/define-data-property/",\ "packageDependencies": [\ - ["deepmerge", "npm:4.2.2"]\ + ["define-data-property", "npm:1.1.1"],\ + ["get-intrinsic", "npm:1.2.1"],\ + ["gopd", "npm:1.0.1"],\ + ["has-property-descriptors", "npm:1.0.0"]\ ],\ "linkType": "HARD"\ }]\ @@ -3516,6 +4323,15 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["object-keys", "npm:1.1.1"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:1.2.0", {\ + "packageLocation": "./.yarn/cache/define-properties-npm-1.2.0-3547cd0fd2-e60aee6a19.zip/node_modules/define-properties/",\ + "packageDependencies": [\ + ["define-properties", "npm:1.2.0"],\ + ["has-property-descriptors", "npm:1.0.0"],\ + ["object-keys", "npm:1.1.1"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["delaunator", [\ @@ -3569,13 +4385,11 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["detect-libc", "npm:1.0.3"]\ ],\ "linkType": "HARD"\ - }]\ - ]],\ - ["diff-sequences", [\ - ["npm:27.5.1", {\ - "packageLocation": "./.yarn/cache/diff-sequences-npm-27.5.1-29338362fa-a00db5554c.zip/node_modules/diff-sequences/",\ + }],\ + ["npm:2.0.2", {\ + "packageLocation": "./.yarn/cache/detect-libc-npm-2.0.2-03afa59137-2b2cd3649b.zip/node_modules/detect-libc/",\ "packageDependencies": [\ - ["diff-sequences", "npm:27.5.1"]\ + ["detect-libc", "npm:2.0.2"]\ ],\ "linkType": "HARD"\ }]\ @@ -3668,6 +4482,15 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["eastasianwidth", [\ + ["npm:0.2.0", {\ + "packageLocation": "./.yarn/cache/eastasianwidth-npm-0.2.0-c37eb16bd1-7d00d7cd8e.zip/node_modules/eastasianwidth/",\ + "packageDependencies": [\ + ["eastasianwidth", "npm:0.2.0"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["ee-first", [\ ["npm:1.1.1", {\ "packageLocation": "./.yarn/cache/ee-first-npm-1.1.1-33f8535b39-1b4cac778d.zip/node_modules/ee-first/",\ @@ -3693,6 +4516,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["emoji-regex", "npm:8.0.0"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:9.2.2", {\ + "packageLocation": "./.yarn/cache/emoji-regex-npm-9.2.2-e6fac8d058-8487182da7.zip/node_modules/emoji-regex/",\ + "packageDependencies": [\ + ["emoji-regex", "npm:9.2.2"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["encodeurl", [\ @@ -3715,307 +4545,165 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["entities", [\ - ["npm:2.2.0", {\ - "packageLocation": "./.yarn/cache/entities-npm-2.2.0-0fc8d5b2f7-19010dacaf.zip/node_modules/entities/",\ - "packageDependencies": [\ - ["entities", "npm:2.2.0"]\ - ],\ - "linkType": "HARD"\ - }],\ - ["npm:3.0.1", {\ - "packageLocation": "./.yarn/cache/entities-npm-3.0.1-21eeb201ba-aaf7f12033.zip/node_modules/entities/",\ - "packageDependencies": [\ - ["entities", "npm:3.0.1"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["env-paths", [\ - ["npm:2.2.1", {\ - "packageLocation": "./.yarn/cache/env-paths-npm-2.2.1-7c7577428c-65b5df55a8.zip/node_modules/env-paths/",\ - "packageDependencies": [\ - ["env-paths", "npm:2.2.1"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["err-code", [\ - ["npm:2.0.3", {\ - "packageLocation": "./.yarn/cache/err-code-npm-2.0.3-082e0ff9a7-8b7b1be20d.zip/node_modules/err-code/",\ - "packageDependencies": [\ - ["err-code", "npm:2.0.3"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["error-ex", [\ - ["npm:1.3.2", {\ - "packageLocation": "./.yarn/cache/error-ex-npm-1.3.2-5654f80c0f-c1c2b8b65f.zip/node_modules/error-ex/",\ - "packageDependencies": [\ - ["error-ex", "npm:1.3.2"],\ - ["is-arrayish", "npm:0.2.1"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["es-abstract", [\ - ["npm:1.19.5", {\ - "packageLocation": "./.yarn/cache/es-abstract-npm-1.19.5-524a87d262-55199b0f17.zip/node_modules/es-abstract/",\ - "packageDependencies": [\ - ["es-abstract", "npm:1.19.5"],\ - ["call-bind", "npm:1.0.2"],\ - ["es-to-primitive", "npm:1.2.1"],\ - ["function-bind", "npm:1.1.1"],\ - ["get-intrinsic", "npm:1.1.1"],\ - ["get-symbol-description", "npm:1.0.0"],\ - ["has", "npm:1.0.3"],\ - ["has-symbols", "npm:1.0.3"],\ - ["internal-slot", "npm:1.0.3"],\ - ["is-callable", "npm:1.2.4"],\ - ["is-negative-zero", "npm:2.0.2"],\ - ["is-regex", "npm:1.1.4"],\ - ["is-shared-array-buffer", "npm:1.0.2"],\ - ["is-string", "npm:1.0.7"],\ - ["is-weakref", "npm:1.0.2"],\ - ["object-inspect", "npm:1.12.0"],\ - ["object-keys", "npm:1.1.1"],\ - ["object.assign", "npm:4.1.2"],\ - ["string.prototype.trimend", "npm:1.0.4"],\ - ["string.prototype.trimstart", "npm:1.0.4"],\ - ["unbox-primitive", "npm:1.0.2"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["es-shim-unscopables", [\ - ["npm:1.0.0", {\ - "packageLocation": "./.yarn/cache/es-shim-unscopables-npm-1.0.0-06186593f1-83e95cadbb.zip/node_modules/es-shim-unscopables/",\ - "packageDependencies": [\ - ["es-shim-unscopables", "npm:1.0.0"],\ - ["has", "npm:1.0.3"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["es-to-primitive", [\ - ["npm:1.2.1", {\ - "packageLocation": "./.yarn/cache/es-to-primitive-npm-1.2.1-b7a7eac6c5-4ead6671a2.zip/node_modules/es-to-primitive/",\ - "packageDependencies": [\ - ["es-to-primitive", "npm:1.2.1"],\ - ["is-callable", "npm:1.2.4"],\ - ["is-date-object", "npm:1.0.5"],\ - ["is-symbol", "npm:1.0.4"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["esbuild", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-npm-0.15.11-352cc4ec35/node_modules/esbuild/",\ - "packageDependencies": [\ - ["esbuild", "npm:0.15.11"],\ - ["@esbuild/android-arm", "npm:0.15.11"],\ - ["@esbuild/linux-loong64", "npm:0.15.11"],\ - ["esbuild-android-64", "npm:0.15.11"],\ - ["esbuild-android-arm64", "npm:0.15.11"],\ - ["esbuild-darwin-64", "npm:0.15.11"],\ - ["esbuild-darwin-arm64", "npm:0.15.11"],\ - ["esbuild-freebsd-64", "npm:0.15.11"],\ - ["esbuild-freebsd-arm64", "npm:0.15.11"],\ - ["esbuild-linux-32", "npm:0.15.11"],\ - ["esbuild-linux-64", "npm:0.15.11"],\ - ["esbuild-linux-arm", "npm:0.15.11"],\ - ["esbuild-linux-arm64", "npm:0.15.11"],\ - ["esbuild-linux-mips64le", "npm:0.15.11"],\ - ["esbuild-linux-ppc64le", "npm:0.15.11"],\ - ["esbuild-linux-riscv64", "npm:0.15.11"],\ - ["esbuild-linux-s390x", "npm:0.15.11"],\ - ["esbuild-netbsd-64", "npm:0.15.11"],\ - ["esbuild-openbsd-64", "npm:0.15.11"],\ - ["esbuild-sunos-64", "npm:0.15.11"],\ - ["esbuild-windows-32", "npm:0.15.11"],\ - ["esbuild-windows-64", "npm:0.15.11"],\ - ["esbuild-windows-arm64", "npm:0.15.11"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["esbuild-android-64", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-android-64-npm-0.15.11-66ed633ea9/node_modules/esbuild-android-64/",\ - "packageDependencies": [\ - ["esbuild-android-64", "npm:0.15.11"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["esbuild-android-arm64", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-android-arm64-npm-0.15.11-982c392fd9/node_modules/esbuild-android-arm64/",\ - "packageDependencies": [\ - ["esbuild-android-arm64", "npm:0.15.11"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["esbuild-darwin-64", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-darwin-64-npm-0.15.11-0ccb211fdf/node_modules/esbuild-darwin-64/",\ - "packageDependencies": [\ - ["esbuild-darwin-64", "npm:0.15.11"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["esbuild-darwin-arm64", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-darwin-arm64-npm-0.15.11-cbb0a8549f/node_modules/esbuild-darwin-arm64/",\ - "packageDependencies": [\ - ["esbuild-darwin-arm64", "npm:0.15.11"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["esbuild-freebsd-64", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-freebsd-64-npm-0.15.11-dc712f7982/node_modules/esbuild-freebsd-64/",\ - "packageDependencies": [\ - ["esbuild-freebsd-64", "npm:0.15.11"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["esbuild-freebsd-arm64", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-freebsd-arm64-npm-0.15.11-9827ea52c4/node_modules/esbuild-freebsd-arm64/",\ - "packageDependencies": [\ - ["esbuild-freebsd-arm64", "npm:0.15.11"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["esbuild-linux-32", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-linux-32-npm-0.15.11-5eed199a3d/node_modules/esbuild-linux-32/",\ - "packageDependencies": [\ - ["esbuild-linux-32", "npm:0.15.11"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["esbuild-linux-64", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-linux-64-npm-0.15.11-fd176c9400/node_modules/esbuild-linux-64/",\ - "packageDependencies": [\ - ["esbuild-linux-64", "npm:0.15.11"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["esbuild-linux-arm", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-linux-arm-npm-0.15.11-f249b7f5c4/node_modules/esbuild-linux-arm/",\ - "packageDependencies": [\ - ["esbuild-linux-arm", "npm:0.15.11"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["esbuild-linux-arm64", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-linux-arm64-npm-0.15.11-eb05503e3f/node_modules/esbuild-linux-arm64/",\ + ["npm:2.2.0", {\ + "packageLocation": "./.yarn/cache/entities-npm-2.2.0-0fc8d5b2f7-19010dacaf.zip/node_modules/entities/",\ "packageDependencies": [\ - ["esbuild-linux-arm64", "npm:0.15.11"]\ + ["entities", "npm:2.2.0"]\ ],\ "linkType": "HARD"\ - }]\ - ]],\ - ["esbuild-linux-mips64le", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-linux-mips64le-npm-0.15.11-49f5ad3b5c/node_modules/esbuild-linux-mips64le/",\ + }],\ + ["npm:3.0.1", {\ + "packageLocation": "./.yarn/cache/entities-npm-3.0.1-21eeb201ba-aaf7f12033.zip/node_modules/entities/",\ "packageDependencies": [\ - ["esbuild-linux-mips64le", "npm:0.15.11"]\ + ["entities", "npm:3.0.1"]\ ],\ "linkType": "HARD"\ - }]\ - ]],\ - ["esbuild-linux-ppc64le", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-linux-ppc64le-npm-0.15.11-d3d910dc0b/node_modules/esbuild-linux-ppc64le/",\ + }],\ + ["npm:4.5.0", {\ + "packageLocation": "./.yarn/cache/entities-npm-4.5.0-7cdb83b832-853f8ebd5b.zip/node_modules/entities/",\ "packageDependencies": [\ - ["esbuild-linux-ppc64le", "npm:0.15.11"]\ + ["entities", "npm:4.5.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ - ["esbuild-linux-riscv64", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-linux-riscv64-npm-0.15.11-28216a85bb/node_modules/esbuild-linux-riscv64/",\ + ["env-paths", [\ + ["npm:2.2.1", {\ + "packageLocation": "./.yarn/cache/env-paths-npm-2.2.1-7c7577428c-65b5df55a8.zip/node_modules/env-paths/",\ "packageDependencies": [\ - ["esbuild-linux-riscv64", "npm:0.15.11"]\ + ["env-paths", "npm:2.2.1"]\ ],\ "linkType": "HARD"\ }]\ ]],\ - ["esbuild-linux-s390x", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-linux-s390x-npm-0.15.11-44b07b13bf/node_modules/esbuild-linux-s390x/",\ + ["err-code", [\ + ["npm:2.0.3", {\ + "packageLocation": "./.yarn/cache/err-code-npm-2.0.3-082e0ff9a7-8b7b1be20d.zip/node_modules/err-code/",\ "packageDependencies": [\ - ["esbuild-linux-s390x", "npm:0.15.11"]\ + ["err-code", "npm:2.0.3"]\ ],\ "linkType": "HARD"\ }]\ ]],\ - ["esbuild-netbsd-64", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-netbsd-64-npm-0.15.11-cb3838318d/node_modules/esbuild-netbsd-64/",\ + ["error-ex", [\ + ["npm:1.3.2", {\ + "packageLocation": "./.yarn/cache/error-ex-npm-1.3.2-5654f80c0f-c1c2b8b65f.zip/node_modules/error-ex/",\ "packageDependencies": [\ - ["esbuild-netbsd-64", "npm:0.15.11"]\ + ["error-ex", "npm:1.3.2"],\ + ["is-arrayish", "npm:0.2.1"]\ ],\ "linkType": "HARD"\ }]\ ]],\ - ["esbuild-openbsd-64", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-openbsd-64-npm-0.15.11-f2b2da055a/node_modules/esbuild-openbsd-64/",\ - "packageDependencies": [\ - ["esbuild-openbsd-64", "npm:0.15.11"]\ + ["es-abstract", [\ + ["npm:1.22.3", {\ + "packageLocation": "./.yarn/cache/es-abstract-npm-1.22.3-15a58832e5-b1bdc96285.zip/node_modules/es-abstract/",\ + "packageDependencies": [\ + ["es-abstract", "npm:1.22.3"],\ + ["array-buffer-byte-length", "npm:1.0.0"],\ + ["arraybuffer.prototype.slice", "npm:1.0.2"],\ + ["available-typed-arrays", "npm:1.0.5"],\ + ["call-bind", "npm:1.0.5"],\ + ["es-set-tostringtag", "npm:2.0.1"],\ + ["es-to-primitive", "npm:1.2.1"],\ + ["function.prototype.name", "npm:1.1.6"],\ + ["get-intrinsic", "npm:1.2.2"],\ + ["get-symbol-description", "npm:1.0.0"],\ + ["globalthis", "npm:1.0.3"],\ + ["gopd", "npm:1.0.1"],\ + ["has-property-descriptors", "npm:1.0.0"],\ + ["has-proto", "npm:1.0.1"],\ + ["has-symbols", "npm:1.0.3"],\ + ["hasown", "npm:2.0.0"],\ + ["internal-slot", "npm:1.0.5"],\ + ["is-array-buffer", "npm:3.0.2"],\ + ["is-callable", "npm:1.2.7"],\ + ["is-negative-zero", "npm:2.0.2"],\ + ["is-regex", "npm:1.1.4"],\ + ["is-shared-array-buffer", "npm:1.0.2"],\ + ["is-string", "npm:1.0.7"],\ + ["is-typed-array", "npm:1.1.12"],\ + ["is-weakref", "npm:1.0.2"],\ + ["object-inspect", "npm:1.13.1"],\ + ["object-keys", "npm:1.1.1"],\ + ["object.assign", "npm:4.1.4"],\ + ["regexp.prototype.flags", "npm:1.5.1"],\ + ["safe-array-concat", "npm:1.0.1"],\ + ["safe-regex-test", "npm:1.0.0"],\ + ["string.prototype.trim", "npm:1.2.8"],\ + ["string.prototype.trimend", "npm:1.0.7"],\ + ["string.prototype.trimstart", "npm:1.0.7"],\ + ["typed-array-buffer", "npm:1.0.0"],\ + ["typed-array-byte-length", "npm:1.0.0"],\ + ["typed-array-byte-offset", "npm:1.0.0"],\ + ["typed-array-length", "npm:1.0.4"],\ + ["unbox-primitive", "npm:1.0.2"],\ + ["which-typed-array", "npm:1.1.13"]\ ],\ "linkType": "HARD"\ }]\ ]],\ - ["esbuild-sunos-64", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-sunos-64-npm-0.15.11-8bc506ee74/node_modules/esbuild-sunos-64/",\ + ["es-set-tostringtag", [\ + ["npm:2.0.1", {\ + "packageLocation": "./.yarn/cache/es-set-tostringtag-npm-2.0.1-c87b5de872-ec416a1294.zip/node_modules/es-set-tostringtag/",\ "packageDependencies": [\ - ["esbuild-sunos-64", "npm:0.15.11"]\ + ["es-set-tostringtag", "npm:2.0.1"],\ + ["get-intrinsic", "npm:1.2.0"],\ + ["has", "npm:1.0.3"],\ + ["has-tostringtag", "npm:1.0.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ - ["esbuild-windows-32", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-windows-32-npm-0.15.11-532e8c6b39/node_modules/esbuild-windows-32/",\ + ["es-shim-unscopables", [\ + ["npm:1.0.0", {\ + "packageLocation": "./.yarn/cache/es-shim-unscopables-npm-1.0.0-06186593f1-83e95cadbb.zip/node_modules/es-shim-unscopables/",\ "packageDependencies": [\ - ["esbuild-windows-32", "npm:0.15.11"]\ + ["es-shim-unscopables", "npm:1.0.0"],\ + ["has", "npm:1.0.3"]\ ],\ "linkType": "HARD"\ }]\ ]],\ - ["esbuild-windows-64", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-windows-64-npm-0.15.11-a6a42a35c8/node_modules/esbuild-windows-64/",\ + ["es-to-primitive", [\ + ["npm:1.2.1", {\ + "packageLocation": "./.yarn/cache/es-to-primitive-npm-1.2.1-b7a7eac6c5-4ead6671a2.zip/node_modules/es-to-primitive/",\ "packageDependencies": [\ - ["esbuild-windows-64", "npm:0.15.11"]\ + ["es-to-primitive", "npm:1.2.1"],\ + ["is-callable", "npm:1.2.4"],\ + ["is-date-object", "npm:1.0.5"],\ + ["is-symbol", "npm:1.0.4"]\ ],\ "linkType": "HARD"\ }]\ ]],\ - ["esbuild-windows-arm64", [\ - ["npm:0.15.11", {\ - "packageLocation": "./.yarn/unplugged/esbuild-windows-arm64-npm-0.15.11-d36b5e4f06/node_modules/esbuild-windows-arm64/",\ - "packageDependencies": [\ - ["esbuild-windows-arm64", "npm:0.15.11"]\ + ["esbuild", [\ + ["npm:0.18.20", {\ + "packageLocation": "./.yarn/unplugged/esbuild-npm-0.18.20-004a76d281/node_modules/esbuild/",\ + "packageDependencies": [\ + ["esbuild", "npm:0.18.20"],\ + ["@esbuild/android-arm", "npm:0.18.20"],\ + ["@esbuild/android-arm64", "npm:0.18.20"],\ + ["@esbuild/android-x64", "npm:0.18.20"],\ + ["@esbuild/darwin-arm64", "npm:0.18.20"],\ + ["@esbuild/darwin-x64", "npm:0.18.20"],\ + ["@esbuild/freebsd-arm64", "npm:0.18.20"],\ + ["@esbuild/freebsd-x64", "npm:0.18.20"],\ + ["@esbuild/linux-arm", "npm:0.18.20"],\ + ["@esbuild/linux-arm64", "npm:0.18.20"],\ + ["@esbuild/linux-ia32", "npm:0.18.20"],\ + ["@esbuild/linux-loong64", "npm:0.18.20"],\ + ["@esbuild/linux-mips64el", "npm:0.18.20"],\ + ["@esbuild/linux-ppc64", "npm:0.18.20"],\ + ["@esbuild/linux-riscv64", "npm:0.18.20"],\ + ["@esbuild/linux-s390x", "npm:0.18.20"],\ + ["@esbuild/linux-x64", "npm:0.18.20"],\ + ["@esbuild/netbsd-x64", "npm:0.18.20"],\ + ["@esbuild/openbsd-x64", "npm:0.18.20"],\ + ["@esbuild/sunos-x64", "npm:0.18.20"],\ + ["@esbuild/win32-arm64", "npm:0.18.20"],\ + ["@esbuild/win32-ia32", "npm:0.18.20"],\ + ["@esbuild/win32-x64", "npm:0.18.20"]\ ],\ "linkType": "HARD"\ }]\ @@ -4055,72 +4743,93 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["eslint", [\ - ["npm:8.31.0", {\ - "packageLocation": "./.yarn/cache/eslint-npm-8.31.0-da99c7e469-5e5688bb86.zip/node_modules/eslint/",\ - "packageDependencies": [\ - ["eslint", "npm:8.31.0"],\ - ["@eslint/eslintrc", "npm:1.4.1"],\ - ["@humanwhocodes/config-array", "npm:0.11.8"],\ + ["npm:8.57.0", {\ + "packageLocation": "./.yarn/cache/eslint-npm-8.57.0-4286e12a3a-3a48d7ff85.zip/node_modules/eslint/",\ + "packageDependencies": [\ + ["eslint", "npm:8.57.0"],\ + ["@eslint-community/eslint-utils", "virtual:4286e12a3a0f74af013bc8f16c6d8fdde823cfbf6389660266b171e551f576c805b0a7a8eb2a7087a5cee7dfe6ebb6e1ea3808d93daf915edc95656907a381bb#npm:4.4.0"],\ + ["@eslint-community/regexpp", "npm:4.8.0"],\ + ["@eslint/eslintrc", "npm:2.1.4"],\ + ["@eslint/js", "npm:8.57.0"],\ + ["@humanwhocodes/config-array", "npm:0.11.14"],\ ["@humanwhocodes/module-importer", "npm:1.0.1"],\ ["@nodelib/fs.walk", "npm:1.2.8"],\ + ["@ungap/structured-clone", "npm:1.2.0"],\ ["ajv", "npm:6.12.6"],\ ["chalk", "npm:4.1.2"],\ ["cross-spawn", "npm:7.0.3"],\ ["debug", "virtual:b86a9fb34323a98c6519528ed55faa0d9b44ca8879307c0b29aa384bde47ff59a7d0c9051b31246f14521dfb71ba3c5d6d0b35c29fffc17bf875aa6ad977d9e8#npm:4.3.4"],\ ["doctrine", "npm:3.0.0"],\ ["escape-string-regexp", "npm:4.0.0"],\ - ["eslint-scope", "npm:7.1.1"],\ - ["eslint-utils", "virtual:da99c7e4695a5fca5898fad31d2872f7fe46acc9506643f9acaa2915e169293c951732a239e9f38aa784a4d9e0c1021cee45cb2a8bcf8008a11e40816a14adaa#npm:3.0.0"],\ - ["eslint-visitor-keys", "npm:3.3.0"],\ - ["espree", "npm:9.4.0"],\ - ["esquery", "npm:1.4.0"],\ + ["eslint-scope", "npm:7.2.2"],\ + ["eslint-visitor-keys", "npm:3.4.3"],\ + ["espree", "npm:9.6.1"],\ + ["esquery", "npm:1.5.0"],\ ["esutils", "npm:2.0.3"],\ ["fast-deep-equal", "npm:3.1.3"],\ ["file-entry-cache", "npm:6.0.1"],\ ["find-up", "npm:5.0.0"],\ ["glob-parent", "npm:6.0.2"],\ ["globals", "npm:13.19.0"],\ - ["grapheme-splitter", "npm:1.0.4"],\ + ["graphemer", "npm:1.4.0"],\ ["ignore", "npm:5.2.0"],\ - ["import-fresh", "npm:3.3.0"],\ ["imurmurhash", "npm:0.1.4"],\ ["is-glob", "npm:4.0.3"],\ ["is-path-inside", "npm:3.0.3"],\ - ["js-sdsl", "npm:4.1.5"],\ ["js-yaml", "npm:4.1.0"],\ ["json-stable-stringify-without-jsonify", "npm:1.0.1"],\ ["levn", "npm:0.4.1"],\ ["lodash.merge", "npm:4.6.2"],\ ["minimatch", "npm:3.1.2"],\ ["natural-compare", "npm:1.4.0"],\ - ["optionator", "npm:0.9.1"],\ - ["regexpp", "npm:3.2.0"],\ + ["optionator", "npm:0.9.3"],\ ["strip-ansi", "npm:6.0.1"],\ - ["strip-json-comments", "npm:3.1.1"],\ ["text-table", "npm:0.2.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ + ["eslint-compat-utils", [\ + ["npm:0.1.2", {\ + "packageLocation": "./.yarn/cache/eslint-compat-utils-npm-0.1.2-361c6992b1-2315d9db81.zip/node_modules/eslint-compat-utils/",\ + "packageDependencies": [\ + ["eslint-compat-utils", "npm:0.1.2"]\ + ],\ + "linkType": "SOFT"\ + }],\ + ["virtual:ff64d06f93654b25d9cae47199e62d111efde9ee7d408664ae44397cd2ddf7906aefd54fcc2557f4d5619d92da3af68c7898126469c2a57c381e05b06491f0da#npm:0.1.2", {\ + "packageLocation": "./.yarn/__virtual__/eslint-compat-utils-virtual-a5f7e6147b/0/cache/eslint-compat-utils-npm-0.1.2-361c6992b1-2315d9db81.zip/node_modules/eslint-compat-utils/",\ + "packageDependencies": [\ + ["eslint-compat-utils", "virtual:ff64d06f93654b25d9cae47199e62d111efde9ee7d408664ae44397cd2ddf7906aefd54fcc2557f4d5619d92da3af68c7898126469c2a57c381e05b06491f0da#npm:0.1.2"],\ + ["@types/eslint", null],\ + ["eslint", "npm:8.57.0"]\ + ],\ + "packagePeers": [\ + "@types/eslint",\ + "eslint"\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["eslint-config-standard", [\ - ["npm:17.0.0", {\ - "packageLocation": "./.yarn/cache/eslint-config-standard-npm-17.0.0-2803f6a79a-dc0ed51e18.zip/node_modules/eslint-config-standard/",\ + ["npm:17.1.0", {\ + "packageLocation": "./.yarn/cache/eslint-config-standard-npm-17.1.0-e72fd623cc-8ed14ffe42.zip/node_modules/eslint-config-standard/",\ "packageDependencies": [\ - ["eslint-config-standard", "npm:17.0.0"]\ + ["eslint-config-standard", "npm:17.1.0"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:17.0.0", {\ - "packageLocation": "./.yarn/__virtual__/eslint-config-standard-virtual-5de208ba69/0/cache/eslint-config-standard-npm-17.0.0-2803f6a79a-dc0ed51e18.zip/node_modules/eslint-config-standard/",\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:17.1.0", {\ + "packageLocation": "./.yarn/__virtual__/eslint-config-standard-virtual-a273ec9ea6/0/cache/eslint-config-standard-npm-17.1.0-e72fd623cc-8ed14ffe42.zip/node_modules/eslint-config-standard/",\ "packageDependencies": [\ - ["eslint-config-standard", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:17.0.0"],\ + ["eslint-config-standard", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:17.1.0"],\ ["@types/eslint", null],\ ["@types/eslint-plugin-import", null],\ ["@types/eslint-plugin-n", null],\ ["@types/eslint-plugin-promise", null],\ - ["eslint", "npm:8.31.0"],\ - ["eslint-plugin-import", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.26.0"],\ - ["eslint-plugin-n", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:15.6.1"],\ + ["eslint", "npm:8.57.0"],\ + ["eslint-plugin-import", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.29.1"],\ + ["eslint-plugin-n", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:16.6.2"],\ ["eslint-plugin-promise", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.1"]\ ],\ "packagePeers": [\ @@ -4137,67 +4846,71 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["eslint-import-resolver-node", [\ - ["npm:0.3.6", {\ - "packageLocation": "./.yarn/cache/eslint-import-resolver-node-npm-0.3.6-d9426786c6-6266733af1.zip/node_modules/eslint-import-resolver-node/",\ + ["npm:0.3.9", {\ + "packageLocation": "./.yarn/cache/eslint-import-resolver-node-npm-0.3.9-2a426afc4b-439b912712.zip/node_modules/eslint-import-resolver-node/",\ "packageDependencies": [\ - ["eslint-import-resolver-node", "npm:0.3.6"],\ - ["debug", "virtual:d9426786c635bc4b52511d6cc4b56156f50d780a698c0e20fc6caf10d3be51cbf176e79cff882f4d42a23ff4d0f89fe94222849578214e7fbae0f2754c82af02#npm:3.2.7"],\ - ["resolve", "patch:resolve@npm%3A1.22.0#~builtin::version=1.22.0&hash=07638b"]\ + ["eslint-import-resolver-node", "npm:0.3.9"],\ + ["debug", "virtual:2a426afc4b2eef43db12a540d29c2b5476640459bfcd5c24f86bb401cf8cce97e63bd81794d206a5643057e7f662643afd5ce3dfc4d4bfd8e706006c6309c5fa#npm:3.2.7"],\ + ["is-core-module", "npm:2.13.0"],\ + ["resolve", "patch:resolve@npm%3A1.22.8#~builtin::version=1.22.8&hash=07638b"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["eslint-module-utils", [\ - ["npm:2.7.3", {\ - "packageLocation": "./.yarn/cache/eslint-module-utils-npm-2.7.3-ccd32fe6fd-77048263f3.zip/node_modules/eslint-module-utils/",\ + ["npm:2.8.0", {\ + "packageLocation": "./.yarn/cache/eslint-module-utils-npm-2.8.0-05e42bcab0-74c6dfea76.zip/node_modules/eslint-module-utils/",\ "packageDependencies": [\ - ["eslint-module-utils", "npm:2.7.3"]\ + ["eslint-module-utils", "npm:2.8.0"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:c0858ad0a599e687a7d876de5591e3b098ca550f5c1ad46e7d0e2b6f5720a919cb228a47405daf7d626be1747e41a5b93e4b4d748f16d5e7c36c433aed618452#npm:2.7.3", {\ - "packageLocation": "./.yarn/__virtual__/eslint-module-utils-virtual-e944405700/0/cache/eslint-module-utils-npm-2.7.3-ccd32fe6fd-77048263f3.zip/node_modules/eslint-module-utils/",\ + ["virtual:caddce79266c9767570f5c081ff9adaab1d8b040965749cfca6a3f3f4fbd011bf36f7d755f18ef80e67a5402a33b10c9e1ffc34efb6909461044fc5d60cfbcd0#npm:2.8.0", {\ + "packageLocation": "./.yarn/__virtual__/eslint-module-utils-virtual-d80573de1e/0/cache/eslint-module-utils-npm-2.8.0-05e42bcab0-74c6dfea76.zip/node_modules/eslint-module-utils/",\ "packageDependencies": [\ - ["eslint-module-utils", "virtual:c0858ad0a599e687a7d876de5591e3b098ca550f5c1ad46e7d0e2b6f5720a919cb228a47405daf7d626be1747e41a5b93e4b4d748f16d5e7c36c433aed618452#npm:2.7.3"],\ + ["eslint-module-utils", "virtual:caddce79266c9767570f5c081ff9adaab1d8b040965749cfca6a3f3f4fbd011bf36f7d755f18ef80e67a5402a33b10c9e1ffc34efb6909461044fc5d60cfbcd0#npm:2.8.0"],\ + ["@types/eslint", null],\ ["@types/eslint-import-resolver-node", null],\ ["@types/eslint-import-resolver-typescript", null],\ ["@types/eslint-import-resolver-webpack", null],\ ["@types/typescript-eslint__parser", null],\ ["@typescript-eslint/parser", null],\ - ["debug", "virtual:d9426786c635bc4b52511d6cc4b56156f50d780a698c0e20fc6caf10d3be51cbf176e79cff882f4d42a23ff4d0f89fe94222849578214e7fbae0f2754c82af02#npm:3.2.7"],\ - ["eslint-import-resolver-node", "npm:0.3.6"],\ + ["debug", "virtual:2a426afc4b2eef43db12a540d29c2b5476640459bfcd5c24f86bb401cf8cce97e63bd81794d206a5643057e7f662643afd5ce3dfc4d4bfd8e706006c6309c5fa#npm:3.2.7"],\ + ["eslint", "npm:8.57.0"],\ + ["eslint-import-resolver-node", "npm:0.3.9"],\ ["eslint-import-resolver-typescript", null],\ - ["eslint-import-resolver-webpack", null],\ - ["find-up", "npm:2.1.0"]\ + ["eslint-import-resolver-webpack", null]\ ],\ "packagePeers": [\ "@types/eslint-import-resolver-node",\ "@types/eslint-import-resolver-typescript",\ "@types/eslint-import-resolver-webpack",\ + "@types/eslint",\ "@types/typescript-eslint__parser",\ "@typescript-eslint/parser",\ "eslint-import-resolver-node",\ "eslint-import-resolver-typescript",\ - "eslint-import-resolver-webpack"\ + "eslint-import-resolver-webpack",\ + "eslint"\ ],\ "linkType": "HARD"\ }]\ ]],\ ["eslint-plugin-cypress", [\ - ["npm:2.12.1", {\ - "packageLocation": "./.yarn/cache/eslint-plugin-cypress-npm-2.12.1-6681f582fa-1f1c36e149.zip/node_modules/eslint-plugin-cypress/",\ + ["npm:2.15.1", {\ + "packageLocation": "./.yarn/cache/eslint-plugin-cypress-npm-2.15.1-90f777d9bd-3e66fa9a94.zip/node_modules/eslint-plugin-cypress/",\ "packageDependencies": [\ - ["eslint-plugin-cypress", "npm:2.12.1"]\ + ["eslint-plugin-cypress", "npm:2.15.1"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.12.1", {\ - "packageLocation": "./.yarn/__virtual__/eslint-plugin-cypress-virtual-3210db2eaa/0/cache/eslint-plugin-cypress-npm-2.12.1-6681f582fa-1f1c36e149.zip/node_modules/eslint-plugin-cypress/",\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.15.1", {\ + "packageLocation": "./.yarn/__virtual__/eslint-plugin-cypress-virtual-33ce75aabf/0/cache/eslint-plugin-cypress-npm-2.15.1-90f777d9bd-3e66fa9a94.zip/node_modules/eslint-plugin-cypress/",\ "packageDependencies": [\ - ["eslint-plugin-cypress", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.12.1"],\ + ["eslint-plugin-cypress", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.15.1"],\ ["@types/eslint", null],\ - ["eslint", "npm:8.31.0"],\ - ["globals", "npm:11.12.0"]\ + ["eslint", "npm:8.57.0"],\ + ["globals", "npm:13.21.0"]\ ],\ "packagePeers": [\ "@types/eslint",\ @@ -4214,19 +4927,12 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "SOFT"\ }],\ - ["npm:4.1.0", {\ - "packageLocation": "./.yarn/cache/eslint-plugin-es-npm-4.1.0-a4cf26d3cd-26b87a216d.zip/node_modules/eslint-plugin-es/",\ - "packageDependencies": [\ - ["eslint-plugin-es", "npm:4.1.0"]\ - ],\ - "linkType": "SOFT"\ - }],\ ["virtual:5cccaf00e87dfff96dbbb5eaf7a3055373358b8114d6a1adfb32f54ed6b40ba06068d3aa1fdd8062899a0cad040f68c17cc6b72bac2cdbe9700f3d6330d112f3#npm:3.0.1", {\ "packageLocation": "./.yarn/__virtual__/eslint-plugin-es-virtual-9a126af2f5/0/cache/eslint-plugin-es-npm-3.0.1-95e8015220-e57592c523.zip/node_modules/eslint-plugin-es/",\ "packageDependencies": [\ ["eslint-plugin-es", "virtual:5cccaf00e87dfff96dbbb5eaf7a3055373358b8114d6a1adfb32f54ed6b40ba06068d3aa1fdd8062899a0cad040f68c17cc6b72bac2cdbe9700f3d6330d112f3#npm:3.0.1"],\ ["@types/eslint", null],\ - ["eslint", "npm:8.31.0"],\ + ["eslint", "npm:8.57.0"],\ ["eslint-utils", "npm:2.1.0"],\ ["regexpp", "npm:3.2.0"]\ ],\ @@ -4235,15 +4941,25 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "eslint"\ ],\ "linkType": "HARD"\ + }]\ + ]],\ + ["eslint-plugin-es-x", [\ + ["npm:7.5.0", {\ + "packageLocation": "./.yarn/cache/eslint-plugin-es-x-npm-7.5.0-77e84d6e5d-e770e57df7.zip/node_modules/eslint-plugin-es-x/",\ + "packageDependencies": [\ + ["eslint-plugin-es-x", "npm:7.5.0"]\ + ],\ + "linkType": "SOFT"\ }],\ - ["virtual:75c7efde890fd11a780800acb1cde469ae49edf8842af9f91484f93337ad61713574973ccf7c8eb0a0a15f7ed6c372fd23697f1d4d519b2d7a42f92c04a56cff#npm:4.1.0", {\ - "packageLocation": "./.yarn/__virtual__/eslint-plugin-es-virtual-c640ae95c4/0/cache/eslint-plugin-es-npm-4.1.0-a4cf26d3cd-26b87a216d.zip/node_modules/eslint-plugin-es/",\ + ["virtual:e72a0a9306438b1033938dd0da350cf9f4ec062648c9360382edaa21499b6290430f07b640481cdb3f67c818af79a821eb8f3071ebf7284ab09c47cb982d8502#npm:7.5.0", {\ + "packageLocation": "./.yarn/__virtual__/eslint-plugin-es-x-virtual-ff64d06f93/0/cache/eslint-plugin-es-x-npm-7.5.0-77e84d6e5d-e770e57df7.zip/node_modules/eslint-plugin-es-x/",\ "packageDependencies": [\ - ["eslint-plugin-es", "virtual:75c7efde890fd11a780800acb1cde469ae49edf8842af9f91484f93337ad61713574973ccf7c8eb0a0a15f7ed6c372fd23697f1d4d519b2d7a42f92c04a56cff#npm:4.1.0"],\ + ["eslint-plugin-es-x", "virtual:e72a0a9306438b1033938dd0da350cf9f4ec062648c9360382edaa21499b6290430f07b640481cdb3f67c818af79a821eb8f3071ebf7284ab09c47cb982d8502#npm:7.5.0"],\ + ["@eslint-community/eslint-utils", "virtual:4286e12a3a0f74af013bc8f16c6d8fdde823cfbf6389660266b171e551f576c805b0a7a8eb2a7087a5cee7dfe6ebb6e1ea3808d93daf915edc95656907a381bb#npm:4.4.0"],\ + ["@eslint-community/regexpp", "npm:4.10.0"],\ ["@types/eslint", null],\ - ["eslint", "npm:8.31.0"],\ - ["eslint-utils", "npm:2.1.0"],\ - ["regexpp", "npm:3.2.0"]\ + ["eslint", "npm:8.57.0"],\ + ["eslint-compat-utils", "virtual:ff64d06f93654b25d9cae47199e62d111efde9ee7d408664ae44397cd2ddf7906aefd54fcc2557f4d5619d92da3af68c7898126469c2a57c381e05b06491f0da#npm:0.1.2"]\ ],\ "packagePeers": [\ "@types/eslint",\ @@ -4253,34 +4969,38 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["eslint-plugin-import", [\ - ["npm:2.26.0", {\ - "packageLocation": "./.yarn/cache/eslint-plugin-import-npm-2.26.0-959fe14a01-0bf77ad803.zip/node_modules/eslint-plugin-import/",\ + ["npm:2.29.1", {\ + "packageLocation": "./.yarn/cache/eslint-plugin-import-npm-2.29.1-b94305f7dc-e65159aef8.zip/node_modules/eslint-plugin-import/",\ "packageDependencies": [\ - ["eslint-plugin-import", "npm:2.26.0"]\ + ["eslint-plugin-import", "npm:2.29.1"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.26.0", {\ - "packageLocation": "./.yarn/__virtual__/eslint-plugin-import-virtual-c0858ad0a5/0/cache/eslint-plugin-import-npm-2.26.0-959fe14a01-0bf77ad803.zip/node_modules/eslint-plugin-import/",\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.29.1", {\ + "packageLocation": "./.yarn/__virtual__/eslint-plugin-import-virtual-caddce7926/0/cache/eslint-plugin-import-npm-2.29.1-b94305f7dc-e65159aef8.zip/node_modules/eslint-plugin-import/",\ "packageDependencies": [\ - ["eslint-plugin-import", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.26.0"],\ + ["eslint-plugin-import", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.29.1"],\ ["@types/eslint", null],\ ["@types/typescript-eslint__parser", null],\ ["@typescript-eslint/parser", null],\ - ["array-includes", "npm:3.1.4"],\ - ["array.prototype.flat", "npm:1.3.0"],\ - ["debug", "virtual:faadf6353f98b703db6d695690b392666015d2aab4b710ea086196f4598c68e2b84944d3717503cadb554811494ac27c376eca728086556897f6a7cdb35eaef5#npm:2.6.9"],\ + ["array-includes", "npm:3.1.7"],\ + ["array.prototype.findlastindex", "npm:1.2.3"],\ + ["array.prototype.flat", "npm:1.3.2"],\ + ["array.prototype.flatmap", "npm:1.3.2"],\ + ["debug", "virtual:2a426afc4b2eef43db12a540d29c2b5476640459bfcd5c24f86bb401cf8cce97e63bd81794d206a5643057e7f662643afd5ce3dfc4d4bfd8e706006c6309c5fa#npm:3.2.7"],\ ["doctrine", "npm:2.1.0"],\ - ["eslint", "npm:8.31.0"],\ - ["eslint-import-resolver-node", "npm:0.3.6"],\ - ["eslint-module-utils", "virtual:c0858ad0a599e687a7d876de5591e3b098ca550f5c1ad46e7d0e2b6f5720a919cb228a47405daf7d626be1747e41a5b93e4b4d748f16d5e7c36c433aed618452#npm:2.7.3"],\ - ["has", "npm:1.0.3"],\ - ["is-core-module", "npm:2.9.0"],\ + ["eslint", "npm:8.57.0"],\ + ["eslint-import-resolver-node", "npm:0.3.9"],\ + ["eslint-module-utils", "virtual:caddce79266c9767570f5c081ff9adaab1d8b040965749cfca6a3f3f4fbd011bf36f7d755f18ef80e67a5402a33b10c9e1ffc34efb6909461044fc5d60cfbcd0#npm:2.8.0"],\ + ["hasown", "npm:2.0.0"],\ + ["is-core-module", "npm:2.13.1"],\ ["is-glob", "npm:4.0.3"],\ ["minimatch", "npm:3.1.2"],\ - ["object.values", "npm:1.1.5"],\ - ["resolve", "patch:resolve@npm%3A1.22.0#~builtin::version=1.22.0&hash=07638b"],\ - ["tsconfig-paths", "npm:3.14.1"]\ + ["object.fromentries", "npm:2.0.7"],\ + ["object.groupby", "npm:1.0.1"],\ + ["object.values", "npm:1.1.7"],\ + ["semver", "npm:6.3.1"],\ + ["tsconfig-paths", "npm:3.15.0"]\ ],\ "packagePeers": [\ "@types/eslint",\ @@ -4292,27 +5012,30 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["eslint-plugin-n", [\ - ["npm:15.6.1", {\ - "packageLocation": "./.yarn/cache/eslint-plugin-n-npm-15.6.1-e4ab4703b3-269d6f2896.zip/node_modules/eslint-plugin-n/",\ + ["npm:16.6.2", {\ + "packageLocation": "./.yarn/cache/eslint-plugin-n-npm-16.6.2-77775852d0-3b468da003.zip/node_modules/eslint-plugin-n/",\ "packageDependencies": [\ - ["eslint-plugin-n", "npm:15.6.1"]\ + ["eslint-plugin-n", "npm:16.6.2"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:15.6.1", {\ - "packageLocation": "./.yarn/__virtual__/eslint-plugin-n-virtual-75c7efde89/0/cache/eslint-plugin-n-npm-15.6.1-e4ab4703b3-269d6f2896.zip/node_modules/eslint-plugin-n/",\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:16.6.2", {\ + "packageLocation": "./.yarn/__virtual__/eslint-plugin-n-virtual-e72a0a9306/0/cache/eslint-plugin-n-npm-16.6.2-77775852d0-3b468da003.zip/node_modules/eslint-plugin-n/",\ "packageDependencies": [\ - ["eslint-plugin-n", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:15.6.1"],\ + ["eslint-plugin-n", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:16.6.2"],\ + ["@eslint-community/eslint-utils", "virtual:4286e12a3a0f74af013bc8f16c6d8fdde823cfbf6389660266b171e551f576c805b0a7a8eb2a7087a5cee7dfe6ebb6e1ea3808d93daf915edc95656907a381bb#npm:4.4.0"],\ ["@types/eslint", null],\ ["builtins", "npm:5.0.1"],\ - ["eslint", "npm:8.31.0"],\ - ["eslint-plugin-es", "virtual:75c7efde890fd11a780800acb1cde469ae49edf8842af9f91484f93337ad61713574973ccf7c8eb0a0a15f7ed6c372fd23697f1d4d519b2d7a42f92c04a56cff#npm:4.1.0"],\ - ["eslint-utils", "virtual:da99c7e4695a5fca5898fad31d2872f7fe46acc9506643f9acaa2915e169293c951732a239e9f38aa784a4d9e0c1021cee45cb2a8bcf8008a11e40816a14adaa#npm:3.0.0"],\ - ["ignore", "npm:5.2.0"],\ - ["is-core-module", "npm:2.11.0"],\ + ["eslint", "npm:8.57.0"],\ + ["eslint-plugin-es-x", "virtual:e72a0a9306438b1033938dd0da350cf9f4ec062648c9360382edaa21499b6290430f07b640481cdb3f67c818af79a821eb8f3071ebf7284ab09c47cb982d8502#npm:7.5.0"],\ + ["get-tsconfig", "npm:4.7.2"],\ + ["globals", "npm:13.24.0"],\ + ["ignore", "npm:5.2.4"],\ + ["is-builtin-module", "npm:3.2.1"],\ + ["is-core-module", "npm:2.12.1"],\ ["minimatch", "npm:3.1.2"],\ - ["resolve", "patch:resolve@npm%3A1.22.1#~builtin::version=1.22.1&hash=07638b"],\ - ["semver", "npm:7.3.8"]\ + ["resolve", "patch:resolve@npm%3A1.22.3#~builtin::version=1.22.3&hash=07638b"],\ + ["semver", "npm:7.5.3"]\ ],\ "packagePeers": [\ "@types/eslint",\ @@ -4334,7 +5057,7 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "packageDependencies": [\ ["eslint-plugin-node", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:11.1.0"],\ ["@types/eslint", null],\ - ["eslint", "npm:8.31.0"],\ + ["eslint", "npm:8.57.0"],\ ["eslint-plugin-es", "virtual:5cccaf00e87dfff96dbbb5eaf7a3055373358b8114d6a1adfb32f54ed6b40ba06068d3aa1fdd8062899a0cad040f68c17cc6b72bac2cdbe9700f3d6330d112f3#npm:3.0.1"],\ ["eslint-utils", "npm:2.1.0"],\ ["ignore", "npm:5.2.0"],\ @@ -4362,7 +5085,7 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "packageDependencies": [\ ["eslint-plugin-promise", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.1"],\ ["@types/eslint", null],\ - ["eslint", "npm:8.31.0"]\ + ["eslint", "npm:8.57.0"]\ ],\ "packagePeers": [\ "@types/eslint",\ @@ -4372,25 +5095,26 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["eslint-plugin-vue", [\ - ["npm:9.8.0", {\ - "packageLocation": "./.yarn/cache/eslint-plugin-vue-npm-9.8.0-ad98dd7e70-f3fc36512f.zip/node_modules/eslint-plugin-vue/",\ + ["npm:9.24.0", {\ + "packageLocation": "./.yarn/cache/eslint-plugin-vue-npm-9.24.0-4c6dba51bf-2309b919d8.zip/node_modules/eslint-plugin-vue/",\ "packageDependencies": [\ - ["eslint-plugin-vue", "npm:9.8.0"]\ + ["eslint-plugin-vue", "npm:9.24.0"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:9.8.0", {\ - "packageLocation": "./.yarn/__virtual__/eslint-plugin-vue-virtual-b1db986ed3/0/cache/eslint-plugin-vue-npm-9.8.0-ad98dd7e70-f3fc36512f.zip/node_modules/eslint-plugin-vue/",\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:9.24.0", {\ + "packageLocation": "./.yarn/__virtual__/eslint-plugin-vue-virtual-e080dd5dc6/0/cache/eslint-plugin-vue-npm-9.24.0-4c6dba51bf-2309b919d8.zip/node_modules/eslint-plugin-vue/",\ "packageDependencies": [\ - ["eslint-plugin-vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:9.8.0"],\ + ["eslint-plugin-vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:9.24.0"],\ + ["@eslint-community/eslint-utils", "virtual:4286e12a3a0f74af013bc8f16c6d8fdde823cfbf6389660266b171e551f576c805b0a7a8eb2a7087a5cee7dfe6ebb6e1ea3808d93daf915edc95656907a381bb#npm:4.4.0"],\ ["@types/eslint", null],\ - ["eslint", "npm:8.31.0"],\ - ["eslint-utils", "virtual:da99c7e4695a5fca5898fad31d2872f7fe46acc9506643f9acaa2915e169293c951732a239e9f38aa784a4d9e0c1021cee45cb2a8bcf8008a11e40816a14adaa#npm:3.0.0"],\ + ["eslint", "npm:8.57.0"],\ + ["globals", "npm:13.24.0"],\ ["natural-compare", "npm:1.4.0"],\ ["nth-check", "npm:2.1.1"],\ - ["postcss-selector-parser", "npm:6.0.10"],\ - ["semver", "npm:7.3.7"],\ - ["vue-eslint-parser", "virtual:b1db986ed39a80a226fbfa4d653465b173f10958eed2dd2079c16424a052a437a3b029f3b0228d6df47b4f864f76b70bc3e54b0b37aaac5dc8f4ae6650567a27#npm:9.0.3"],\ + ["postcss-selector-parser", "npm:6.0.15"],\ + ["semver", "npm:7.6.0"],\ + ["vue-eslint-parser", "virtual:e080dd5dc65fb3541eb98fd929c3a1d3733f3aff4bb24b09a6b5cce9fba4a29aca07e286ef93079f2144caa0fd33bb6545549286d3a9f2b9a211caa1f4b68ff9#npm:9.4.2"],\ ["xml-name-validator", "npm:4.0.0"]\ ],\ "packagePeers": [\ @@ -4409,6 +5133,15 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["estraverse", "npm:5.3.0"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:7.2.2", {\ + "packageLocation": "./.yarn/cache/eslint-scope-npm-7.2.2-53cb0df8e8-ec97dbf5fb.zip/node_modules/eslint-scope/",\ + "packageDependencies": [\ + ["eslint-scope", "npm:7.2.2"],\ + ["esrecurse", "npm:4.3.0"],\ + ["estraverse", "npm:5.3.0"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["eslint-utils", [\ @@ -4419,27 +5152,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["eslint-visitor-keys", "npm:1.3.0"]\ ],\ "linkType": "HARD"\ - }],\ - ["npm:3.0.0", {\ - "packageLocation": "./.yarn/cache/eslint-utils-npm-3.0.0-630b3a4013-0668fe02f5.zip/node_modules/eslint-utils/",\ - "packageDependencies": [\ - ["eslint-utils", "npm:3.0.0"]\ - ],\ - "linkType": "SOFT"\ - }],\ - ["virtual:da99c7e4695a5fca5898fad31d2872f7fe46acc9506643f9acaa2915e169293c951732a239e9f38aa784a4d9e0c1021cee45cb2a8bcf8008a11e40816a14adaa#npm:3.0.0", {\ - "packageLocation": "./.yarn/__virtual__/eslint-utils-virtual-e17c8db5e4/0/cache/eslint-utils-npm-3.0.0-630b3a4013-0668fe02f5.zip/node_modules/eslint-utils/",\ - "packageDependencies": [\ - ["eslint-utils", "virtual:da99c7e4695a5fca5898fad31d2872f7fe46acc9506643f9acaa2915e169293c951732a239e9f38aa784a4d9e0c1021cee45cb2a8bcf8008a11e40816a14adaa#npm:3.0.0"],\ - ["@types/eslint", null],\ - ["eslint", "npm:8.31.0"],\ - ["eslint-visitor-keys", "npm:2.1.0"]\ - ],\ - "packagePeers": [\ - "@types/eslint",\ - "eslint"\ - ],\ - "linkType": "HARD"\ }]\ ]],\ ["eslint-visitor-keys", [\ @@ -4450,17 +5162,24 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "HARD"\ }],\ - ["npm:2.1.0", {\ - "packageLocation": "./.yarn/cache/eslint-visitor-keys-npm-2.1.0-c31806b6b9-e3081d7dd2.zip/node_modules/eslint-visitor-keys/",\ + ["npm:3.3.0", {\ + "packageLocation": "./.yarn/cache/eslint-visitor-keys-npm-3.3.0-d329af7c8c-d59e68a7c5.zip/node_modules/eslint-visitor-keys/",\ "packageDependencies": [\ - ["eslint-visitor-keys", "npm:2.1.0"]\ + ["eslint-visitor-keys", "npm:3.3.0"]\ ],\ "linkType": "HARD"\ }],\ - ["npm:3.3.0", {\ - "packageLocation": "./.yarn/cache/eslint-visitor-keys-npm-3.3.0-d329af7c8c-d59e68a7c5.zip/node_modules/eslint-visitor-keys/",\ + ["npm:3.4.1", {\ + "packageLocation": "./.yarn/cache/eslint-visitor-keys-npm-3.4.1-a5d0a58208-f05121d868.zip/node_modules/eslint-visitor-keys/",\ "packageDependencies": [\ - ["eslint-visitor-keys", "npm:3.3.0"]\ + ["eslint-visitor-keys", "npm:3.4.1"]\ + ],\ + "linkType": "HARD"\ + }],\ + ["npm:3.4.3", {\ + "packageLocation": "./.yarn/cache/eslint-visitor-keys-npm-3.4.3-a356ac7e46-36e9ef87fc.zip/node_modules/eslint-visitor-keys/",\ + "packageDependencies": [\ + ["eslint-visitor-keys", "npm:3.4.3"]\ ],\ "linkType": "HARD"\ }]\ @@ -4476,13 +5195,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "HARD"\ }],\ - ["npm:9.4.0", {\ - "packageLocation": "./.yarn/cache/espree-npm-9.4.0-0371ef3614-2e3020dde6.zip/node_modules/espree/",\ + ["npm:9.6.1", {\ + "packageLocation": "./.yarn/cache/espree-npm-9.6.1-a50722a5a9-eb8c149c7a.zip/node_modules/espree/",\ "packageDependencies": [\ - ["espree", "npm:9.4.0"],\ - ["acorn", "npm:8.8.0"],\ - ["acorn-jsx", "virtual:0371ef3614c1182c8fcb05e5954a5fd9c124be4b821bd43f5ef7bdb1bf9603ab5ec16aa2a2a1de93fa397b424774f98f00c5e66d99eec56be0bd9f2a1ab2c75f#npm:5.3.2"],\ - ["eslint-visitor-keys", "npm:3.3.0"]\ + ["espree", "npm:9.6.1"],\ + ["acorn", "npm:8.10.0"],\ + ["acorn-jsx", "virtual:a50722a5a9326b6a5f12350c494c4db3aa0f4caeac45e3e9e5fe071da20014ecfe738fe2ebe2c9c98abae81a4ea86b42f56d776b3bd5ec37f9ad3670c242b242#npm:5.3.2"],\ + ["eslint-visitor-keys", "npm:3.4.1"]\ ],\ "linkType": "HARD"\ }]\ @@ -4495,6 +5214,14 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["estraverse", "npm:5.3.0"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:1.5.0", {\ + "packageLocation": "./.yarn/cache/esquery-npm-1.5.0-d8f8a06879-aefb0d2596.zip/node_modules/esquery/",\ + "packageDependencies": [\ + ["esquery", "npm:1.5.0"],\ + ["estraverse", "npm:5.3.0"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["esrecurse", [\ @@ -4626,14 +5353,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["find-up", [\ - ["npm:2.1.0", {\ - "packageLocation": "./.yarn/cache/find-up-npm-2.1.0-9f6cb1765c-43284fe4da.zip/node_modules/find-up/",\ - "packageDependencies": [\ - ["find-up", "npm:2.1.0"],\ - ["locate-path", "npm:2.0.0"]\ - ],\ - "linkType": "HARD"\ - }],\ ["npm:5.0.0", {\ "packageLocation": "./.yarn/cache/find-up-npm-5.0.0-e03e9b796d-07955e3573.zip/node_modules/find-up/",\ "packageDependencies": [\ @@ -4664,13 +5383,23 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["for-each", [\ + ["npm:0.3.3", {\ + "packageLocation": "./.yarn/cache/for-each-npm-0.3.3-0010ca8cdd-6c48ff2bc6.zip/node_modules/for-each/",\ + "packageDependencies": [\ + ["for-each", "npm:0.3.3"],\ + ["is-callable", "npm:1.2.7"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["foreground-child", [\ - ["npm:2.0.0", {\ - "packageLocation": "./.yarn/cache/foreground-child-npm-2.0.0-80c976b61e-f77ec9aff6.zip/node_modules/foreground-child/",\ + ["npm:3.1.1", {\ + "packageLocation": "./.yarn/cache/foreground-child-npm-3.1.1-77e78ed774-139d270bc8.zip/node_modules/foreground-child/",\ "packageDependencies": [\ - ["foreground-child", "npm:2.0.0"],\ + ["foreground-child", "npm:3.1.1"],\ ["cross-spawn", "npm:7.0.3"],\ - ["signal-exit", "npm:3.0.7"]\ + ["signal-exit", "npm:4.0.2"]\ ],\ "linkType": "HARD"\ }]\ @@ -4720,6 +5449,35 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["function-bind", "npm:1.1.1"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:1.1.2", {\ + "packageLocation": "./.yarn/cache/function-bind-npm-1.1.2-7a55be9b03-2b0ff4ce70.zip/node_modules/function-bind/",\ + "packageDependencies": [\ + ["function-bind", "npm:1.1.2"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["function.prototype.name", [\ + ["npm:1.1.6", {\ + "packageLocation": "./.yarn/cache/function.prototype.name-npm-1.1.6-fd3a6a5cdd-7a3f9bd98a.zip/node_modules/function.prototype.name/",\ + "packageDependencies": [\ + ["function.prototype.name", "npm:1.1.6"],\ + ["call-bind", "npm:1.0.2"],\ + ["define-properties", "npm:1.2.0"],\ + ["es-abstract", "npm:1.22.3"],\ + ["functions-have-names", "npm:1.2.3"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["functions-have-names", [\ + ["npm:1.2.3", {\ + "packageLocation": "./.yarn/cache/functions-have-names-npm-1.2.3-e5cf1e2208-c3f1f5ba20.zip/node_modules/functions-have-names/",\ + "packageDependencies": [\ + ["functions-have-names", "npm:1.2.3"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["gauge", [\ @@ -4758,6 +5516,38 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["has-symbols", "npm:1.0.3"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:1.2.0", {\ + "packageLocation": "./.yarn/cache/get-intrinsic-npm-1.2.0-eb08ea9b1d-78fc0487b7.zip/node_modules/get-intrinsic/",\ + "packageDependencies": [\ + ["get-intrinsic", "npm:1.2.0"],\ + ["function-bind", "npm:1.1.1"],\ + ["has", "npm:1.0.3"],\ + ["has-symbols", "npm:1.0.3"]\ + ],\ + "linkType": "HARD"\ + }],\ + ["npm:1.2.1", {\ + "packageLocation": "./.yarn/cache/get-intrinsic-npm-1.2.1-ae857fd610-5b61d88552.zip/node_modules/get-intrinsic/",\ + "packageDependencies": [\ + ["get-intrinsic", "npm:1.2.1"],\ + ["function-bind", "npm:1.1.1"],\ + ["has", "npm:1.0.3"],\ + ["has-proto", "npm:1.0.1"],\ + ["has-symbols", "npm:1.0.3"]\ + ],\ + "linkType": "HARD"\ + }],\ + ["npm:1.2.2", {\ + "packageLocation": "./.yarn/cache/get-intrinsic-npm-1.2.2-3f446d8847-447ff0724d.zip/node_modules/get-intrinsic/",\ + "packageDependencies": [\ + ["get-intrinsic", "npm:1.2.2"],\ + ["function-bind", "npm:1.1.2"],\ + ["has-proto", "npm:1.0.1"],\ + ["has-symbols", "npm:1.0.3"],\ + ["hasown", "npm:2.0.0"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["get-port", [\ @@ -4780,7 +5570,29 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["get-tsconfig", [\ + ["npm:4.7.2", {\ + "packageLocation": "./.yarn/cache/get-tsconfig-npm-4.7.2-8fbccd9fcf-1723589032.zip/node_modules/get-tsconfig/",\ + "packageDependencies": [\ + ["get-tsconfig", "npm:4.7.2"],\ + ["resolve-pkg-maps", "npm:1.0.0"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["glob", [\ + ["npm:10.2.4", {\ + "packageLocation": "./.yarn/cache/glob-npm-10.2.4-49f715fccc-29845faaa1.zip/node_modules/glob/",\ + "packageDependencies": [\ + ["glob", "npm:10.2.4"],\ + ["foreground-child", "npm:3.1.1"],\ + ["jackspeak", "npm:2.2.0"],\ + ["minimatch", "npm:9.0.0"],\ + ["minipass", "npm:6.0.1"],\ + ["path-scurry", "npm:1.9.1"]\ + ],\ + "linkType": "HARD"\ + }],\ ["npm:7.2.3", {\ "packageLocation": "./.yarn/cache/glob-npm-7.2.3-2d866d17a5-29452e97b3.zip/node_modules/glob/",\ "packageDependencies": [\ @@ -4826,13 +5638,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["globals", [\ - ["npm:11.12.0", {\ - "packageLocation": "./.yarn/cache/globals-npm-11.12.0-1fa7f41a6c-67051a45ec.zip/node_modules/globals/",\ - "packageDependencies": [\ - ["globals", "npm:11.12.0"]\ - ],\ - "linkType": "HARD"\ - }],\ ["npm:13.15.0", {\ "packageLocation": "./.yarn/cache/globals-npm-13.15.0-c0b0c83a7a-383ade0873.zip/node_modules/globals/",\ "packageDependencies": [\ @@ -4848,6 +5653,42 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["type-fest", "npm:0.20.2"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:13.21.0", {\ + "packageLocation": "./.yarn/cache/globals-npm-13.21.0-c0829ce1cb-86c92ca8a0.zip/node_modules/globals/",\ + "packageDependencies": [\ + ["globals", "npm:13.21.0"],\ + ["type-fest", "npm:0.20.2"]\ + ],\ + "linkType": "HARD"\ + }],\ + ["npm:13.24.0", {\ + "packageLocation": "./.yarn/cache/globals-npm-13.24.0-cc7713139c-56066ef058.zip/node_modules/globals/",\ + "packageDependencies": [\ + ["globals", "npm:13.24.0"],\ + ["type-fest", "npm:0.20.2"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["globalthis", [\ + ["npm:1.0.3", {\ + "packageLocation": "./.yarn/cache/globalthis-npm-1.0.3-96cd56020d-fbd7d760dc.zip/node_modules/globalthis/",\ + "packageDependencies": [\ + ["globalthis", "npm:1.0.3"],\ + ["define-properties", "npm:1.1.4"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["gopd", [\ + ["npm:1.0.1", {\ + "packageLocation": "./.yarn/cache/gopd-npm-1.0.1-10c1d0b534-a5ccfb8806.zip/node_modules/gopd/",\ + "packageDependencies": [\ + ["gopd", "npm:1.0.1"],\ + ["get-intrinsic", "npm:1.2.0"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["graceful-fs", [\ @@ -4859,11 +5700,11 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ - ["grapheme-splitter", [\ - ["npm:1.0.4", {\ - "packageLocation": "./.yarn/cache/grapheme-splitter-npm-1.0.4-648f2bf509-0c22ec54de.zip/node_modules/grapheme-splitter/",\ + ["graphemer", [\ + ["npm:1.4.0", {\ + "packageLocation": "./.yarn/cache/graphemer-npm-1.4.0-0627732d35-bab8f0be9b.zip/node_modules/graphemer/",\ "packageDependencies": [\ - ["grapheme-splitter", "npm:1.0.4"]\ + ["graphemer", "npm:1.4.0"]\ ],\ "linkType": "HARD"\ }]\ @@ -4913,6 +5754,15 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["has-proto", [\ + ["npm:1.0.1", {\ + "packageLocation": "./.yarn/cache/has-proto-npm-1.0.1-631ea9d820-febc5b5b53.zip/node_modules/has-proto/",\ + "packageDependencies": [\ + ["has-proto", "npm:1.0.1"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["has-symbols", [\ ["npm:1.0.3", {\ "packageLocation": "./.yarn/cache/has-symbols-npm-1.0.3-1986bff2c4-a054c40c63.zip/node_modules/has-symbols/",\ @@ -4941,20 +5791,30 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["hasown", [\ + ["npm:2.0.0", {\ + "packageLocation": "./.yarn/cache/hasown-npm-2.0.0-78b794ceef-6151c75ca1.zip/node_modules/hasown/",\ + "packageDependencies": [\ + ["hasown", "npm:2.0.0"],\ + ["function-bind", "npm:1.1.2"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["highcharts", [\ - ["npm:10.3.2", {\ - "packageLocation": "./.yarn/cache/highcharts-npm-10.3.2-1672942f09-43cb42b24c.zip/node_modules/highcharts/",\ + ["npm:11.4.0", {\ + "packageLocation": "./.yarn/cache/highcharts-npm-11.4.0-8a1f46b545-873e661914.zip/node_modules/highcharts/",\ "packageDependencies": [\ - ["highcharts", "npm:10.3.2"]\ + ["highcharts", "npm:11.4.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["highlight.js", [\ - ["npm:11.5.1", {\ - "packageLocation": "./.yarn/cache/highlight.js-npm-11.5.1-0fb1167640-bff556101d.zip/node_modules/highlight.js/",\ + ["npm:11.9.0", {\ + "packageLocation": "./.yarn/cache/highlight.js-npm-11.9.0-ec99f7b12f-4043d31c5d.zip/node_modules/highlight.js/",\ "packageDependencies": [\ - ["highlight.js", "npm:11.5.1"]\ + ["highlight.js", "npm:11.9.0"]\ ],\ "linkType": "HARD"\ }]\ @@ -4969,44 +5829,46 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["html-validate", [\ - ["npm:7.12.2", {\ - "packageLocation": "./.yarn/cache/html-validate-npm-7.12.2-4a7c5f12a3-04bccd1680.zip/node_modules/html-validate/",\ + ["npm:8.18.1", {\ + "packageLocation": "./.yarn/cache/html-validate-npm-8.18.1-c5271a0fb9-53479bf75b.zip/node_modules/html-validate/",\ "packageDependencies": [\ - ["html-validate", "npm:7.12.2"]\ + ["html-validate", "npm:8.18.1"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:7.12.2", {\ - "packageLocation": "./.yarn/__virtual__/html-validate-virtual-ef05135489/0/cache/html-validate-npm-7.12.2-4a7c5f12a3-04bccd1680.zip/node_modules/html-validate/",\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:8.18.1", {\ + "packageLocation": "./.yarn/__virtual__/html-validate-virtual-640261ed3b/0/cache/html-validate-npm-8.18.1-c5271a0fb9-53479bf75b.zip/node_modules/html-validate/",\ "packageDependencies": [\ - ["html-validate", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:7.12.2"],\ + ["html-validate", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:8.18.1"],\ ["@babel/code-frame", "npm:7.16.7"],\ - ["@html-validate/stylish", "npm:3.0.0"],\ - ["@sidvind/better-ajv-errors", "virtual:ef051354894d83cc0cda7d7a4a6eca604681f35eb8f2a25fe305435ac599a4d71af04fcaa26daa3f5aa064e73442d202ba9dfbda5c3f8c7b1daf472f3a17da86#npm:2.0.0"],\ + ["@html-validate/stylish", "npm:4.1.0"],\ + ["@sidvind/better-ajv-errors", "virtual:640261ed3b7a9880a388cc504caacf8ea790dd52f1cb31fbc3be445cb2adc6e73fc87097de620863105eb917510145ef2457d30000c7361456ab67ec0b895136#npm:2.1.3"],\ ["@types/jest", null],\ ["@types/jest-diff", null],\ ["@types/jest-snapshot", null],\ - ["acorn-walk", "npm:8.2.0"],\ + ["@types/vitest", null],\ ["ajv", "npm:8.11.0"],\ - ["deepmerge", "npm:4.2.2"],\ - ["espree", "npm:9.3.2"],\ - ["glob", "npm:8.0.3"],\ - ["ignore", "npm:5.2.0"],\ + ["deepmerge", "npm:4.3.1"],\ + ["glob", "npm:10.2.4"],\ + ["ignore", "npm:5.3.1"],\ ["jest", null],\ ["jest-diff", null],\ ["jest-snapshot", null],\ ["kleur", "npm:4.1.4"],\ ["minimist", "npm:1.2.6"],\ ["prompts", "npm:2.4.2"],\ - ["semver", "npm:7.3.7"]\ + ["semver", "npm:7.3.7"],\ + ["vitest", null]\ ],\ "packagePeers": [\ "@types/jest-diff",\ "@types/jest-snapshot",\ "@types/jest",\ + "@types/vitest",\ "jest-diff",\ "jest-snapshot",\ - "jest"\ + "jest",\ + "vitest"\ ],\ "linkType": "HARD"\ }]\ @@ -5019,10 +5881,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "SOFT"\ }],\ - ["virtual:989bccf2aa3cb6d741c7ce5643ed4e817accfff58f84aaf6345d18a48f60699e8de9c81c8278d6bdf8deed7d9e463b2cc20d78724ce4bd72d1bbf84cb8c02220#npm:2.0.2", {\ - "packageLocation": "./.yarn/__virtual__/htmlnano-virtual-6bf87520a5/0/cache/htmlnano-npm-2.0.2-a89803bfeb-41f9e0c0e5.zip/node_modules/htmlnano/",\ + ["virtual:cdd2835c1202e86fad55b2266578ff3755267672440481af37bdfff670fd205f561469a10385c20d1ff403af7fad49006bc71ffff21d12592a8ebd0c8be79c0c#npm:2.0.2", {\ + "packageLocation": "./.yarn/__virtual__/htmlnano-virtual-d2bb6df599/0/cache/htmlnano-npm-2.0.2-a89803bfeb-41f9e0c0e5.zip/node_modules/htmlnano/",\ "packageDependencies": [\ - ["htmlnano", "virtual:989bccf2aa3cb6d741c7ce5643ed4e817accfff58f84aaf6345d18a48f60699e8de9c81c8278d6bdf8deed7d9e463b2cc20d78724ce4bd72d1bbf84cb8c02220#npm:2.0.2"],\ + ["htmlnano", "virtual:cdd2835c1202e86fad55b2266578ff3755267672440481af37bdfff670fd205f561469a10385c20d1ff403af7fad49006bc71ffff21d12592a8ebd0c8be79c0c#npm:2.0.2"],\ ["@types/cssnano", null],\ ["@types/postcss", null],\ ["@types/purgecss", null],\ @@ -5078,10 +5940,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["http-cache-semantics", [\ - ["npm:4.1.0", {\ - "packageLocation": "./.yarn/cache/http-cache-semantics-npm-4.1.0-860520a31f-974de94a81.zip/node_modules/http-cache-semantics/",\ + ["npm:4.1.1", {\ + "packageLocation": "./.yarn/cache/http-cache-semantics-npm-4.1.1-1120131375-83ac0bc60b.zip/node_modules/http-cache-semantics/",\ "packageDependencies": [\ - ["http-cache-semantics", "npm:4.1.0"]\ + ["http-cache-semantics", "npm:4.1.1"]\ ],\ "linkType": "HARD"\ }]\ @@ -5133,6 +5995,15 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["ical.js", [\ + ["npm:1.5.0", {\ + "packageLocation": "./.yarn/cache/ical.js-npm-1.5.0-5ba1c69420-51df7a01f4.zip/node_modules/ical.js/",\ + "packageDependencies": [\ + ["ical.js", "npm:1.5.0"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["iconv-lite", [\ ["npm:0.6.3", {\ "packageLocation": "./.yarn/cache/iconv-lite-npm-0.6.3-24b8aae27e-3f60d47a5c.zip/node_modules/iconv-lite/",\ @@ -5147,7 +6018,21 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["npm:5.2.0", {\ "packageLocation": "./.yarn/cache/ignore-npm-5.2.0-fc4b58a4f3-6b1f926792.zip/node_modules/ignore/",\ "packageDependencies": [\ - ["ignore", "npm:5.2.0"]\ + ["ignore", "npm:5.2.0"]\ + ],\ + "linkType": "HARD"\ + }],\ + ["npm:5.2.4", {\ + "packageLocation": "./.yarn/cache/ignore-npm-5.2.4-fbe6e989e5-3d4c309c60.zip/node_modules/ignore/",\ + "packageDependencies": [\ + ["ignore", "npm:5.2.4"]\ + ],\ + "linkType": "HARD"\ + }],\ + ["npm:5.3.1", {\ + "packageLocation": "./.yarn/cache/ignore-npm-5.3.1-f6947c5df7-71d7bb4c1d.zip/node_modules/ignore/",\ + "packageDependencies": [\ + ["ignore", "npm:5.3.1"]\ ],\ "linkType": "HARD"\ }]\ @@ -5220,11 +6105,11 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["internal-slot", [\ - ["npm:1.0.3", {\ - "packageLocation": "./.yarn/cache/internal-slot-npm-1.0.3-9e05eea002-1944f92e98.zip/node_modules/internal-slot/",\ + ["npm:1.0.5", {\ + "packageLocation": "./.yarn/cache/internal-slot-npm-1.0.5-a2241f3e66-97e84046bf.zip/node_modules/internal-slot/",\ "packageDependencies": [\ - ["internal-slot", "npm:1.0.3"],\ - ["get-intrinsic", "npm:1.1.1"],\ + ["internal-slot", "npm:1.0.5"],\ + ["get-intrinsic", "npm:1.2.1"],\ ["has", "npm:1.0.3"],\ ["side-channel", "npm:1.0.4"]\ ],\ @@ -5249,6 +6134,28 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["is-array-buffer", [\ + ["npm:3.0.1", {\ + "packageLocation": "./.yarn/cache/is-array-buffer-npm-3.0.1-3e93b14326-f26ab87448.zip/node_modules/is-array-buffer/",\ + "packageDependencies": [\ + ["is-array-buffer", "npm:3.0.1"],\ + ["call-bind", "npm:1.0.2"],\ + ["get-intrinsic", "npm:1.2.0"],\ + ["is-typed-array", "npm:1.1.10"]\ + ],\ + "linkType": "HARD"\ + }],\ + ["npm:3.0.2", {\ + "packageLocation": "./.yarn/cache/is-array-buffer-npm-3.0.2-0dec897785-dcac9dda66.zip/node_modules/is-array-buffer/",\ + "packageDependencies": [\ + ["is-array-buffer", "npm:3.0.2"],\ + ["call-bind", "npm:1.0.2"],\ + ["get-intrinsic", "npm:1.2.1"],\ + ["is-typed-array", "npm:1.1.10"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["is-arrayish", [\ ["npm:0.2.1", {\ "packageLocation": "./.yarn/cache/is-arrayish-npm-0.2.1-23927dfb15-eef4417e3c.zip/node_modules/is-arrayish/",\ @@ -5289,6 +6196,16 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["is-builtin-module", [\ + ["npm:3.2.1", {\ + "packageLocation": "./.yarn/cache/is-builtin-module-npm-3.2.1-2f92a5d353-e8f0ffc19a.zip/node_modules/is-builtin-module/",\ + "packageDependencies": [\ + ["is-builtin-module", "npm:3.2.1"],\ + ["builtin-modules", "npm:3.3.0"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["is-callable", [\ ["npm:1.2.4", {\ "packageLocation": "./.yarn/cache/is-callable-npm-1.2.4-03fc17459c-1a28d57dc4.zip/node_modules/is-callable/",\ @@ -5296,25 +6213,40 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["is-callable", "npm:1.2.4"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:1.2.7", {\ + "packageLocation": "./.yarn/cache/is-callable-npm-1.2.7-808a303e61-61fd57d03b.zip/node_modules/is-callable/",\ + "packageDependencies": [\ + ["is-callable", "npm:1.2.7"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["is-core-module", [\ - ["npm:2.10.0", {\ - "packageLocation": "./.yarn/cache/is-core-module-npm-2.10.0-6dff9310aa-0f3f77811f.zip/node_modules/is-core-module/",\ + ["npm:2.12.1", {\ + "packageLocation": "./.yarn/cache/is-core-module-npm-2.12.1-ce74e89160-f04ea30533.zip/node_modules/is-core-module/",\ "packageDependencies": [\ - ["is-core-module", "npm:2.10.0"],\ + ["is-core-module", "npm:2.12.1"],\ ["has", "npm:1.0.3"]\ ],\ "linkType": "HARD"\ }],\ - ["npm:2.11.0", {\ - "packageLocation": "./.yarn/cache/is-core-module-npm-2.11.0-70061e141a-f96fd490c6.zip/node_modules/is-core-module/",\ + ["npm:2.13.0", {\ + "packageLocation": "./.yarn/cache/is-core-module-npm-2.13.0-e444c50225-053ab101fb.zip/node_modules/is-core-module/",\ "packageDependencies": [\ - ["is-core-module", "npm:2.11.0"],\ + ["is-core-module", "npm:2.13.0"],\ ["has", "npm:1.0.3"]\ ],\ "linkType": "HARD"\ }],\ + ["npm:2.13.1", {\ + "packageLocation": "./.yarn/cache/is-core-module-npm-2.13.1-36e17434f9-256559ee8a.zip/node_modules/is-core-module/",\ + "packageDependencies": [\ + ["is-core-module", "npm:2.13.1"],\ + ["hasown", "npm:2.0.0"]\ + ],\ + "linkType": "HARD"\ + }],\ ["npm:2.9.0", {\ "packageLocation": "./.yarn/cache/is-core-module-npm-2.9.0-5ba77c35ae-b27034318b.zip/node_modules/is-core-module/",\ "packageDependencies": [\ @@ -5478,6 +6410,28 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["is-typed-array", [\ + ["npm:1.1.10", {\ + "packageLocation": "./.yarn/cache/is-typed-array-npm-1.1.10-fe4ef83cdc-aac6ecb59d.zip/node_modules/is-typed-array/",\ + "packageDependencies": [\ + ["is-typed-array", "npm:1.1.10"],\ + ["available-typed-arrays", "npm:1.0.5"],\ + ["call-bind", "npm:1.0.2"],\ + ["for-each", "npm:0.3.3"],\ + ["gopd", "npm:1.0.1"],\ + ["has-tostringtag", "npm:1.0.0"]\ + ],\ + "linkType": "HARD"\ + }],\ + ["npm:1.1.12", {\ + "packageLocation": "./.yarn/cache/is-typed-array-npm-1.1.12-6135c91b1a-4c89c4a3be.zip/node_modules/is-typed-array/",\ + "packageDependencies": [\ + ["is-typed-array", "npm:1.1.12"],\ + ["which-typed-array", "npm:1.1.13"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["is-weakref", [\ ["npm:1.0.2", {\ "packageLocation": "./.yarn/cache/is-weakref-npm-1.0.2-ff80e8c314-95bd9a57cd.zip/node_modules/is-weakref/",\ @@ -5488,6 +6442,15 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["isarray", [\ + ["npm:2.0.5", {\ + "packageLocation": "./.yarn/cache/isarray-npm-2.0.5-4ba522212d-bd5bbe4104.zip/node_modules/isarray/",\ + "packageDependencies": [\ + ["isarray", "npm:2.0.5"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["isbinaryfile", [\ ["npm:4.0.10", {\ "packageLocation": "./.yarn/cache/isbinaryfile-npm-4.0.10-91d1251522-a6b28db7e2.zip/node_modules/isbinaryfile/",\ @@ -5525,84 +6488,63 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["supports-color", "npm:7.2.0"]\ ],\ "linkType": "HARD"\ - }]\ - ]],\ - ["istanbul-reports", [\ - ["npm:3.1.4", {\ - "packageLocation": "./.yarn/cache/istanbul-reports-npm-3.1.4-5faaa9636c-2132983355.zip/node_modules/istanbul-reports/",\ - "packageDependencies": [\ - ["istanbul-reports", "npm:3.1.4"],\ - ["html-escaper", "npm:2.0.2"],\ - ["istanbul-lib-report", "npm:3.0.0"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["jest-diff", [\ - ["npm:27.5.1", {\ - "packageLocation": "./.yarn/cache/jest-diff-npm-27.5.1-818e549196-8be27c1e1e.zip/node_modules/jest-diff/",\ + }],\ + ["npm:3.0.1", {\ + "packageLocation": "./.yarn/cache/istanbul-lib-report-npm-3.0.1-b17446ab24-fd17a1b879.zip/node_modules/istanbul-lib-report/",\ "packageDependencies": [\ - ["jest-diff", "npm:27.5.1"],\ - ["chalk", "npm:4.1.2"],\ - ["diff-sequences", "npm:27.5.1"],\ - ["jest-get-type", "npm:27.5.1"],\ - ["pretty-format", "npm:27.5.1"]\ + ["istanbul-lib-report", "npm:3.0.1"],\ + ["istanbul-lib-coverage", "npm:3.2.0"],\ + ["make-dir", "npm:4.0.0"],\ + ["supports-color", "npm:7.2.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ - ["jest-get-type", [\ - ["npm:27.5.1", {\ - "packageLocation": "./.yarn/cache/jest-get-type-npm-27.5.1-980fbf7a43-63064ab701.zip/node_modules/jest-get-type/",\ + ["istanbul-reports", [\ + ["npm:3.1.6", {\ + "packageLocation": "./.yarn/cache/istanbul-reports-npm-3.1.6-66918eb97f-44c4c0582f.zip/node_modules/istanbul-reports/",\ "packageDependencies": [\ - ["jest-get-type", "npm:27.5.1"]\ + ["istanbul-reports", "npm:3.1.6"],\ + ["html-escaper", "npm:2.0.2"],\ + ["istanbul-lib-report", "npm:3.0.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ - ["jest-matcher-utils", [\ - ["npm:27.5.1", {\ - "packageLocation": "./.yarn/cache/jest-matcher-utils-npm-27.5.1-0c47b071fb-bb2135fc48.zip/node_modules/jest-matcher-utils/",\ + ["jackspeak", [\ + ["npm:2.2.0", {\ + "packageLocation": "./.yarn/cache/jackspeak-npm-2.2.0-5383861524-d8cd5be4f0.zip/node_modules/jackspeak/",\ "packageDependencies": [\ - ["jest-matcher-utils", "npm:27.5.1"],\ - ["chalk", "npm:4.1.2"],\ - ["jest-diff", "npm:27.5.1"],\ - ["jest-get-type", "npm:27.5.1"],\ - ["pretty-format", "npm:27.5.1"]\ + ["jackspeak", "npm:2.2.0"],\ + ["@isaacs/cliui", "npm:8.0.2"],\ + ["@pkgjs/parseargs", "npm:0.11.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["jquery", [\ - ["npm:3.6.0", {\ - "packageLocation": "./.yarn/cache/jquery-npm-3.6.0-ca7872bdbb-8fd5fef4aa.zip/node_modules/jquery/",\ - "packageDependencies": [\ - ["jquery", "npm:3.6.0"]\ - ],\ - "linkType": "HARD"\ - }],\ - ["npm:3.6.3", {\ - "packageLocation": "./.yarn/cache/jquery-npm-3.6.3-cbc34d2330-0fd366bdca.zip/node_modules/jquery/",\ + ["npm:3.7.1", {\ + "packageLocation": "./.yarn/cache/jquery-npm-3.7.1-eeeac0f21e-4370b8139d.zip/node_modules/jquery/",\ "packageDependencies": [\ - ["jquery", "npm:3.6.3"]\ + ["jquery", "npm:3.7.1"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["jquery-migrate", [\ - ["npm:3.4.0", {\ - "packageLocation": "./.yarn/cache/jquery-migrate-npm-3.4.0-88c209e61f-7431685c56.zip/node_modules/jquery-migrate/",\ + ["npm:3.4.1", {\ + "packageLocation": "./.yarn/cache/jquery-migrate-npm-3.4.1-c842b6adb7-d2cb17d055.zip/node_modules/jquery-migrate/",\ "packageDependencies": [\ - ["jquery-migrate", "npm:3.4.0"]\ + ["jquery-migrate", "npm:3.4.1"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.0", {\ - "packageLocation": "./.yarn/__virtual__/jquery-migrate-virtual-68c5ec0b7a/0/cache/jquery-migrate-npm-3.4.0-88c209e61f-7431685c56.zip/node_modules/jquery-migrate/",\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.1", {\ + "packageLocation": "./.yarn/__virtual__/jquery-migrate-virtual-e23c9912e5/0/cache/jquery-migrate-npm-3.4.1-c842b6adb7-d2cb17d055.zip/node_modules/jquery-migrate/",\ "packageDependencies": [\ - ["jquery-migrate", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.0"],\ + ["jquery-migrate", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.1"],\ ["@types/jquery", null],\ - ["jquery", "npm:3.6.3"]\ + ["jquery", "npm:3.7.1"]\ ],\ "packagePeers": [\ "@types/jquery",\ @@ -5611,30 +6553,11 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ - ["jquery-ui-dist", [\ - ["npm:1.13.2", {\ - "packageLocation": "./.yarn/cache/jquery-ui-dist-npm-1.13.2-86225c0ce7-4f3a3a2ff8.zip/node_modules/jquery-ui-dist/",\ - "packageDependencies": [\ - ["jquery-ui-dist", "npm:1.13.2"],\ - ["jquery", "npm:3.6.0"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ ["js-cookie", [\ - ["npm:3.0.1", {\ - "packageLocation": "./.yarn/cache/js-cookie-npm-3.0.1-04c7177de1-bb48de67e2.zip/node_modules/js-cookie/",\ - "packageDependencies": [\ - ["js-cookie", "npm:3.0.1"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["js-sdsl", [\ - ["npm:4.1.5", {\ - "packageLocation": "./.yarn/cache/js-sdsl-npm-4.1.5-66fcf4f580-695f657ddc.zip/node_modules/js-sdsl/",\ + ["npm:3.0.5", {\ + "packageLocation": "./.yarn/cache/js-cookie-npm-3.0.5-8fc8fcc9b4-2dbd2809c6.zip/node_modules/js-cookie/",\ "packageDependencies": [\ - ["js-sdsl", "npm:4.1.5"]\ + ["js-cookie", "npm:3.0.5"]\ ],\ "linkType": "HARD"\ }]\ @@ -5702,10 +6625,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["json5", [\ - ["npm:1.0.1", {\ - "packageLocation": "./.yarn/cache/json5-npm-1.0.1-647fc8794b-e76ea23dbb.zip/node_modules/json5/",\ + ["npm:1.0.2", {\ + "packageLocation": "./.yarn/cache/json5-npm-1.0.2-9607f93e30-866458a8c5.zip/node_modules/json5/",\ "packageDependencies": [\ - ["json5", "npm:1.0.1"],\ + ["json5", "npm:1.0.2"],\ ["minimist", "npm:1.2.6"]\ ],\ "linkType": "HARD"\ @@ -5884,18 +6807,28 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["weak-lru-cache", "npm:1.2.2"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:2.8.5", {\ + "packageLocation": "./.yarn/unplugged/lmdb-npm-2.8.5-e5fdd937dd/node_modules/lmdb/",\ + "packageDependencies": [\ + ["lmdb", "npm:2.8.5"],\ + ["@lmdb/lmdb-darwin-arm64", "npm:2.8.5"],\ + ["@lmdb/lmdb-darwin-x64", "npm:2.8.5"],\ + ["@lmdb/lmdb-linux-arm", "npm:2.8.5"],\ + ["@lmdb/lmdb-linux-arm64", "npm:2.8.5"],\ + ["@lmdb/lmdb-linux-x64", "npm:2.8.5"],\ + ["@lmdb/lmdb-win32-x64", "npm:2.8.5"],\ + ["msgpackr", "npm:1.9.9"],\ + ["node-addon-api", "npm:6.1.0"],\ + ["node-gyp", "npm:9.0.0"],\ + ["node-gyp-build-optional-packages", "npm:5.1.1"],\ + ["ordered-binary", "npm:1.4.1"],\ + ["weak-lru-cache", "npm:1.2.2"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["locate-path", [\ - ["npm:2.0.0", {\ - "packageLocation": "./.yarn/cache/locate-path-npm-2.0.0-673d28b0ea-02d581edbb.zip/node_modules/locate-path/",\ - "packageDependencies": [\ - ["locate-path", "npm:2.0.0"],\ - ["p-locate", "npm:2.0.0"],\ - ["path-exists", "npm:3.0.0"]\ - ],\ - "linkType": "HARD"\ - }],\ ["npm:6.0.0", {\ "packageLocation": "./.yarn/cache/locate-path-npm-6.0.0-06a1e4c528-72eb661788.zip/node_modules/locate-path/",\ "packageDependencies": [\ @@ -5932,15 +6865,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ - ["lodash.sortby", [\ - ["npm:4.7.0", {\ - "packageLocation": "./.yarn/cache/lodash.sortby-npm-4.7.0-fda8ab950d-db170c9396.zip/node_modules/lodash.sortby/",\ - "packageDependencies": [\ - ["lodash.sortby", "npm:4.7.0"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ ["lru-cache", [\ ["npm:6.0.0", {\ "packageLocation": "./.yarn/cache/lru-cache-npm-6.0.0-b4c8668fe1-f97f499f89.zip/node_modules/lru-cache/",\ @@ -5956,23 +6880,30 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["lru-cache", "npm:7.10.1"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:9.1.1", {\ + "packageLocation": "./.yarn/cache/lru-cache-npm-9.1.1-765199cb01-4d703bb9b6.zip/node_modules/lru-cache/",\ + "packageDependencies": [\ + ["lru-cache", "npm:9.1.1"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["luxon", [\ - ["npm:3.2.1", {\ - "packageLocation": "./.yarn/cache/luxon-npm-3.2.1-56f8d97395-3fa3def2c5.zip/node_modules/luxon/",\ + ["npm:3.4.4", {\ + "packageLocation": "./.yarn/cache/luxon-npm-3.4.4-c93f95dde8-36c1f99c47.zip/node_modules/luxon/",\ "packageDependencies": [\ - ["luxon", "npm:3.2.1"]\ + ["luxon", "npm:3.4.4"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["magic-string", [\ - ["npm:0.25.9", {\ - "packageLocation": "./.yarn/cache/magic-string-npm-0.25.9-0b51c0ea50-9a0e55a15c.zip/node_modules/magic-string/",\ + ["npm:0.30.7", {\ + "packageLocation": "./.yarn/cache/magic-string-npm-0.30.7-0bb5819095-bdf102e36a.zip/node_modules/magic-string/",\ "packageDependencies": [\ - ["magic-string", "npm:0.25.9"],\ - ["sourcemap-codec", "npm:1.4.8"]\ + ["magic-string", "npm:0.30.7"],\ + ["@jridgewell/sourcemap-codec", "npm:1.4.15"]\ ],\ "linkType": "HARD"\ }]\ @@ -5985,6 +6916,14 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["semver", "npm:6.3.0"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:4.0.0", {\ + "packageLocation": "./.yarn/cache/make-dir-npm-4.0.0-ec3cd921cc-bf0731a2dd.zip/node_modules/make-dir/",\ + "packageDependencies": [\ + ["make-dir", "npm:4.0.0"],\ + ["semver", "npm:7.5.3"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["make-fetch-happen", [\ @@ -5994,7 +6933,7 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["make-fetch-happen", "npm:10.1.5"],\ ["agentkeepalive", "npm:4.2.1"],\ ["cacache", "npm:16.1.0"],\ - ["http-cache-semantics", "npm:4.1.0"],\ + ["http-cache-semantics", "npm:4.1.1"],\ ["http-proxy-agent", "npm:5.0.0"],\ ["https-proxy-agent", "npm:5.0.1"],\ ["is-lambda", "npm:1.0.1"],\ @@ -6053,6 +6992,14 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["brace-expansion", "npm:2.0.1"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:9.0.0", {\ + "packageLocation": "./.yarn/cache/minimatch-npm-9.0.0-c6737cb1be-7bd57899ed.zip/node_modules/minimatch/",\ + "packageDependencies": [\ + ["minimatch", "npm:9.0.0"],\ + ["brace-expansion", "npm:2.0.1"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["minimist", [\ @@ -6072,6 +7019,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["yallist", "npm:4.0.0"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:6.0.1", {\ + "packageLocation": "./.yarn/cache/minipass-npm-6.0.1-634723433e-1df70bb565.zip/node_modules/minipass/",\ + "packageDependencies": [\ + ["minipass", "npm:6.0.1"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["minipass-collect", [\ @@ -6148,27 +7102,27 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["moment", [\ - ["npm:2.29.3", {\ - "packageLocation": "./.yarn/cache/moment-npm-2.29.3-fe4ba99bae-2e780e36d9.zip/node_modules/moment/",\ + ["npm:2.29.4", {\ + "packageLocation": "./.yarn/cache/moment-npm-2.29.4-902943305d-0ec3f9c2bc.zip/node_modules/moment/",\ "packageDependencies": [\ - ["moment", "npm:2.29.3"]\ + ["moment", "npm:2.29.4"]\ ],\ "linkType": "HARD"\ }],\ - ["npm:2.29.4", {\ - "packageLocation": "./.yarn/cache/moment-npm-2.29.4-902943305d-0ec3f9c2bc.zip/node_modules/moment/",\ + ["npm:2.30.1", {\ + "packageLocation": "./.yarn/cache/moment-npm-2.30.1-1c51a5c631-859236bab1.zip/node_modules/moment/",\ "packageDependencies": [\ - ["moment", "npm:2.29.4"]\ + ["moment", "npm:2.30.1"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["moment-timezone", [\ - ["npm:0.5.40", {\ - "packageLocation": "./.yarn/cache/moment-timezone-npm-0.5.40-873e898229-6f6be5412b.zip/node_modules/moment-timezone/",\ + ["npm:0.5.45", {\ + "packageLocation": "./.yarn/cache/moment-timezone-npm-0.5.45-2df3ad72a4-a22e9f983f.zip/node_modules/moment-timezone/",\ "packageDependencies": [\ - ["moment-timezone", "npm:0.5.40"],\ - ["moment", "npm:2.29.3"]\ + ["moment-timezone", "npm:0.5.45"],\ + ["moment", "npm:2.29.4"]\ ],\ "linkType": "HARD"\ }]\ @@ -6197,6 +7151,14 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["msgpackr", [\ + ["npm:1.10.1", {\ + "packageLocation": "./.yarn/cache/msgpackr-npm-1.10.1-5c5ff5c553-e422d18b01.zip/node_modules/msgpackr/",\ + "packageDependencies": [\ + ["msgpackr", "npm:1.10.1"],\ + ["msgpackr-extract", "npm:3.0.2"]\ + ],\ + "linkType": "HARD"\ + }],\ ["npm:1.6.0", {\ "packageLocation": "./.yarn/cache/msgpackr-npm-1.6.0-de9303a46e-7f94acbe93.zip/node_modules/msgpackr/",\ "packageDependencies": [\ @@ -6204,6 +7166,14 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["msgpackr-extract", "npm:2.0.2"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:1.9.9", {\ + "packageLocation": "./.yarn/cache/msgpackr-npm-1.9.9-75b366d55f-b63182d99f.zip/node_modules/msgpackr/",\ + "packageDependencies": [\ + ["msgpackr", "npm:1.9.9"],\ + ["msgpackr-extract", "npm:3.0.2"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["msgpackr-extract", [\ @@ -6221,6 +7191,30 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["node-gyp-build-optional-packages", "npm:5.0.2"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:3.0.2", {\ + "packageLocation": "./.yarn/unplugged/msgpackr-extract-npm-3.0.2-93e8773fad/node_modules/msgpackr-extract/",\ + "packageDependencies": [\ + ["msgpackr-extract", "npm:3.0.2"],\ + ["@msgpackr-extract/msgpackr-extract-darwin-arm64", "npm:3.0.2"],\ + ["@msgpackr-extract/msgpackr-extract-darwin-x64", "npm:3.0.2"],\ + ["@msgpackr-extract/msgpackr-extract-linux-arm", "npm:3.0.2"],\ + ["@msgpackr-extract/msgpackr-extract-linux-arm64", "npm:3.0.2"],\ + ["@msgpackr-extract/msgpackr-extract-linux-x64", "npm:3.0.2"],\ + ["@msgpackr-extract/msgpackr-extract-win32-x64", "npm:3.0.2"],\ + ["node-gyp", "npm:9.0.0"],\ + ["node-gyp-build-optional-packages", "npm:5.0.7"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["muggle-string", [\ + ["npm:0.4.1", {\ + "packageLocation": "./.yarn/cache/muggle-string-npm-0.4.1-fe3c825cc2-85fe1766d1.zip/node_modules/muggle-string/",\ + "packageDependencies": [\ + ["muggle-string", "npm:0.4.1"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["murmurhash-js", [\ @@ -6233,37 +7227,38 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["naive-ui", [\ - ["npm:2.34.3", {\ - "packageLocation": "./.yarn/cache/naive-ui-npm-2.34.3-ba2dfb08d8-792d9e6c51.zip/node_modules/naive-ui/",\ + ["npm:2.38.1", {\ + "packageLocation": "./.yarn/cache/naive-ui-npm-2.38.1-0edd2e5816-88a8f981de.zip/node_modules/naive-ui/",\ "packageDependencies": [\ - ["naive-ui", "npm:2.34.3"]\ + ["naive-ui", "npm:2.38.1"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.34.3", {\ - "packageLocation": "./.yarn/__virtual__/naive-ui-virtual-4fa5810747/0/cache/naive-ui-npm-2.34.3-ba2dfb08d8-792d9e6c51.zip/node_modules/naive-ui/",\ - "packageDependencies": [\ - ["naive-ui", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.34.3"],\ - ["@css-render/plugin-bem", "virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:0.15.10"],\ - ["@css-render/vue3-ssr", "virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:0.15.10"],\ - ["@types/katex", "npm:0.14.0"],\ - ["@types/lodash", "npm:4.14.182"],\ - ["@types/lodash-es", "npm:4.17.6"],\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.38.1", {\ + "packageLocation": "./.yarn/__virtual__/naive-ui-virtual-32fd9c861d/0/cache/naive-ui-npm-2.38.1-0edd2e5816-88a8f981de.zip/node_modules/naive-ui/",\ + "packageDependencies": [\ + ["naive-ui", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.38.1"],\ + ["@css-render/plugin-bem", "virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:0.15.12"],\ + ["@css-render/vue3-ssr", "virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:0.15.12"],\ + ["@types/katex", "npm:0.16.5"],\ + ["@types/lodash", "npm:4.14.200"],\ + ["@types/lodash-es", "npm:4.17.10"],\ ["@types/vue", null],\ - ["async-validator", "npm:4.1.1"],\ - ["css-render", "npm:0.15.10"],\ - ["date-fns", "npm:2.28.0"],\ - ["date-fns-tz", "virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:1.3.3"],\ + ["async-validator", "npm:4.2.5"],\ + ["css-render", "npm:0.15.12"],\ + ["csstype", "npm:3.1.3"],\ + ["date-fns", "npm:2.30.0"],\ + ["date-fns-tz", "virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:2.0.0"],\ ["evtd", "npm:0.2.4"],\ - ["highlight.js", "npm:11.5.1"],\ + ["highlight.js", "npm:11.9.0"],\ ["lodash", "npm:4.17.21"],\ ["lodash-es", "npm:4.17.21"],\ - ["seemly", "npm:0.3.6"],\ + ["seemly", "npm:0.3.8"],\ ["treemate", "npm:0.3.11"],\ - ["vdirs", "virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:0.1.8"],\ - ["vooks", "virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:0.2.12"],\ - ["vue", "npm:3.2.45"],\ - ["vueuc", "virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:0.4.47"]\ + ["vdirs", "virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:0.1.8"],\ + ["vooks", "virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:0.2.12"],\ + ["vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.21"],\ + ["vueuc", "virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:0.4.58"]\ ],\ "packagePeers": [\ "@types/vue",\ @@ -6273,17 +7268,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["nanoid", [\ - ["npm:3.3.3", {\ - "packageLocation": "./.yarn/cache/nanoid-npm-3.3.3-25d865be84-ada019402a.zip/node_modules/nanoid/",\ - "packageDependencies": [\ - ["nanoid", "npm:3.3.3"]\ - ],\ - "linkType": "HARD"\ - }],\ - ["npm:3.3.4", {\ - "packageLocation": "./.yarn/cache/nanoid-npm-3.3.4-3d250377d6-2fddd6dee9.zip/node_modules/nanoid/",\ + ["npm:3.3.7", {\ + "packageLocation": "./.yarn/cache/nanoid-npm-3.3.7-98824ba130-d36c427e53.zip/node_modules/nanoid/",\ "packageDependencies": [\ - ["nanoid", "npm:3.3.4"]\ + ["nanoid", "npm:3.3.7"]\ ],\ "linkType": "HARD"\ }]\ @@ -6322,6 +7310,14 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["node-gyp", "npm:9.0.0"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:6.1.0", {\ + "packageLocation": "./.yarn/unplugged/node-addon-api-npm-6.1.0-634c545b39/node_modules/node-addon-api/",\ + "packageDependencies": [\ + ["node-addon-api", "npm:6.1.0"],\ + ["node-gyp", "npm:9.0.0"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["node-gyp", [\ @@ -6366,6 +7362,21 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["node-gyp-build-optional-packages", "npm:5.0.3"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:5.0.7", {\ + "packageLocation": "./.yarn/cache/node-gyp-build-optional-packages-npm-5.0.7-40f21a5d68-bcb4537af1.zip/node_modules/node-gyp-build-optional-packages/",\ + "packageDependencies": [\ + ["node-gyp-build-optional-packages", "npm:5.0.7"]\ + ],\ + "linkType": "HARD"\ + }],\ + ["npm:5.1.1", {\ + "packageLocation": "./.yarn/cache/node-gyp-build-optional-packages-npm-5.1.1-ff11e179dd-f3cb197862.zip/node_modules/node-gyp-build-optional-packages/",\ + "packageDependencies": [\ + ["node-gyp-build-optional-packages", "npm:5.1.1"],\ + ["detect-libc", "npm:2.0.2"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["node-releases", [\ @@ -6444,6 +7455,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["object-inspect", "npm:1.12.0"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:1.13.1", {\ + "packageLocation": "./.yarn/cache/object-inspect-npm-1.13.1-fd038a2f0a-7d9fa9221d.zip/node_modules/object-inspect/",\ + "packageDependencies": [\ + ["object-inspect", "npm:1.13.1"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["object-keys", [\ @@ -6456,10 +7474,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["object.assign", [\ - ["npm:4.1.2", {\ - "packageLocation": "./.yarn/cache/object.assign-npm-4.1.2-d52edada1c-d621d832ed.zip/node_modules/object.assign/",\ + ["npm:4.1.4", {\ + "packageLocation": "./.yarn/cache/object.assign-npm-4.1.4-fb3deb1c3a-76cab513a5.zip/node_modules/object.assign/",\ "packageDependencies": [\ - ["object.assign", "npm:4.1.2"],\ + ["object.assign", "npm:4.1.4"],\ ["call-bind", "npm:1.0.2"],\ ["define-properties", "npm:1.1.4"],\ ["has-symbols", "npm:1.0.3"],\ @@ -6468,14 +7486,39 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["object.fromentries", [\ + ["npm:2.0.7", {\ + "packageLocation": "./.yarn/cache/object.fromentries-npm-2.0.7-2e38392540-7341ce246e.zip/node_modules/object.fromentries/",\ + "packageDependencies": [\ + ["object.fromentries", "npm:2.0.7"],\ + ["call-bind", "npm:1.0.2"],\ + ["define-properties", "npm:1.2.0"],\ + ["es-abstract", "npm:1.22.3"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["object.groupby", [\ + ["npm:1.0.1", {\ + "packageLocation": "./.yarn/cache/object.groupby-npm-1.0.1-fc268391fe-d7959d6eaa.zip/node_modules/object.groupby/",\ + "packageDependencies": [\ + ["object.groupby", "npm:1.0.1"],\ + ["call-bind", "npm:1.0.2"],\ + ["define-properties", "npm:1.2.0"],\ + ["es-abstract", "npm:1.22.3"],\ + ["get-intrinsic", "npm:1.2.1"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["object.values", [\ - ["npm:1.1.5", {\ - "packageLocation": "./.yarn/cache/object.values-npm-1.1.5-f1de7f3742-0f17e99741.zip/node_modules/object.values/",\ + ["npm:1.1.7", {\ + "packageLocation": "./.yarn/cache/object.values-npm-1.1.7-deae619f88-f3e4ae4f21.zip/node_modules/object.values/",\ "packageDependencies": [\ - ["object.values", "npm:1.1.5"],\ + ["object.values", "npm:1.1.7"],\ ["call-bind", "npm:1.0.2"],\ - ["define-properties", "npm:1.1.4"],\ - ["es-abstract", "npm:1.19.5"]\ + ["define-properties", "npm:1.2.0"],\ + ["es-abstract", "npm:1.22.3"]\ ],\ "linkType": "HARD"\ }]\ @@ -6501,16 +7544,16 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["optionator", [\ - ["npm:0.9.1", {\ - "packageLocation": "./.yarn/cache/optionator-npm-0.9.1-577e397aae-dbc6fa0656.zip/node_modules/optionator/",\ + ["npm:0.9.3", {\ + "packageLocation": "./.yarn/cache/optionator-npm-0.9.3-56c3a4bf80-0928199944.zip/node_modules/optionator/",\ "packageDependencies": [\ - ["optionator", "npm:0.9.1"],\ + ["optionator", "npm:0.9.3"],\ + ["@aashutoshrathi/word-wrap", "npm:1.2.6"],\ ["deep-is", "npm:0.1.4"],\ ["fast-levenshtein", "npm:2.0.6"],\ ["levn", "npm:0.4.1"],\ ["prelude-ls", "npm:1.2.1"],\ - ["type-check", "npm:0.4.0"],\ - ["word-wrap", "npm:1.2.3"]\ + ["type-check", "npm:0.4.0"]\ ],\ "linkType": "HARD"\ }]\ @@ -6522,17 +7565,16 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["ordered-binary", "npm:1.2.5"]\ ],\ "linkType": "HARD"\ - }]\ - ]],\ - ["p-limit", [\ - ["npm:1.3.0", {\ - "packageLocation": "./.yarn/cache/p-limit-npm-1.3.0-fdb471d864-281c1c0b8c.zip/node_modules/p-limit/",\ + }],\ + ["npm:1.4.1", {\ + "packageLocation": "./.yarn/cache/ordered-binary-npm-1.4.1-9ad6b7c6b5-274940b4ef.zip/node_modules/ordered-binary/",\ "packageDependencies": [\ - ["p-limit", "npm:1.3.0"],\ - ["p-try", "npm:1.0.0"]\ + ["ordered-binary", "npm:1.4.1"]\ ],\ "linkType": "HARD"\ - }],\ + }]\ + ]],\ + ["p-limit", [\ ["npm:3.1.0", {\ "packageLocation": "./.yarn/cache/p-limit-npm-3.1.0-05d2ede37f-7c3690c4db.zip/node_modules/p-limit/",\ "packageDependencies": [\ @@ -6543,14 +7585,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["p-locate", [\ - ["npm:2.0.0", {\ - "packageLocation": "./.yarn/cache/p-locate-npm-2.0.0-3a2ee263dd-e2dceb9b49.zip/node_modules/p-locate/",\ - "packageDependencies": [\ - ["p-locate", "npm:2.0.0"],\ - ["p-limit", "npm:1.3.0"]\ - ],\ - "linkType": "HARD"\ - }],\ ["npm:5.0.0", {\ "packageLocation": "./.yarn/cache/p-locate-npm-5.0.0-92cc7c7a3e-1623088f36.zip/node_modules/p-locate/",\ "packageDependencies": [\ @@ -6570,42 +7604,33 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ - ["p-try", [\ - ["npm:1.0.0", {\ - "packageLocation": "./.yarn/cache/p-try-npm-1.0.0-7373139e40-3b5303f77e.zip/node_modules/p-try/",\ - "packageDependencies": [\ - ["p-try", "npm:1.0.0"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ ["parcel", [\ - ["npm:2.8.2", {\ - "packageLocation": "./.yarn/cache/parcel-npm-2.8.2-7cad55fa52-b95ef40bad.zip/node_modules/parcel/",\ + ["npm:2.12.0", {\ + "packageLocation": "./.yarn/cache/parcel-npm-2.12.0-96a4bb6cc3-d8e6cb690a.zip/node_modules/parcel/",\ "packageDependencies": [\ - ["parcel", "npm:2.8.2"]\ + ["parcel", "npm:2.12.0"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.8.2", {\ - "packageLocation": "./.yarn/__virtual__/parcel-virtual-7ff1d17261/0/cache/parcel-npm-2.8.2-7cad55fa52-b95ef40bad.zip/node_modules/parcel/",\ - "packageDependencies": [\ - ["parcel", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.8.2"],\ - ["@parcel/config-default", "virtual:7ff1d17261e888ed45770ab8af325407870f62c23cd6264c3a2830a9a45cf064c4196c0c92d06cfbc9b69b3bbe2e4c7776dce9f9d39af95141f0808c9e3cc9ec#npm:2.8.2"],\ - ["@parcel/core", "npm:2.8.2"],\ - ["@parcel/diagnostic", "npm:2.8.2"],\ - ["@parcel/events", "npm:2.8.2"],\ - ["@parcel/fs", "virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2"],\ - ["@parcel/logger", "npm:2.8.2"],\ - ["@parcel/package-manager", "virtual:7ac9ecd9f9189cad6b089bc5c3344f5cd08cd89057c51a89b1d7e9a7387c7270d501c42325a9d3479c01962ba802b947b105e454958aa797e67af85ee1b8bbf3#npm:2.8.2"],\ - ["@parcel/reporter-cli", "npm:2.8.2"],\ - ["@parcel/reporter-dev-server", "npm:2.8.2"],\ - ["@parcel/utils", "npm:2.8.2"],\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.12.0", {\ + "packageLocation": "./.yarn/__virtual__/parcel-virtual-fdd74b573c/0/cache/parcel-npm-2.12.0-96a4bb6cc3-d8e6cb690a.zip/node_modules/parcel/",\ + "packageDependencies": [\ + ["parcel", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.12.0"],\ + ["@parcel/config-default", "virtual:fdd74b573cf769bcde15fb47c39fbe0d73f59838182900fd59d3d43b2214ea01b1d45084fb49d0c192fc3e8a49adea5782afcb7fe14e09c63bedaf09f4939e35#npm:2.12.0"],\ + ["@parcel/core", "npm:2.12.0"],\ + ["@parcel/diagnostic", "npm:2.12.0"],\ + ["@parcel/events", "npm:2.12.0"],\ + ["@parcel/fs", "virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0"],\ + ["@parcel/logger", "npm:2.12.0"],\ + ["@parcel/package-manager", "virtual:8f08b883d4cc438aa2ec719eb5cec278f9ea627197c55f35530bcaf9cd4e4738e04be8abe946bd2702b3f5c94b812f529f1b87c05c7d6de04e1ade9b3f3e00f6#npm:2.12.0"],\ + ["@parcel/reporter-cli", "npm:2.12.0"],\ + ["@parcel/reporter-dev-server", "npm:2.12.0"],\ + ["@parcel/reporter-tracer", "npm:2.12.0"],\ + ["@parcel/utils", "npm:2.12.0"],\ ["@types/parcel__core", null],\ ["chalk", "npm:4.1.2"],\ ["commander", "npm:7.2.0"],\ - ["get-port", "npm:4.2.0"],\ - ["v8-compile-cache", "npm:2.3.0"]\ + ["get-port", "npm:4.2.0"]\ ],\ "packagePeers": [\ "@types/parcel__core"\ @@ -6637,13 +7662,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["path-exists", [\ - ["npm:3.0.0", {\ - "packageLocation": "./.yarn/cache/path-exists-npm-3.0.0-e80371aa68-96e92643aa.zip/node_modules/path-exists/",\ - "packageDependencies": [\ - ["path-exists", "npm:3.0.0"]\ - ],\ - "linkType": "HARD"\ - }],\ ["npm:4.0.0", {\ "packageLocation": "./.yarn/cache/path-exists-npm-4.0.0-e9e4f63eb0-505807199d.zip/node_modules/path-exists/",\ "packageDependencies": [\ @@ -6679,6 +7697,17 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["path-scurry", [\ + ["npm:1.9.1", {\ + "packageLocation": "./.yarn/cache/path-scurry-npm-1.9.1-b9d6b1c5bf-28caa788f1.zip/node_modules/path-scurry/",\ + "packageDependencies": [\ + ["path-scurry", "npm:1.9.1"],\ + ["lru-cache", "npm:9.1.1"],\ + ["minipass", "npm:6.0.1"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["path-type", [\ ["npm:4.0.0", {\ "packageLocation": "./.yarn/cache/path-type-npm-4.0.0-10d47fc86a-5b1e2daa24.zip/node_modules/path-type/",\ @@ -6707,25 +7736,25 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["pinia", [\ - ["npm:2.0.28", {\ - "packageLocation": "./.yarn/cache/pinia-npm-2.0.28-9b0289223e-d515cd6220.zip/node_modules/pinia/",\ + ["npm:2.1.7", {\ + "packageLocation": "./.yarn/cache/pinia-npm-2.1.7-195409c154-1b7882aab2.zip/node_modules/pinia/",\ "packageDependencies": [\ - ["pinia", "npm:2.0.28"]\ + ["pinia", "npm:2.1.7"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.0.28", {\ - "packageLocation": "./.yarn/__virtual__/pinia-virtual-e1f94167f7/0/cache/pinia-npm-2.0.28-9b0289223e-d515cd6220.zip/node_modules/pinia/",\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.1.7", {\ + "packageLocation": "./.yarn/__virtual__/pinia-virtual-cf6f7439ee/0/cache/pinia-npm-2.1.7-195409c154-1b7882aab2.zip/node_modules/pinia/",\ "packageDependencies": [\ - ["pinia", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.0.28"],\ + ["pinia", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.1.7"],\ ["@types/typescript", null],\ ["@types/vue", null],\ ["@types/vue__composition-api", null],\ ["@vue/composition-api", null],\ - ["@vue/devtools-api", "npm:6.4.5"],\ + ["@vue/devtools-api", "npm:6.5.0"],\ ["typescript", null],\ - ["vue", "npm:3.2.45"],\ - ["vue-demi", "virtual:e1f94167f73dcd0042012cf3a978b5096a058d81d2d3813652fe465a5fe7d84e810c9bea1e50af574784bc4294f4f838b7ba55cf745fd1337e972f45d3531d5e#npm:0.13.1"]\ + ["vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.21"],\ + ["vue-demi", "virtual:cf6f7439ee76dfd2e7f8f2565ae847d76901434fc49c65702190cdf3d1c61e61c701a5c45b514c4bdeacb8f4bcac9c8a98bd4db3d0bc8e403d9e8db2cf14372a#npm:0.14.5"]\ ],\ "packagePeers": [\ "@types/typescript",\ @@ -6754,8 +7783,8 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@types/vue", null],\ ["@types/vue__composition-api", null],\ ["@vue/composition-api", null],\ - ["pinia", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.0.28"],\ - ["vue", "npm:3.2.45"],\ + ["pinia", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.1.7"],\ + ["vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.21"],\ ["vue-demi", "virtual:f56fcf19bbebc2ada1b28955da8cc216b1e9a569a1a7337d2d1926c1ebd1bc7a5bd91aedae1d05c15c8562f33caf7c59bd3020a667340f6bdc6a7b13fc2ba847#npm:0.12.5"]\ ],\ "packagePeers": [\ @@ -6770,21 +7799,21 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["postcss", [\ - ["npm:8.4.12", {\ - "packageLocation": "./.yarn/cache/postcss-npm-8.4.12-e941d78a98-248e3d0f9b.zip/node_modules/postcss/",\ + ["npm:8.4.33", {\ + "packageLocation": "./.yarn/cache/postcss-npm-8.4.33-6ba8157009-6f98b2af4b.zip/node_modules/postcss/",\ "packageDependencies": [\ - ["postcss", "npm:8.4.12"],\ - ["nanoid", "npm:3.3.3"],\ + ["postcss", "npm:8.4.33"],\ + ["nanoid", "npm:3.3.7"],\ ["picocolors", "npm:1.0.0"],\ ["source-map-js", "npm:1.0.2"]\ ],\ "linkType": "HARD"\ }],\ - ["npm:8.4.18", {\ - "packageLocation": "./.yarn/cache/postcss-npm-8.4.18-f1d73c0a84-9349fd9984.zip/node_modules/postcss/",\ + ["npm:8.4.35", {\ + "packageLocation": "./.yarn/cache/postcss-npm-8.4.35-6bc1848fff-cf3c3124d3.zip/node_modules/postcss/",\ "packageDependencies": [\ - ["postcss", "npm:8.4.18"],\ - ["nanoid", "npm:3.3.4"],\ + ["postcss", "npm:8.4.35"],\ + ["nanoid", "npm:3.3.7"],\ ["picocolors", "npm:1.0.0"],\ ["source-map-js", "npm:1.0.2"]\ ],\ @@ -6792,10 +7821,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["postcss-selector-parser", [\ - ["npm:6.0.10", {\ - "packageLocation": "./.yarn/cache/postcss-selector-parser-npm-6.0.10-a4d7aaa270-46afaa60e3.zip/node_modules/postcss-selector-parser/",\ + ["npm:6.0.15", {\ + "packageLocation": "./.yarn/cache/postcss-selector-parser-npm-6.0.15-0ec4819b4e-57decb9415.zip/node_modules/postcss-selector-parser/",\ "packageDependencies": [\ - ["postcss-selector-parser", "npm:6.0.10"],\ + ["postcss-selector-parser", "npm:6.0.15"],\ ["cssesc", "npm:3.0.0"],\ ["util-deprecate", "npm:1.0.2"]\ ],\ @@ -6851,10 +7880,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["preact", [\ - ["npm:10.7.2", {\ - "packageLocation": "./.yarn/cache/preact-npm-10.7.2-dffb68bd4b-2f0655e043.zip/node_modules/preact/",\ + ["npm:10.12.1", {\ + "packageLocation": "./.yarn/cache/preact-npm-10.12.1-fdb903e9a5-0de99f4775.zip/node_modules/preact/",\ "packageDependencies": [\ - ["preact", "npm:10.7.2"]\ + ["preact", "npm:10.12.1"]\ ],\ "linkType": "HARD"\ }]\ @@ -6868,18 +7897,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ - ["pretty-format", [\ - ["npm:27.5.1", {\ - "packageLocation": "./.yarn/cache/pretty-format-npm-27.5.1-cd7d49696f-cf610cffcb.zip/node_modules/pretty-format/",\ - "packageDependencies": [\ - ["pretty-format", "npm:27.5.1"],\ - ["ansi-regex", "npm:5.0.1"],\ - ["ansi-styles", "npm:5.2.0"],\ - ["react-is", "npm:17.0.2"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ ["promise", [\ ["npm:7.3.1", {\ "packageLocation": "./.yarn/cache/promise-npm-7.3.1-5d81d474c0-475bb06913.zip/node_modules/promise/",\ @@ -7112,15 +8129,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ - ["react-is", [\ - ["npm:17.0.2", {\ - "packageLocation": "./.yarn/cache/react-is-npm-17.0.2-091bbb8db6-9d6d111d89.zip/node_modules/react-is/",\ - "packageDependencies": [\ - ["react-is", "npm:17.0.2"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ ["react-refresh", [\ ["npm:0.9.0", {\ "packageLocation": "./.yarn/cache/react-refresh-npm-0.9.0-02c61ee045-6440146176.zip/node_modules/react-refresh/",\ @@ -7146,17 +8154,36 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["npm:3.6.0", {\ "packageLocation": "./.yarn/cache/readdirp-npm-3.6.0-f950cc74ab-1ced032e6e.zip/node_modules/readdirp/",\ "packageDependencies": [\ - ["readdirp", "npm:3.6.0"],\ - ["picomatch", "npm:2.3.1"]\ + ["readdirp", "npm:3.6.0"],\ + ["picomatch", "npm:2.3.1"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["regenerator-runtime", [\ + ["npm:0.13.9", {\ + "packageLocation": "./.yarn/cache/regenerator-runtime-npm-0.13.9-6d02340eec-65ed455fe5.zip/node_modules/regenerator-runtime/",\ + "packageDependencies": [\ + ["regenerator-runtime", "npm:0.13.9"]\ + ],\ + "linkType": "HARD"\ + }],\ + ["npm:0.14.0", {\ + "packageLocation": "./.yarn/cache/regenerator-runtime-npm-0.14.0-e060897cf7-1c977ad82a.zip/node_modules/regenerator-runtime/",\ + "packageDependencies": [\ + ["regenerator-runtime", "npm:0.14.0"]\ ],\ "linkType": "HARD"\ }]\ ]],\ - ["regenerator-runtime", [\ - ["npm:0.13.9", {\ - "packageLocation": "./.yarn/cache/regenerator-runtime-npm-0.13.9-6d02340eec-65ed455fe5.zip/node_modules/regenerator-runtime/",\ + ["regexp.prototype.flags", [\ + ["npm:1.5.1", {\ + "packageLocation": "./.yarn/cache/regexp.prototype.flags-npm-1.5.1-b8faeee306-869edff002.zip/node_modules/regexp.prototype.flags/",\ "packageDependencies": [\ - ["regenerator-runtime", "npm:0.13.9"]\ + ["regexp.prototype.flags", "npm:1.5.1"],\ + ["call-bind", "npm:1.0.2"],\ + ["define-properties", "npm:1.2.0"],\ + ["set-function-name", "npm:2.0.1"]\ ],\ "linkType": "HARD"\ }]\ @@ -7199,11 +8226,21 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "HARD"\ }],\ - ["patch:resolve@npm%3A1.22.1#~builtin::version=1.22.1&hash=07638b", {\ - "packageLocation": "./.yarn/cache/resolve-patch-46f9469d0d-5656f4d0be.zip/node_modules/resolve/",\ + ["patch:resolve@npm%3A1.22.3#~builtin::version=1.22.3&hash=07638b", {\ + "packageLocation": "./.yarn/cache/resolve-patch-8df1eb26d0-ad59734723.zip/node_modules/resolve/",\ "packageDependencies": [\ - ["resolve", "patch:resolve@npm%3A1.22.1#~builtin::version=1.22.1&hash=07638b"],\ - ["is-core-module", "npm:2.10.0"],\ + ["resolve", "patch:resolve@npm%3A1.22.3#~builtin::version=1.22.3&hash=07638b"],\ + ["is-core-module", "npm:2.12.1"],\ + ["path-parse", "npm:1.0.7"],\ + ["supports-preserve-symlinks-flag", "npm:1.0.0"]\ + ],\ + "linkType": "HARD"\ + }],\ + ["patch:resolve@npm%3A1.22.8#~builtin::version=1.22.8&hash=07638b", {\ + "packageLocation": "./.yarn/cache/resolve-patch-f6b5304cab-5479b7d431.zip/node_modules/resolve/",\ + "packageDependencies": [\ + ["resolve", "patch:resolve@npm%3A1.22.8#~builtin::version=1.22.8&hash=07638b"],\ + ["is-core-module", "npm:2.13.0"],\ ["path-parse", "npm:1.0.7"],\ ["supports-preserve-symlinks-flag", "npm:1.0.0"]\ ],\ @@ -7219,6 +8256,15 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["resolve-pkg-maps", [\ + ["npm:1.0.0", {\ + "packageLocation": "./.yarn/cache/resolve-pkg-maps-npm-1.0.0-135b70c854-1012afc566.zip/node_modules/resolve-pkg-maps/",\ + "packageDependencies": [\ + ["resolve-pkg-maps", "npm:1.0.0"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["retry", [\ ["npm:0.12.0", {\ "packageLocation": "./.yarn/cache/retry-npm-0.12.0-72ac7fb4cc-623bd7d2e5.zip/node_modules/retry/",\ @@ -7257,10 +8303,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["rollup", [\ - ["npm:2.79.1", {\ - "packageLocation": "./.yarn/cache/rollup-npm-2.79.1-94e707a9a3-6a2bf167b3.zip/node_modules/rollup/",\ + ["npm:3.29.4", {\ + "packageLocation": "./.yarn/cache/rollup-npm-3.29.4-5e5e5f2087-8bb20a39c8.zip/node_modules/rollup/",\ "packageDependencies": [\ - ["rollup", "npm:2.79.1"],\ + ["rollup", "npm:3.29.4"],\ ["fsevents", "patch:fsevents@npm%3A2.3.2#~builtin::version=2.3.2&hash=18f3a7"]\ ],\ "linkType": "HARD"\ @@ -7271,68 +8317,70 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "packageLocation": "./",\ "packageDependencies": [\ ["root-workspace-0b6124", "workspace:."],\ - ["@faker-js/faker", "npm:7.6.0"],\ - ["@fullcalendar/bootstrap5", "npm:5.11.3"],\ - ["@fullcalendar/core", "npm:5.11.3"],\ - ["@fullcalendar/daygrid", "npm:5.11.3"],\ - ["@fullcalendar/interaction", "npm:5.11.3"],\ - ["@fullcalendar/list", "npm:5.11.3"],\ - ["@fullcalendar/luxon2", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.11.3"],\ - ["@fullcalendar/timegrid", "npm:5.11.3"],\ - ["@fullcalendar/vue3", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.11.3"],\ - ["@parcel/optimizer-data-url", "npm:2.8.2"],\ - ["@parcel/transformer-inline-string", "npm:2.8.2"],\ - ["@parcel/transformer-sass", "npm:2.8.2"],\ - ["@popperjs/core", "npm:2.11.6"],\ - ["@rollup/pluginutils", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.0.2"],\ + ["@fullcalendar/bootstrap5", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/core", "npm:6.1.11"],\ + ["@fullcalendar/daygrid", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/icalendar", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/interaction", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/list", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/luxon3", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/timegrid", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@fullcalendar/vue3", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.11"],\ + ["@parcel/optimizer-data-url", "npm:2.12.0"],\ + ["@parcel/transformer-inline-string", "npm:2.12.0"],\ + ["@parcel/transformer-sass", "npm:2.12.0"],\ + ["@popperjs/core", "npm:2.11.8"],\ + ["@rollup/pluginutils", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.1.0"],\ ["@twuni/emojify", "npm:1.0.2"],\ - ["@vitejs/plugin-vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.2.0"],\ - ["bootstrap", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.2.3"],\ - ["bootstrap-icons", "npm:1.10.3"],\ - ["browser-fs-access", "npm:0.31.1"],\ + ["@vitejs/plugin-vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:4.6.2"],\ + ["@vue/language-plugin-pug", "npm:2.0.7"],\ + ["bootstrap", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:5.3.3"],\ + ["bootstrap-icons", "npm:1.11.3"],\ + ["browser-fs-access", "npm:0.35.0"],\ ["browserlist", "npm:1.0.1"],\ - ["c8", "npm:7.12.0"],\ - ["caniuse-lite", "npm:1.0.30001442"],\ - ["d3", "npm:7.8.0"],\ - ["eslint", "npm:8.31.0"],\ - ["eslint-config-standard", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:17.0.0"],\ - ["eslint-plugin-cypress", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.12.1"],\ - ["eslint-plugin-import", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.26.0"],\ - ["eslint-plugin-n", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:15.6.1"],\ + ["c8", "npm:9.1.0"],\ + ["caniuse-lite", "npm:1.0.30001603"],\ + ["d3", "npm:7.9.0"],\ + ["eslint", "npm:8.57.0"],\ + ["eslint-config-standard", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:17.1.0"],\ + ["eslint-plugin-cypress", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.15.1"],\ + ["eslint-plugin-import", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.29.1"],\ + ["eslint-plugin-n", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:16.6.2"],\ ["eslint-plugin-node", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:11.1.0"],\ ["eslint-plugin-promise", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:6.1.1"],\ - ["eslint-plugin-vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:9.8.0"],\ + ["eslint-plugin-vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:9.24.0"],\ ["file-saver", "npm:2.0.5"],\ - ["highcharts", "npm:10.3.2"],\ - ["html-validate", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:7.12.2"],\ - ["jquery", "npm:3.6.3"],\ - ["jquery-migrate", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.0"],\ - ["jquery-ui-dist", "npm:1.13.2"],\ - ["js-cookie", "npm:3.0.1"],\ + ["highcharts", "npm:11.4.0"],\ + ["html-validate", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:8.18.1"],\ + ["ical.js", "npm:1.5.0"],\ + ["jquery", "npm:3.7.1"],\ + ["jquery-migrate", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.1"],\ + ["js-cookie", "npm:3.0.5"],\ ["list.js", "npm:2.3.1"],\ ["lodash", "npm:4.17.21"],\ ["lodash-es", "npm:4.17.21"],\ - ["luxon", "npm:3.2.1"],\ - ["moment", "npm:2.29.4"],\ - ["moment-timezone", "npm:0.5.40"],\ + ["luxon", "npm:3.4.4"],\ + ["moment", "npm:2.30.1"],\ + ["moment-timezone", "npm:0.5.45"],\ ["ms", "npm:2.1.3"],\ ["murmurhash-js", "npm:1.0.0"],\ - ["naive-ui", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.34.3"],\ - ["parcel", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.8.2"],\ - ["pinia", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.0.28"],\ + ["naive-ui", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.38.1"],\ + ["parcel", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.12.0"],\ + ["pinia", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:2.1.7"],\ ["pinia-plugin-persist", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:1.0.0"],\ ["pug", "npm:3.0.2"],\ - ["sass", "npm:1.57.1"],\ + ["sass", "npm:1.72.0"],\ ["seedrandom", "npm:3.0.5"],\ ["select2", "npm:4.1.0-rc.0"],\ ["select2-bootstrap-5-theme", "npm:1.3.0"],\ ["send", "npm:0.18.0"],\ - ["shepherd.js", "npm:10.0.1"],\ - ["slugify", "npm:1.6.5"],\ - ["sortablejs", "npm:1.15.0"],\ - ["vite", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.2.5"],\ - ["vue", "npm:3.2.45"],\ - ["vue-router", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:4.1.6"],\ + ["shepherd.js", "npm:11.2.0"],\ + ["slugify", "npm:1.6.6"],\ + ["sortablejs", "npm:1.15.2"],\ + ["vanillajs-datepicker", "npm:1.3.4"],\ + ["vite", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:4.5.3"],\ + ["vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.21"],\ + ["vue-router", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:4.3.0"],\ ["zxcvbn", "npm:4.4.2"]\ ],\ "linkType": "SOFT"\ @@ -7357,6 +8405,19 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["safe-array-concat", [\ + ["npm:1.0.1", {\ + "packageLocation": "./.yarn/cache/safe-array-concat-npm-1.0.1-8a42907bbf-001ecf1d8a.zip/node_modules/safe-array-concat/",\ + "packageDependencies": [\ + ["safe-array-concat", "npm:1.0.1"],\ + ["call-bind", "npm:1.0.2"],\ + ["get-intrinsic", "npm:1.2.1"],\ + ["has-symbols", "npm:1.0.3"],\ + ["isarray", "npm:2.0.5"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["safe-buffer", [\ ["npm:5.1.2", {\ "packageLocation": "./.yarn/cache/safe-buffer-npm-5.1.2-c27fedf6c4-f2f1f7943c.zip/node_modules/safe-buffer/",\ @@ -7373,6 +8434,18 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["safe-regex-test", [\ + ["npm:1.0.0", {\ + "packageLocation": "./.yarn/cache/safe-regex-test-npm-1.0.0-e94a09b84e-bc566d8beb.zip/node_modules/safe-regex-test/",\ + "packageDependencies": [\ + ["safe-regex-test", "npm:1.0.0"],\ + ["call-bind", "npm:1.0.2"],\ + ["get-intrinsic", "npm:1.2.0"],\ + ["is-regex", "npm:1.1.4"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["safer-buffer", [\ ["npm:2.1.2", {\ "packageLocation": "./.yarn/cache/safer-buffer-npm-2.1.2-8d5c0b705e-cab8f25ae6.zip/node_modules/safer-buffer/",\ @@ -7393,10 +8466,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "HARD"\ }],\ - ["npm:1.57.1", {\ - "packageLocation": "./.yarn/cache/sass-npm-1.57.1-bafdba484f-734a08781b.zip/node_modules/sass/",\ + ["npm:1.72.0", {\ + "packageLocation": "./.yarn/cache/sass-npm-1.72.0-fb38bb530c-f420079c7d.zip/node_modules/sass/",\ "packageDependencies": [\ - ["sass", "npm:1.57.1"],\ + ["sass", "npm:1.72.0"],\ ["chokidar", "npm:3.5.3"],\ ["immutable", "npm:4.0.0"],\ ["source-map-js", "npm:1.0.2"]\ @@ -7414,18 +8487,17 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["seemly", [\ - ["npm:0.3.3", {\ - "packageLocation": "./.yarn/cache/seemly-npm-0.3.3-1df3254399-b6445553f8.zip/node_modules/seemly/",\ + ["npm:0.3.6", {\ + "packageLocation": "./.yarn/cache/seemly-npm-0.3.6-87ae398976-56d0472d99.zip/node_modules/seemly/",\ "packageDependencies": [\ - ["seemly", "npm:0.3.3"],\ - ["@types/jest", "npm:27.4.1"]\ + ["seemly", "npm:0.3.6"]\ ],\ "linkType": "HARD"\ }],\ - ["npm:0.3.6", {\ - "packageLocation": "./.yarn/cache/seemly-npm-0.3.6-87ae398976-56d0472d99.zip/node_modules/seemly/",\ + ["npm:0.3.8", {\ + "packageLocation": "./.yarn/cache/seemly-npm-0.3.8-4940336497-98171fd4d9.zip/node_modules/seemly/",\ "packageDependencies": [\ - ["seemly", "npm:0.3.6"]\ + ["seemly", "npm:0.3.8"]\ ],\ "linkType": "HARD"\ }]\ @@ -7465,6 +8537,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "HARD"\ }],\ + ["npm:6.3.1", {\ + "packageLocation": "./.yarn/cache/semver-npm-6.3.1-bcba31fdbe-ae47d06de2.zip/node_modules/semver/",\ + "packageDependencies": [\ + ["semver", "npm:6.3.1"]\ + ],\ + "linkType": "HARD"\ + }],\ ["npm:7.3.7", {\ "packageLocation": "./.yarn/cache/semver-npm-7.3.7-3bfe704194-2fa3e87756.zip/node_modules/semver/",\ "packageDependencies": [\ @@ -7473,10 +8552,26 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "HARD"\ }],\ - ["npm:7.3.8", {\ - "packageLocation": "./.yarn/cache/semver-npm-7.3.8-25a996cb4f-ba9c7cbbf2.zip/node_modules/semver/",\ + ["npm:7.5.3", {\ + "packageLocation": "./.yarn/cache/semver-npm-7.5.3-275095dbf3-9d58db1652.zip/node_modules/semver/",\ + "packageDependencies": [\ + ["semver", "npm:7.5.3"],\ + ["lru-cache", "npm:6.0.0"]\ + ],\ + "linkType": "HARD"\ + }],\ + ["npm:7.5.4", {\ + "packageLocation": "./.yarn/cache/semver-npm-7.5.4-c4ad957fcd-12d8ad952f.zip/node_modules/semver/",\ + "packageDependencies": [\ + ["semver", "npm:7.5.4"],\ + ["lru-cache", "npm:6.0.0"]\ + ],\ + "linkType": "HARD"\ + }],\ + ["npm:7.6.0", {\ + "packageLocation": "./.yarn/cache/semver-npm-7.6.0-f4630729f6-7427f05b70.zip/node_modules/semver/",\ "packageDependencies": [\ - ["semver", "npm:7.3.8"],\ + ["semver", "npm:7.6.0"],\ ["lru-cache", "npm:6.0.0"]\ ],\ "linkType": "HARD"\ @@ -7513,6 +8608,31 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["set-function-length", [\ + ["npm:1.1.1", {\ + "packageLocation": "./.yarn/cache/set-function-length-npm-1.1.1-d362bf8221-c131d7569c.zip/node_modules/set-function-length/",\ + "packageDependencies": [\ + ["set-function-length", "npm:1.1.1"],\ + ["define-data-property", "npm:1.1.1"],\ + ["get-intrinsic", "npm:1.2.1"],\ + ["gopd", "npm:1.0.1"],\ + ["has-property-descriptors", "npm:1.0.0"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["set-function-name", [\ + ["npm:2.0.1", {\ + "packageLocation": "./.yarn/cache/set-function-name-npm-2.0.1-a9f970eea0-4975d17d90.zip/node_modules/set-function-name/",\ + "packageDependencies": [\ + ["set-function-name", "npm:2.0.1"],\ + ["define-data-property", "npm:1.1.1"],\ + ["functions-have-names", "npm:1.2.3"],\ + ["has-property-descriptors", "npm:1.0.0"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["setprototypeof", [\ ["npm:1.2.0", {\ "packageLocation": "./.yarn/cache/setprototypeof-npm-1.2.0-0fedbdcd3a-be18cbbf70.zip/node_modules/setprototypeof/",\ @@ -7542,13 +8662,12 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["shepherd.js", [\ - ["npm:10.0.1", {\ - "packageLocation": "./.yarn/cache/shepherd.js-npm-10.0.1-64acc35968-be51f42734.zip/node_modules/shepherd.js/",\ + ["npm:11.2.0", {\ + "packageLocation": "./.yarn/cache/shepherd.js-npm-11.2.0-94b9af1487-0e71e63e51.zip/node_modules/shepherd.js/",\ "packageDependencies": [\ - ["shepherd.js", "npm:10.0.1"],\ - ["@popperjs/core", "npm:2.11.6"],\ - ["deepmerge", "npm:4.2.2"],\ - ["smoothscroll-polyfill", "npm:0.4.4"]\ + ["shepherd.js", "npm:11.2.0"],\ + ["@floating-ui/dom", "npm:1.5.2"],\ + ["deepmerge", "npm:4.3.1"]\ ],\ "linkType": "HARD"\ }]\ @@ -7572,6 +8691,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["signal-exit", "npm:3.0.7"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:4.0.2", {\ + "packageLocation": "./.yarn/cache/signal-exit-npm-4.0.2-e3f0e8ed25-41f5928431.zip/node_modules/signal-exit/",\ + "packageDependencies": [\ + ["signal-exit", "npm:4.0.2"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["sisteransi", [\ @@ -7584,10 +8710,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["slugify", [\ - ["npm:1.6.5", {\ - "packageLocation": "./.yarn/cache/slugify-npm-1.6.5-6db25d7016-a955a1b600.zip/node_modules/slugify/",\ + ["npm:1.6.6", {\ + "packageLocation": "./.yarn/cache/slugify-npm-1.6.6-7ce458677d-04773c2d3b.zip/node_modules/slugify/",\ "packageDependencies": [\ - ["slugify", "npm:1.6.5"]\ + ["slugify", "npm:1.6.6"]\ ],\ "linkType": "HARD"\ }]\ @@ -7601,15 +8727,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ - ["smoothscroll-polyfill", [\ - ["npm:0.4.4", {\ - "packageLocation": "./.yarn/cache/smoothscroll-polyfill-npm-0.4.4-69b5bb4bf7-b99ff7d916.zip/node_modules/smoothscroll-polyfill/",\ - "packageDependencies": [\ - ["smoothscroll-polyfill", "npm:0.4.4"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ ["socks", [\ ["npm:2.6.2", {\ "packageLocation": "./.yarn/cache/socks-npm-2.6.2-94c1dcb8b8-dd91942930.zip/node_modules/socks/",\ @@ -7634,10 +8751,10 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["sortablejs", [\ - ["npm:1.15.0", {\ - "packageLocation": "./.yarn/cache/sortablejs-npm-1.15.0-f3a393abcc-bb82223a66.zip/node_modules/sortablejs/",\ + ["npm:1.15.2", {\ + "packageLocation": "./.yarn/cache/sortablejs-npm-1.15.2-73347ae85a-36b20b144f.zip/node_modules/sortablejs/",\ "packageDependencies": [\ - ["sortablejs", "npm:1.15.0"]\ + ["sortablejs", "npm:1.15.2"]\ ],\ "linkType": "HARD"\ }]\ @@ -7649,14 +8766,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["source-map", "npm:0.6.1"]\ ],\ "linkType": "HARD"\ - }],\ - ["npm:0.8.0-beta.0", {\ - "packageLocation": "./.yarn/cache/source-map-npm-0.8.0-beta.0-688a309e94-e94169be64.zip/node_modules/source-map/",\ - "packageDependencies": [\ - ["source-map", "npm:0.8.0-beta.0"],\ - ["whatwg-url", "npm:7.1.0"]\ - ],\ - "linkType": "HARD"\ }]\ ]],\ ["source-map-js", [\ @@ -7668,22 +8777,11 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ - ["source-map-support", [\ - ["npm:0.5.21", {\ - "packageLocation": "./.yarn/cache/source-map-support-npm-0.5.21-09ca99e250-43e98d700d.zip/node_modules/source-map-support/",\ - "packageDependencies": [\ - ["source-map-support", "npm:0.5.21"],\ - ["buffer-from", "npm:1.1.2"],\ - ["source-map", "npm:0.6.1"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["sourcemap-codec", [\ - ["npm:1.4.8", {\ - "packageLocation": "./.yarn/cache/sourcemap-codec-npm-1.4.8-3a1a9e60b1-b57981c056.zip/node_modules/sourcemap-codec/",\ + ["srcset", [\ + ["npm:4.0.0", {\ + "packageLocation": "./.yarn/cache/srcset-npm-4.0.0-4e99d43236-aceb898c92.zip/node_modules/srcset/",\ "packageDependencies": [\ - ["sourcemap-codec", "npm:1.4.8"]\ + ["srcset", "npm:4.0.0"]\ ],\ "linkType": "HARD"\ }]\ @@ -7735,26 +8833,50 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["strip-ansi", "npm:6.0.1"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:5.1.2", {\ + "packageLocation": "./.yarn/cache/string-width-npm-5.1.2-bf60531341-7369deaa29.zip/node_modules/string-width/",\ + "packageDependencies": [\ + ["string-width", "npm:5.1.2"],\ + ["eastasianwidth", "npm:0.2.0"],\ + ["emoji-regex", "npm:9.2.2"],\ + ["strip-ansi", "npm:7.0.1"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["string.prototype.trim", [\ + ["npm:1.2.8", {\ + "packageLocation": "./.yarn/cache/string.prototype.trim-npm-1.2.8-7ed4517ce8-49eb1a862a.zip/node_modules/string.prototype.trim/",\ + "packageDependencies": [\ + ["string.prototype.trim", "npm:1.2.8"],\ + ["call-bind", "npm:1.0.2"],\ + ["define-properties", "npm:1.2.0"],\ + ["es-abstract", "npm:1.22.3"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["string.prototype.trimend", [\ - ["npm:1.0.4", {\ - "packageLocation": "./.yarn/cache/string.prototype.trimend-npm-1.0.4-a656b8fe24-17e5aa45c3.zip/node_modules/string.prototype.trimend/",\ + ["npm:1.0.7", {\ + "packageLocation": "./.yarn/cache/string.prototype.trimend-npm-1.0.7-159b9dcfbc-2375516272.zip/node_modules/string.prototype.trimend/",\ "packageDependencies": [\ - ["string.prototype.trimend", "npm:1.0.4"],\ + ["string.prototype.trimend", "npm:1.0.7"],\ ["call-bind", "npm:1.0.2"],\ - ["define-properties", "npm:1.1.4"]\ + ["define-properties", "npm:1.2.0"],\ + ["es-abstract", "npm:1.22.3"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["string.prototype.trimstart", [\ - ["npm:1.0.4", {\ - "packageLocation": "./.yarn/cache/string.prototype.trimstart-npm-1.0.4-b31f5e7c85-3fb06818d3.zip/node_modules/string.prototype.trimstart/",\ + ["npm:1.0.7", {\ + "packageLocation": "./.yarn/cache/string.prototype.trimstart-npm-1.0.7-ae2f803b78-13d0c2cb0d.zip/node_modules/string.prototype.trimstart/",\ "packageDependencies": [\ - ["string.prototype.trimstart", "npm:1.0.4"],\ + ["string.prototype.trimstart", "npm:1.0.7"],\ ["call-bind", "npm:1.0.2"],\ - ["define-properties", "npm:1.1.4"]\ + ["define-properties", "npm:1.2.0"],\ + ["es-abstract", "npm:1.22.3"]\ ],\ "linkType": "HARD"\ }]\ @@ -7777,6 +8899,14 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["ansi-regex", "npm:5.0.1"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:7.0.1", {\ + "packageLocation": "./.yarn/cache/strip-ansi-npm-7.0.1-668c121204-257f78fa43.zip/node_modules/strip-ansi/",\ + "packageDependencies": [\ + ["strip-ansi", "npm:7.0.1"],\ + ["ansi-regex", "npm:6.0.1"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["strip-bom", [\ @@ -7864,19 +8994,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ - ["terser", [\ - ["npm:5.13.1", {\ - "packageLocation": "./.yarn/cache/terser-npm-5.13.1-c7df10bd07-0b1f5043cf.zip/node_modules/terser/",\ - "packageDependencies": [\ - ["terser", "npm:5.13.1"],\ - ["acorn", "npm:8.7.1"],\ - ["commander", "npm:2.20.3"],\ - ["source-map", "npm:0.8.0-beta.0"],\ - ["source-map-support", "npm:0.5.21"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ ["test-exclude", [\ ["npm:6.0.0", {\ "packageLocation": "./.yarn/cache/test-exclude-npm-6.0.0-3fb03d69df-3b34a3d771.zip/node_modules/test-exclude/",\ @@ -7944,16 +9061,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ - ["tr46", [\ - ["npm:1.0.1", {\ - "packageLocation": "./.yarn/cache/tr46-npm-1.0.1-9547f343a4-96d4ed46bc.zip/node_modules/tr46/",\ - "packageDependencies": [\ - ["tr46", "npm:1.0.1"],\ - ["punycode", "npm:2.1.1"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ ["treemate", [\ ["npm:0.3.11", {\ "packageLocation": "./.yarn/cache/treemate-npm-0.3.11-7be66c23fc-0c6ccbc6c5.zip/node_modules/treemate/",\ @@ -7964,12 +9071,12 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["tsconfig-paths", [\ - ["npm:3.14.1", {\ - "packageLocation": "./.yarn/cache/tsconfig-paths-npm-3.14.1-17a815b5c5-8afa01c673.zip/node_modules/tsconfig-paths/",\ + ["npm:3.15.0", {\ + "packageLocation": "./.yarn/cache/tsconfig-paths-npm-3.15.0-ff68930e0e-59f35407a3.zip/node_modules/tsconfig-paths/",\ "packageDependencies": [\ - ["tsconfig-paths", "npm:3.14.1"],\ + ["tsconfig-paths", "npm:3.15.0"],\ ["@types/json5", "npm:0.0.29"],\ - ["json5", "npm:1.0.1"],\ + ["json5", "npm:1.0.2"],\ ["minimist", "npm:1.2.6"],\ ["strip-bom", "npm:3.0.0"]\ ],\ @@ -8004,6 +9111,57 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["typed-array-buffer", [\ + ["npm:1.0.0", {\ + "packageLocation": "./.yarn/cache/typed-array-buffer-npm-1.0.0-95cb610310-3e0281c79b.zip/node_modules/typed-array-buffer/",\ + "packageDependencies": [\ + ["typed-array-buffer", "npm:1.0.0"],\ + ["call-bind", "npm:1.0.2"],\ + ["get-intrinsic", "npm:1.2.1"],\ + ["is-typed-array", "npm:1.1.10"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["typed-array-byte-length", [\ + ["npm:1.0.0", {\ + "packageLocation": "./.yarn/cache/typed-array-byte-length-npm-1.0.0-94d79975ca-b03db16458.zip/node_modules/typed-array-byte-length/",\ + "packageDependencies": [\ + ["typed-array-byte-length", "npm:1.0.0"],\ + ["call-bind", "npm:1.0.2"],\ + ["for-each", "npm:0.3.3"],\ + ["has-proto", "npm:1.0.1"],\ + ["is-typed-array", "npm:1.1.10"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["typed-array-byte-offset", [\ + ["npm:1.0.0", {\ + "packageLocation": "./.yarn/cache/typed-array-byte-offset-npm-1.0.0-8cbb911cf5-04f6f02d0e.zip/node_modules/typed-array-byte-offset/",\ + "packageDependencies": [\ + ["typed-array-byte-offset", "npm:1.0.0"],\ + ["available-typed-arrays", "npm:1.0.5"],\ + ["call-bind", "npm:1.0.2"],\ + ["for-each", "npm:0.3.3"],\ + ["has-proto", "npm:1.0.1"],\ + ["is-typed-array", "npm:1.1.10"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["typed-array-length", [\ + ["npm:1.0.4", {\ + "packageLocation": "./.yarn/cache/typed-array-length-npm-1.0.4-92771b81fc-2228febc93.zip/node_modules/typed-array-length/",\ + "packageDependencies": [\ + ["typed-array-length", "npm:1.0.4"],\ + ["call-bind", "npm:1.0.2"],\ + ["for-each", "npm:0.3.3"],\ + ["is-typed-array", "npm:1.1.10"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["unbox-primitive", [\ ["npm:1.0.2", {\ "packageLocation": "./.yarn/cache/unbox-primitive-npm-1.0.2-cb56a05066-b7a1cf5862.zip/node_modules/unbox-primitive/",\ @@ -8065,15 +9223,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ - ["v8-compile-cache", [\ - ["npm:2.3.0", {\ - "packageLocation": "./.yarn/cache/v8-compile-cache-npm-2.3.0-961375f150-adb0a271ea.zip/node_modules/v8-compile-cache/",\ - "packageDependencies": [\ - ["v8-compile-cache", "npm:2.3.0"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ ["v8-to-istanbul", [\ ["npm:9.0.1", {\ "packageLocation": "./.yarn/cache/v8-to-istanbul-npm-9.0.1-58bbce7857-a49c34bf0a.zip/node_modules/v8-to-istanbul/",\ @@ -8086,6 +9235,15 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["vanillajs-datepicker", [\ + ["npm:1.3.4", {\ + "packageLocation": "./.yarn/cache/vanillajs-datepicker-npm-1.3.4-bc86e15a9c-830958f8af.zip/node_modules/vanillajs-datepicker/",\ + "packageDependencies": [\ + ["vanillajs-datepicker", "npm:1.3.4"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["vdirs", [\ ["npm:0.1.8", {\ "packageLocation": "./.yarn/cache/vdirs-npm-0.1.8-59a32a98d6-a7be8ccad3.zip/node_modules/vdirs/",\ @@ -8094,13 +9252,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "SOFT"\ }],\ - ["virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:0.1.8", {\ - "packageLocation": "./.yarn/__virtual__/vdirs-virtual-f3d9f623fd/0/cache/vdirs-npm-0.1.8-59a32a98d6-a7be8ccad3.zip/node_modules/vdirs/",\ + ["virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:0.1.8", {\ + "packageLocation": "./.yarn/__virtual__/vdirs-virtual-6e8e27ef7d/0/cache/vdirs-npm-0.1.8-59a32a98d6-a7be8ccad3.zip/node_modules/vdirs/",\ "packageDependencies": [\ - ["vdirs", "virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:0.1.8"],\ + ["vdirs", "virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:0.1.8"],\ ["@types/vue", null],\ ["evtd", "npm:0.2.3"],\ - ["vue", "npm:3.2.45"]\ + ["vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.21"]\ ],\ "packagePeers": [\ "@types/vue",\ @@ -8110,42 +9268,45 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["vite", [\ - ["npm:3.2.5", {\ - "packageLocation": "./.yarn/cache/vite-npm-3.2.5-f23b9ecb5b-ad35b7008c.zip/node_modules/vite/",\ + ["npm:4.5.3", {\ + "packageLocation": "./.yarn/cache/vite-npm-4.5.3-5cedc7cb8f-fd3f512ce4.zip/node_modules/vite/",\ "packageDependencies": [\ - ["vite", "npm:3.2.5"]\ + ["vite", "npm:4.5.3"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.2.5", {\ - "packageLocation": "./.yarn/__virtual__/vite-virtual-4387dfb6db/0/cache/vite-npm-3.2.5-f23b9ecb5b-ad35b7008c.zip/node_modules/vite/",\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:4.5.3", {\ + "packageLocation": "./.yarn/__virtual__/vite-virtual-69c30fd9fd/0/cache/vite-npm-4.5.3-5cedc7cb8f-fd3f512ce4.zip/node_modules/vite/",\ "packageDependencies": [\ - ["vite", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.2.5"],\ + ["vite", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:4.5.3"],\ ["@types/less", null],\ + ["@types/lightningcss", null],\ ["@types/node", null],\ ["@types/sass", null],\ ["@types/stylus", null],\ ["@types/sugarss", null],\ ["@types/terser", null],\ - ["esbuild", "npm:0.15.11"],\ + ["esbuild", "npm:0.18.20"],\ ["fsevents", "patch:fsevents@npm%3A2.3.2#~builtin::version=2.3.2&hash=18f3a7"],\ ["less", null],\ - ["postcss", "npm:8.4.18"],\ - ["resolve", "patch:resolve@npm%3A1.22.1#~builtin::version=1.22.1&hash=07638b"],\ - ["rollup", "npm:2.79.1"],\ - ["sass", "npm:1.57.1"],\ + ["lightningcss", null],\ + ["postcss", "npm:8.4.33"],\ + ["rollup", "npm:3.29.4"],\ + ["sass", "npm:1.72.0"],\ ["stylus", null],\ ["sugarss", null],\ ["terser", null]\ ],\ "packagePeers": [\ "@types/less",\ + "@types/lightningcss",\ "@types/node",\ "@types/sass",\ "@types/stylus",\ "@types/sugarss",\ "@types/terser",\ "less",\ + "lightningcss",\ "sass",\ "stylus",\ "sugarss",\ @@ -8163,6 +9324,46 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["volar-service-html", [\ + ["npm:0.0.34", {\ + "packageLocation": "./.yarn/cache/volar-service-html-npm-0.0.34-32b6d24136-83b50cd805.zip/node_modules/volar-service-html/",\ + "packageDependencies": [\ + ["volar-service-html", "npm:0.0.34"]\ + ],\ + "linkType": "SOFT"\ + }],\ + ["virtual:6f5429e17c4ecd390af605a4e97ecc7b34f2f1374a5e30c21f0a978cbdc904738a42d0d6f5d44d2e969250218b3c205853d6afefd88b87bcda877286d12bef83#npm:0.0.34", {\ + "packageLocation": "./.yarn/__virtual__/volar-service-html-virtual-5a9107a24d/0/cache/volar-service-html-npm-0.0.34-32b6d24136-83b50cd805.zip/node_modules/volar-service-html/",\ + "packageDependencies": [\ + ["volar-service-html", "virtual:6f5429e17c4ecd390af605a4e97ecc7b34f2f1374a5e30c21f0a978cbdc904738a42d0d6f5d44d2e969250218b3c205853d6afefd88b87bcda877286d12bef83#npm:0.0.34"],\ + ["@types/volar__language-service", null],\ + ["@volar/language-service", "npm:2.1.4"],\ + ["vscode-html-languageservice", "npm:5.1.2"],\ + ["vscode-languageserver-textdocument", "npm:1.0.11"],\ + ["vscode-uri", "npm:3.0.8"]\ + ],\ + "packagePeers": [\ + "@types/volar__language-service",\ + "@volar/language-service"\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["volar-service-pug", [\ + ["npm:0.0.34", {\ + "packageLocation": "./.yarn/cache/volar-service-pug-npm-0.0.34-6f5429e17c-4691aa1c8e.zip/node_modules/volar-service-pug/",\ + "packageDependencies": [\ + ["volar-service-pug", "npm:0.0.34"],\ + ["@volar/language-service", "npm:2.1.4"],\ + ["pug-lexer", "npm:5.0.1"],\ + ["pug-parser", "npm:6.0.0"],\ + ["volar-service-html", "virtual:6f5429e17c4ecd390af605a4e97ecc7b34f2f1374a5e30c21f0a978cbdc904738a42d0d6f5d44d2e969250218b3c205853d6afefd88b87bcda877286d12bef83#npm:0.0.34"],\ + ["vscode-html-languageservice", "npm:5.1.2"],\ + ["vscode-languageserver-textdocument", "npm:1.0.11"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["vooks", [\ ["npm:0.2.12", {\ "packageLocation": "./.yarn/cache/vooks-npm-0.2.12-0d1a2d856b-e6841ec5b6.zip/node_modules/vooks/",\ @@ -8171,13 +9372,13 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "SOFT"\ }],\ - ["virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:0.2.12", {\ - "packageLocation": "./.yarn/__virtual__/vooks-virtual-f44f5b55b4/0/cache/vooks-npm-0.2.12-0d1a2d856b-e6841ec5b6.zip/node_modules/vooks/",\ + ["virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:0.2.12", {\ + "packageLocation": "./.yarn/__virtual__/vooks-virtual-ca0a47c4bf/0/cache/vooks-npm-0.2.12-0d1a2d856b-e6841ec5b6.zip/node_modules/vooks/",\ "packageDependencies": [\ - ["vooks", "virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:0.2.12"],\ + ["vooks", "virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:0.2.12"],\ ["@types/vue", null],\ ["evtd", "npm:0.2.3"],\ - ["vue", "npm:3.2.45"]\ + ["vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.21"]\ ],\ "packagePeers": [\ "@types/vue",\ @@ -8186,16 +9387,89 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["vscode-html-languageservice", [\ + ["npm:5.1.2", {\ + "packageLocation": "./.yarn/cache/vscode-html-languageservice-npm-5.1.2-2ea2618bdd-3a2a5ee5ad.zip/node_modules/vscode-html-languageservice/",\ + "packageDependencies": [\ + ["vscode-html-languageservice", "npm:5.1.2"],\ + ["@vscode/l10n", "npm:0.0.18"],\ + ["vscode-languageserver-textdocument", "npm:1.0.11"],\ + ["vscode-languageserver-types", "npm:3.17.5"],\ + ["vscode-uri", "npm:3.0.8"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["vscode-jsonrpc", [\ + ["npm:8.2.0", {\ + "packageLocation": "./.yarn/cache/vscode-jsonrpc-npm-8.2.0-b7d2e5b553-f302a01e59.zip/node_modules/vscode-jsonrpc/",\ + "packageDependencies": [\ + ["vscode-jsonrpc", "npm:8.2.0"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["vscode-languageserver-protocol", [\ + ["npm:3.17.5", {\ + "packageLocation": "./.yarn/cache/vscode-languageserver-protocol-npm-3.17.5-2b07e16989-dfb42d276d.zip/node_modules/vscode-languageserver-protocol/",\ + "packageDependencies": [\ + ["vscode-languageserver-protocol", "npm:3.17.5"],\ + ["vscode-jsonrpc", "npm:8.2.0"],\ + ["vscode-languageserver-types", "npm:3.17.5"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["vscode-languageserver-textdocument", [\ + ["npm:1.0.11", {\ + "packageLocation": "./.yarn/cache/vscode-languageserver-textdocument-npm-1.0.11-6fc94d2b7b-ea7cdc9d4f.zip/node_modules/vscode-languageserver-textdocument/",\ + "packageDependencies": [\ + ["vscode-languageserver-textdocument", "npm:1.0.11"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["vscode-languageserver-types", [\ + ["npm:3.17.5", {\ + "packageLocation": "./.yarn/cache/vscode-languageserver-types-npm-3.17.5-aca3b71a5a-79b420e757.zip/node_modules/vscode-languageserver-types/",\ + "packageDependencies": [\ + ["vscode-languageserver-types", "npm:3.17.5"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ + ["vscode-uri", [\ + ["npm:3.0.8", {\ + "packageLocation": "./.yarn/cache/vscode-uri-npm-3.0.8-56f46b9d24-5142491268.zip/node_modules/vscode-uri/",\ + "packageDependencies": [\ + ["vscode-uri", "npm:3.0.8"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["vue", [\ - ["npm:3.2.45", {\ - "packageLocation": "./.yarn/cache/vue-npm-3.2.45-06b4b60efe-df60ca80cb.zip/node_modules/vue/",\ + ["npm:3.4.21", {\ + "packageLocation": "./.yarn/cache/vue-npm-3.4.21-02110aa6d9-3c477982a0.zip/node_modules/vue/",\ + "packageDependencies": [\ + ["vue", "npm:3.4.21"]\ + ],\ + "linkType": "SOFT"\ + }],\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.21", {\ + "packageLocation": "./.yarn/__virtual__/vue-virtual-b79af6274d/0/cache/vue-npm-3.4.21-02110aa6d9-3c477982a0.zip/node_modules/vue/",\ "packageDependencies": [\ - ["vue", "npm:3.2.45"],\ - ["@vue/compiler-dom", "npm:3.2.45"],\ - ["@vue/compiler-sfc", "npm:3.2.45"],\ - ["@vue/runtime-dom", "npm:3.2.45"],\ - ["@vue/server-renderer", "virtual:06b4b60efe017e0fdf8405fef0e04a64d7dc86cf13e04f05fb681889d996c75efd151033b2e012c72dfb9a0b8d1a647b6d3a8115078891aebe2fa1a4e81f50bf#npm:3.2.45"],\ - ["@vue/shared", "npm:3.2.45"]\ + ["vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.21"],\ + ["@types/typescript", null],\ + ["@vue/compiler-dom", "npm:3.4.21"],\ + ["@vue/compiler-sfc", "npm:3.4.21"],\ + ["@vue/runtime-dom", "npm:3.4.21"],\ + ["@vue/server-renderer", "virtual:b79af6274dddda2b283f42be2b827e30c3e5389bce2938ee73bdb74ee9781811fc079c6836719e57940708d59b3beeb14d9e3c12f37f2d22582a53e6c32e4c97#npm:3.4.21"],\ + ["@vue/shared", "npm:3.4.21"],\ + ["typescript", null]\ + ],\ + "packagePeers": [\ + "@types/typescript",\ + "typescript"\ ],\ "linkType": "HARD"\ }]\ @@ -8208,21 +9482,21 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ],\ "linkType": "SOFT"\ }],\ - ["npm:0.13.1", {\ - "packageLocation": "./.yarn/unplugged/vue-demi-virtual-8581c55d48/node_modules/vue-demi/",\ + ["npm:0.14.5", {\ + "packageLocation": "./.yarn/unplugged/vue-demi-virtual-b0e571907e/node_modules/vue-demi/",\ "packageDependencies": [\ - ["vue-demi", "npm:0.13.1"]\ + ["vue-demi", "npm:0.14.5"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:e1f94167f73dcd0042012cf3a978b5096a058d81d2d3813652fe465a5fe7d84e810c9bea1e50af574784bc4294f4f838b7ba55cf745fd1337e972f45d3531d5e#npm:0.13.1", {\ - "packageLocation": "./.yarn/unplugged/vue-demi-virtual-8581c55d48/node_modules/vue-demi/",\ + ["virtual:cf6f7439ee76dfd2e7f8f2565ae847d76901434fc49c65702190cdf3d1c61e61c701a5c45b514c4bdeacb8f4bcac9c8a98bd4db3d0bc8e403d9e8db2cf14372a#npm:0.14.5", {\ + "packageLocation": "./.yarn/unplugged/vue-demi-virtual-b0e571907e/node_modules/vue-demi/",\ "packageDependencies": [\ - ["vue-demi", "virtual:e1f94167f73dcd0042012cf3a978b5096a058d81d2d3813652fe465a5fe7d84e810c9bea1e50af574784bc4294f4f838b7ba55cf745fd1337e972f45d3531d5e#npm:0.13.1"],\ + ["vue-demi", "virtual:cf6f7439ee76dfd2e7f8f2565ae847d76901434fc49c65702190cdf3d1c61e61c701a5c45b514c4bdeacb8f4bcac9c8a98bd4db3d0bc8e403d9e8db2cf14372a#npm:0.14.5"],\ ["@types/vue", null],\ ["@types/vue__composition-api", null],\ ["@vue/composition-api", null],\ - ["vue", "npm:3.2.45"]\ + ["vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.21"]\ ],\ "packagePeers": [\ "@types/vue",\ @@ -8239,7 +9513,7 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["@types/vue", null],\ ["@types/vue__composition-api", null],\ ["@vue/composition-api", null],\ - ["vue", "npm:3.2.45"]\ + ["vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.21"]\ ],\ "packagePeers": [\ "@types/vue",\ @@ -8251,20 +9525,20 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["vue-eslint-parser", [\ - ["npm:9.0.3", {\ - "packageLocation": "./.yarn/cache/vue-eslint-parser-npm-9.0.3-1d52721799-61248eb504.zip/node_modules/vue-eslint-parser/",\ + ["npm:9.4.2", {\ + "packageLocation": "./.yarn/cache/vue-eslint-parser-npm-9.4.2-3e4e696025-67f14c8ea1.zip/node_modules/vue-eslint-parser/",\ "packageDependencies": [\ - ["vue-eslint-parser", "npm:9.0.3"]\ + ["vue-eslint-parser", "npm:9.4.2"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:b1db986ed39a80a226fbfa4d653465b173f10958eed2dd2079c16424a052a437a3b029f3b0228d6df47b4f864f76b70bc3e54b0b37aaac5dc8f4ae6650567a27#npm:9.0.3", {\ - "packageLocation": "./.yarn/__virtual__/vue-eslint-parser-virtual-77554ae6bd/0/cache/vue-eslint-parser-npm-9.0.3-1d52721799-61248eb504.zip/node_modules/vue-eslint-parser/",\ + ["virtual:e080dd5dc65fb3541eb98fd929c3a1d3733f3aff4bb24b09a6b5cce9fba4a29aca07e286ef93079f2144caa0fd33bb6545549286d3a9f2b9a211caa1f4b68ff9#npm:9.4.2", {\ + "packageLocation": "./.yarn/__virtual__/vue-eslint-parser-virtual-f703c550a2/0/cache/vue-eslint-parser-npm-9.4.2-3e4e696025-67f14c8ea1.zip/node_modules/vue-eslint-parser/",\ "packageDependencies": [\ - ["vue-eslint-parser", "virtual:b1db986ed39a80a226fbfa4d653465b173f10958eed2dd2079c16424a052a437a3b029f3b0228d6df47b4f864f76b70bc3e54b0b37aaac5dc8f4ae6650567a27#npm:9.0.3"],\ + ["vue-eslint-parser", "virtual:e080dd5dc65fb3541eb98fd929c3a1d3733f3aff4bb24b09a6b5cce9fba4a29aca07e286ef93079f2144caa0fd33bb6545549286d3a9f2b9a211caa1f4b68ff9#npm:9.4.2"],\ ["@types/eslint", null],\ ["debug", "virtual:b86a9fb34323a98c6519528ed55faa0d9b44ca8879307c0b29aa384bde47ff59a7d0c9051b31246f14521dfb71ba3c5d6d0b35c29fffc17bf875aa6ad977d9e8#npm:4.3.4"],\ - ["eslint", "npm:8.31.0"],\ + ["eslint", "npm:8.57.0"],\ ["eslint-scope", "npm:7.1.1"],\ ["eslint-visitor-keys", "npm:3.3.0"],\ ["espree", "npm:9.3.2"],\ @@ -8280,20 +9554,20 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["vue-router", [\ - ["npm:4.1.6", {\ - "packageLocation": "./.yarn/cache/vue-router-npm-4.1.6-ccab7109e1-c7f0156ac0.zip/node_modules/vue-router/",\ + ["npm:4.3.0", {\ + "packageLocation": "./.yarn/cache/vue-router-npm-4.3.0-b765d40138-0059261d39.zip/node_modules/vue-router/",\ "packageDependencies": [\ - ["vue-router", "npm:4.1.6"]\ + ["vue-router", "npm:4.3.0"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:4.1.6", {\ - "packageLocation": "./.yarn/__virtual__/vue-router-virtual-670ec833a5/0/cache/vue-router-npm-4.1.6-ccab7109e1-c7f0156ac0.zip/node_modules/vue-router/",\ + ["virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:4.3.0", {\ + "packageLocation": "./.yarn/__virtual__/vue-router-virtual-82f54143bf/0/cache/vue-router-npm-4.3.0-b765d40138-0059261d39.zip/node_modules/vue-router/",\ "packageDependencies": [\ - ["vue-router", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:4.1.6"],\ + ["vue-router", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:4.3.0"],\ ["@types/vue", null],\ - ["@vue/devtools-api", "npm:6.4.5"],\ - ["vue", "npm:3.2.45"]\ + ["@vue/devtools-api", "npm:6.6.1"],\ + ["vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.21"]\ ],\ "packagePeers": [\ "@types/vue",\ @@ -8303,26 +9577,26 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["vueuc", [\ - ["npm:0.4.47", {\ - "packageLocation": "./.yarn/cache/vueuc-npm-0.4.47-ad081ddd15-b82b77a882.zip/node_modules/vueuc/",\ + ["npm:0.4.58", {\ + "packageLocation": "./.yarn/cache/vueuc-npm-0.4.58-be5584770c-fb0b9a69be.zip/node_modules/vueuc/",\ "packageDependencies": [\ - ["vueuc", "npm:0.4.47"]\ + ["vueuc", "npm:0.4.58"]\ ],\ "linkType": "SOFT"\ }],\ - ["virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:0.4.47", {\ - "packageLocation": "./.yarn/__virtual__/vueuc-virtual-326ef48c83/0/cache/vueuc-npm-0.4.47-ad081ddd15-b82b77a882.zip/node_modules/vueuc/",\ + ["virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:0.4.58", {\ + "packageLocation": "./.yarn/__virtual__/vueuc-virtual-2366be83ef/0/cache/vueuc-npm-0.4.58-be5584770c-fb0b9a69be.zip/node_modules/vueuc/",\ "packageDependencies": [\ - ["vueuc", "virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:0.4.47"],\ - ["@css-render/vue3-ssr", "virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:0.15.10"],\ + ["vueuc", "virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:0.4.58"],\ + ["@css-render/vue3-ssr", "virtual:2366be83ef58a728ebb5a5e9ed4600f4465f98b2a844262fcfbe89415361d5d5f9e964ec3b9a72d6a5004f37c1024d017c65e67473dd9cc39cd61f51768c65e6#npm:0.15.10"],\ ["@juggle/resize-observer", "npm:3.3.1"],\ ["@types/vue", null],\ ["css-render", "npm:0.15.10"],\ - ["evtd", "npm:0.2.3"],\ - ["seemly", "npm:0.3.3"],\ - ["vdirs", "virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:0.1.8"],\ - ["vooks", "virtual:4fa5810747c131dcf20cc36c17e835ba5fa42265b3b1f99a6712420e8672195f31ac73d077125c82af29bde6232df6ecc61b1f3ff0dde337b62085ef712e7d6a#npm:0.2.12"],\ - ["vue", "npm:3.2.45"]\ + ["evtd", "npm:0.2.4"],\ + ["seemly", "npm:0.3.6"],\ + ["vdirs", "virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:0.1.8"],\ + ["vooks", "virtual:32fd9c861d759cd42dabb479e4fd652286369e629cc7ef63c9cf4f1af5387c64be25fafc985023ea8534b1ec1f4cc92e6c918c7f3b594aa0f8acad026c671a6a#npm:0.2.12"],\ + ["vue", "virtual:dc3fc578bfa5e06182a4d2be39ede0bc5b74940b1ffe0d70c26892ab140a4699787750fba175dc306292e80b4aa2c8c5f68c2a821e69b2c37e360c0dff36ff66#npm:3.4.21"]\ ],\ "packagePeers": [\ "@types/vue",\ @@ -8340,27 +9614,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ - ["webidl-conversions", [\ - ["npm:4.0.2", {\ - "packageLocation": "./.yarn/cache/webidl-conversions-npm-4.0.2-1d159e6409-c93d8dfe90.zip/node_modules/webidl-conversions/",\ - "packageDependencies": [\ - ["webidl-conversions", "npm:4.0.2"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ - ["whatwg-url", [\ - ["npm:7.1.0", {\ - "packageLocation": "./.yarn/cache/whatwg-url-npm-7.1.0-d6cae01571-fecb07c872.zip/node_modules/whatwg-url/",\ - "packageDependencies": [\ - ["whatwg-url", "npm:7.1.0"],\ - ["lodash.sortby", "npm:4.7.0"],\ - ["tr46", "npm:1.0.1"],\ - ["webidl-conversions", "npm:4.0.2"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ ["which", [\ ["npm:2.0.2", {\ "packageLocation": "./.yarn/cache/which-npm-2.0.2-320ddf72f7-1a5c563d3c.zip/node_modules/which/",\ @@ -8385,6 +9638,20 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ + ["which-typed-array", [\ + ["npm:1.1.13", {\ + "packageLocation": "./.yarn/cache/which-typed-array-npm-1.1.13-92c18b4878-3828a0d5d7.zip/node_modules/which-typed-array/",\ + "packageDependencies": [\ + ["which-typed-array", "npm:1.1.13"],\ + ["available-typed-arrays", "npm:1.0.5"],\ + ["call-bind", "npm:1.0.5"],\ + ["for-each", "npm:0.3.3"],\ + ["gopd", "npm:1.0.1"],\ + ["has-tostringtag", "npm:1.0.0"]\ + ],\ + "linkType": "HARD"\ + }]\ + ]],\ ["wide-align", [\ ["npm:1.1.5", {\ "packageLocation": "./.yarn/cache/wide-align-npm-1.1.5-889d77e592-d5fc37cd56.zip/node_modules/wide-align/",\ @@ -8408,15 +9675,6 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { "linkType": "HARD"\ }]\ ]],\ - ["word-wrap", [\ - ["npm:1.2.3", {\ - "packageLocation": "./.yarn/cache/word-wrap-npm-1.2.3-7fb15ab002-30b48f91fc.zip/node_modules/word-wrap/",\ - "packageDependencies": [\ - ["word-wrap", "npm:1.2.3"]\ - ],\ - "linkType": "HARD"\ - }]\ - ]],\ ["wrap-ansi", [\ ["npm:7.0.0", {\ "packageLocation": "./.yarn/cache/wrap-ansi-npm-7.0.0-ad6e1a0554-a790b846fd.zip/node_modules/wrap-ansi/",\ @@ -8427,6 +9685,16 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { ["strip-ansi", "npm:6.0.1"]\ ],\ "linkType": "HARD"\ + }],\ + ["npm:8.1.0", {\ + "packageLocation": "./.yarn/cache/wrap-ansi-npm-8.1.0-26a4e6ae28-371733296d.zip/node_modules/wrap-ansi/",\ + "packageDependencies": [\ + ["wrap-ansi", "npm:8.1.0"],\ + ["ansi-styles", "npm:6.2.1"],\ + ["string-width", "npm:5.1.2"],\ + ["strip-ansi", "npm:7.0.1"]\ + ],\ + "linkType": "HARD"\ }]\ ]],\ ["wrappy", [\ @@ -8484,26 +9752,26 @@ function $$SETUP_STATE(hydrateRuntimeState, basePath) { }]\ ]],\ ["yargs", [\ - ["npm:16.2.0", {\ - "packageLocation": "./.yarn/cache/yargs-npm-16.2.0-547873d425-b14afbb51e.zip/node_modules/yargs/",\ + ["npm:17.7.2", {\ + "packageLocation": "./.yarn/cache/yargs-npm-17.7.2-80b62638e1-73b572e863.zip/node_modules/yargs/",\ "packageDependencies": [\ - ["yargs", "npm:16.2.0"],\ - ["cliui", "npm:7.0.4"],\ + ["yargs", "npm:17.7.2"],\ + ["cliui", "npm:8.0.1"],\ ["escalade", "npm:3.1.1"],\ ["get-caller-file", "npm:2.0.5"],\ ["require-directory", "npm:2.1.1"],\ ["string-width", "npm:4.2.3"],\ ["y18n", "npm:5.0.8"],\ - ["yargs-parser", "npm:20.2.9"]\ + ["yargs-parser", "npm:21.1.1"]\ ],\ "linkType": "HARD"\ }]\ ]],\ ["yargs-parser", [\ - ["npm:20.2.9", {\ - "packageLocation": "./.yarn/cache/yargs-parser-npm-20.2.9-a1d19e598d-8bb69015f2.zip/node_modules/yargs-parser/",\ + ["npm:21.1.1", {\ + "packageLocation": "./.yarn/cache/yargs-parser-npm-21.1.1-8fdc003314-ed2d96a616.zip/node_modules/yargs-parser/",\ "packageDependencies": [\ - ["yargs-parser", "npm:20.2.9"]\ + ["yargs-parser", "npm:21.1.1"]\ ],\ "linkType": "HARD"\ }]\ diff --git a/.pylintrc b/.pylintrc index c9a33fcec2..008f89e454 100644 --- a/.pylintrc +++ b/.pylintrc @@ -405,4 +405,4 @@ analyse-fallback-blocks=no # Exceptions that will emit a warning when being caught. Defaults to # "Exception" -overgeneral-exceptions=Exception +overgeneral-exceptions=builtins.Exception diff --git a/.vscode/launch.json b/.vscode/launch.json index 8dfc1b9b7a..227eb8f615 100644 --- a/.vscode/launch.json +++ b/.vscode/launch.json @@ -12,7 +12,7 @@ "program": "${workspaceFolder}/ietf/manage.py", "args": [ "runserver", - "0.0.0.0:8000", + "0.0.0.0:8001", "--settings=settings_local" ], "django": true, @@ -30,7 +30,7 @@ "program": "${workspaceFolder}/ietf/manage.py", "args": [ "runserver", - "0.0.0.0:8000", + "0.0.0.0:8001", "--settings=settings_local_vite" ], "django": true, @@ -48,7 +48,7 @@ "program": "${workspaceFolder}/ietf/manage.py", "args": [ "runserver", - "0.0.0.0:8000", + "0.0.0.0:8001", "--settings=settings_local_debug" ], "django": true, diff --git a/.vscode/settings.json b/.vscode/settings.json index 6acc641264..b323cd02f7 100644 --- a/.vscode/settings.json +++ b/.vscode/settings.json @@ -1,55 +1,61 @@ { - "taskExplorer.exclude": [ - "**/.vscode-test/**", - "**/bin/**", - "**/build/**", - "**/CompiledOutput/**", - "**/dist/**", - "**/doc/**", - "**/ext/**", - "**/out/**", - "**/output/**", - "**/packages/**", - "**/release/**", - "**/releases/**", - "**/samples/**", - "**/sdks/**", - "**/static/**", - "**/target/**", - "**/test/**", - "**/third_party/**", - "**/vendor/**", - "**/work/**", - "/workspace/bootstrap/nuget/MyGet.ps1" - ], - "taskExplorer.enableAnt": false, - "taskExplorer.enableAppPublisher": false, - "taskExplorer.enablePipenv": false, - "taskExplorer.enableBash": false, - "taskExplorer.enableBatch": false, - "taskExplorer.enableGradle": false, - "taskExplorer.enableGrunt": false, - "taskExplorer.enableGulp": false, - "taskExplorer.enablePerl": false, - "taskExplorer.enableMake": false, - "taskExplorer.enableMaven": false, - "taskExplorer.enableNsis": false, - "taskExplorer.enableNpm": false, - "taskExplorer.enablePowershell": false, - "taskExplorer.enablePython": false, - "taskExplorer.enableRuby": false, - "taskExplorer.enableTsc": false, - "taskExplorer.enableWorkspace": true, - "taskExplorer.enableExplorerView": false, - "taskExplorer.enableSideBar": true, - "search.exclude": { - "**/.yarn": true, - "**/.pnp.*": true - }, - "eslint.nodePath": ".yarn/sdks", - "eslint.validate": [ - "javascript", - "javascriptreact", - "vue" - ] + "taskExplorer.exclude": [ + "**/.vscode-test/**", + "**/bin/**", + "**/build/**", + "**/CompiledOutput/**", + "**/dist/**", + "**/doc/**", + "**/ext/**", + "**/out/**", + "**/output/**", + "**/packages/**", + "**/release/**", + "**/releases/**", + "**/samples/**", + "**/sdks/**", + "**/static/**", + "**/target/**", + "**/test/**", + "**/third_party/**", + "**/vendor/**", + "**/work/**", + "/workspace/bootstrap/nuget/MyGet.ps1" + ], + "taskExplorer.enabledTasks": { + "ant": false, + "bash": false, + "batch": false, + "composer": false, + "gradle": false, + "grunt": false, + "gulp": false, + "make": false, + "maven": false, + "npm": false, + "perl": false, + "pipenv": false, + "powershell": false, + "python": false, + "ruby": false, + "tsc": false + }, + "taskExplorer.enableExplorerView": false, + "taskExplorer.enableSideBar": true, + "taskExplorer.showLastTasks": false, + "search.exclude": { + "**/.yarn": true, + "**/.pnp.*": true + }, + "eslint.nodePath": ".yarn/sdks", + "eslint.validate": [ + "javascript", + "javascriptreact", + "vue" + ], + "python.linting.pylintArgs": ["--load-plugins", "pylint_django"], + "python.testing.pytestEnabled": false, + "python.testing.unittestEnabled": false, + "python.linting.enabled": true, + "python.terminal.shellIntegration.enabled": false } diff --git a/.vscode/tasks.json b/.vscode/tasks.json index 5490af335c..8b36b0e6ac 100644 --- a/.vscode/tasks.json +++ b/.vscode/tasks.json @@ -48,7 +48,7 @@ "args": [ "${workspaceFolder}/ietf/manage.py", "test", - "--settings=settings_local_sqlitetest" + "--settings=settings_test" ], "group": "test", "presentation": { @@ -68,7 +68,7 @@ "args": [ "${workspaceFolder}/ietf/manage.py", "test", - "--settings=settings_local_sqlitetest", + "--settings=settings_test", "--pattern=tests_js.py" ], "group": "test", @@ -105,10 +105,11 @@ "command": "/usr/local/bin/python", "args": [ "-m", - "smtpd", + "aiosmtpd", "-n", "-c", - "DebuggingServer", + "ietf.utils.aiosmtpd.DevDebuggingHandler", + "-l", "localhost:2025" ], "presentation": { diff --git a/.yarn/cache/@aashutoshrathi-word-wrap-npm-1.2.6-5b1d95e487-ada901b9e7.zip b/.yarn/cache/@aashutoshrathi-word-wrap-npm-1.2.6-5b1d95e487-ada901b9e7.zip new file mode 100644 index 0000000000..9334304c2a Binary files /dev/null and b/.yarn/cache/@aashutoshrathi-word-wrap-npm-1.2.6-5b1d95e487-ada901b9e7.zip differ diff --git a/.yarn/cache/@babel-parser-npm-7.23.9-720a0b56cb-e7cd4960ac.zip b/.yarn/cache/@babel-parser-npm-7.23.9-720a0b56cb-e7cd4960ac.zip new file mode 100644 index 0000000000..7b6c44fc3f Binary files /dev/null and b/.yarn/cache/@babel-parser-npm-7.23.9-720a0b56cb-e7cd4960ac.zip differ diff --git a/.yarn/cache/@babel-runtime-npm-7.23.2-d013d6cf7e-6c4df4839e.zip b/.yarn/cache/@babel-runtime-npm-7.23.2-d013d6cf7e-6c4df4839e.zip new file mode 100644 index 0000000000..f0d4497857 Binary files /dev/null and b/.yarn/cache/@babel-runtime-npm-7.23.2-d013d6cf7e-6c4df4839e.zip differ diff --git a/.yarn/cache/@css-render-plugin-bem-npm-0.15.10-41ccecaa2f-cbab72a7b5.zip b/.yarn/cache/@css-render-plugin-bem-npm-0.15.10-41ccecaa2f-cbab72a7b5.zip deleted file mode 100644 index 7df7772871..0000000000 Binary files a/.yarn/cache/@css-render-plugin-bem-npm-0.15.10-41ccecaa2f-cbab72a7b5.zip and /dev/null differ diff --git a/.yarn/cache/@css-render-plugin-bem-npm-0.15.12-bf8b43dc1f-9fa7ddd62b.zip b/.yarn/cache/@css-render-plugin-bem-npm-0.15.12-bf8b43dc1f-9fa7ddd62b.zip new file mode 100644 index 0000000000..7145fae118 Binary files /dev/null and b/.yarn/cache/@css-render-plugin-bem-npm-0.15.12-bf8b43dc1f-9fa7ddd62b.zip differ diff --git a/.yarn/cache/@css-render-vue3-ssr-npm-0.15.12-a130f4db3a-a5505ae161.zip b/.yarn/cache/@css-render-vue3-ssr-npm-0.15.12-a130f4db3a-a5505ae161.zip new file mode 100644 index 0000000000..0e75f8a4c7 Binary files /dev/null and b/.yarn/cache/@css-render-vue3-ssr-npm-0.15.12-a130f4db3a-a5505ae161.zip differ diff --git a/.yarn/cache/@esbuild-darwin-arm64-npm-0.18.20-00b3504077-8.zip b/.yarn/cache/@esbuild-darwin-arm64-npm-0.18.20-00b3504077-8.zip new file mode 100644 index 0000000000..dfd7b76554 Binary files /dev/null and b/.yarn/cache/@esbuild-darwin-arm64-npm-0.18.20-00b3504077-8.zip differ diff --git a/.yarn/cache/@esbuild-darwin-x64-npm-0.18.20-767fe27d1b-8.zip b/.yarn/cache/@esbuild-darwin-x64-npm-0.18.20-767fe27d1b-8.zip new file mode 100644 index 0000000000..432802b69e Binary files /dev/null and b/.yarn/cache/@esbuild-darwin-x64-npm-0.18.20-767fe27d1b-8.zip differ diff --git a/.yarn/cache/@esbuild-linux-arm64-npm-0.18.20-7b48b328fe-8.zip b/.yarn/cache/@esbuild-linux-arm64-npm-0.18.20-7b48b328fe-8.zip new file mode 100644 index 0000000000..6eb51fcc99 Binary files /dev/null and b/.yarn/cache/@esbuild-linux-arm64-npm-0.18.20-7b48b328fe-8.zip differ diff --git a/.yarn/cache/@esbuild-linux-x64-npm-0.18.20-de8e99b449-8.zip b/.yarn/cache/@esbuild-linux-x64-npm-0.18.20-de8e99b449-8.zip new file mode 100644 index 0000000000..bcbc77a84f Binary files /dev/null and b/.yarn/cache/@esbuild-linux-x64-npm-0.18.20-de8e99b449-8.zip differ diff --git a/.yarn/cache/@esbuild-win32-arm64-npm-0.18.20-a58fe6c6a3-8.zip b/.yarn/cache/@esbuild-win32-arm64-npm-0.18.20-a58fe6c6a3-8.zip new file mode 100644 index 0000000000..cf9c15613b Binary files /dev/null and b/.yarn/cache/@esbuild-win32-arm64-npm-0.18.20-a58fe6c6a3-8.zip differ diff --git a/.yarn/cache/@esbuild-win32-x64-npm-0.18.20-37a9ab2bda-8.zip b/.yarn/cache/@esbuild-win32-x64-npm-0.18.20-37a9ab2bda-8.zip new file mode 100644 index 0000000000..768cc68f13 Binary files /dev/null and b/.yarn/cache/@esbuild-win32-x64-npm-0.18.20-37a9ab2bda-8.zip differ diff --git a/.yarn/cache/@eslint-community-eslint-utils-npm-4.4.0-d1791bd5a3-cdfe3ae42b.zip b/.yarn/cache/@eslint-community-eslint-utils-npm-4.4.0-d1791bd5a3-cdfe3ae42b.zip new file mode 100644 index 0000000000..4e48357020 Binary files /dev/null and b/.yarn/cache/@eslint-community-eslint-utils-npm-4.4.0-d1791bd5a3-cdfe3ae42b.zip differ diff --git a/.yarn/cache/@eslint-community-regexpp-npm-4.10.0-6bfb984c81-2a6e345429.zip b/.yarn/cache/@eslint-community-regexpp-npm-4.10.0-6bfb984c81-2a6e345429.zip new file mode 100644 index 0000000000..7ef5a48973 Binary files /dev/null and b/.yarn/cache/@eslint-community-regexpp-npm-4.10.0-6bfb984c81-2a6e345429.zip differ diff --git a/.yarn/cache/@eslint-community-regexpp-npm-4.8.0-92ece47e3d-601e6d033d.zip b/.yarn/cache/@eslint-community-regexpp-npm-4.8.0-92ece47e3d-601e6d033d.zip new file mode 100644 index 0000000000..0cbfbf8d84 Binary files /dev/null and b/.yarn/cache/@eslint-community-regexpp-npm-4.8.0-92ece47e3d-601e6d033d.zip differ diff --git a/.yarn/cache/@eslint-eslintrc-npm-1.4.1-007f670de2-cd3e5a8683.zip b/.yarn/cache/@eslint-eslintrc-npm-1.4.1-007f670de2-cd3e5a8683.zip deleted file mode 100755 index b51019b7a1..0000000000 Binary files a/.yarn/cache/@eslint-eslintrc-npm-1.4.1-007f670de2-cd3e5a8683.zip and /dev/null differ diff --git a/.yarn/cache/@eslint-eslintrc-npm-2.1.4-1ff4b5f908-10957c7592.zip b/.yarn/cache/@eslint-eslintrc-npm-2.1.4-1ff4b5f908-10957c7592.zip new file mode 100644 index 0000000000..58788ff7a6 Binary files /dev/null and b/.yarn/cache/@eslint-eslintrc-npm-2.1.4-1ff4b5f908-10957c7592.zip differ diff --git a/.yarn/cache/@eslint-js-npm-8.57.0-00ead3710a-315dc65b0e.zip b/.yarn/cache/@eslint-js-npm-8.57.0-00ead3710a-315dc65b0e.zip new file mode 100644 index 0000000000..82eab16e7c Binary files /dev/null and b/.yarn/cache/@eslint-js-npm-8.57.0-00ead3710a-315dc65b0e.zip differ diff --git a/.yarn/cache/@faker-js-faker-npm-7.6.0-fa135883e9-942af62217.zip b/.yarn/cache/@faker-js-faker-npm-7.6.0-fa135883e9-942af62217.zip deleted file mode 100644 index 1317dcc1a8..0000000000 Binary files a/.yarn/cache/@faker-js-faker-npm-7.6.0-fa135883e9-942af62217.zip and /dev/null differ diff --git a/.yarn/cache/@floating-ui-core-npm-1.4.1-fe89c45d92-be4ab864fe.zip b/.yarn/cache/@floating-ui-core-npm-1.4.1-fe89c45d92-be4ab864fe.zip new file mode 100644 index 0000000000..e8ce36ae61 Binary files /dev/null and b/.yarn/cache/@floating-ui-core-npm-1.4.1-fe89c45d92-be4ab864fe.zip differ diff --git a/.yarn/cache/@floating-ui-dom-npm-1.5.2-f1b8ca0c30-3c71eed50b.zip b/.yarn/cache/@floating-ui-dom-npm-1.5.2-f1b8ca0c30-3c71eed50b.zip new file mode 100644 index 0000000000..a984181e2c Binary files /dev/null and b/.yarn/cache/@floating-ui-dom-npm-1.5.2-f1b8ca0c30-3c71eed50b.zip differ diff --git a/.yarn/cache/@floating-ui-utils-npm-0.1.2-22eefe56f0-3e29fd3c69.zip b/.yarn/cache/@floating-ui-utils-npm-0.1.2-22eefe56f0-3e29fd3c69.zip new file mode 100644 index 0000000000..ada2c49e44 Binary files /dev/null and b/.yarn/cache/@floating-ui-utils-npm-0.1.2-22eefe56f0-3e29fd3c69.zip differ diff --git a/.yarn/cache/@fullcalendar-bootstrap5-npm-5.11.3-3e86f39d7d-a63a500d72.zip b/.yarn/cache/@fullcalendar-bootstrap5-npm-5.11.3-3e86f39d7d-a63a500d72.zip deleted file mode 100644 index 8571502c87..0000000000 Binary files a/.yarn/cache/@fullcalendar-bootstrap5-npm-5.11.3-3e86f39d7d-a63a500d72.zip and /dev/null differ diff --git a/.yarn/cache/@fullcalendar-bootstrap5-npm-6.1.11-6e0fbf281a-a0c3b94346.zip b/.yarn/cache/@fullcalendar-bootstrap5-npm-6.1.11-6e0fbf281a-a0c3b94346.zip new file mode 100644 index 0000000000..edc7da3b25 Binary files /dev/null and b/.yarn/cache/@fullcalendar-bootstrap5-npm-6.1.11-6e0fbf281a-a0c3b94346.zip differ diff --git a/.yarn/cache/@fullcalendar-common-npm-5.11.3-6268994b76-be4b365dca.zip b/.yarn/cache/@fullcalendar-common-npm-5.11.3-6268994b76-be4b365dca.zip deleted file mode 100644 index 2e6ed4393f..0000000000 Binary files a/.yarn/cache/@fullcalendar-common-npm-5.11.3-6268994b76-be4b365dca.zip and /dev/null differ diff --git a/.yarn/cache/@fullcalendar-core-npm-5.11.3-ed98a1ea9f-2774d0fa18.zip b/.yarn/cache/@fullcalendar-core-npm-5.11.3-ed98a1ea9f-2774d0fa18.zip deleted file mode 100644 index c6ee2043b4..0000000000 Binary files a/.yarn/cache/@fullcalendar-core-npm-5.11.3-ed98a1ea9f-2774d0fa18.zip and /dev/null differ diff --git a/.yarn/cache/@fullcalendar-core-npm-6.1.11-ae049c8ace-0078a6f96b.zip b/.yarn/cache/@fullcalendar-core-npm-6.1.11-ae049c8ace-0078a6f96b.zip new file mode 100644 index 0000000000..c9eee67d63 Binary files /dev/null and b/.yarn/cache/@fullcalendar-core-npm-6.1.11-ae049c8ace-0078a6f96b.zip differ diff --git a/.yarn/cache/@fullcalendar-daygrid-npm-5.11.3-b387dff934-426b53c5bb.zip b/.yarn/cache/@fullcalendar-daygrid-npm-5.11.3-b387dff934-426b53c5bb.zip deleted file mode 100644 index f8b5af4de7..0000000000 Binary files a/.yarn/cache/@fullcalendar-daygrid-npm-5.11.3-b387dff934-426b53c5bb.zip and /dev/null differ diff --git a/.yarn/cache/@fullcalendar-daygrid-npm-6.1.11-2187ca1b8f-6eb5606de5.zip b/.yarn/cache/@fullcalendar-daygrid-npm-6.1.11-2187ca1b8f-6eb5606de5.zip new file mode 100644 index 0000000000..3a7449a3a8 Binary files /dev/null and b/.yarn/cache/@fullcalendar-daygrid-npm-6.1.11-2187ca1b8f-6eb5606de5.zip differ diff --git a/.yarn/cache/@fullcalendar-icalendar-npm-6.1.11-73807e790d-4e6eff15a8.zip b/.yarn/cache/@fullcalendar-icalendar-npm-6.1.11-73807e790d-4e6eff15a8.zip new file mode 100644 index 0000000000..861ed1b366 Binary files /dev/null and b/.yarn/cache/@fullcalendar-icalendar-npm-6.1.11-73807e790d-4e6eff15a8.zip differ diff --git a/.yarn/cache/@fullcalendar-interaction-npm-5.11.3-15335cb10a-e8a1b49f2f.zip b/.yarn/cache/@fullcalendar-interaction-npm-5.11.3-15335cb10a-e8a1b49f2f.zip deleted file mode 100644 index be7b594e35..0000000000 Binary files a/.yarn/cache/@fullcalendar-interaction-npm-5.11.3-15335cb10a-e8a1b49f2f.zip and /dev/null differ diff --git a/.yarn/cache/@fullcalendar-interaction-npm-6.1.11-39630596c7-c67d4cfa0b.zip b/.yarn/cache/@fullcalendar-interaction-npm-6.1.11-39630596c7-c67d4cfa0b.zip new file mode 100644 index 0000000000..b04343467b Binary files /dev/null and b/.yarn/cache/@fullcalendar-interaction-npm-6.1.11-39630596c7-c67d4cfa0b.zip differ diff --git a/.yarn/cache/@fullcalendar-list-npm-5.11.3-6174d0e1da-976da49b12.zip b/.yarn/cache/@fullcalendar-list-npm-5.11.3-6174d0e1da-976da49b12.zip deleted file mode 100644 index 089e8d7c95..0000000000 Binary files a/.yarn/cache/@fullcalendar-list-npm-5.11.3-6174d0e1da-976da49b12.zip and /dev/null differ diff --git a/.yarn/cache/@fullcalendar-list-npm-6.1.11-8f1846f302-84a8cd6e63.zip b/.yarn/cache/@fullcalendar-list-npm-6.1.11-8f1846f302-84a8cd6e63.zip new file mode 100644 index 0000000000..93cd34af81 Binary files /dev/null and b/.yarn/cache/@fullcalendar-list-npm-6.1.11-8f1846f302-84a8cd6e63.zip differ diff --git a/.yarn/cache/@fullcalendar-luxon2-npm-5.11.3-ccde7500a8-7533018590.zip b/.yarn/cache/@fullcalendar-luxon2-npm-5.11.3-ccde7500a8-7533018590.zip deleted file mode 100644 index bb23c44bf3..0000000000 Binary files a/.yarn/cache/@fullcalendar-luxon2-npm-5.11.3-ccde7500a8-7533018590.zip and /dev/null differ diff --git a/.yarn/cache/@fullcalendar-luxon3-npm-6.1.11-3e90656a71-8e7f45aab2.zip b/.yarn/cache/@fullcalendar-luxon3-npm-6.1.11-3e90656a71-8e7f45aab2.zip new file mode 100644 index 0000000000..6e717b3495 Binary files /dev/null and b/.yarn/cache/@fullcalendar-luxon3-npm-6.1.11-3e90656a71-8e7f45aab2.zip differ diff --git a/.yarn/cache/@fullcalendar-timegrid-npm-5.11.3-4075b09051-ce675eca7d.zip b/.yarn/cache/@fullcalendar-timegrid-npm-5.11.3-4075b09051-ce675eca7d.zip deleted file mode 100644 index ebe13b989a..0000000000 Binary files a/.yarn/cache/@fullcalendar-timegrid-npm-5.11.3-4075b09051-ce675eca7d.zip and /dev/null differ diff --git a/.yarn/cache/@fullcalendar-timegrid-npm-6.1.11-1d43455bfd-4a11e6dd90.zip b/.yarn/cache/@fullcalendar-timegrid-npm-6.1.11-1d43455bfd-4a11e6dd90.zip new file mode 100644 index 0000000000..917beeda69 Binary files /dev/null and b/.yarn/cache/@fullcalendar-timegrid-npm-6.1.11-1d43455bfd-4a11e6dd90.zip differ diff --git a/.yarn/cache/@fullcalendar-vue3-npm-5.11.3-047b9981f6-13a648a0c5.zip b/.yarn/cache/@fullcalendar-vue3-npm-5.11.3-047b9981f6-13a648a0c5.zip deleted file mode 100644 index 56420f95cd..0000000000 Binary files a/.yarn/cache/@fullcalendar-vue3-npm-5.11.3-047b9981f6-13a648a0c5.zip and /dev/null differ diff --git a/.yarn/cache/@fullcalendar-vue3-npm-6.1.11-f6b8b48da4-5891a596e9.zip b/.yarn/cache/@fullcalendar-vue3-npm-6.1.11-f6b8b48da4-5891a596e9.zip new file mode 100644 index 0000000000..3054aa761f Binary files /dev/null and b/.yarn/cache/@fullcalendar-vue3-npm-6.1.11-f6b8b48da4-5891a596e9.zip differ diff --git a/.yarn/cache/@html-validate-stylish-npm-3.0.0-6d9dccafda-818efd25ac.zip b/.yarn/cache/@html-validate-stylish-npm-3.0.0-6d9dccafda-818efd25ac.zip deleted file mode 100644 index 5ed78286fb..0000000000 Binary files a/.yarn/cache/@html-validate-stylish-npm-3.0.0-6d9dccafda-818efd25ac.zip and /dev/null differ diff --git a/.yarn/cache/@html-validate-stylish-npm-4.1.0-aba0cf2d6c-4af90db4f9.zip b/.yarn/cache/@html-validate-stylish-npm-4.1.0-aba0cf2d6c-4af90db4f9.zip new file mode 100644 index 0000000000..d56d9f34cf Binary files /dev/null and b/.yarn/cache/@html-validate-stylish-npm-4.1.0-aba0cf2d6c-4af90db4f9.zip differ diff --git a/.yarn/cache/@humanwhocodes-config-array-npm-0.11.14-94a02fcc87-861ccce9ea.zip b/.yarn/cache/@humanwhocodes-config-array-npm-0.11.14-94a02fcc87-861ccce9ea.zip new file mode 100644 index 0000000000..166fee4b82 Binary files /dev/null and b/.yarn/cache/@humanwhocodes-config-array-npm-0.11.14-94a02fcc87-861ccce9ea.zip differ diff --git a/.yarn/cache/@humanwhocodes-config-array-npm-0.11.8-7955bfecc2-0fd6b3c54f.zip b/.yarn/cache/@humanwhocodes-config-array-npm-0.11.8-7955bfecc2-0fd6b3c54f.zip deleted file mode 100755 index dc21af1cbc..0000000000 Binary files a/.yarn/cache/@humanwhocodes-config-array-npm-0.11.8-7955bfecc2-0fd6b3c54f.zip and /dev/null differ diff --git a/.yarn/cache/@humanwhocodes-object-schema-npm-1.2.1-eb622b5d0e-a824a1ec31.zip b/.yarn/cache/@humanwhocodes-object-schema-npm-1.2.1-eb622b5d0e-a824a1ec31.zip deleted file mode 100644 index 2b79104af5..0000000000 Binary files a/.yarn/cache/@humanwhocodes-object-schema-npm-1.2.1-eb622b5d0e-a824a1ec31.zip and /dev/null differ diff --git a/.yarn/cache/@humanwhocodes-object-schema-npm-2.0.2-77b42018f9-2fc1150336.zip b/.yarn/cache/@humanwhocodes-object-schema-npm-2.0.2-77b42018f9-2fc1150336.zip new file mode 100644 index 0000000000..cf6847cf44 Binary files /dev/null and b/.yarn/cache/@humanwhocodes-object-schema-npm-2.0.2-77b42018f9-2fc1150336.zip differ diff --git a/.yarn/cache/@isaacs-cliui-npm-8.0.2-f4364666d5-4a473b9b32.zip b/.yarn/cache/@isaacs-cliui-npm-8.0.2-f4364666d5-4a473b9b32.zip new file mode 100644 index 0000000000..d19176fadd Binary files /dev/null and b/.yarn/cache/@isaacs-cliui-npm-8.0.2-f4364666d5-4a473b9b32.zip differ diff --git a/.yarn/cache/@jridgewell-sourcemap-codec-npm-1.4.15-a055fb62cf-b881c7e503.zip b/.yarn/cache/@jridgewell-sourcemap-codec-npm-1.4.15-a055fb62cf-b881c7e503.zip new file mode 100644 index 0000000000..402f52b7ae Binary files /dev/null and b/.yarn/cache/@jridgewell-sourcemap-codec-npm-1.4.15-a055fb62cf-b881c7e503.zip differ diff --git a/.yarn/cache/@lmdb-lmdb-darwin-arm64-npm-2.8.5-a9ab00615c-8.zip b/.yarn/cache/@lmdb-lmdb-darwin-arm64-npm-2.8.5-a9ab00615c-8.zip new file mode 100644 index 0000000000..6df931b4af Binary files /dev/null and b/.yarn/cache/@lmdb-lmdb-darwin-arm64-npm-2.8.5-a9ab00615c-8.zip differ diff --git a/.yarn/cache/@lmdb-lmdb-darwin-x64-npm-2.8.5-080b8c9329-8.zip b/.yarn/cache/@lmdb-lmdb-darwin-x64-npm-2.8.5-080b8c9329-8.zip new file mode 100644 index 0000000000..db77cafaea Binary files /dev/null and b/.yarn/cache/@lmdb-lmdb-darwin-x64-npm-2.8.5-080b8c9329-8.zip differ diff --git a/.yarn/cache/@lmdb-lmdb-linux-arm64-npm-2.8.5-9dfda9f24f-8.zip b/.yarn/cache/@lmdb-lmdb-linux-arm64-npm-2.8.5-9dfda9f24f-8.zip new file mode 100644 index 0000000000..d4522df85e Binary files /dev/null and b/.yarn/cache/@lmdb-lmdb-linux-arm64-npm-2.8.5-9dfda9f24f-8.zip differ diff --git a/.yarn/cache/@lmdb-lmdb-linux-x64-npm-2.8.5-0f668ba9a7-8.zip b/.yarn/cache/@lmdb-lmdb-linux-x64-npm-2.8.5-0f668ba9a7-8.zip new file mode 100644 index 0000000000..8820ec421f Binary files /dev/null and b/.yarn/cache/@lmdb-lmdb-linux-x64-npm-2.8.5-0f668ba9a7-8.zip differ diff --git a/.yarn/cache/@lmdb-lmdb-win32-x64-npm-2.8.5-3702de4edb-8.zip b/.yarn/cache/@lmdb-lmdb-win32-x64-npm-2.8.5-3702de4edb-8.zip new file mode 100644 index 0000000000..201d7cb1f1 Binary files /dev/null and b/.yarn/cache/@lmdb-lmdb-win32-x64-npm-2.8.5-3702de4edb-8.zip differ diff --git a/.yarn/cache/@msgpackr-extract-msgpackr-extract-darwin-arm64-npm-3.0.2-18ac236cc4-8.zip b/.yarn/cache/@msgpackr-extract-msgpackr-extract-darwin-arm64-npm-3.0.2-18ac236cc4-8.zip new file mode 100644 index 0000000000..06cbbf0cff Binary files /dev/null and b/.yarn/cache/@msgpackr-extract-msgpackr-extract-darwin-arm64-npm-3.0.2-18ac236cc4-8.zip differ diff --git a/.yarn/cache/@msgpackr-extract-msgpackr-extract-darwin-x64-npm-3.0.2-39dd07082a-8.zip b/.yarn/cache/@msgpackr-extract-msgpackr-extract-darwin-x64-npm-3.0.2-39dd07082a-8.zip new file mode 100644 index 0000000000..110c956115 Binary files /dev/null and b/.yarn/cache/@msgpackr-extract-msgpackr-extract-darwin-x64-npm-3.0.2-39dd07082a-8.zip differ diff --git a/.yarn/cache/@msgpackr-extract-msgpackr-extract-linux-arm64-npm-3.0.2-cfbf50d4c6-8.zip b/.yarn/cache/@msgpackr-extract-msgpackr-extract-linux-arm64-npm-3.0.2-cfbf50d4c6-8.zip new file mode 100644 index 0000000000..ab2c36a442 Binary files /dev/null and b/.yarn/cache/@msgpackr-extract-msgpackr-extract-linux-arm64-npm-3.0.2-cfbf50d4c6-8.zip differ diff --git a/.yarn/cache/@msgpackr-extract-msgpackr-extract-linux-x64-npm-3.0.2-262fca760d-8.zip b/.yarn/cache/@msgpackr-extract-msgpackr-extract-linux-x64-npm-3.0.2-262fca760d-8.zip new file mode 100644 index 0000000000..2fa6ef4f77 Binary files /dev/null and b/.yarn/cache/@msgpackr-extract-msgpackr-extract-linux-x64-npm-3.0.2-262fca760d-8.zip differ diff --git a/.yarn/cache/@msgpackr-extract-msgpackr-extract-win32-x64-npm-3.0.2-c627beab89-8.zip b/.yarn/cache/@msgpackr-extract-msgpackr-extract-win32-x64-npm-3.0.2-c627beab89-8.zip new file mode 100644 index 0000000000..b63546421d Binary files /dev/null and b/.yarn/cache/@msgpackr-extract-msgpackr-extract-win32-x64-npm-3.0.2-c627beab89-8.zip differ diff --git a/.yarn/cache/@parcel-bundler-default-npm-2.12.0-9ba57d919c-f211a76f55.zip b/.yarn/cache/@parcel-bundler-default-npm-2.12.0-9ba57d919c-f211a76f55.zip new file mode 100644 index 0000000000..024e036391 Binary files /dev/null and b/.yarn/cache/@parcel-bundler-default-npm-2.12.0-9ba57d919c-f211a76f55.zip differ diff --git a/.yarn/cache/@parcel-bundler-default-npm-2.8.2-497641ec3a-8330a76248.zip b/.yarn/cache/@parcel-bundler-default-npm-2.8.2-497641ec3a-8330a76248.zip deleted file mode 100755 index 1ef5a897a5..0000000000 Binary files a/.yarn/cache/@parcel-bundler-default-npm-2.8.2-497641ec3a-8330a76248.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-cache-npm-2.12.0-3389909f2c-a45e799809.zip b/.yarn/cache/@parcel-cache-npm-2.12.0-3389909f2c-a45e799809.zip new file mode 100644 index 0000000000..a358668eb7 Binary files /dev/null and b/.yarn/cache/@parcel-cache-npm-2.12.0-3389909f2c-a45e799809.zip differ diff --git a/.yarn/cache/@parcel-cache-npm-2.8.2-4957caf228-7d1c951e3f.zip b/.yarn/cache/@parcel-cache-npm-2.8.2-4957caf228-7d1c951e3f.zip deleted file mode 100755 index d593bf8fb2..0000000000 Binary files a/.yarn/cache/@parcel-cache-npm-2.8.2-4957caf228-7d1c951e3f.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-codeframe-npm-2.12.0-aa8027940e-265c4d7ebe.zip b/.yarn/cache/@parcel-codeframe-npm-2.12.0-aa8027940e-265c4d7ebe.zip new file mode 100644 index 0000000000..f4239d8ba7 Binary files /dev/null and b/.yarn/cache/@parcel-codeframe-npm-2.12.0-aa8027940e-265c4d7ebe.zip differ diff --git a/.yarn/cache/@parcel-codeframe-npm-2.8.2-77f4dce4ad-a2638353c6.zip b/.yarn/cache/@parcel-codeframe-npm-2.8.2-77f4dce4ad-a2638353c6.zip deleted file mode 100755 index 1bcdf805a0..0000000000 Binary files a/.yarn/cache/@parcel-codeframe-npm-2.8.2-77f4dce4ad-a2638353c6.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-compressor-raw-npm-2.12.0-19f313c172-16c56704f3.zip b/.yarn/cache/@parcel-compressor-raw-npm-2.12.0-19f313c172-16c56704f3.zip new file mode 100644 index 0000000000..da57625381 Binary files /dev/null and b/.yarn/cache/@parcel-compressor-raw-npm-2.12.0-19f313c172-16c56704f3.zip differ diff --git a/.yarn/cache/@parcel-compressor-raw-npm-2.8.2-0d385dde76-61a1299615.zip b/.yarn/cache/@parcel-compressor-raw-npm-2.8.2-0d385dde76-61a1299615.zip deleted file mode 100755 index 862a956d61..0000000000 Binary files a/.yarn/cache/@parcel-compressor-raw-npm-2.8.2-0d385dde76-61a1299615.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-config-default-npm-2.12.0-aefd3c699e-72877c5dc4.zip b/.yarn/cache/@parcel-config-default-npm-2.12.0-aefd3c699e-72877c5dc4.zip new file mode 100644 index 0000000000..a4934d017e Binary files /dev/null and b/.yarn/cache/@parcel-config-default-npm-2.12.0-aefd3c699e-72877c5dc4.zip differ diff --git a/.yarn/cache/@parcel-config-default-npm-2.8.2-89026bc258-035db3ab37.zip b/.yarn/cache/@parcel-config-default-npm-2.8.2-89026bc258-035db3ab37.zip deleted file mode 100755 index 3dad00b056..0000000000 Binary files a/.yarn/cache/@parcel-config-default-npm-2.8.2-89026bc258-035db3ab37.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-core-npm-2.12.0-8f08b883d4-5bf6746308.zip b/.yarn/cache/@parcel-core-npm-2.12.0-8f08b883d4-5bf6746308.zip new file mode 100644 index 0000000000..42c39ebe36 Binary files /dev/null and b/.yarn/cache/@parcel-core-npm-2.12.0-8f08b883d4-5bf6746308.zip differ diff --git a/.yarn/cache/@parcel-core-npm-2.8.2-7ac9ecd9f9-0c989ef087.zip b/.yarn/cache/@parcel-core-npm-2.8.2-7ac9ecd9f9-0c989ef087.zip deleted file mode 100755 index 85040b3944..0000000000 Binary files a/.yarn/cache/@parcel-core-npm-2.8.2-7ac9ecd9f9-0c989ef087.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-diagnostic-npm-2.12.0-6e89ddad28-a4b918c1a0.zip b/.yarn/cache/@parcel-diagnostic-npm-2.12.0-6e89ddad28-a4b918c1a0.zip new file mode 100644 index 0000000000..a8e890bf5c Binary files /dev/null and b/.yarn/cache/@parcel-diagnostic-npm-2.12.0-6e89ddad28-a4b918c1a0.zip differ diff --git a/.yarn/cache/@parcel-diagnostic-npm-2.8.2-7f2dfb035e-91ca29cce4.zip b/.yarn/cache/@parcel-diagnostic-npm-2.8.2-7f2dfb035e-91ca29cce4.zip deleted file mode 100755 index ee9bdb7e45..0000000000 Binary files a/.yarn/cache/@parcel-diagnostic-npm-2.8.2-7f2dfb035e-91ca29cce4.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-events-npm-2.12.0-e6eff18c8c-136a8a2921.zip b/.yarn/cache/@parcel-events-npm-2.12.0-e6eff18c8c-136a8a2921.zip new file mode 100644 index 0000000000..b806eb99ac Binary files /dev/null and b/.yarn/cache/@parcel-events-npm-2.12.0-e6eff18c8c-136a8a2921.zip differ diff --git a/.yarn/cache/@parcel-events-npm-2.8.2-ddf12da1ba-99aad2e735.zip b/.yarn/cache/@parcel-events-npm-2.8.2-ddf12da1ba-99aad2e735.zip deleted file mode 100755 index 054d30dd82..0000000000 Binary files a/.yarn/cache/@parcel-events-npm-2.8.2-ddf12da1ba-99aad2e735.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-fs-npm-2.12.0-3c46842e62-43d454d55d.zip b/.yarn/cache/@parcel-fs-npm-2.12.0-3c46842e62-43d454d55d.zip new file mode 100644 index 0000000000..52cbc5f7f1 Binary files /dev/null and b/.yarn/cache/@parcel-fs-npm-2.12.0-3c46842e62-43d454d55d.zip differ diff --git a/.yarn/cache/@parcel-fs-npm-2.8.2-97422ca16d-c25408fe2d.zip b/.yarn/cache/@parcel-fs-npm-2.8.2-97422ca16d-c25408fe2d.zip deleted file mode 100755 index dff053a7f1..0000000000 Binary files a/.yarn/cache/@parcel-fs-npm-2.8.2-97422ca16d-c25408fe2d.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-fs-search-npm-2.8.2-a3c70b64fe-b6b5956cc1.zip b/.yarn/cache/@parcel-fs-search-npm-2.8.2-a3c70b64fe-b6b5956cc1.zip deleted file mode 100755 index 77295640bb..0000000000 Binary files a/.yarn/cache/@parcel-fs-search-npm-2.8.2-a3c70b64fe-b6b5956cc1.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-graph-npm-2.8.2-039d19c5f3-d503597911.zip b/.yarn/cache/@parcel-graph-npm-2.8.2-039d19c5f3-d503597911.zip deleted file mode 100755 index f2da21bcf8..0000000000 Binary files a/.yarn/cache/@parcel-graph-npm-2.8.2-039d19c5f3-d503597911.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-graph-npm-3.2.0-92821d4289-b4d31624fc.zip b/.yarn/cache/@parcel-graph-npm-3.2.0-92821d4289-b4d31624fc.zip new file mode 100644 index 0000000000..27f3718928 Binary files /dev/null and b/.yarn/cache/@parcel-graph-npm-3.2.0-92821d4289-b4d31624fc.zip differ diff --git a/.yarn/cache/@parcel-hash-npm-2.8.2-4189a2e2e3-03f11563d2.zip b/.yarn/cache/@parcel-hash-npm-2.8.2-4189a2e2e3-03f11563d2.zip deleted file mode 100755 index cd1b5bf8ee..0000000000 Binary files a/.yarn/cache/@parcel-hash-npm-2.8.2-4189a2e2e3-03f11563d2.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-logger-npm-2.12.0-7d2f85a906-be3fe9d9ea.zip b/.yarn/cache/@parcel-logger-npm-2.12.0-7d2f85a906-be3fe9d9ea.zip new file mode 100644 index 0000000000..7231e4c65d Binary files /dev/null and b/.yarn/cache/@parcel-logger-npm-2.12.0-7d2f85a906-be3fe9d9ea.zip differ diff --git a/.yarn/cache/@parcel-logger-npm-2.8.2-0b40fa2df8-8d9b4264cb.zip b/.yarn/cache/@parcel-logger-npm-2.8.2-0b40fa2df8-8d9b4264cb.zip deleted file mode 100755 index 25b177e4a0..0000000000 Binary files a/.yarn/cache/@parcel-logger-npm-2.8.2-0b40fa2df8-8d9b4264cb.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-markdown-ansi-npm-2.12.0-6b0fe453df-850ee665d9.zip b/.yarn/cache/@parcel-markdown-ansi-npm-2.12.0-6b0fe453df-850ee665d9.zip new file mode 100644 index 0000000000..22582b46fa Binary files /dev/null and b/.yarn/cache/@parcel-markdown-ansi-npm-2.12.0-6b0fe453df-850ee665d9.zip differ diff --git a/.yarn/cache/@parcel-markdown-ansi-npm-2.8.2-3a4b50f123-aaff302f12.zip b/.yarn/cache/@parcel-markdown-ansi-npm-2.8.2-3a4b50f123-aaff302f12.zip deleted file mode 100755 index d96d871f7a..0000000000 Binary files a/.yarn/cache/@parcel-markdown-ansi-npm-2.8.2-3a4b50f123-aaff302f12.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-namer-default-npm-2.12.0-28980cfd47-dc92ec0945.zip b/.yarn/cache/@parcel-namer-default-npm-2.12.0-28980cfd47-dc92ec0945.zip new file mode 100644 index 0000000000..7db7fb405c Binary files /dev/null and b/.yarn/cache/@parcel-namer-default-npm-2.12.0-28980cfd47-dc92ec0945.zip differ diff --git a/.yarn/cache/@parcel-namer-default-npm-2.8.2-d3e74161c0-c9592f4022.zip b/.yarn/cache/@parcel-namer-default-npm-2.8.2-d3e74161c0-c9592f4022.zip deleted file mode 100755 index 72c3c2ca7b..0000000000 Binary files a/.yarn/cache/@parcel-namer-default-npm-2.8.2-d3e74161c0-c9592f4022.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-node-resolver-core-npm-2.8.2-5629a9b021-92f0e2bf4b.zip b/.yarn/cache/@parcel-node-resolver-core-npm-2.8.2-5629a9b021-92f0e2bf4b.zip deleted file mode 100755 index 30731a9684..0000000000 Binary files a/.yarn/cache/@parcel-node-resolver-core-npm-2.8.2-5629a9b021-92f0e2bf4b.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-node-resolver-core-npm-3.3.0-53804df663-acc3721678.zip b/.yarn/cache/@parcel-node-resolver-core-npm-3.3.0-53804df663-acc3721678.zip new file mode 100644 index 0000000000..76a69962a6 Binary files /dev/null and b/.yarn/cache/@parcel-node-resolver-core-npm-3.3.0-53804df663-acc3721678.zip differ diff --git a/.yarn/cache/@parcel-optimizer-css-npm-2.12.0-f95bd4d060-abcdf58c29.zip b/.yarn/cache/@parcel-optimizer-css-npm-2.12.0-f95bd4d060-abcdf58c29.zip new file mode 100644 index 0000000000..f1c61749b9 Binary files /dev/null and b/.yarn/cache/@parcel-optimizer-css-npm-2.12.0-f95bd4d060-abcdf58c29.zip differ diff --git a/.yarn/cache/@parcel-optimizer-css-npm-2.8.2-6de222af5e-8298155bac.zip b/.yarn/cache/@parcel-optimizer-css-npm-2.8.2-6de222af5e-8298155bac.zip deleted file mode 100755 index 9ac5db4ec2..0000000000 Binary files a/.yarn/cache/@parcel-optimizer-css-npm-2.8.2-6de222af5e-8298155bac.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-optimizer-data-url-npm-2.12.0-dad3731170-0397293961.zip b/.yarn/cache/@parcel-optimizer-data-url-npm-2.12.0-dad3731170-0397293961.zip new file mode 100644 index 0000000000..28497d3327 Binary files /dev/null and b/.yarn/cache/@parcel-optimizer-data-url-npm-2.12.0-dad3731170-0397293961.zip differ diff --git a/.yarn/cache/@parcel-optimizer-data-url-npm-2.8.2-2b95b0c045-e0966a5e18.zip b/.yarn/cache/@parcel-optimizer-data-url-npm-2.8.2-2b95b0c045-e0966a5e18.zip deleted file mode 100755 index 39f7814cf8..0000000000 Binary files a/.yarn/cache/@parcel-optimizer-data-url-npm-2.8.2-2b95b0c045-e0966a5e18.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-optimizer-htmlnano-npm-2.12.0-cdd2835c12-64e571f56f.zip b/.yarn/cache/@parcel-optimizer-htmlnano-npm-2.12.0-cdd2835c12-64e571f56f.zip new file mode 100644 index 0000000000..4089a870fb Binary files /dev/null and b/.yarn/cache/@parcel-optimizer-htmlnano-npm-2.12.0-cdd2835c12-64e571f56f.zip differ diff --git a/.yarn/cache/@parcel-optimizer-htmlnano-npm-2.8.2-989bccf2aa-3913b51ccd.zip b/.yarn/cache/@parcel-optimizer-htmlnano-npm-2.8.2-989bccf2aa-3913b51ccd.zip deleted file mode 100755 index a2ca7c933d..0000000000 Binary files a/.yarn/cache/@parcel-optimizer-htmlnano-npm-2.8.2-989bccf2aa-3913b51ccd.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-optimizer-image-npm-2.12.0-4cbc56f72d-7d28379bf1.zip b/.yarn/cache/@parcel-optimizer-image-npm-2.12.0-4cbc56f72d-7d28379bf1.zip new file mode 100644 index 0000000000..8b0a44e756 Binary files /dev/null and b/.yarn/cache/@parcel-optimizer-image-npm-2.12.0-4cbc56f72d-7d28379bf1.zip differ diff --git a/.yarn/cache/@parcel-optimizer-image-npm-2.8.2-eb7453ba87-7e45b2698b.zip b/.yarn/cache/@parcel-optimizer-image-npm-2.8.2-eb7453ba87-7e45b2698b.zip deleted file mode 100755 index fa4ad53b6d..0000000000 Binary files a/.yarn/cache/@parcel-optimizer-image-npm-2.8.2-eb7453ba87-7e45b2698b.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-optimizer-svgo-npm-2.12.0-08c0f1b17f-d3a4d2de9f.zip b/.yarn/cache/@parcel-optimizer-svgo-npm-2.12.0-08c0f1b17f-d3a4d2de9f.zip new file mode 100644 index 0000000000..441bead99b Binary files /dev/null and b/.yarn/cache/@parcel-optimizer-svgo-npm-2.12.0-08c0f1b17f-d3a4d2de9f.zip differ diff --git a/.yarn/cache/@parcel-optimizer-svgo-npm-2.8.2-d86f49e88e-608179fb18.zip b/.yarn/cache/@parcel-optimizer-svgo-npm-2.8.2-d86f49e88e-608179fb18.zip deleted file mode 100755 index 7ee372b8fb..0000000000 Binary files a/.yarn/cache/@parcel-optimizer-svgo-npm-2.8.2-d86f49e88e-608179fb18.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-optimizer-swc-npm-2.12.0-fb535e4283-0b7fdf3df1.zip b/.yarn/cache/@parcel-optimizer-swc-npm-2.12.0-fb535e4283-0b7fdf3df1.zip new file mode 100644 index 0000000000..8b137cf673 Binary files /dev/null and b/.yarn/cache/@parcel-optimizer-swc-npm-2.12.0-fb535e4283-0b7fdf3df1.zip differ diff --git a/.yarn/cache/@parcel-optimizer-terser-npm-2.8.2-8af8c43b6e-e5cc9ef648.zip b/.yarn/cache/@parcel-optimizer-terser-npm-2.8.2-8af8c43b6e-e5cc9ef648.zip deleted file mode 100755 index 561a11976b..0000000000 Binary files a/.yarn/cache/@parcel-optimizer-terser-npm-2.8.2-8af8c43b6e-e5cc9ef648.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-package-manager-npm-2.12.0-fc90aacf70-a517e9efe1.zip b/.yarn/cache/@parcel-package-manager-npm-2.12.0-fc90aacf70-a517e9efe1.zip new file mode 100644 index 0000000000..1e757bdf2f Binary files /dev/null and b/.yarn/cache/@parcel-package-manager-npm-2.12.0-fc90aacf70-a517e9efe1.zip differ diff --git a/.yarn/cache/@parcel-package-manager-npm-2.8.2-40215edd8a-99d022d3fa.zip b/.yarn/cache/@parcel-package-manager-npm-2.8.2-40215edd8a-99d022d3fa.zip deleted file mode 100755 index 6598ca6df7..0000000000 Binary files a/.yarn/cache/@parcel-package-manager-npm-2.8.2-40215edd8a-99d022d3fa.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-packager-css-npm-2.12.0-b1c27a8323-684aaa1d85.zip b/.yarn/cache/@parcel-packager-css-npm-2.12.0-b1c27a8323-684aaa1d85.zip new file mode 100644 index 0000000000..4cf7815f57 Binary files /dev/null and b/.yarn/cache/@parcel-packager-css-npm-2.12.0-b1c27a8323-684aaa1d85.zip differ diff --git a/.yarn/cache/@parcel-packager-css-npm-2.8.2-63302c1b3b-18ba8e43b3.zip b/.yarn/cache/@parcel-packager-css-npm-2.8.2-63302c1b3b-18ba8e43b3.zip deleted file mode 100755 index d53942f2a1..0000000000 Binary files a/.yarn/cache/@parcel-packager-css-npm-2.8.2-63302c1b3b-18ba8e43b3.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-packager-html-npm-2.12.0-ad361b1265-ee558ad616.zip b/.yarn/cache/@parcel-packager-html-npm-2.12.0-ad361b1265-ee558ad616.zip new file mode 100644 index 0000000000..989402a62c Binary files /dev/null and b/.yarn/cache/@parcel-packager-html-npm-2.12.0-ad361b1265-ee558ad616.zip differ diff --git a/.yarn/cache/@parcel-packager-html-npm-2.8.2-b901dd589c-e4975a4869.zip b/.yarn/cache/@parcel-packager-html-npm-2.8.2-b901dd589c-e4975a4869.zip deleted file mode 100755 index 1e9a71d8bd..0000000000 Binary files a/.yarn/cache/@parcel-packager-html-npm-2.8.2-b901dd589c-e4975a4869.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-packager-js-npm-2.12.0-093e3200cd-2189b7ff15.zip b/.yarn/cache/@parcel-packager-js-npm-2.12.0-093e3200cd-2189b7ff15.zip new file mode 100644 index 0000000000..461ec50d28 Binary files /dev/null and b/.yarn/cache/@parcel-packager-js-npm-2.12.0-093e3200cd-2189b7ff15.zip differ diff --git a/.yarn/cache/@parcel-packager-js-npm-2.8.2-9730c3d7a1-5c4a74e9b2.zip b/.yarn/cache/@parcel-packager-js-npm-2.8.2-9730c3d7a1-5c4a74e9b2.zip deleted file mode 100755 index 145d0be176..0000000000 Binary files a/.yarn/cache/@parcel-packager-js-npm-2.8.2-9730c3d7a1-5c4a74e9b2.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-packager-raw-npm-2.12.0-b7f15635f8-39ce2fc7ae.zip b/.yarn/cache/@parcel-packager-raw-npm-2.12.0-b7f15635f8-39ce2fc7ae.zip new file mode 100644 index 0000000000..e27b5ed1e3 Binary files /dev/null and b/.yarn/cache/@parcel-packager-raw-npm-2.12.0-b7f15635f8-39ce2fc7ae.zip differ diff --git a/.yarn/cache/@parcel-packager-raw-npm-2.8.2-e7b417ac32-198984e93e.zip b/.yarn/cache/@parcel-packager-raw-npm-2.8.2-e7b417ac32-198984e93e.zip deleted file mode 100755 index cdaae52829..0000000000 Binary files a/.yarn/cache/@parcel-packager-raw-npm-2.8.2-e7b417ac32-198984e93e.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-packager-svg-npm-2.12.0-fa921ce522-436ac9ea39.zip b/.yarn/cache/@parcel-packager-svg-npm-2.12.0-fa921ce522-436ac9ea39.zip new file mode 100644 index 0000000000..f3d37303b0 Binary files /dev/null and b/.yarn/cache/@parcel-packager-svg-npm-2.12.0-fa921ce522-436ac9ea39.zip differ diff --git a/.yarn/cache/@parcel-packager-svg-npm-2.8.2-a7884bf9a1-7e10546425.zip b/.yarn/cache/@parcel-packager-svg-npm-2.8.2-a7884bf9a1-7e10546425.zip deleted file mode 100755 index 16435c0f00..0000000000 Binary files a/.yarn/cache/@parcel-packager-svg-npm-2.8.2-a7884bf9a1-7e10546425.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-packager-wasm-npm-2.12.0-ec551a9e29-a10e1cd988.zip b/.yarn/cache/@parcel-packager-wasm-npm-2.12.0-ec551a9e29-a10e1cd988.zip new file mode 100644 index 0000000000..5b569f2004 Binary files /dev/null and b/.yarn/cache/@parcel-packager-wasm-npm-2.12.0-ec551a9e29-a10e1cd988.zip differ diff --git a/.yarn/cache/@parcel-plugin-npm-2.12.0-947dec85d3-0b52f1dd06.zip b/.yarn/cache/@parcel-plugin-npm-2.12.0-947dec85d3-0b52f1dd06.zip new file mode 100644 index 0000000000..667d7230e6 Binary files /dev/null and b/.yarn/cache/@parcel-plugin-npm-2.12.0-947dec85d3-0b52f1dd06.zip differ diff --git a/.yarn/cache/@parcel-plugin-npm-2.8.2-1747a062e1-5c9f0ec6ff.zip b/.yarn/cache/@parcel-plugin-npm-2.8.2-1747a062e1-5c9f0ec6ff.zip deleted file mode 100755 index ee49844e5c..0000000000 Binary files a/.yarn/cache/@parcel-plugin-npm-2.8.2-1747a062e1-5c9f0ec6ff.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-profiler-npm-2.12.0-69720a23ab-b683b74e10.zip b/.yarn/cache/@parcel-profiler-npm-2.12.0-69720a23ab-b683b74e10.zip new file mode 100644 index 0000000000..1cacc84571 Binary files /dev/null and b/.yarn/cache/@parcel-profiler-npm-2.12.0-69720a23ab-b683b74e10.zip differ diff --git a/.yarn/cache/@parcel-reporter-cli-npm-2.12.0-b3e4c5fe19-8cc524fa15.zip b/.yarn/cache/@parcel-reporter-cli-npm-2.12.0-b3e4c5fe19-8cc524fa15.zip new file mode 100644 index 0000000000..f6e625d396 Binary files /dev/null and b/.yarn/cache/@parcel-reporter-cli-npm-2.12.0-b3e4c5fe19-8cc524fa15.zip differ diff --git a/.yarn/cache/@parcel-reporter-cli-npm-2.8.2-57fd49365f-5ac5cbb7c3.zip b/.yarn/cache/@parcel-reporter-cli-npm-2.8.2-57fd49365f-5ac5cbb7c3.zip deleted file mode 100755 index 86ad8728b0..0000000000 Binary files a/.yarn/cache/@parcel-reporter-cli-npm-2.8.2-57fd49365f-5ac5cbb7c3.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-reporter-dev-server-npm-2.12.0-aed1d2c68c-43957b4656.zip b/.yarn/cache/@parcel-reporter-dev-server-npm-2.12.0-aed1d2c68c-43957b4656.zip new file mode 100644 index 0000000000..f1fb1818e9 Binary files /dev/null and b/.yarn/cache/@parcel-reporter-dev-server-npm-2.12.0-aed1d2c68c-43957b4656.zip differ diff --git a/.yarn/cache/@parcel-reporter-dev-server-npm-2.8.2-55972e618f-1efff76ed9.zip b/.yarn/cache/@parcel-reporter-dev-server-npm-2.8.2-55972e618f-1efff76ed9.zip deleted file mode 100755 index 22da914e3c..0000000000 Binary files a/.yarn/cache/@parcel-reporter-dev-server-npm-2.8.2-55972e618f-1efff76ed9.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-reporter-tracer-npm-2.12.0-5cec9ab2d5-24cddacd19.zip b/.yarn/cache/@parcel-reporter-tracer-npm-2.12.0-5cec9ab2d5-24cddacd19.zip new file mode 100644 index 0000000000..2196f5407c Binary files /dev/null and b/.yarn/cache/@parcel-reporter-tracer-npm-2.12.0-5cec9ab2d5-24cddacd19.zip differ diff --git a/.yarn/cache/@parcel-resolver-default-npm-2.12.0-8da790891c-f3652eea09.zip b/.yarn/cache/@parcel-resolver-default-npm-2.12.0-8da790891c-f3652eea09.zip new file mode 100644 index 0000000000..8022d04651 Binary files /dev/null and b/.yarn/cache/@parcel-resolver-default-npm-2.12.0-8da790891c-f3652eea09.zip differ diff --git a/.yarn/cache/@parcel-resolver-default-npm-2.8.2-f0fe8ef74c-66e0233ed6.zip b/.yarn/cache/@parcel-resolver-default-npm-2.8.2-f0fe8ef74c-66e0233ed6.zip deleted file mode 100755 index 49e8bbd505..0000000000 Binary files a/.yarn/cache/@parcel-resolver-default-npm-2.8.2-f0fe8ef74c-66e0233ed6.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-runtime-browser-hmr-npm-2.12.0-6f0da66673-bbba57ecee.zip b/.yarn/cache/@parcel-runtime-browser-hmr-npm-2.12.0-6f0da66673-bbba57ecee.zip new file mode 100644 index 0000000000..f71de2152b Binary files /dev/null and b/.yarn/cache/@parcel-runtime-browser-hmr-npm-2.12.0-6f0da66673-bbba57ecee.zip differ diff --git a/.yarn/cache/@parcel-runtime-browser-hmr-npm-2.8.2-bfd277b18f-64543de8cf.zip b/.yarn/cache/@parcel-runtime-browser-hmr-npm-2.8.2-bfd277b18f-64543de8cf.zip deleted file mode 100755 index 21eb8bd3c8..0000000000 Binary files a/.yarn/cache/@parcel-runtime-browser-hmr-npm-2.8.2-bfd277b18f-64543de8cf.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-runtime-js-npm-2.12.0-e21acc0f42-6afa3e7eb2.zip b/.yarn/cache/@parcel-runtime-js-npm-2.12.0-e21acc0f42-6afa3e7eb2.zip new file mode 100644 index 0000000000..be9c7d7e4b Binary files /dev/null and b/.yarn/cache/@parcel-runtime-js-npm-2.12.0-e21acc0f42-6afa3e7eb2.zip differ diff --git a/.yarn/cache/@parcel-runtime-js-npm-2.8.2-171208460f-a5c0c7d2ad.zip b/.yarn/cache/@parcel-runtime-js-npm-2.8.2-171208460f-a5c0c7d2ad.zip deleted file mode 100755 index 0be8126f63..0000000000 Binary files a/.yarn/cache/@parcel-runtime-js-npm-2.8.2-171208460f-a5c0c7d2ad.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-runtime-react-refresh-npm-2.12.0-2b09615691-41aee9a874.zip b/.yarn/cache/@parcel-runtime-react-refresh-npm-2.12.0-2b09615691-41aee9a874.zip new file mode 100644 index 0000000000..8dc8e5281c Binary files /dev/null and b/.yarn/cache/@parcel-runtime-react-refresh-npm-2.12.0-2b09615691-41aee9a874.zip differ diff --git a/.yarn/cache/@parcel-runtime-react-refresh-npm-2.8.2-2b20ac8c6d-6483b8ed55.zip b/.yarn/cache/@parcel-runtime-react-refresh-npm-2.8.2-2b20ac8c6d-6483b8ed55.zip deleted file mode 100755 index 83e38937b9..0000000000 Binary files a/.yarn/cache/@parcel-runtime-react-refresh-npm-2.8.2-2b20ac8c6d-6483b8ed55.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-runtime-service-worker-npm-2.12.0-7d227ff0bf-c71246428e.zip b/.yarn/cache/@parcel-runtime-service-worker-npm-2.12.0-7d227ff0bf-c71246428e.zip new file mode 100644 index 0000000000..18682c22ae Binary files /dev/null and b/.yarn/cache/@parcel-runtime-service-worker-npm-2.12.0-7d227ff0bf-c71246428e.zip differ diff --git a/.yarn/cache/@parcel-runtime-service-worker-npm-2.8.2-1ec24cff9d-4b52703d3b.zip b/.yarn/cache/@parcel-runtime-service-worker-npm-2.8.2-1ec24cff9d-4b52703d3b.zip deleted file mode 100755 index bccda1201b..0000000000 Binary files a/.yarn/cache/@parcel-runtime-service-worker-npm-2.8.2-1ec24cff9d-4b52703d3b.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-rust-npm-2.12.0-0cf943f3e5-51c5b67b9e.zip b/.yarn/cache/@parcel-rust-npm-2.12.0-0cf943f3e5-51c5b67b9e.zip new file mode 100644 index 0000000000..d5fe4206c9 Binary files /dev/null and b/.yarn/cache/@parcel-rust-npm-2.12.0-0cf943f3e5-51c5b67b9e.zip differ diff --git a/.yarn/cache/@parcel-transformer-babel-npm-2.12.0-953de52432-b8c457c0be.zip b/.yarn/cache/@parcel-transformer-babel-npm-2.12.0-953de52432-b8c457c0be.zip new file mode 100644 index 0000000000..9286325c9e Binary files /dev/null and b/.yarn/cache/@parcel-transformer-babel-npm-2.12.0-953de52432-b8c457c0be.zip differ diff --git a/.yarn/cache/@parcel-transformer-babel-npm-2.8.2-94dae9d0e8-4b2064aaba.zip b/.yarn/cache/@parcel-transformer-babel-npm-2.8.2-94dae9d0e8-4b2064aaba.zip deleted file mode 100755 index eaaf19f45f..0000000000 Binary files a/.yarn/cache/@parcel-transformer-babel-npm-2.8.2-94dae9d0e8-4b2064aaba.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-transformer-css-npm-2.12.0-24ddc31ae3-3a6f16321d.zip b/.yarn/cache/@parcel-transformer-css-npm-2.12.0-24ddc31ae3-3a6f16321d.zip new file mode 100644 index 0000000000..f3e0520c71 Binary files /dev/null and b/.yarn/cache/@parcel-transformer-css-npm-2.12.0-24ddc31ae3-3a6f16321d.zip differ diff --git a/.yarn/cache/@parcel-transformer-css-npm-2.8.2-283cfa7f07-d0d3121d2b.zip b/.yarn/cache/@parcel-transformer-css-npm-2.8.2-283cfa7f07-d0d3121d2b.zip deleted file mode 100755 index e0449519e0..0000000000 Binary files a/.yarn/cache/@parcel-transformer-css-npm-2.8.2-283cfa7f07-d0d3121d2b.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-transformer-html-npm-2.12.0-be2b9ee40c-7fcfac62ca.zip b/.yarn/cache/@parcel-transformer-html-npm-2.12.0-be2b9ee40c-7fcfac62ca.zip new file mode 100644 index 0000000000..3628f3f90d Binary files /dev/null and b/.yarn/cache/@parcel-transformer-html-npm-2.12.0-be2b9ee40c-7fcfac62ca.zip differ diff --git a/.yarn/cache/@parcel-transformer-html-npm-2.8.2-998bc39b95-e3bead4866.zip b/.yarn/cache/@parcel-transformer-html-npm-2.8.2-998bc39b95-e3bead4866.zip deleted file mode 100755 index 944b6f4f10..0000000000 Binary files a/.yarn/cache/@parcel-transformer-html-npm-2.8.2-998bc39b95-e3bead4866.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-transformer-image-npm-2.12.0-53f04e21c0-0a1581eacc.zip b/.yarn/cache/@parcel-transformer-image-npm-2.12.0-53f04e21c0-0a1581eacc.zip new file mode 100644 index 0000000000..3a78e4e070 Binary files /dev/null and b/.yarn/cache/@parcel-transformer-image-npm-2.12.0-53f04e21c0-0a1581eacc.zip differ diff --git a/.yarn/cache/@parcel-transformer-image-npm-2.8.2-c8f5d0643b-acfe6e06f3.zip b/.yarn/cache/@parcel-transformer-image-npm-2.8.2-c8f5d0643b-acfe6e06f3.zip deleted file mode 100755 index f9e3582bfa..0000000000 Binary files a/.yarn/cache/@parcel-transformer-image-npm-2.8.2-c8f5d0643b-acfe6e06f3.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-transformer-inline-string-npm-2.12.0-a33f10bafa-5f63c08695.zip b/.yarn/cache/@parcel-transformer-inline-string-npm-2.12.0-a33f10bafa-5f63c08695.zip new file mode 100644 index 0000000000..0c4f3341c8 Binary files /dev/null and b/.yarn/cache/@parcel-transformer-inline-string-npm-2.12.0-a33f10bafa-5f63c08695.zip differ diff --git a/.yarn/cache/@parcel-transformer-inline-string-npm-2.8.2-3a03397064-5f6f4be447.zip b/.yarn/cache/@parcel-transformer-inline-string-npm-2.8.2-3a03397064-5f6f4be447.zip deleted file mode 100755 index febf8502bb..0000000000 Binary files a/.yarn/cache/@parcel-transformer-inline-string-npm-2.8.2-3a03397064-5f6f4be447.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-transformer-js-npm-2.12.0-404d54db18-b9fe4c887b.zip b/.yarn/cache/@parcel-transformer-js-npm-2.12.0-404d54db18-b9fe4c887b.zip new file mode 100644 index 0000000000..1ce667ac8d Binary files /dev/null and b/.yarn/cache/@parcel-transformer-js-npm-2.12.0-404d54db18-b9fe4c887b.zip differ diff --git a/.yarn/cache/@parcel-transformer-js-npm-2.8.2-79df2d6c4f-2ccbe5f98e.zip b/.yarn/cache/@parcel-transformer-js-npm-2.8.2-79df2d6c4f-2ccbe5f98e.zip deleted file mode 100755 index 0b6c7808e2..0000000000 Binary files a/.yarn/cache/@parcel-transformer-js-npm-2.8.2-79df2d6c4f-2ccbe5f98e.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-transformer-json-npm-2.12.0-652d8d99d2-a711cb65a8.zip b/.yarn/cache/@parcel-transformer-json-npm-2.12.0-652d8d99d2-a711cb65a8.zip new file mode 100644 index 0000000000..926c01eb81 Binary files /dev/null and b/.yarn/cache/@parcel-transformer-json-npm-2.12.0-652d8d99d2-a711cb65a8.zip differ diff --git a/.yarn/cache/@parcel-transformer-json-npm-2.8.2-98e2e0cf80-b22a609ae9.zip b/.yarn/cache/@parcel-transformer-json-npm-2.8.2-98e2e0cf80-b22a609ae9.zip deleted file mode 100755 index 755721c310..0000000000 Binary files a/.yarn/cache/@parcel-transformer-json-npm-2.8.2-98e2e0cf80-b22a609ae9.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-transformer-postcss-npm-2.12.0-f0cfb95fac-b210044a7f.zip b/.yarn/cache/@parcel-transformer-postcss-npm-2.12.0-f0cfb95fac-b210044a7f.zip new file mode 100644 index 0000000000..3bbacafa81 Binary files /dev/null and b/.yarn/cache/@parcel-transformer-postcss-npm-2.12.0-f0cfb95fac-b210044a7f.zip differ diff --git a/.yarn/cache/@parcel-transformer-postcss-npm-2.8.2-547cd470da-ee152a91fb.zip b/.yarn/cache/@parcel-transformer-postcss-npm-2.8.2-547cd470da-ee152a91fb.zip deleted file mode 100755 index fd2e63121d..0000000000 Binary files a/.yarn/cache/@parcel-transformer-postcss-npm-2.8.2-547cd470da-ee152a91fb.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-transformer-posthtml-npm-2.12.0-41c570db12-b62582ae7e.zip b/.yarn/cache/@parcel-transformer-posthtml-npm-2.12.0-41c570db12-b62582ae7e.zip new file mode 100644 index 0000000000..e912a09713 Binary files /dev/null and b/.yarn/cache/@parcel-transformer-posthtml-npm-2.12.0-41c570db12-b62582ae7e.zip differ diff --git a/.yarn/cache/@parcel-transformer-posthtml-npm-2.8.2-76f67e31b6-4865968546.zip b/.yarn/cache/@parcel-transformer-posthtml-npm-2.8.2-76f67e31b6-4865968546.zip deleted file mode 100755 index b67b15f4ce..0000000000 Binary files a/.yarn/cache/@parcel-transformer-posthtml-npm-2.8.2-76f67e31b6-4865968546.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-transformer-raw-npm-2.12.0-bd2cb66ddf-de6681e2e7.zip b/.yarn/cache/@parcel-transformer-raw-npm-2.12.0-bd2cb66ddf-de6681e2e7.zip new file mode 100644 index 0000000000..40b7e2d3c4 Binary files /dev/null and b/.yarn/cache/@parcel-transformer-raw-npm-2.12.0-bd2cb66ddf-de6681e2e7.zip differ diff --git a/.yarn/cache/@parcel-transformer-raw-npm-2.8.2-a43c4fa2f7-386f64445a.zip b/.yarn/cache/@parcel-transformer-raw-npm-2.8.2-a43c4fa2f7-386f64445a.zip deleted file mode 100755 index fe5decf967..0000000000 Binary files a/.yarn/cache/@parcel-transformer-raw-npm-2.8.2-a43c4fa2f7-386f64445a.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-transformer-react-refresh-wrap-npm-2.12.0-59ed68910f-9aba8c1ab0.zip b/.yarn/cache/@parcel-transformer-react-refresh-wrap-npm-2.12.0-59ed68910f-9aba8c1ab0.zip new file mode 100644 index 0000000000..23210becb7 Binary files /dev/null and b/.yarn/cache/@parcel-transformer-react-refresh-wrap-npm-2.12.0-59ed68910f-9aba8c1ab0.zip differ diff --git a/.yarn/cache/@parcel-transformer-react-refresh-wrap-npm-2.8.2-62e6dd04c2-d091ab4a25.zip b/.yarn/cache/@parcel-transformer-react-refresh-wrap-npm-2.8.2-62e6dd04c2-d091ab4a25.zip deleted file mode 100755 index 07ed434f06..0000000000 Binary files a/.yarn/cache/@parcel-transformer-react-refresh-wrap-npm-2.8.2-62e6dd04c2-d091ab4a25.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-transformer-sass-npm-2.12.0-ef787eef35-ce6b4d329b.zip b/.yarn/cache/@parcel-transformer-sass-npm-2.12.0-ef787eef35-ce6b4d329b.zip new file mode 100644 index 0000000000..d62c342067 Binary files /dev/null and b/.yarn/cache/@parcel-transformer-sass-npm-2.12.0-ef787eef35-ce6b4d329b.zip differ diff --git a/.yarn/cache/@parcel-transformer-sass-npm-2.8.2-4e0c2f2900-42bbfa9401.zip b/.yarn/cache/@parcel-transformer-sass-npm-2.8.2-4e0c2f2900-42bbfa9401.zip deleted file mode 100755 index a1ae546d87..0000000000 Binary files a/.yarn/cache/@parcel-transformer-sass-npm-2.8.2-4e0c2f2900-42bbfa9401.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-transformer-svg-npm-2.12.0-f41b181676-92b7c65894.zip b/.yarn/cache/@parcel-transformer-svg-npm-2.12.0-f41b181676-92b7c65894.zip new file mode 100644 index 0000000000..01af21f6a3 Binary files /dev/null and b/.yarn/cache/@parcel-transformer-svg-npm-2.12.0-f41b181676-92b7c65894.zip differ diff --git a/.yarn/cache/@parcel-transformer-svg-npm-2.8.2-c8870e67e5-e4522b69e3.zip b/.yarn/cache/@parcel-transformer-svg-npm-2.8.2-c8870e67e5-e4522b69e3.zip deleted file mode 100755 index a697b835d5..0000000000 Binary files a/.yarn/cache/@parcel-transformer-svg-npm-2.8.2-c8870e67e5-e4522b69e3.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-types-npm-2.12.0-ffe47febbf-250f95580c.zip b/.yarn/cache/@parcel-types-npm-2.12.0-ffe47febbf-250f95580c.zip new file mode 100644 index 0000000000..ea6decc566 Binary files /dev/null and b/.yarn/cache/@parcel-types-npm-2.12.0-ffe47febbf-250f95580c.zip differ diff --git a/.yarn/cache/@parcel-types-npm-2.8.2-4a1952be09-04b3d5f199.zip b/.yarn/cache/@parcel-types-npm-2.8.2-4a1952be09-04b3d5f199.zip deleted file mode 100755 index d524b92234..0000000000 Binary files a/.yarn/cache/@parcel-types-npm-2.8.2-4a1952be09-04b3d5f199.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-utils-npm-2.12.0-d8a9a48a66-ba80a60fed.zip b/.yarn/cache/@parcel-utils-npm-2.12.0-d8a9a48a66-ba80a60fed.zip new file mode 100644 index 0000000000..8eda598941 Binary files /dev/null and b/.yarn/cache/@parcel-utils-npm-2.12.0-d8a9a48a66-ba80a60fed.zip differ diff --git a/.yarn/cache/@parcel-utils-npm-2.8.2-8c378b4d3a-fcbc70426e.zip b/.yarn/cache/@parcel-utils-npm-2.8.2-8c378b4d3a-fcbc70426e.zip deleted file mode 100755 index 7509c388e1..0000000000 Binary files a/.yarn/cache/@parcel-utils-npm-2.8.2-8c378b4d3a-fcbc70426e.zip and /dev/null differ diff --git a/.yarn/cache/@parcel-workers-npm-2.12.0-3ddd4664bc-e19c3c0a66.zip b/.yarn/cache/@parcel-workers-npm-2.12.0-3ddd4664bc-e19c3c0a66.zip new file mode 100644 index 0000000000..53f28c9470 Binary files /dev/null and b/.yarn/cache/@parcel-workers-npm-2.12.0-3ddd4664bc-e19c3c0a66.zip differ diff --git a/.yarn/cache/@parcel-workers-npm-2.8.2-48e612dc38-df3f793301.zip b/.yarn/cache/@parcel-workers-npm-2.8.2-48e612dc38-df3f793301.zip deleted file mode 100755 index 735f899f16..0000000000 Binary files a/.yarn/cache/@parcel-workers-npm-2.8.2-48e612dc38-df3f793301.zip and /dev/null differ diff --git a/.yarn/cache/@pkgjs-parseargs-npm-0.11.0-cd2a3fe948-6ad6a00fc4.zip b/.yarn/cache/@pkgjs-parseargs-npm-0.11.0-cd2a3fe948-6ad6a00fc4.zip new file mode 100644 index 0000000000..96f576f7de Binary files /dev/null and b/.yarn/cache/@pkgjs-parseargs-npm-0.11.0-cd2a3fe948-6ad6a00fc4.zip differ diff --git a/.yarn/cache/@popperjs-core-npm-2.11.6-5bcdc104bd-47fb328cec.zip b/.yarn/cache/@popperjs-core-npm-2.11.6-5bcdc104bd-47fb328cec.zip deleted file mode 100644 index e372ae30af..0000000000 Binary files a/.yarn/cache/@popperjs-core-npm-2.11.6-5bcdc104bd-47fb328cec.zip and /dev/null differ diff --git a/.yarn/cache/@popperjs-core-npm-2.11.8-f1692e11a0-e5c69fdebf.zip b/.yarn/cache/@popperjs-core-npm-2.11.8-f1692e11a0-e5c69fdebf.zip new file mode 100644 index 0000000000..a5eef4b227 Binary files /dev/null and b/.yarn/cache/@popperjs-core-npm-2.11.8-f1692e11a0-e5c69fdebf.zip differ diff --git a/.yarn/cache/@rollup-pluginutils-npm-5.0.2-6aa9d0ddd4-edea15e543.zip b/.yarn/cache/@rollup-pluginutils-npm-5.0.2-6aa9d0ddd4-edea15e543.zip deleted file mode 100644 index d898c5035c..0000000000 Binary files a/.yarn/cache/@rollup-pluginutils-npm-5.0.2-6aa9d0ddd4-edea15e543.zip and /dev/null differ diff --git a/.yarn/cache/@rollup-pluginutils-npm-5.1.0-6939820ef8-3cc5a6d914.zip b/.yarn/cache/@rollup-pluginutils-npm-5.1.0-6939820ef8-3cc5a6d914.zip new file mode 100644 index 0000000000..923a7a91a8 Binary files /dev/null and b/.yarn/cache/@rollup-pluginutils-npm-5.1.0-6939820ef8-3cc5a6d914.zip differ diff --git a/.yarn/cache/@sidvind-better-ajv-errors-npm-2.0.0-3531bddef9-12b0d87855.zip b/.yarn/cache/@sidvind-better-ajv-errors-npm-2.0.0-3531bddef9-12b0d87855.zip deleted file mode 100644 index 5990cc7604..0000000000 Binary files a/.yarn/cache/@sidvind-better-ajv-errors-npm-2.0.0-3531bddef9-12b0d87855.zip and /dev/null differ diff --git a/.yarn/cache/@sidvind-better-ajv-errors-npm-2.1.3-e3d1c524a8-949cb805a1.zip b/.yarn/cache/@sidvind-better-ajv-errors-npm-2.1.3-e3d1c524a8-949cb805a1.zip new file mode 100644 index 0000000000..ad36770e19 Binary files /dev/null and b/.yarn/cache/@sidvind-better-ajv-errors-npm-2.1.3-e3d1c524a8-949cb805a1.zip differ diff --git a/.yarn/cache/@swc-core-darwin-arm64-npm-1.3.62-b4af5d9b32-8.zip b/.yarn/cache/@swc-core-darwin-arm64-npm-1.3.62-b4af5d9b32-8.zip new file mode 100644 index 0000000000..ad2ff12c7f Binary files /dev/null and b/.yarn/cache/@swc-core-darwin-arm64-npm-1.3.62-b4af5d9b32-8.zip differ diff --git a/.yarn/cache/@swc-core-darwin-x64-npm-1.3.62-7d7bc99502-8.zip b/.yarn/cache/@swc-core-darwin-x64-npm-1.3.62-7d7bc99502-8.zip new file mode 100644 index 0000000000..7edd14afd0 Binary files /dev/null and b/.yarn/cache/@swc-core-darwin-x64-npm-1.3.62-7d7bc99502-8.zip differ diff --git a/.yarn/cache/@swc-core-linux-arm64-gnu-npm-1.3.62-7b527a3356-8.zip b/.yarn/cache/@swc-core-linux-arm64-gnu-npm-1.3.62-7b527a3356-8.zip new file mode 100644 index 0000000000..87afaa9285 Binary files /dev/null and b/.yarn/cache/@swc-core-linux-arm64-gnu-npm-1.3.62-7b527a3356-8.zip differ diff --git a/.yarn/cache/@swc-core-linux-x64-gnu-npm-1.3.62-1fc43a8907-8.zip b/.yarn/cache/@swc-core-linux-x64-gnu-npm-1.3.62-1fc43a8907-8.zip new file mode 100644 index 0000000000..a1aa6dbae5 Binary files /dev/null and b/.yarn/cache/@swc-core-linux-x64-gnu-npm-1.3.62-1fc43a8907-8.zip differ diff --git a/.yarn/cache/@swc-core-npm-1.3.62-9a4c32739d-a7a0d9ffdb.zip b/.yarn/cache/@swc-core-npm-1.3.62-9a4c32739d-a7a0d9ffdb.zip new file mode 100644 index 0000000000..dc6b151bf1 Binary files /dev/null and b/.yarn/cache/@swc-core-npm-1.3.62-9a4c32739d-a7a0d9ffdb.zip differ diff --git a/.yarn/cache/@swc-core-win32-arm64-msvc-npm-1.3.62-f4199145ca-8.zip b/.yarn/cache/@swc-core-win32-arm64-msvc-npm-1.3.62-f4199145ca-8.zip new file mode 100644 index 0000000000..bb62885e7f Binary files /dev/null and b/.yarn/cache/@swc-core-win32-arm64-msvc-npm-1.3.62-f4199145ca-8.zip differ diff --git a/.yarn/cache/@swc-core-win32-x64-msvc-npm-1.3.62-200450bac0-8.zip b/.yarn/cache/@swc-core-win32-x64-msvc-npm-1.3.62-200450bac0-8.zip new file mode 100644 index 0000000000..f306c1e943 Binary files /dev/null and b/.yarn/cache/@swc-core-win32-x64-msvc-npm-1.3.62-200450bac0-8.zip differ diff --git a/.yarn/cache/@swc-helpers-npm-0.4.14-f806c3fb16-273fd3f3fc.zip b/.yarn/cache/@swc-helpers-npm-0.4.14-f806c3fb16-273fd3f3fc.zip deleted file mode 100644 index 9cf5ea445e..0000000000 Binary files a/.yarn/cache/@swc-helpers-npm-0.4.14-f806c3fb16-273fd3f3fc.zip and /dev/null differ diff --git a/.yarn/cache/@swc-helpers-npm-0.5.1-424376f311-71e0e27234.zip b/.yarn/cache/@swc-helpers-npm-0.5.1-424376f311-71e0e27234.zip new file mode 100644 index 0000000000..36ed12e7cb Binary files /dev/null and b/.yarn/cache/@swc-helpers-npm-0.5.1-424376f311-71e0e27234.zip differ diff --git a/.yarn/cache/@types-jest-npm-27.4.1-31d07cd0d8-5184f3eef4.zip b/.yarn/cache/@types-jest-npm-27.4.1-31d07cd0d8-5184f3eef4.zip deleted file mode 100644 index 28e1c1b124..0000000000 Binary files a/.yarn/cache/@types-jest-npm-27.4.1-31d07cd0d8-5184f3eef4.zip and /dev/null differ diff --git a/.yarn/cache/@types-katex-npm-0.14.0-acd5bc3e87-330e0d0337.zip b/.yarn/cache/@types-katex-npm-0.14.0-acd5bc3e87-330e0d0337.zip deleted file mode 100644 index 79825772d5..0000000000 Binary files a/.yarn/cache/@types-katex-npm-0.14.0-acd5bc3e87-330e0d0337.zip and /dev/null differ diff --git a/.yarn/cache/@types-katex-npm-0.16.5-ff9336f176-a1ce22cd87.zip b/.yarn/cache/@types-katex-npm-0.16.5-ff9336f176-a1ce22cd87.zip new file mode 100644 index 0000000000..92aafc4818 Binary files /dev/null and b/.yarn/cache/@types-katex-npm-0.16.5-ff9336f176-a1ce22cd87.zip differ diff --git a/.yarn/cache/@types-lodash-es-npm-4.17.10-a7dae21818-129e9dde83.zip b/.yarn/cache/@types-lodash-es-npm-4.17.10-a7dae21818-129e9dde83.zip new file mode 100644 index 0000000000..d0043c3a60 Binary files /dev/null and b/.yarn/cache/@types-lodash-es-npm-4.17.10-a7dae21818-129e9dde83.zip differ diff --git a/.yarn/cache/@types-lodash-es-npm-4.17.6-fd5abbdc74-9bd239dd52.zip b/.yarn/cache/@types-lodash-es-npm-4.17.6-fd5abbdc74-9bd239dd52.zip deleted file mode 100644 index 3bd29bcbad..0000000000 Binary files a/.yarn/cache/@types-lodash-es-npm-4.17.6-fd5abbdc74-9bd239dd52.zip and /dev/null differ diff --git a/.yarn/cache/@types-lodash-npm-4.14.200-8559f51fce-6471f8bb5d.zip b/.yarn/cache/@types-lodash-npm-4.14.200-8559f51fce-6471f8bb5d.zip new file mode 100644 index 0000000000..ae8b2ba4c0 Binary files /dev/null and b/.yarn/cache/@types-lodash-npm-4.14.200-8559f51fce-6471f8bb5d.zip differ diff --git a/.yarn/cache/@ungap-structured-clone-npm-1.2.0-648f0b82e0-4f656b7b46.zip b/.yarn/cache/@ungap-structured-clone-npm-1.2.0-648f0b82e0-4f656b7b46.zip new file mode 100644 index 0000000000..598a36e085 Binary files /dev/null and b/.yarn/cache/@ungap-structured-clone-npm-1.2.0-648f0b82e0-4f656b7b46.zip differ diff --git a/.yarn/cache/@vitejs-plugin-vue-npm-3.2.0-d467fde943-64774f770e.zip b/.yarn/cache/@vitejs-plugin-vue-npm-3.2.0-d467fde943-64774f770e.zip deleted file mode 100644 index 69de575d01..0000000000 Binary files a/.yarn/cache/@vitejs-plugin-vue-npm-3.2.0-d467fde943-64774f770e.zip and /dev/null differ diff --git a/.yarn/cache/@vitejs-plugin-vue-npm-4.6.2-d7ace53203-01bc4ed643.zip b/.yarn/cache/@vitejs-plugin-vue-npm-4.6.2-d7ace53203-01bc4ed643.zip new file mode 100644 index 0000000000..7cf07fbe2d Binary files /dev/null and b/.yarn/cache/@vitejs-plugin-vue-npm-4.6.2-d7ace53203-01bc4ed643.zip differ diff --git a/.yarn/cache/@volar-language-core-npm-2.1.4-18ee1a037d-7430f65143.zip b/.yarn/cache/@volar-language-core-npm-2.1.4-18ee1a037d-7430f65143.zip new file mode 100644 index 0000000000..25e6d3f94d Binary files /dev/null and b/.yarn/cache/@volar-language-core-npm-2.1.4-18ee1a037d-7430f65143.zip differ diff --git a/.yarn/cache/@volar-language-service-npm-2.1.4-2d34cb628f-06cdcfacf0.zip b/.yarn/cache/@volar-language-service-npm-2.1.4-2d34cb628f-06cdcfacf0.zip new file mode 100644 index 0000000000..5f494d902e Binary files /dev/null and b/.yarn/cache/@volar-language-service-npm-2.1.4-2d34cb628f-06cdcfacf0.zip differ diff --git a/.yarn/cache/@volar-source-map-npm-2.1.4-5963b1701f-e2f65bcfd6.zip b/.yarn/cache/@volar-source-map-npm-2.1.4-5963b1701f-e2f65bcfd6.zip new file mode 100644 index 0000000000..0ea96c4d97 Binary files /dev/null and b/.yarn/cache/@volar-source-map-npm-2.1.4-5963b1701f-e2f65bcfd6.zip differ diff --git a/.yarn/cache/@vscode-l10n-npm-0.0.18-8a12efe4b5-c33876cebd.zip b/.yarn/cache/@vscode-l10n-npm-0.0.18-8a12efe4b5-c33876cebd.zip new file mode 100644 index 0000000000..2d6533a204 Binary files /dev/null and b/.yarn/cache/@vscode-l10n-npm-0.0.18-8a12efe4b5-c33876cebd.zip differ diff --git a/.yarn/cache/@vue-compiler-core-npm-3.2.45-2a68bebbd0-e3c687b24c.zip b/.yarn/cache/@vue-compiler-core-npm-3.2.45-2a68bebbd0-e3c687b24c.zip deleted file mode 100644 index ed284ac632..0000000000 Binary files a/.yarn/cache/@vue-compiler-core-npm-3.2.45-2a68bebbd0-e3c687b24c.zip and /dev/null differ diff --git a/.yarn/cache/@vue-compiler-core-npm-3.4.21-ec7f24d7f5-0d6b7732bc.zip b/.yarn/cache/@vue-compiler-core-npm-3.4.21-ec7f24d7f5-0d6b7732bc.zip new file mode 100644 index 0000000000..ba6ec89e54 Binary files /dev/null and b/.yarn/cache/@vue-compiler-core-npm-3.4.21-ec7f24d7f5-0d6b7732bc.zip differ diff --git a/.yarn/cache/@vue-compiler-dom-npm-3.2.45-e742186d0b-8911553863.zip b/.yarn/cache/@vue-compiler-dom-npm-3.2.45-e742186d0b-8911553863.zip deleted file mode 100644 index f8b1d3cb1e..0000000000 Binary files a/.yarn/cache/@vue-compiler-dom-npm-3.2.45-e742186d0b-8911553863.zip and /dev/null differ diff --git a/.yarn/cache/@vue-compiler-dom-npm-3.4.21-3d49f99020-f53e4f4e0a.zip b/.yarn/cache/@vue-compiler-dom-npm-3.4.21-3d49f99020-f53e4f4e0a.zip new file mode 100644 index 0000000000..4d0c8cd01f Binary files /dev/null and b/.yarn/cache/@vue-compiler-dom-npm-3.4.21-3d49f99020-f53e4f4e0a.zip differ diff --git a/.yarn/cache/@vue-compiler-sfc-npm-3.2.45-f1fe8426df-bec375faa0.zip b/.yarn/cache/@vue-compiler-sfc-npm-3.2.45-f1fe8426df-bec375faa0.zip deleted file mode 100644 index 43158bc255..0000000000 Binary files a/.yarn/cache/@vue-compiler-sfc-npm-3.2.45-f1fe8426df-bec375faa0.zip and /dev/null differ diff --git a/.yarn/cache/@vue-compiler-sfc-npm-3.4.21-c2b76ee1ff-226dc404be.zip b/.yarn/cache/@vue-compiler-sfc-npm-3.4.21-c2b76ee1ff-226dc404be.zip new file mode 100644 index 0000000000..95e0d0d70c Binary files /dev/null and b/.yarn/cache/@vue-compiler-sfc-npm-3.4.21-c2b76ee1ff-226dc404be.zip differ diff --git a/.yarn/cache/@vue-compiler-ssr-npm-3.2.45-0b951e5028-830c475506.zip b/.yarn/cache/@vue-compiler-ssr-npm-3.2.45-0b951e5028-830c475506.zip deleted file mode 100644 index 062ff3555f..0000000000 Binary files a/.yarn/cache/@vue-compiler-ssr-npm-3.2.45-0b951e5028-830c475506.zip and /dev/null differ diff --git a/.yarn/cache/@vue-compiler-ssr-npm-3.4.21-e6f043341e-c510bee68b.zip b/.yarn/cache/@vue-compiler-ssr-npm-3.4.21-e6f043341e-c510bee68b.zip new file mode 100644 index 0000000000..f03e17b080 Binary files /dev/null and b/.yarn/cache/@vue-compiler-ssr-npm-3.4.21-e6f043341e-c510bee68b.zip differ diff --git a/.yarn/cache/@vue-devtools-api-npm-6.4.5-bcd56e5fec-40c5adc878.zip b/.yarn/cache/@vue-devtools-api-npm-6.4.5-bcd56e5fec-40c5adc878.zip deleted file mode 100644 index 2b07edd0a9..0000000000 Binary files a/.yarn/cache/@vue-devtools-api-npm-6.4.5-bcd56e5fec-40c5adc878.zip and /dev/null differ diff --git a/.yarn/cache/@vue-devtools-api-npm-6.5.0-0dc0468299-ec819ef3a4.zip b/.yarn/cache/@vue-devtools-api-npm-6.5.0-0dc0468299-ec819ef3a4.zip new file mode 100644 index 0000000000..c8a187e6d2 Binary files /dev/null and b/.yarn/cache/@vue-devtools-api-npm-6.5.0-0dc0468299-ec819ef3a4.zip differ diff --git a/.yarn/cache/@vue-devtools-api-npm-6.6.1-ef3c82703e-cf12b5ebcc.zip b/.yarn/cache/@vue-devtools-api-npm-6.6.1-ef3c82703e-cf12b5ebcc.zip new file mode 100644 index 0000000000..f14e2cdac7 Binary files /dev/null and b/.yarn/cache/@vue-devtools-api-npm-6.6.1-ef3c82703e-cf12b5ebcc.zip differ diff --git a/.yarn/cache/@vue-language-plugin-pug-npm-2.0.7-547300c7e0-11cc96eb5f.zip b/.yarn/cache/@vue-language-plugin-pug-npm-2.0.7-547300c7e0-11cc96eb5f.zip new file mode 100644 index 0000000000..e637e5f556 Binary files /dev/null and b/.yarn/cache/@vue-language-plugin-pug-npm-2.0.7-547300c7e0-11cc96eb5f.zip differ diff --git a/.yarn/cache/@vue-reactivity-npm-3.2.45-bc3378a52c-4ba609744a.zip b/.yarn/cache/@vue-reactivity-npm-3.2.45-bc3378a52c-4ba609744a.zip deleted file mode 100644 index 87e4130294..0000000000 Binary files a/.yarn/cache/@vue-reactivity-npm-3.2.45-bc3378a52c-4ba609744a.zip and /dev/null differ diff --git a/.yarn/cache/@vue-reactivity-npm-3.4.21-fd3e254d08-79c7ebe3ec.zip b/.yarn/cache/@vue-reactivity-npm-3.4.21-fd3e254d08-79c7ebe3ec.zip new file mode 100644 index 0000000000..adc965a473 Binary files /dev/null and b/.yarn/cache/@vue-reactivity-npm-3.4.21-fd3e254d08-79c7ebe3ec.zip differ diff --git a/.yarn/cache/@vue-reactivity-transform-npm-3.2.45-05914b9134-4010408189.zip b/.yarn/cache/@vue-reactivity-transform-npm-3.2.45-05914b9134-4010408189.zip deleted file mode 100644 index bcecaa12fb..0000000000 Binary files a/.yarn/cache/@vue-reactivity-transform-npm-3.2.45-05914b9134-4010408189.zip and /dev/null differ diff --git a/.yarn/cache/@vue-runtime-core-npm-3.2.45-084482e779-0ac376a760.zip b/.yarn/cache/@vue-runtime-core-npm-3.2.45-084482e779-0ac376a760.zip deleted file mode 100644 index 37599d4331..0000000000 Binary files a/.yarn/cache/@vue-runtime-core-npm-3.2.45-084482e779-0ac376a760.zip and /dev/null differ diff --git a/.yarn/cache/@vue-runtime-core-npm-3.4.21-7bf985040b-4eb9b5d91f.zip b/.yarn/cache/@vue-runtime-core-npm-3.4.21-7bf985040b-4eb9b5d91f.zip new file mode 100644 index 0000000000..ffb48a907a Binary files /dev/null and b/.yarn/cache/@vue-runtime-core-npm-3.4.21-7bf985040b-4eb9b5d91f.zip differ diff --git a/.yarn/cache/@vue-runtime-dom-npm-3.2.45-6ab018299f-c66c71a2fc.zip b/.yarn/cache/@vue-runtime-dom-npm-3.2.45-6ab018299f-c66c71a2fc.zip deleted file mode 100644 index 05fd0c423c..0000000000 Binary files a/.yarn/cache/@vue-runtime-dom-npm-3.2.45-6ab018299f-c66c71a2fc.zip and /dev/null differ diff --git a/.yarn/cache/@vue-runtime-dom-npm-3.4.21-40f99cf9a2-ebfdaa081f.zip b/.yarn/cache/@vue-runtime-dom-npm-3.4.21-40f99cf9a2-ebfdaa081f.zip new file mode 100644 index 0000000000..c65601f0a9 Binary files /dev/null and b/.yarn/cache/@vue-runtime-dom-npm-3.4.21-40f99cf9a2-ebfdaa081f.zip differ diff --git a/.yarn/cache/@vue-server-renderer-npm-3.2.45-dbee798520-062812235c.zip b/.yarn/cache/@vue-server-renderer-npm-3.2.45-dbee798520-062812235c.zip deleted file mode 100644 index 1b0080fc0f..0000000000 Binary files a/.yarn/cache/@vue-server-renderer-npm-3.2.45-dbee798520-062812235c.zip and /dev/null differ diff --git a/.yarn/cache/@vue-server-renderer-npm-3.4.21-bf6b2daebb-faa3dc4876.zip b/.yarn/cache/@vue-server-renderer-npm-3.4.21-bf6b2daebb-faa3dc4876.zip new file mode 100644 index 0000000000..4da755254b Binary files /dev/null and b/.yarn/cache/@vue-server-renderer-npm-3.4.21-bf6b2daebb-faa3dc4876.zip differ diff --git a/.yarn/cache/@vue-shared-npm-3.2.45-1855c9c551-ff3205056c.zip b/.yarn/cache/@vue-shared-npm-3.2.45-1855c9c551-ff3205056c.zip deleted file mode 100644 index 1b4a2dc3ce..0000000000 Binary files a/.yarn/cache/@vue-shared-npm-3.2.45-1855c9c551-ff3205056c.zip and /dev/null differ diff --git a/.yarn/cache/@vue-shared-npm-3.4.21-2aee4ae0bc-5f30a40891.zip b/.yarn/cache/@vue-shared-npm-3.4.21-2aee4ae0bc-5f30a40891.zip new file mode 100644 index 0000000000..01c52809b3 Binary files /dev/null and b/.yarn/cache/@vue-shared-npm-3.4.21-2aee4ae0bc-5f30a40891.zip differ diff --git a/.yarn/cache/acorn-npm-8.10.0-2230c9e83e-538ba38af0.zip b/.yarn/cache/acorn-npm-8.10.0-2230c9e83e-538ba38af0.zip new file mode 100644 index 0000000000..6820207002 Binary files /dev/null and b/.yarn/cache/acorn-npm-8.10.0-2230c9e83e-538ba38af0.zip differ diff --git a/.yarn/cache/acorn-npm-8.8.0-9ef399ab45-7270ca82b2.zip b/.yarn/cache/acorn-npm-8.8.0-9ef399ab45-7270ca82b2.zip deleted file mode 100644 index b5376b1392..0000000000 Binary files a/.yarn/cache/acorn-npm-8.8.0-9ef399ab45-7270ca82b2.zip and /dev/null differ diff --git a/.yarn/cache/acorn-walk-npm-8.2.0-2f2cac3177-1715e76c01.zip b/.yarn/cache/acorn-walk-npm-8.2.0-2f2cac3177-1715e76c01.zip deleted file mode 100644 index f140c4ab5c..0000000000 Binary files a/.yarn/cache/acorn-walk-npm-8.2.0-2f2cac3177-1715e76c01.zip and /dev/null differ diff --git a/.yarn/cache/ansi-regex-npm-6.0.1-8d663a607d-1ff8b7667c.zip b/.yarn/cache/ansi-regex-npm-6.0.1-8d663a607d-1ff8b7667c.zip new file mode 100644 index 0000000000..088e552d0f Binary files /dev/null and b/.yarn/cache/ansi-regex-npm-6.0.1-8d663a607d-1ff8b7667c.zip differ diff --git a/.yarn/cache/ansi-styles-npm-5.2.0-72fc7003e3-d7f4e97ce0.zip b/.yarn/cache/ansi-styles-npm-5.2.0-72fc7003e3-d7f4e97ce0.zip deleted file mode 100644 index 62c09039bd..0000000000 Binary files a/.yarn/cache/ansi-styles-npm-5.2.0-72fc7003e3-d7f4e97ce0.zip and /dev/null differ diff --git a/.yarn/cache/ansi-styles-npm-6.2.1-d43647018c-ef940f2f0c.zip b/.yarn/cache/ansi-styles-npm-6.2.1-d43647018c-ef940f2f0c.zip new file mode 100644 index 0000000000..aa1bdfde18 Binary files /dev/null and b/.yarn/cache/ansi-styles-npm-6.2.1-d43647018c-ef940f2f0c.zip differ diff --git a/.yarn/cache/array-buffer-byte-length-npm-1.0.0-331671f28a-044e101ce1.zip b/.yarn/cache/array-buffer-byte-length-npm-1.0.0-331671f28a-044e101ce1.zip new file mode 100644 index 0000000000..d2d609a667 Binary files /dev/null and b/.yarn/cache/array-buffer-byte-length-npm-1.0.0-331671f28a-044e101ce1.zip differ diff --git a/.yarn/cache/array-includes-npm-3.1.4-79bb883109-69967c38c5.zip b/.yarn/cache/array-includes-npm-3.1.4-79bb883109-69967c38c5.zip deleted file mode 100644 index c88aec7c23..0000000000 Binary files a/.yarn/cache/array-includes-npm-3.1.4-79bb883109-69967c38c5.zip and /dev/null differ diff --git a/.yarn/cache/array-includes-npm-3.1.7-d32a5ee179-06f9e4598f.zip b/.yarn/cache/array-includes-npm-3.1.7-d32a5ee179-06f9e4598f.zip new file mode 100644 index 0000000000..1f7fc2c577 Binary files /dev/null and b/.yarn/cache/array-includes-npm-3.1.7-d32a5ee179-06f9e4598f.zip differ diff --git a/.yarn/cache/array.prototype.findlastindex-npm-1.2.3-2a36f4417b-31f35d7b37.zip b/.yarn/cache/array.prototype.findlastindex-npm-1.2.3-2a36f4417b-31f35d7b37.zip new file mode 100644 index 0000000000..8aaa4a956a Binary files /dev/null and b/.yarn/cache/array.prototype.findlastindex-npm-1.2.3-2a36f4417b-31f35d7b37.zip differ diff --git a/.yarn/cache/array.prototype.flat-npm-1.3.0-6c5c4292bd-2a652b3e8d.zip b/.yarn/cache/array.prototype.flat-npm-1.3.0-6c5c4292bd-2a652b3e8d.zip deleted file mode 100644 index 66f81fb0ab..0000000000 Binary files a/.yarn/cache/array.prototype.flat-npm-1.3.0-6c5c4292bd-2a652b3e8d.zip and /dev/null differ diff --git a/.yarn/cache/array.prototype.flat-npm-1.3.2-350729f7f4-5d6b4bf102.zip b/.yarn/cache/array.prototype.flat-npm-1.3.2-350729f7f4-5d6b4bf102.zip new file mode 100644 index 0000000000..7720137d70 Binary files /dev/null and b/.yarn/cache/array.prototype.flat-npm-1.3.2-350729f7f4-5d6b4bf102.zip differ diff --git a/.yarn/cache/array.prototype.flatmap-npm-1.3.2-5c6a4af226-ce09fe21dc.zip b/.yarn/cache/array.prototype.flatmap-npm-1.3.2-5c6a4af226-ce09fe21dc.zip new file mode 100644 index 0000000000..2553a317f1 Binary files /dev/null and b/.yarn/cache/array.prototype.flatmap-npm-1.3.2-5c6a4af226-ce09fe21dc.zip differ diff --git a/.yarn/cache/arraybuffer.prototype.slice-npm-1.0.2-4eda52ad8c-c200faf437.zip b/.yarn/cache/arraybuffer.prototype.slice-npm-1.0.2-4eda52ad8c-c200faf437.zip new file mode 100644 index 0000000000..559e55f81a Binary files /dev/null and b/.yarn/cache/arraybuffer.prototype.slice-npm-1.0.2-4eda52ad8c-c200faf437.zip differ diff --git a/.yarn/cache/async-validator-npm-4.1.1-470b8d5b59-88590ab8ad.zip b/.yarn/cache/async-validator-npm-4.1.1-470b8d5b59-88590ab8ad.zip deleted file mode 100644 index 71730f2e23..0000000000 Binary files a/.yarn/cache/async-validator-npm-4.1.1-470b8d5b59-88590ab8ad.zip and /dev/null differ diff --git a/.yarn/cache/async-validator-npm-4.2.5-4d61110c66-3e3d891a2e.zip b/.yarn/cache/async-validator-npm-4.2.5-4d61110c66-3e3d891a2e.zip new file mode 100644 index 0000000000..36bedd6286 Binary files /dev/null and b/.yarn/cache/async-validator-npm-4.2.5-4d61110c66-3e3d891a2e.zip differ diff --git a/.yarn/cache/available-typed-arrays-npm-1.0.5-88f321e4d3-20eb47b3ce.zip b/.yarn/cache/available-typed-arrays-npm-1.0.5-88f321e4d3-20eb47b3ce.zip new file mode 100644 index 0000000000..62f8601d5b Binary files /dev/null and b/.yarn/cache/available-typed-arrays-npm-1.0.5-88f321e4d3-20eb47b3ce.zip differ diff --git a/.yarn/cache/bootstrap-icons-npm-1.10.3-d595a95ca4-bef7f83ce0.zip b/.yarn/cache/bootstrap-icons-npm-1.10.3-d595a95ca4-bef7f83ce0.zip deleted file mode 100755 index 90f1604758..0000000000 Binary files a/.yarn/cache/bootstrap-icons-npm-1.10.3-d595a95ca4-bef7f83ce0.zip and /dev/null differ diff --git a/.yarn/cache/bootstrap-icons-npm-1.11.3-8d5387bef2-d5cdb90fe3.zip b/.yarn/cache/bootstrap-icons-npm-1.11.3-8d5387bef2-d5cdb90fe3.zip new file mode 100644 index 0000000000..e20ab2ecb3 Binary files /dev/null and b/.yarn/cache/bootstrap-icons-npm-1.11.3-8d5387bef2-d5cdb90fe3.zip differ diff --git a/.yarn/cache/bootstrap-npm-5.2.3-7458283a23-0211805dec.zip b/.yarn/cache/bootstrap-npm-5.2.3-7458283a23-0211805dec.zip deleted file mode 100644 index 24c59290c5..0000000000 Binary files a/.yarn/cache/bootstrap-npm-5.2.3-7458283a23-0211805dec.zip and /dev/null differ diff --git a/.yarn/cache/bootstrap-npm-5.3.3-da08e2f0fe-537b68db30.zip b/.yarn/cache/bootstrap-npm-5.3.3-da08e2f0fe-537b68db30.zip new file mode 100644 index 0000000000..ca3961acc1 Binary files /dev/null and b/.yarn/cache/bootstrap-npm-5.3.3-da08e2f0fe-537b68db30.zip differ diff --git a/.yarn/cache/browser-fs-access-npm-0.31.1-c276b62f78-4a5c88839e.zip b/.yarn/cache/browser-fs-access-npm-0.31.1-c276b62f78-4a5c88839e.zip deleted file mode 100644 index 0b50bb2da6..0000000000 Binary files a/.yarn/cache/browser-fs-access-npm-0.31.1-c276b62f78-4a5c88839e.zip and /dev/null differ diff --git a/.yarn/cache/browser-fs-access-npm-0.35.0-1577b5a7ba-5f3bf1ec17.zip b/.yarn/cache/browser-fs-access-npm-0.35.0-1577b5a7ba-5f3bf1ec17.zip new file mode 100644 index 0000000000..2202ffed8e Binary files /dev/null and b/.yarn/cache/browser-fs-access-npm-0.35.0-1577b5a7ba-5f3bf1ec17.zip differ diff --git a/.yarn/cache/buffer-from-npm-1.1.2-03d2f20d7e-0448524a56.zip b/.yarn/cache/buffer-from-npm-1.1.2-03d2f20d7e-0448524a56.zip deleted file mode 100644 index efe1b76380..0000000000 Binary files a/.yarn/cache/buffer-from-npm-1.1.2-03d2f20d7e-0448524a56.zip and /dev/null differ diff --git a/.yarn/cache/builtin-modules-npm-3.3.0-db4f3d32de-db021755d7.zip b/.yarn/cache/builtin-modules-npm-3.3.0-db4f3d32de-db021755d7.zip new file mode 100644 index 0000000000..c7e20444c6 Binary files /dev/null and b/.yarn/cache/builtin-modules-npm-3.3.0-db4f3d32de-db021755d7.zip differ diff --git a/.yarn/cache/c8-npm-7.12.0-c808cac509-3b7fa9ad7c.zip b/.yarn/cache/c8-npm-7.12.0-c808cac509-3b7fa9ad7c.zip deleted file mode 100644 index 4558e5b9f9..0000000000 Binary files a/.yarn/cache/c8-npm-7.12.0-c808cac509-3b7fa9ad7c.zip and /dev/null differ diff --git a/.yarn/cache/c8-npm-9.1.0-92c3d37f46-c5249bf9c3.zip b/.yarn/cache/c8-npm-9.1.0-92c3d37f46-c5249bf9c3.zip new file mode 100644 index 0000000000..1e5812b784 Binary files /dev/null and b/.yarn/cache/c8-npm-9.1.0-92c3d37f46-c5249bf9c3.zip differ diff --git a/.yarn/cache/call-bind-npm-1.0.5-65600fae47-449e83ecbd.zip b/.yarn/cache/call-bind-npm-1.0.5-65600fae47-449e83ecbd.zip new file mode 100644 index 0000000000..29854c129a Binary files /dev/null and b/.yarn/cache/call-bind-npm-1.0.5-65600fae47-449e83ecbd.zip differ diff --git a/.yarn/cache/caniuse-lite-npm-1.0.30001442-4206643829-c1bff65bd4.zip b/.yarn/cache/caniuse-lite-npm-1.0.30001442-4206643829-c1bff65bd4.zip deleted file mode 100755 index 698cd7ec96..0000000000 Binary files a/.yarn/cache/caniuse-lite-npm-1.0.30001442-4206643829-c1bff65bd4.zip and /dev/null differ diff --git a/.yarn/cache/caniuse-lite-npm-1.0.30001603-77af81f60b-e66e0d24b8.zip b/.yarn/cache/caniuse-lite-npm-1.0.30001603-77af81f60b-e66e0d24b8.zip new file mode 100644 index 0000000000..f3bd2d06bc Binary files /dev/null and b/.yarn/cache/caniuse-lite-npm-1.0.30001603-77af81f60b-e66e0d24b8.zip differ diff --git a/.yarn/cache/cliui-npm-7.0.4-d6b8a9edb6-ce2e8f578a.zip b/.yarn/cache/cliui-npm-7.0.4-d6b8a9edb6-ce2e8f578a.zip deleted file mode 100644 index 24f58564e4..0000000000 Binary files a/.yarn/cache/cliui-npm-7.0.4-d6b8a9edb6-ce2e8f578a.zip and /dev/null differ diff --git a/.yarn/cache/cliui-npm-8.0.1-3b029092cf-79648b3b00.zip b/.yarn/cache/cliui-npm-8.0.1-3b029092cf-79648b3b00.zip new file mode 100644 index 0000000000..a90643c5e5 Binary files /dev/null and b/.yarn/cache/cliui-npm-8.0.1-3b029092cf-79648b3b00.zip differ diff --git a/.yarn/cache/commander-npm-2.20.3-d8dcbaa39b-ab8c07884e.zip b/.yarn/cache/commander-npm-2.20.3-d8dcbaa39b-ab8c07884e.zip deleted file mode 100644 index 6a14adf507..0000000000 Binary files a/.yarn/cache/commander-npm-2.20.3-d8dcbaa39b-ab8c07884e.zip and /dev/null differ diff --git a/.yarn/cache/css-render-npm-0.15.12-ff93ab2bdd-80265c5055.zip b/.yarn/cache/css-render-npm-0.15.12-ff93ab2bdd-80265c5055.zip new file mode 100644 index 0000000000..a23ef5e7b9 Binary files /dev/null and b/.yarn/cache/css-render-npm-0.15.12-ff93ab2bdd-80265c5055.zip differ diff --git a/.yarn/cache/csstype-npm-2.6.20-7c929732a1-cb5d5ded49.zip b/.yarn/cache/csstype-npm-2.6.20-7c929732a1-cb5d5ded49.zip deleted file mode 100644 index 59ddf4f69f..0000000000 Binary files a/.yarn/cache/csstype-npm-2.6.20-7c929732a1-cb5d5ded49.zip and /dev/null differ diff --git a/.yarn/cache/csstype-npm-3.1.3-e9a1c85013-8db785cc92.zip b/.yarn/cache/csstype-npm-3.1.3-e9a1c85013-8db785cc92.zip new file mode 100644 index 0000000000..9853f0cf0b Binary files /dev/null and b/.yarn/cache/csstype-npm-3.1.3-e9a1c85013-8db785cc92.zip differ diff --git a/.yarn/cache/d3-npm-7.8.0-43f19bbccd-383d2c8aa6.zip b/.yarn/cache/d3-npm-7.8.0-43f19bbccd-383d2c8aa6.zip deleted file mode 100755 index c4600b4d0f..0000000000 Binary files a/.yarn/cache/d3-npm-7.8.0-43f19bbccd-383d2c8aa6.zip and /dev/null differ diff --git a/.yarn/cache/d3-npm-7.9.0-d293821ce6-1c0e9135f1.zip b/.yarn/cache/d3-npm-7.9.0-d293821ce6-1c0e9135f1.zip new file mode 100644 index 0000000000..e78ffffee5 Binary files /dev/null and b/.yarn/cache/d3-npm-7.9.0-d293821ce6-1c0e9135f1.zip differ diff --git a/.yarn/cache/date-fns-npm-2.28.0-c19c5add1b-a0516b2e4f.zip b/.yarn/cache/date-fns-npm-2.28.0-c19c5add1b-a0516b2e4f.zip deleted file mode 100644 index 1e88493b72..0000000000 Binary files a/.yarn/cache/date-fns-npm-2.28.0-c19c5add1b-a0516b2e4f.zip and /dev/null differ diff --git a/.yarn/cache/date-fns-npm-2.30.0-895c790e0f-f7be015232.zip b/.yarn/cache/date-fns-npm-2.30.0-895c790e0f-f7be015232.zip new file mode 100644 index 0000000000..f51ffd3ec9 Binary files /dev/null and b/.yarn/cache/date-fns-npm-2.30.0-895c790e0f-f7be015232.zip differ diff --git a/.yarn/cache/date-fns-tz-npm-1.3.3-4b42de3dcf-52111dffb4.zip b/.yarn/cache/date-fns-tz-npm-1.3.3-4b42de3dcf-52111dffb4.zip deleted file mode 100644 index 856e44240d..0000000000 Binary files a/.yarn/cache/date-fns-tz-npm-1.3.3-4b42de3dcf-52111dffb4.zip and /dev/null differ diff --git a/.yarn/cache/date-fns-tz-npm-2.0.0-9b7996f292-a6553603a9.zip b/.yarn/cache/date-fns-tz-npm-2.0.0-9b7996f292-a6553603a9.zip new file mode 100644 index 0000000000..337d3f2fd4 Binary files /dev/null and b/.yarn/cache/date-fns-tz-npm-2.0.0-9b7996f292-a6553603a9.zip differ diff --git a/.yarn/cache/deepmerge-npm-4.2.2-112165ced2-a8c43a1ed8.zip b/.yarn/cache/deepmerge-npm-4.2.2-112165ced2-a8c43a1ed8.zip deleted file mode 100644 index 3e07a61c47..0000000000 Binary files a/.yarn/cache/deepmerge-npm-4.2.2-112165ced2-a8c43a1ed8.zip and /dev/null differ diff --git a/.yarn/cache/deepmerge-npm-4.3.1-4f751a0844-2024c6a980.zip b/.yarn/cache/deepmerge-npm-4.3.1-4f751a0844-2024c6a980.zip new file mode 100644 index 0000000000..93a5246287 Binary files /dev/null and b/.yarn/cache/deepmerge-npm-4.3.1-4f751a0844-2024c6a980.zip differ diff --git a/.yarn/cache/define-data-property-npm-1.1.1-2b5156d112-a29855ad3f.zip b/.yarn/cache/define-data-property-npm-1.1.1-2b5156d112-a29855ad3f.zip new file mode 100644 index 0000000000..75936e2374 Binary files /dev/null and b/.yarn/cache/define-data-property-npm-1.1.1-2b5156d112-a29855ad3f.zip differ diff --git a/.yarn/cache/define-properties-npm-1.2.0-3547cd0fd2-e60aee6a19.zip b/.yarn/cache/define-properties-npm-1.2.0-3547cd0fd2-e60aee6a19.zip new file mode 100644 index 0000000000..bcbfcf6e68 Binary files /dev/null and b/.yarn/cache/define-properties-npm-1.2.0-3547cd0fd2-e60aee6a19.zip differ diff --git a/.yarn/cache/detect-libc-npm-2.0.2-03afa59137-2b2cd3649b.zip b/.yarn/cache/detect-libc-npm-2.0.2-03afa59137-2b2cd3649b.zip new file mode 100644 index 0000000000..1db92146ba Binary files /dev/null and b/.yarn/cache/detect-libc-npm-2.0.2-03afa59137-2b2cd3649b.zip differ diff --git a/.yarn/cache/diff-sequences-npm-27.5.1-29338362fa-a00db5554c.zip b/.yarn/cache/diff-sequences-npm-27.5.1-29338362fa-a00db5554c.zip deleted file mode 100644 index ddfadea458..0000000000 Binary files a/.yarn/cache/diff-sequences-npm-27.5.1-29338362fa-a00db5554c.zip and /dev/null differ diff --git a/.yarn/cache/eastasianwidth-npm-0.2.0-c37eb16bd1-7d00d7cd8e.zip b/.yarn/cache/eastasianwidth-npm-0.2.0-c37eb16bd1-7d00d7cd8e.zip new file mode 100644 index 0000000000..10385995a6 Binary files /dev/null and b/.yarn/cache/eastasianwidth-npm-0.2.0-c37eb16bd1-7d00d7cd8e.zip differ diff --git a/.yarn/cache/emoji-regex-npm-9.2.2-e6fac8d058-8487182da7.zip b/.yarn/cache/emoji-regex-npm-9.2.2-e6fac8d058-8487182da7.zip new file mode 100644 index 0000000000..e6b0ab4d80 Binary files /dev/null and b/.yarn/cache/emoji-regex-npm-9.2.2-e6fac8d058-8487182da7.zip differ diff --git a/.yarn/cache/entities-npm-4.5.0-7cdb83b832-853f8ebd5b.zip b/.yarn/cache/entities-npm-4.5.0-7cdb83b832-853f8ebd5b.zip new file mode 100644 index 0000000000..3772a4510c Binary files /dev/null and b/.yarn/cache/entities-npm-4.5.0-7cdb83b832-853f8ebd5b.zip differ diff --git a/.yarn/cache/es-abstract-npm-1.19.5-524a87d262-55199b0f17.zip b/.yarn/cache/es-abstract-npm-1.19.5-524a87d262-55199b0f17.zip deleted file mode 100644 index 9c6cf6749b..0000000000 Binary files a/.yarn/cache/es-abstract-npm-1.19.5-524a87d262-55199b0f17.zip and /dev/null differ diff --git a/.yarn/cache/es-abstract-npm-1.22.3-15a58832e5-b1bdc96285.zip b/.yarn/cache/es-abstract-npm-1.22.3-15a58832e5-b1bdc96285.zip new file mode 100644 index 0000000000..f72f30d6f5 Binary files /dev/null and b/.yarn/cache/es-abstract-npm-1.22.3-15a58832e5-b1bdc96285.zip differ diff --git a/.yarn/cache/es-set-tostringtag-npm-2.0.1-c87b5de872-ec416a1294.zip b/.yarn/cache/es-set-tostringtag-npm-2.0.1-c87b5de872-ec416a1294.zip new file mode 100644 index 0000000000..af638f13cd Binary files /dev/null and b/.yarn/cache/es-set-tostringtag-npm-2.0.1-c87b5de872-ec416a1294.zip differ diff --git a/.yarn/cache/esbuild-darwin-64-npm-0.15.11-0ccb211fdf-8.zip b/.yarn/cache/esbuild-darwin-64-npm-0.15.11-0ccb211fdf-8.zip deleted file mode 100644 index 229af81ff5..0000000000 Binary files a/.yarn/cache/esbuild-darwin-64-npm-0.15.11-0ccb211fdf-8.zip and /dev/null differ diff --git a/.yarn/cache/esbuild-darwin-arm64-npm-0.15.11-cbb0a8549f-8.zip b/.yarn/cache/esbuild-darwin-arm64-npm-0.15.11-cbb0a8549f-8.zip deleted file mode 100644 index a00782efe7..0000000000 Binary files a/.yarn/cache/esbuild-darwin-arm64-npm-0.15.11-cbb0a8549f-8.zip and /dev/null differ diff --git a/.yarn/cache/esbuild-linux-64-npm-0.15.11-fd176c9400-8.zip b/.yarn/cache/esbuild-linux-64-npm-0.15.11-fd176c9400-8.zip deleted file mode 100644 index 91b8b5764d..0000000000 Binary files a/.yarn/cache/esbuild-linux-64-npm-0.15.11-fd176c9400-8.zip and /dev/null differ diff --git a/.yarn/cache/esbuild-linux-arm64-npm-0.15.11-eb05503e3f-8.zip b/.yarn/cache/esbuild-linux-arm64-npm-0.15.11-eb05503e3f-8.zip deleted file mode 100644 index d7dd67cd83..0000000000 Binary files a/.yarn/cache/esbuild-linux-arm64-npm-0.15.11-eb05503e3f-8.zip and /dev/null differ diff --git a/.yarn/cache/esbuild-npm-0.15.11-352cc4ec35-afe5f2e6fb.zip b/.yarn/cache/esbuild-npm-0.15.11-352cc4ec35-afe5f2e6fb.zip deleted file mode 100644 index 484adfc5a1..0000000000 Binary files a/.yarn/cache/esbuild-npm-0.15.11-352cc4ec35-afe5f2e6fb.zip and /dev/null differ diff --git a/.yarn/cache/esbuild-npm-0.18.20-004a76d281-5d253614e5.zip b/.yarn/cache/esbuild-npm-0.18.20-004a76d281-5d253614e5.zip new file mode 100644 index 0000000000..74931c9be3 Binary files /dev/null and b/.yarn/cache/esbuild-npm-0.18.20-004a76d281-5d253614e5.zip differ diff --git a/.yarn/cache/esbuild-windows-64-npm-0.15.11-a6a42a35c8-8.zip b/.yarn/cache/esbuild-windows-64-npm-0.15.11-a6a42a35c8-8.zip deleted file mode 100644 index bee49b7092..0000000000 Binary files a/.yarn/cache/esbuild-windows-64-npm-0.15.11-a6a42a35c8-8.zip and /dev/null differ diff --git a/.yarn/cache/esbuild-windows-arm64-npm-0.15.11-d36b5e4f06-8.zip b/.yarn/cache/esbuild-windows-arm64-npm-0.15.11-d36b5e4f06-8.zip deleted file mode 100644 index 38d370bffa..0000000000 Binary files a/.yarn/cache/esbuild-windows-arm64-npm-0.15.11-d36b5e4f06-8.zip and /dev/null differ diff --git a/.yarn/cache/eslint-compat-utils-npm-0.1.2-361c6992b1-2315d9db81.zip b/.yarn/cache/eslint-compat-utils-npm-0.1.2-361c6992b1-2315d9db81.zip new file mode 100644 index 0000000000..505e336b08 Binary files /dev/null and b/.yarn/cache/eslint-compat-utils-npm-0.1.2-361c6992b1-2315d9db81.zip differ diff --git a/.yarn/cache/eslint-config-standard-npm-17.0.0-2803f6a79a-dc0ed51e18.zip b/.yarn/cache/eslint-config-standard-npm-17.0.0-2803f6a79a-dc0ed51e18.zip deleted file mode 100644 index 0cb3ae760a..0000000000 Binary files a/.yarn/cache/eslint-config-standard-npm-17.0.0-2803f6a79a-dc0ed51e18.zip and /dev/null differ diff --git a/.yarn/cache/eslint-config-standard-npm-17.1.0-e72fd623cc-8ed14ffe42.zip b/.yarn/cache/eslint-config-standard-npm-17.1.0-e72fd623cc-8ed14ffe42.zip new file mode 100644 index 0000000000..1cc26fbd86 Binary files /dev/null and b/.yarn/cache/eslint-config-standard-npm-17.1.0-e72fd623cc-8ed14ffe42.zip differ diff --git a/.yarn/cache/eslint-import-resolver-node-npm-0.3.6-d9426786c6-6266733af1.zip b/.yarn/cache/eslint-import-resolver-node-npm-0.3.6-d9426786c6-6266733af1.zip deleted file mode 100644 index a4588dad43..0000000000 Binary files a/.yarn/cache/eslint-import-resolver-node-npm-0.3.6-d9426786c6-6266733af1.zip and /dev/null differ diff --git a/.yarn/cache/eslint-import-resolver-node-npm-0.3.9-2a426afc4b-439b912712.zip b/.yarn/cache/eslint-import-resolver-node-npm-0.3.9-2a426afc4b-439b912712.zip new file mode 100644 index 0000000000..f2e17574bd Binary files /dev/null and b/.yarn/cache/eslint-import-resolver-node-npm-0.3.9-2a426afc4b-439b912712.zip differ diff --git a/.yarn/cache/eslint-module-utils-npm-2.7.3-ccd32fe6fd-77048263f3.zip b/.yarn/cache/eslint-module-utils-npm-2.7.3-ccd32fe6fd-77048263f3.zip deleted file mode 100644 index 647dc49600..0000000000 Binary files a/.yarn/cache/eslint-module-utils-npm-2.7.3-ccd32fe6fd-77048263f3.zip and /dev/null differ diff --git a/.yarn/cache/eslint-module-utils-npm-2.8.0-05e42bcab0-74c6dfea76.zip b/.yarn/cache/eslint-module-utils-npm-2.8.0-05e42bcab0-74c6dfea76.zip new file mode 100644 index 0000000000..964bee4e4d Binary files /dev/null and b/.yarn/cache/eslint-module-utils-npm-2.8.0-05e42bcab0-74c6dfea76.zip differ diff --git a/.yarn/cache/eslint-npm-8.31.0-da99c7e469-5e5688bb86.zip b/.yarn/cache/eslint-npm-8.31.0-da99c7e469-5e5688bb86.zip deleted file mode 100755 index f91ee3f7bd..0000000000 Binary files a/.yarn/cache/eslint-npm-8.31.0-da99c7e469-5e5688bb86.zip and /dev/null differ diff --git a/.yarn/cache/eslint-npm-8.57.0-4286e12a3a-3a48d7ff85.zip b/.yarn/cache/eslint-npm-8.57.0-4286e12a3a-3a48d7ff85.zip new file mode 100644 index 0000000000..73f8f9dff6 Binary files /dev/null and b/.yarn/cache/eslint-npm-8.57.0-4286e12a3a-3a48d7ff85.zip differ diff --git a/.yarn/cache/eslint-plugin-cypress-npm-2.12.1-6681f582fa-1f1c36e149.zip b/.yarn/cache/eslint-plugin-cypress-npm-2.12.1-6681f582fa-1f1c36e149.zip deleted file mode 100644 index 53d6f880de..0000000000 Binary files a/.yarn/cache/eslint-plugin-cypress-npm-2.12.1-6681f582fa-1f1c36e149.zip and /dev/null differ diff --git a/.yarn/cache/eslint-plugin-cypress-npm-2.15.1-90f777d9bd-3e66fa9a94.zip b/.yarn/cache/eslint-plugin-cypress-npm-2.15.1-90f777d9bd-3e66fa9a94.zip new file mode 100644 index 0000000000..13724ec234 Binary files /dev/null and b/.yarn/cache/eslint-plugin-cypress-npm-2.15.1-90f777d9bd-3e66fa9a94.zip differ diff --git a/.yarn/cache/eslint-plugin-es-x-npm-7.5.0-77e84d6e5d-e770e57df7.zip b/.yarn/cache/eslint-plugin-es-x-npm-7.5.0-77e84d6e5d-e770e57df7.zip new file mode 100644 index 0000000000..1d334e0a10 Binary files /dev/null and b/.yarn/cache/eslint-plugin-es-x-npm-7.5.0-77e84d6e5d-e770e57df7.zip differ diff --git a/.yarn/cache/eslint-plugin-import-npm-2.26.0-959fe14a01-0bf77ad803.zip b/.yarn/cache/eslint-plugin-import-npm-2.26.0-959fe14a01-0bf77ad803.zip deleted file mode 100644 index 62c5e22fd6..0000000000 Binary files a/.yarn/cache/eslint-plugin-import-npm-2.26.0-959fe14a01-0bf77ad803.zip and /dev/null differ diff --git a/.yarn/cache/eslint-plugin-import-npm-2.29.1-b94305f7dc-e65159aef8.zip b/.yarn/cache/eslint-plugin-import-npm-2.29.1-b94305f7dc-e65159aef8.zip new file mode 100644 index 0000000000..bc424a6a64 Binary files /dev/null and b/.yarn/cache/eslint-plugin-import-npm-2.29.1-b94305f7dc-e65159aef8.zip differ diff --git a/.yarn/cache/eslint-plugin-n-npm-15.6.1-e4ab4703b3-269d6f2896.zip b/.yarn/cache/eslint-plugin-n-npm-15.6.1-e4ab4703b3-269d6f2896.zip deleted file mode 100755 index 7de1a1a973..0000000000 Binary files a/.yarn/cache/eslint-plugin-n-npm-15.6.1-e4ab4703b3-269d6f2896.zip and /dev/null differ diff --git a/.yarn/cache/eslint-plugin-n-npm-16.6.2-77775852d0-3b468da003.zip b/.yarn/cache/eslint-plugin-n-npm-16.6.2-77775852d0-3b468da003.zip new file mode 100644 index 0000000000..9c7224993f Binary files /dev/null and b/.yarn/cache/eslint-plugin-n-npm-16.6.2-77775852d0-3b468da003.zip differ diff --git a/.yarn/cache/eslint-plugin-vue-npm-9.24.0-4c6dba51bf-2309b919d8.zip b/.yarn/cache/eslint-plugin-vue-npm-9.24.0-4c6dba51bf-2309b919d8.zip new file mode 100644 index 0000000000..285d11da2d Binary files /dev/null and b/.yarn/cache/eslint-plugin-vue-npm-9.24.0-4c6dba51bf-2309b919d8.zip differ diff --git a/.yarn/cache/eslint-plugin-vue-npm-9.8.0-ad98dd7e70-f3fc36512f.zip b/.yarn/cache/eslint-plugin-vue-npm-9.8.0-ad98dd7e70-f3fc36512f.zip deleted file mode 100644 index 639ff3c6a5..0000000000 Binary files a/.yarn/cache/eslint-plugin-vue-npm-9.8.0-ad98dd7e70-f3fc36512f.zip and /dev/null differ diff --git a/.yarn/cache/eslint-scope-npm-7.2.2-53cb0df8e8-ec97dbf5fb.zip b/.yarn/cache/eslint-scope-npm-7.2.2-53cb0df8e8-ec97dbf5fb.zip new file mode 100644 index 0000000000..29b002eb98 Binary files /dev/null and b/.yarn/cache/eslint-scope-npm-7.2.2-53cb0df8e8-ec97dbf5fb.zip differ diff --git a/.yarn/cache/eslint-visitor-keys-npm-3.4.1-a5d0a58208-f05121d868.zip b/.yarn/cache/eslint-visitor-keys-npm-3.4.1-a5d0a58208-f05121d868.zip new file mode 100644 index 0000000000..e442ca3b41 Binary files /dev/null and b/.yarn/cache/eslint-visitor-keys-npm-3.4.1-a5d0a58208-f05121d868.zip differ diff --git a/.yarn/cache/eslint-visitor-keys-npm-3.4.3-a356ac7e46-36e9ef87fc.zip b/.yarn/cache/eslint-visitor-keys-npm-3.4.3-a356ac7e46-36e9ef87fc.zip new file mode 100644 index 0000000000..7c61b814bf Binary files /dev/null and b/.yarn/cache/eslint-visitor-keys-npm-3.4.3-a356ac7e46-36e9ef87fc.zip differ diff --git a/.yarn/cache/espree-npm-9.4.0-0371ef3614-2e3020dde6.zip b/.yarn/cache/espree-npm-9.4.0-0371ef3614-2e3020dde6.zip deleted file mode 100644 index 95a79f462f..0000000000 Binary files a/.yarn/cache/espree-npm-9.4.0-0371ef3614-2e3020dde6.zip and /dev/null differ diff --git a/.yarn/cache/espree-npm-9.6.1-a50722a5a9-eb8c149c7a.zip b/.yarn/cache/espree-npm-9.6.1-a50722a5a9-eb8c149c7a.zip new file mode 100644 index 0000000000..0014c0574a Binary files /dev/null and b/.yarn/cache/espree-npm-9.6.1-a50722a5a9-eb8c149c7a.zip differ diff --git a/.yarn/cache/esquery-npm-1.5.0-d8f8a06879-aefb0d2596.zip b/.yarn/cache/esquery-npm-1.5.0-d8f8a06879-aefb0d2596.zip new file mode 100644 index 0000000000..6006b96052 Binary files /dev/null and b/.yarn/cache/esquery-npm-1.5.0-d8f8a06879-aefb0d2596.zip differ diff --git a/.yarn/cache/find-up-npm-2.1.0-9f6cb1765c-43284fe4da.zip b/.yarn/cache/find-up-npm-2.1.0-9f6cb1765c-43284fe4da.zip deleted file mode 100644 index 6b2c2d9da4..0000000000 Binary files a/.yarn/cache/find-up-npm-2.1.0-9f6cb1765c-43284fe4da.zip and /dev/null differ diff --git a/.yarn/cache/for-each-npm-0.3.3-0010ca8cdd-6c48ff2bc6.zip b/.yarn/cache/for-each-npm-0.3.3-0010ca8cdd-6c48ff2bc6.zip new file mode 100644 index 0000000000..7ba7b1639b Binary files /dev/null and b/.yarn/cache/for-each-npm-0.3.3-0010ca8cdd-6c48ff2bc6.zip differ diff --git a/.yarn/cache/foreground-child-npm-2.0.0-80c976b61e-f77ec9aff6.zip b/.yarn/cache/foreground-child-npm-2.0.0-80c976b61e-f77ec9aff6.zip deleted file mode 100644 index d947311d1e..0000000000 Binary files a/.yarn/cache/foreground-child-npm-2.0.0-80c976b61e-f77ec9aff6.zip and /dev/null differ diff --git a/.yarn/cache/foreground-child-npm-3.1.1-77e78ed774-139d270bc8.zip b/.yarn/cache/foreground-child-npm-3.1.1-77e78ed774-139d270bc8.zip new file mode 100644 index 0000000000..a288850fbb Binary files /dev/null and b/.yarn/cache/foreground-child-npm-3.1.1-77e78ed774-139d270bc8.zip differ diff --git a/.yarn/cache/function-bind-npm-1.1.2-7a55be9b03-2b0ff4ce70.zip b/.yarn/cache/function-bind-npm-1.1.2-7a55be9b03-2b0ff4ce70.zip new file mode 100644 index 0000000000..55fbdad3a3 Binary files /dev/null and b/.yarn/cache/function-bind-npm-1.1.2-7a55be9b03-2b0ff4ce70.zip differ diff --git a/.yarn/cache/function.prototype.name-npm-1.1.6-fd3a6a5cdd-7a3f9bd98a.zip b/.yarn/cache/function.prototype.name-npm-1.1.6-fd3a6a5cdd-7a3f9bd98a.zip new file mode 100644 index 0000000000..9c6ff345f9 Binary files /dev/null and b/.yarn/cache/function.prototype.name-npm-1.1.6-fd3a6a5cdd-7a3f9bd98a.zip differ diff --git a/.yarn/cache/functions-have-names-npm-1.2.3-e5cf1e2208-c3f1f5ba20.zip b/.yarn/cache/functions-have-names-npm-1.2.3-e5cf1e2208-c3f1f5ba20.zip new file mode 100644 index 0000000000..931661976f Binary files /dev/null and b/.yarn/cache/functions-have-names-npm-1.2.3-e5cf1e2208-c3f1f5ba20.zip differ diff --git a/.yarn/cache/get-intrinsic-npm-1.2.0-eb08ea9b1d-78fc0487b7.zip b/.yarn/cache/get-intrinsic-npm-1.2.0-eb08ea9b1d-78fc0487b7.zip new file mode 100644 index 0000000000..2ed7c8918f Binary files /dev/null and b/.yarn/cache/get-intrinsic-npm-1.2.0-eb08ea9b1d-78fc0487b7.zip differ diff --git a/.yarn/cache/get-intrinsic-npm-1.2.1-ae857fd610-5b61d88552.zip b/.yarn/cache/get-intrinsic-npm-1.2.1-ae857fd610-5b61d88552.zip new file mode 100644 index 0000000000..687f611165 Binary files /dev/null and b/.yarn/cache/get-intrinsic-npm-1.2.1-ae857fd610-5b61d88552.zip differ diff --git a/.yarn/cache/get-intrinsic-npm-1.2.2-3f446d8847-447ff0724d.zip b/.yarn/cache/get-intrinsic-npm-1.2.2-3f446d8847-447ff0724d.zip new file mode 100644 index 0000000000..510eb5f0ed Binary files /dev/null and b/.yarn/cache/get-intrinsic-npm-1.2.2-3f446d8847-447ff0724d.zip differ diff --git a/.yarn/cache/get-tsconfig-npm-4.7.2-8fbccd9fcf-1723589032.zip b/.yarn/cache/get-tsconfig-npm-4.7.2-8fbccd9fcf-1723589032.zip new file mode 100644 index 0000000000..6580ce4351 Binary files /dev/null and b/.yarn/cache/get-tsconfig-npm-4.7.2-8fbccd9fcf-1723589032.zip differ diff --git a/.yarn/cache/glob-npm-10.2.4-49f715fccc-29845faaa1.zip b/.yarn/cache/glob-npm-10.2.4-49f715fccc-29845faaa1.zip new file mode 100644 index 0000000000..f9f2284bf1 Binary files /dev/null and b/.yarn/cache/glob-npm-10.2.4-49f715fccc-29845faaa1.zip differ diff --git a/.yarn/cache/globals-npm-11.12.0-1fa7f41a6c-67051a45ec.zip b/.yarn/cache/globals-npm-11.12.0-1fa7f41a6c-67051a45ec.zip deleted file mode 100644 index 306b5aacad..0000000000 Binary files a/.yarn/cache/globals-npm-11.12.0-1fa7f41a6c-67051a45ec.zip and /dev/null differ diff --git a/.yarn/cache/globals-npm-13.21.0-c0829ce1cb-86c92ca8a0.zip b/.yarn/cache/globals-npm-13.21.0-c0829ce1cb-86c92ca8a0.zip new file mode 100644 index 0000000000..597f67a92e Binary files /dev/null and b/.yarn/cache/globals-npm-13.21.0-c0829ce1cb-86c92ca8a0.zip differ diff --git a/.yarn/cache/globals-npm-13.24.0-cc7713139c-56066ef058.zip b/.yarn/cache/globals-npm-13.24.0-cc7713139c-56066ef058.zip new file mode 100644 index 0000000000..c8cb0244af Binary files /dev/null and b/.yarn/cache/globals-npm-13.24.0-cc7713139c-56066ef058.zip differ diff --git a/.yarn/cache/globalthis-npm-1.0.3-96cd56020d-fbd7d760dc.zip b/.yarn/cache/globalthis-npm-1.0.3-96cd56020d-fbd7d760dc.zip new file mode 100644 index 0000000000..b82d79dbac Binary files /dev/null and b/.yarn/cache/globalthis-npm-1.0.3-96cd56020d-fbd7d760dc.zip differ diff --git a/.yarn/cache/gopd-npm-1.0.1-10c1d0b534-a5ccfb8806.zip b/.yarn/cache/gopd-npm-1.0.1-10c1d0b534-a5ccfb8806.zip new file mode 100644 index 0000000000..cafca67758 Binary files /dev/null and b/.yarn/cache/gopd-npm-1.0.1-10c1d0b534-a5ccfb8806.zip differ diff --git a/.yarn/cache/grapheme-splitter-npm-1.0.4-648f2bf509-0c22ec54de.zip b/.yarn/cache/grapheme-splitter-npm-1.0.4-648f2bf509-0c22ec54de.zip deleted file mode 100644 index 1eb26cc6a9..0000000000 Binary files a/.yarn/cache/grapheme-splitter-npm-1.0.4-648f2bf509-0c22ec54de.zip and /dev/null differ diff --git a/.yarn/cache/graphemer-npm-1.4.0-0627732d35-bab8f0be9b.zip b/.yarn/cache/graphemer-npm-1.4.0-0627732d35-bab8f0be9b.zip new file mode 100644 index 0000000000..e04f8d3724 Binary files /dev/null and b/.yarn/cache/graphemer-npm-1.4.0-0627732d35-bab8f0be9b.zip differ diff --git a/.yarn/cache/has-proto-npm-1.0.1-631ea9d820-febc5b5b53.zip b/.yarn/cache/has-proto-npm-1.0.1-631ea9d820-febc5b5b53.zip new file mode 100644 index 0000000000..78afc3de42 Binary files /dev/null and b/.yarn/cache/has-proto-npm-1.0.1-631ea9d820-febc5b5b53.zip differ diff --git a/.yarn/cache/hasown-npm-2.0.0-78b794ceef-6151c75ca1.zip b/.yarn/cache/hasown-npm-2.0.0-78b794ceef-6151c75ca1.zip new file mode 100644 index 0000000000..5454406288 Binary files /dev/null and b/.yarn/cache/hasown-npm-2.0.0-78b794ceef-6151c75ca1.zip differ diff --git a/.yarn/cache/highcharts-npm-10.3.2-1672942f09-43cb42b24c.zip b/.yarn/cache/highcharts-npm-10.3.2-1672942f09-43cb42b24c.zip deleted file mode 100644 index 410fae01e8..0000000000 Binary files a/.yarn/cache/highcharts-npm-10.3.2-1672942f09-43cb42b24c.zip and /dev/null differ diff --git a/.yarn/cache/highcharts-npm-11.4.0-8a1f46b545-873e661914.zip b/.yarn/cache/highcharts-npm-11.4.0-8a1f46b545-873e661914.zip new file mode 100644 index 0000000000..9c2f2df154 Binary files /dev/null and b/.yarn/cache/highcharts-npm-11.4.0-8a1f46b545-873e661914.zip differ diff --git a/.yarn/cache/highlight.js-npm-11.5.1-0fb1167640-bff556101d.zip b/.yarn/cache/highlight.js-npm-11.5.1-0fb1167640-bff556101d.zip deleted file mode 100644 index efbd98ed5d..0000000000 Binary files a/.yarn/cache/highlight.js-npm-11.5.1-0fb1167640-bff556101d.zip and /dev/null differ diff --git a/.yarn/cache/highlight.js-npm-11.9.0-ec99f7b12f-4043d31c5d.zip b/.yarn/cache/highlight.js-npm-11.9.0-ec99f7b12f-4043d31c5d.zip new file mode 100644 index 0000000000..7a740063fa Binary files /dev/null and b/.yarn/cache/highlight.js-npm-11.9.0-ec99f7b12f-4043d31c5d.zip differ diff --git a/.yarn/cache/html-validate-npm-7.12.2-4a7c5f12a3-04bccd1680.zip b/.yarn/cache/html-validate-npm-7.12.2-4a7c5f12a3-04bccd1680.zip deleted file mode 100755 index 680577511b..0000000000 Binary files a/.yarn/cache/html-validate-npm-7.12.2-4a7c5f12a3-04bccd1680.zip and /dev/null differ diff --git a/.yarn/cache/html-validate-npm-8.18.1-c5271a0fb9-53479bf75b.zip b/.yarn/cache/html-validate-npm-8.18.1-c5271a0fb9-53479bf75b.zip new file mode 100644 index 0000000000..b2f855af03 Binary files /dev/null and b/.yarn/cache/html-validate-npm-8.18.1-c5271a0fb9-53479bf75b.zip differ diff --git a/.yarn/cache/http-cache-semantics-npm-4.1.0-860520a31f-974de94a81.zip b/.yarn/cache/http-cache-semantics-npm-4.1.0-860520a31f-974de94a81.zip deleted file mode 100644 index ed85c1c4c7..0000000000 Binary files a/.yarn/cache/http-cache-semantics-npm-4.1.0-860520a31f-974de94a81.zip and /dev/null differ diff --git a/.yarn/cache/http-cache-semantics-npm-4.1.1-1120131375-83ac0bc60b.zip b/.yarn/cache/http-cache-semantics-npm-4.1.1-1120131375-83ac0bc60b.zip new file mode 100644 index 0000000000..19f1e0a201 Binary files /dev/null and b/.yarn/cache/http-cache-semantics-npm-4.1.1-1120131375-83ac0bc60b.zip differ diff --git a/.yarn/cache/ical.js-npm-1.5.0-5ba1c69420-51df7a01f4.zip b/.yarn/cache/ical.js-npm-1.5.0-5ba1c69420-51df7a01f4.zip new file mode 100644 index 0000000000..7aa64acf00 Binary files /dev/null and b/.yarn/cache/ical.js-npm-1.5.0-5ba1c69420-51df7a01f4.zip differ diff --git a/.yarn/cache/ignore-npm-5.2.4-fbe6e989e5-3d4c309c60.zip b/.yarn/cache/ignore-npm-5.2.4-fbe6e989e5-3d4c309c60.zip new file mode 100644 index 0000000000..50627d8e10 Binary files /dev/null and b/.yarn/cache/ignore-npm-5.2.4-fbe6e989e5-3d4c309c60.zip differ diff --git a/.yarn/cache/ignore-npm-5.3.1-f6947c5df7-71d7bb4c1d.zip b/.yarn/cache/ignore-npm-5.3.1-f6947c5df7-71d7bb4c1d.zip new file mode 100644 index 0000000000..75ba53a270 Binary files /dev/null and b/.yarn/cache/ignore-npm-5.3.1-f6947c5df7-71d7bb4c1d.zip differ diff --git a/.yarn/cache/internal-slot-npm-1.0.3-9e05eea002-1944f92e98.zip b/.yarn/cache/internal-slot-npm-1.0.3-9e05eea002-1944f92e98.zip deleted file mode 100644 index 18c6edaa91..0000000000 Binary files a/.yarn/cache/internal-slot-npm-1.0.3-9e05eea002-1944f92e98.zip and /dev/null differ diff --git a/.yarn/cache/internal-slot-npm-1.0.5-a2241f3e66-97e84046bf.zip b/.yarn/cache/internal-slot-npm-1.0.5-a2241f3e66-97e84046bf.zip new file mode 100644 index 0000000000..18fccd3ac2 Binary files /dev/null and b/.yarn/cache/internal-slot-npm-1.0.5-a2241f3e66-97e84046bf.zip differ diff --git a/.yarn/cache/is-array-buffer-npm-3.0.1-3e93b14326-f26ab87448.zip b/.yarn/cache/is-array-buffer-npm-3.0.1-3e93b14326-f26ab87448.zip new file mode 100644 index 0000000000..4fb5eb3634 Binary files /dev/null and b/.yarn/cache/is-array-buffer-npm-3.0.1-3e93b14326-f26ab87448.zip differ diff --git a/.yarn/cache/is-array-buffer-npm-3.0.2-0dec897785-dcac9dda66.zip b/.yarn/cache/is-array-buffer-npm-3.0.2-0dec897785-dcac9dda66.zip new file mode 100644 index 0000000000..7556381d45 Binary files /dev/null and b/.yarn/cache/is-array-buffer-npm-3.0.2-0dec897785-dcac9dda66.zip differ diff --git a/.yarn/cache/is-builtin-module-npm-3.2.1-2f92a5d353-e8f0ffc19a.zip b/.yarn/cache/is-builtin-module-npm-3.2.1-2f92a5d353-e8f0ffc19a.zip new file mode 100644 index 0000000000..be908976b5 Binary files /dev/null and b/.yarn/cache/is-builtin-module-npm-3.2.1-2f92a5d353-e8f0ffc19a.zip differ diff --git a/.yarn/cache/is-callable-npm-1.2.7-808a303e61-61fd57d03b.zip b/.yarn/cache/is-callable-npm-1.2.7-808a303e61-61fd57d03b.zip new file mode 100644 index 0000000000..0e383ae51f Binary files /dev/null and b/.yarn/cache/is-callable-npm-1.2.7-808a303e61-61fd57d03b.zip differ diff --git a/.yarn/cache/is-core-module-npm-2.10.0-6dff9310aa-0f3f77811f.zip b/.yarn/cache/is-core-module-npm-2.10.0-6dff9310aa-0f3f77811f.zip deleted file mode 100644 index 42dc0c31eb..0000000000 Binary files a/.yarn/cache/is-core-module-npm-2.10.0-6dff9310aa-0f3f77811f.zip and /dev/null differ diff --git a/.yarn/cache/is-core-module-npm-2.11.0-70061e141a-f96fd490c6.zip b/.yarn/cache/is-core-module-npm-2.11.0-70061e141a-f96fd490c6.zip deleted file mode 100644 index 4b89bc40ec..0000000000 Binary files a/.yarn/cache/is-core-module-npm-2.11.0-70061e141a-f96fd490c6.zip and /dev/null differ diff --git a/.yarn/cache/is-core-module-npm-2.12.1-ce74e89160-f04ea30533.zip b/.yarn/cache/is-core-module-npm-2.12.1-ce74e89160-f04ea30533.zip new file mode 100644 index 0000000000..9512b2ef2b Binary files /dev/null and b/.yarn/cache/is-core-module-npm-2.12.1-ce74e89160-f04ea30533.zip differ diff --git a/.yarn/cache/is-core-module-npm-2.13.0-e444c50225-053ab101fb.zip b/.yarn/cache/is-core-module-npm-2.13.0-e444c50225-053ab101fb.zip new file mode 100644 index 0000000000..636775cb5e Binary files /dev/null and b/.yarn/cache/is-core-module-npm-2.13.0-e444c50225-053ab101fb.zip differ diff --git a/.yarn/cache/is-core-module-npm-2.13.1-36e17434f9-256559ee8a.zip b/.yarn/cache/is-core-module-npm-2.13.1-36e17434f9-256559ee8a.zip new file mode 100644 index 0000000000..897f505685 Binary files /dev/null and b/.yarn/cache/is-core-module-npm-2.13.1-36e17434f9-256559ee8a.zip differ diff --git a/.yarn/cache/is-typed-array-npm-1.1.10-fe4ef83cdc-aac6ecb59d.zip b/.yarn/cache/is-typed-array-npm-1.1.10-fe4ef83cdc-aac6ecb59d.zip new file mode 100644 index 0000000000..b3a4495f94 Binary files /dev/null and b/.yarn/cache/is-typed-array-npm-1.1.10-fe4ef83cdc-aac6ecb59d.zip differ diff --git a/.yarn/cache/is-typed-array-npm-1.1.12-6135c91b1a-4c89c4a3be.zip b/.yarn/cache/is-typed-array-npm-1.1.12-6135c91b1a-4c89c4a3be.zip new file mode 100644 index 0000000000..4a35c2e95f Binary files /dev/null and b/.yarn/cache/is-typed-array-npm-1.1.12-6135c91b1a-4c89c4a3be.zip differ diff --git a/.yarn/cache/isarray-npm-2.0.5-4ba522212d-bd5bbe4104.zip b/.yarn/cache/isarray-npm-2.0.5-4ba522212d-bd5bbe4104.zip new file mode 100644 index 0000000000..f46224f1cc Binary files /dev/null and b/.yarn/cache/isarray-npm-2.0.5-4ba522212d-bd5bbe4104.zip differ diff --git a/.yarn/cache/istanbul-lib-report-npm-3.0.1-b17446ab24-fd17a1b879.zip b/.yarn/cache/istanbul-lib-report-npm-3.0.1-b17446ab24-fd17a1b879.zip new file mode 100644 index 0000000000..b946848afd Binary files /dev/null and b/.yarn/cache/istanbul-lib-report-npm-3.0.1-b17446ab24-fd17a1b879.zip differ diff --git a/.yarn/cache/istanbul-reports-npm-3.1.4-5faaa9636c-2132983355.zip b/.yarn/cache/istanbul-reports-npm-3.1.4-5faaa9636c-2132983355.zip deleted file mode 100644 index c9a9a9c949..0000000000 Binary files a/.yarn/cache/istanbul-reports-npm-3.1.4-5faaa9636c-2132983355.zip and /dev/null differ diff --git a/.yarn/cache/istanbul-reports-npm-3.1.6-66918eb97f-44c4c0582f.zip b/.yarn/cache/istanbul-reports-npm-3.1.6-66918eb97f-44c4c0582f.zip new file mode 100644 index 0000000000..4a337c3397 Binary files /dev/null and b/.yarn/cache/istanbul-reports-npm-3.1.6-66918eb97f-44c4c0582f.zip differ diff --git a/.yarn/cache/jackspeak-npm-2.2.0-5383861524-d8cd5be4f0.zip b/.yarn/cache/jackspeak-npm-2.2.0-5383861524-d8cd5be4f0.zip new file mode 100644 index 0000000000..224a2f61a9 Binary files /dev/null and b/.yarn/cache/jackspeak-npm-2.2.0-5383861524-d8cd5be4f0.zip differ diff --git a/.yarn/cache/jest-diff-npm-27.5.1-818e549196-8be27c1e1e.zip b/.yarn/cache/jest-diff-npm-27.5.1-818e549196-8be27c1e1e.zip deleted file mode 100644 index de55e34a9c..0000000000 Binary files a/.yarn/cache/jest-diff-npm-27.5.1-818e549196-8be27c1e1e.zip and /dev/null differ diff --git a/.yarn/cache/jest-get-type-npm-27.5.1-980fbf7a43-63064ab701.zip b/.yarn/cache/jest-get-type-npm-27.5.1-980fbf7a43-63064ab701.zip deleted file mode 100644 index 50167f4d8d..0000000000 Binary files a/.yarn/cache/jest-get-type-npm-27.5.1-980fbf7a43-63064ab701.zip and /dev/null differ diff --git a/.yarn/cache/jest-matcher-utils-npm-27.5.1-0c47b071fb-bb2135fc48.zip b/.yarn/cache/jest-matcher-utils-npm-27.5.1-0c47b071fb-bb2135fc48.zip deleted file mode 100644 index f4bc56be0f..0000000000 Binary files a/.yarn/cache/jest-matcher-utils-npm-27.5.1-0c47b071fb-bb2135fc48.zip and /dev/null differ diff --git a/.yarn/cache/jquery-migrate-npm-3.4.0-88c209e61f-7431685c56.zip b/.yarn/cache/jquery-migrate-npm-3.4.0-88c209e61f-7431685c56.zip deleted file mode 100644 index 04eb1cbaab..0000000000 Binary files a/.yarn/cache/jquery-migrate-npm-3.4.0-88c209e61f-7431685c56.zip and /dev/null differ diff --git a/.yarn/cache/jquery-migrate-npm-3.4.1-c842b6adb7-d2cb17d055.zip b/.yarn/cache/jquery-migrate-npm-3.4.1-c842b6adb7-d2cb17d055.zip new file mode 100644 index 0000000000..b59ac0e1a3 Binary files /dev/null and b/.yarn/cache/jquery-migrate-npm-3.4.1-c842b6adb7-d2cb17d055.zip differ diff --git a/.yarn/cache/jquery-npm-3.6.0-ca7872bdbb-8fd5fef4aa.zip b/.yarn/cache/jquery-npm-3.6.0-ca7872bdbb-8fd5fef4aa.zip deleted file mode 100644 index 2c23cc857b..0000000000 Binary files a/.yarn/cache/jquery-npm-3.6.0-ca7872bdbb-8fd5fef4aa.zip and /dev/null differ diff --git a/.yarn/cache/jquery-npm-3.6.3-cbc34d2330-0fd366bdca.zip b/.yarn/cache/jquery-npm-3.6.3-cbc34d2330-0fd366bdca.zip deleted file mode 100755 index 5009afaf97..0000000000 Binary files a/.yarn/cache/jquery-npm-3.6.3-cbc34d2330-0fd366bdca.zip and /dev/null differ diff --git a/.yarn/cache/jquery-npm-3.7.1-eeeac0f21e-4370b8139d.zip b/.yarn/cache/jquery-npm-3.7.1-eeeac0f21e-4370b8139d.zip new file mode 100644 index 0000000000..dda19f270f Binary files /dev/null and b/.yarn/cache/jquery-npm-3.7.1-eeeac0f21e-4370b8139d.zip differ diff --git a/.yarn/cache/jquery-ui-dist-npm-1.13.2-86225c0ce7-4f3a3a2ff8.zip b/.yarn/cache/jquery-ui-dist-npm-1.13.2-86225c0ce7-4f3a3a2ff8.zip deleted file mode 100644 index 61d92e7f7b..0000000000 Binary files a/.yarn/cache/jquery-ui-dist-npm-1.13.2-86225c0ce7-4f3a3a2ff8.zip and /dev/null differ diff --git a/.yarn/cache/js-cookie-npm-3.0.1-04c7177de1-bb48de67e2.zip b/.yarn/cache/js-cookie-npm-3.0.1-04c7177de1-bb48de67e2.zip deleted file mode 100644 index 3c0bf5db17..0000000000 Binary files a/.yarn/cache/js-cookie-npm-3.0.1-04c7177de1-bb48de67e2.zip and /dev/null differ diff --git a/.yarn/cache/js-cookie-npm-3.0.5-8fc8fcc9b4-2dbd2809c6.zip b/.yarn/cache/js-cookie-npm-3.0.5-8fc8fcc9b4-2dbd2809c6.zip new file mode 100644 index 0000000000..a8eacc4b1a Binary files /dev/null and b/.yarn/cache/js-cookie-npm-3.0.5-8fc8fcc9b4-2dbd2809c6.zip differ diff --git a/.yarn/cache/js-sdsl-npm-4.1.5-66fcf4f580-695f657ddc.zip b/.yarn/cache/js-sdsl-npm-4.1.5-66fcf4f580-695f657ddc.zip deleted file mode 100644 index c581e91e64..0000000000 Binary files a/.yarn/cache/js-sdsl-npm-4.1.5-66fcf4f580-695f657ddc.zip and /dev/null differ diff --git a/.yarn/cache/json5-npm-1.0.1-647fc8794b-e76ea23dbb.zip b/.yarn/cache/json5-npm-1.0.1-647fc8794b-e76ea23dbb.zip deleted file mode 100644 index cc70df5220..0000000000 Binary files a/.yarn/cache/json5-npm-1.0.1-647fc8794b-e76ea23dbb.zip and /dev/null differ diff --git a/.yarn/cache/json5-npm-1.0.2-9607f93e30-866458a8c5.zip b/.yarn/cache/json5-npm-1.0.2-9607f93e30-866458a8c5.zip new file mode 100644 index 0000000000..aa52eb0458 Binary files /dev/null and b/.yarn/cache/json5-npm-1.0.2-9607f93e30-866458a8c5.zip differ diff --git a/.yarn/cache/lmdb-npm-2.8.5-e5fdd937dd-b1ec76650d.zip b/.yarn/cache/lmdb-npm-2.8.5-e5fdd937dd-b1ec76650d.zip new file mode 100644 index 0000000000..1fe6a6f48d Binary files /dev/null and b/.yarn/cache/lmdb-npm-2.8.5-e5fdd937dd-b1ec76650d.zip differ diff --git a/.yarn/cache/locate-path-npm-2.0.0-673d28b0ea-02d581edbb.zip b/.yarn/cache/locate-path-npm-2.0.0-673d28b0ea-02d581edbb.zip deleted file mode 100644 index 0841fd1c17..0000000000 Binary files a/.yarn/cache/locate-path-npm-2.0.0-673d28b0ea-02d581edbb.zip and /dev/null differ diff --git a/.yarn/cache/lodash.sortby-npm-4.7.0-fda8ab950d-db170c9396.zip b/.yarn/cache/lodash.sortby-npm-4.7.0-fda8ab950d-db170c9396.zip deleted file mode 100644 index 915d1f2fcc..0000000000 Binary files a/.yarn/cache/lodash.sortby-npm-4.7.0-fda8ab950d-db170c9396.zip and /dev/null differ diff --git a/.yarn/cache/lru-cache-npm-9.1.1-765199cb01-4d703bb9b6.zip b/.yarn/cache/lru-cache-npm-9.1.1-765199cb01-4d703bb9b6.zip new file mode 100644 index 0000000000..5d688f763c Binary files /dev/null and b/.yarn/cache/lru-cache-npm-9.1.1-765199cb01-4d703bb9b6.zip differ diff --git a/.yarn/cache/luxon-npm-3.2.1-56f8d97395-3fa3def2c5.zip b/.yarn/cache/luxon-npm-3.2.1-56f8d97395-3fa3def2c5.zip deleted file mode 100755 index e4837bdf78..0000000000 Binary files a/.yarn/cache/luxon-npm-3.2.1-56f8d97395-3fa3def2c5.zip and /dev/null differ diff --git a/.yarn/cache/luxon-npm-3.4.4-c93f95dde8-36c1f99c47.zip b/.yarn/cache/luxon-npm-3.4.4-c93f95dde8-36c1f99c47.zip new file mode 100644 index 0000000000..ed7709ee9e Binary files /dev/null and b/.yarn/cache/luxon-npm-3.4.4-c93f95dde8-36c1f99c47.zip differ diff --git a/.yarn/cache/magic-string-npm-0.25.9-0b51c0ea50-9a0e55a15c.zip b/.yarn/cache/magic-string-npm-0.25.9-0b51c0ea50-9a0e55a15c.zip deleted file mode 100644 index caa6d6b49e..0000000000 Binary files a/.yarn/cache/magic-string-npm-0.25.9-0b51c0ea50-9a0e55a15c.zip and /dev/null differ diff --git a/.yarn/cache/magic-string-npm-0.30.7-0bb5819095-bdf102e36a.zip b/.yarn/cache/magic-string-npm-0.30.7-0bb5819095-bdf102e36a.zip new file mode 100644 index 0000000000..7d9e6ff1d3 Binary files /dev/null and b/.yarn/cache/magic-string-npm-0.30.7-0bb5819095-bdf102e36a.zip differ diff --git a/.yarn/cache/make-dir-npm-4.0.0-ec3cd921cc-bf0731a2dd.zip b/.yarn/cache/make-dir-npm-4.0.0-ec3cd921cc-bf0731a2dd.zip new file mode 100644 index 0000000000..2a141eff65 Binary files /dev/null and b/.yarn/cache/make-dir-npm-4.0.0-ec3cd921cc-bf0731a2dd.zip differ diff --git a/.yarn/cache/minimatch-npm-9.0.0-c6737cb1be-7bd57899ed.zip b/.yarn/cache/minimatch-npm-9.0.0-c6737cb1be-7bd57899ed.zip new file mode 100644 index 0000000000..ef764d9b78 Binary files /dev/null and b/.yarn/cache/minimatch-npm-9.0.0-c6737cb1be-7bd57899ed.zip differ diff --git a/.yarn/cache/minipass-npm-6.0.1-634723433e-1df70bb565.zip b/.yarn/cache/minipass-npm-6.0.1-634723433e-1df70bb565.zip new file mode 100644 index 0000000000..db17726e14 Binary files /dev/null and b/.yarn/cache/minipass-npm-6.0.1-634723433e-1df70bb565.zip differ diff --git a/.yarn/cache/moment-npm-2.29.3-fe4ba99bae-2e780e36d9.zip b/.yarn/cache/moment-npm-2.29.3-fe4ba99bae-2e780e36d9.zip deleted file mode 100644 index 8db6d8b80b..0000000000 Binary files a/.yarn/cache/moment-npm-2.29.3-fe4ba99bae-2e780e36d9.zip and /dev/null differ diff --git a/.yarn/cache/moment-npm-2.30.1-1c51a5c631-859236bab1.zip b/.yarn/cache/moment-npm-2.30.1-1c51a5c631-859236bab1.zip new file mode 100644 index 0000000000..7454cc21af Binary files /dev/null and b/.yarn/cache/moment-npm-2.30.1-1c51a5c631-859236bab1.zip differ diff --git a/.yarn/cache/moment-timezone-npm-0.5.40-873e898229-6f6be5412b.zip b/.yarn/cache/moment-timezone-npm-0.5.40-873e898229-6f6be5412b.zip deleted file mode 100755 index cc364f7d6e..0000000000 Binary files a/.yarn/cache/moment-timezone-npm-0.5.40-873e898229-6f6be5412b.zip and /dev/null differ diff --git a/.yarn/cache/moment-timezone-npm-0.5.45-2df3ad72a4-a22e9f983f.zip b/.yarn/cache/moment-timezone-npm-0.5.45-2df3ad72a4-a22e9f983f.zip new file mode 100644 index 0000000000..4cd7864ca5 Binary files /dev/null and b/.yarn/cache/moment-timezone-npm-0.5.45-2df3ad72a4-a22e9f983f.zip differ diff --git a/.yarn/cache/msgpackr-extract-npm-3.0.2-93e8773fad-5adb809b96.zip b/.yarn/cache/msgpackr-extract-npm-3.0.2-93e8773fad-5adb809b96.zip new file mode 100644 index 0000000000..b9af6cd241 Binary files /dev/null and b/.yarn/cache/msgpackr-extract-npm-3.0.2-93e8773fad-5adb809b96.zip differ diff --git a/.yarn/cache/msgpackr-npm-1.10.1-5c5ff5c553-e422d18b01.zip b/.yarn/cache/msgpackr-npm-1.10.1-5c5ff5c553-e422d18b01.zip new file mode 100644 index 0000000000..12aaa36344 Binary files /dev/null and b/.yarn/cache/msgpackr-npm-1.10.1-5c5ff5c553-e422d18b01.zip differ diff --git a/.yarn/cache/msgpackr-npm-1.9.9-75b366d55f-b63182d99f.zip b/.yarn/cache/msgpackr-npm-1.9.9-75b366d55f-b63182d99f.zip new file mode 100644 index 0000000000..ce927778aa Binary files /dev/null and b/.yarn/cache/msgpackr-npm-1.9.9-75b366d55f-b63182d99f.zip differ diff --git a/.yarn/cache/muggle-string-npm-0.4.1-fe3c825cc2-85fe1766d1.zip b/.yarn/cache/muggle-string-npm-0.4.1-fe3c825cc2-85fe1766d1.zip new file mode 100644 index 0000000000..4cec1b177d Binary files /dev/null and b/.yarn/cache/muggle-string-npm-0.4.1-fe3c825cc2-85fe1766d1.zip differ diff --git a/.yarn/cache/naive-ui-npm-2.34.3-ba2dfb08d8-792d9e6c51.zip b/.yarn/cache/naive-ui-npm-2.34.3-ba2dfb08d8-792d9e6c51.zip deleted file mode 100755 index d025f4e674..0000000000 Binary files a/.yarn/cache/naive-ui-npm-2.34.3-ba2dfb08d8-792d9e6c51.zip and /dev/null differ diff --git a/.yarn/cache/naive-ui-npm-2.38.1-0edd2e5816-88a8f981de.zip b/.yarn/cache/naive-ui-npm-2.38.1-0edd2e5816-88a8f981de.zip new file mode 100644 index 0000000000..fb6dc789a1 Binary files /dev/null and b/.yarn/cache/naive-ui-npm-2.38.1-0edd2e5816-88a8f981de.zip differ diff --git a/.yarn/cache/nanoid-npm-3.3.3-25d865be84-ada019402a.zip b/.yarn/cache/nanoid-npm-3.3.3-25d865be84-ada019402a.zip deleted file mode 100644 index d28e91f1ff..0000000000 Binary files a/.yarn/cache/nanoid-npm-3.3.3-25d865be84-ada019402a.zip and /dev/null differ diff --git a/.yarn/cache/nanoid-npm-3.3.4-3d250377d6-2fddd6dee9.zip b/.yarn/cache/nanoid-npm-3.3.4-3d250377d6-2fddd6dee9.zip deleted file mode 100644 index 740fd4c336..0000000000 Binary files a/.yarn/cache/nanoid-npm-3.3.4-3d250377d6-2fddd6dee9.zip and /dev/null differ diff --git a/.yarn/cache/nanoid-npm-3.3.7-98824ba130-d36c427e53.zip b/.yarn/cache/nanoid-npm-3.3.7-98824ba130-d36c427e53.zip new file mode 100644 index 0000000000..7b2fd6e1b5 Binary files /dev/null and b/.yarn/cache/nanoid-npm-3.3.7-98824ba130-d36c427e53.zip differ diff --git a/.yarn/cache/node-addon-api-npm-6.1.0-634c545b39-3a539510e6.zip b/.yarn/cache/node-addon-api-npm-6.1.0-634c545b39-3a539510e6.zip new file mode 100644 index 0000000000..012df449c0 Binary files /dev/null and b/.yarn/cache/node-addon-api-npm-6.1.0-634c545b39-3a539510e6.zip differ diff --git a/.yarn/cache/node-gyp-build-optional-packages-npm-5.0.7-40f21a5d68-bcb4537af1.zip b/.yarn/cache/node-gyp-build-optional-packages-npm-5.0.7-40f21a5d68-bcb4537af1.zip new file mode 100644 index 0000000000..d023f1a69d Binary files /dev/null and b/.yarn/cache/node-gyp-build-optional-packages-npm-5.0.7-40f21a5d68-bcb4537af1.zip differ diff --git a/.yarn/cache/node-gyp-build-optional-packages-npm-5.1.1-ff11e179dd-f3cb197862.zip b/.yarn/cache/node-gyp-build-optional-packages-npm-5.1.1-ff11e179dd-f3cb197862.zip new file mode 100644 index 0000000000..840821996d Binary files /dev/null and b/.yarn/cache/node-gyp-build-optional-packages-npm-5.1.1-ff11e179dd-f3cb197862.zip differ diff --git a/.yarn/cache/object-inspect-npm-1.13.1-fd038a2f0a-7d9fa9221d.zip b/.yarn/cache/object-inspect-npm-1.13.1-fd038a2f0a-7d9fa9221d.zip new file mode 100644 index 0000000000..1e1bbfbcfa Binary files /dev/null and b/.yarn/cache/object-inspect-npm-1.13.1-fd038a2f0a-7d9fa9221d.zip differ diff --git a/.yarn/cache/object.assign-npm-4.1.2-d52edada1c-d621d832ed.zip b/.yarn/cache/object.assign-npm-4.1.2-d52edada1c-d621d832ed.zip deleted file mode 100644 index 0031b97816..0000000000 Binary files a/.yarn/cache/object.assign-npm-4.1.2-d52edada1c-d621d832ed.zip and /dev/null differ diff --git a/.yarn/cache/object.assign-npm-4.1.4-fb3deb1c3a-76cab513a5.zip b/.yarn/cache/object.assign-npm-4.1.4-fb3deb1c3a-76cab513a5.zip new file mode 100644 index 0000000000..8a1fef0557 Binary files /dev/null and b/.yarn/cache/object.assign-npm-4.1.4-fb3deb1c3a-76cab513a5.zip differ diff --git a/.yarn/cache/object.fromentries-npm-2.0.7-2e38392540-7341ce246e.zip b/.yarn/cache/object.fromentries-npm-2.0.7-2e38392540-7341ce246e.zip new file mode 100644 index 0000000000..a976cc8e0e Binary files /dev/null and b/.yarn/cache/object.fromentries-npm-2.0.7-2e38392540-7341ce246e.zip differ diff --git a/.yarn/cache/object.groupby-npm-1.0.1-fc268391fe-d7959d6eaa.zip b/.yarn/cache/object.groupby-npm-1.0.1-fc268391fe-d7959d6eaa.zip new file mode 100644 index 0000000000..c67f462cfb Binary files /dev/null and b/.yarn/cache/object.groupby-npm-1.0.1-fc268391fe-d7959d6eaa.zip differ diff --git a/.yarn/cache/object.values-npm-1.1.5-f1de7f3742-0f17e99741.zip b/.yarn/cache/object.values-npm-1.1.5-f1de7f3742-0f17e99741.zip deleted file mode 100644 index e03d02d7dd..0000000000 Binary files a/.yarn/cache/object.values-npm-1.1.5-f1de7f3742-0f17e99741.zip and /dev/null differ diff --git a/.yarn/cache/object.values-npm-1.1.7-deae619f88-f3e4ae4f21.zip b/.yarn/cache/object.values-npm-1.1.7-deae619f88-f3e4ae4f21.zip new file mode 100644 index 0000000000..4c12832e02 Binary files /dev/null and b/.yarn/cache/object.values-npm-1.1.7-deae619f88-f3e4ae4f21.zip differ diff --git a/.yarn/cache/optionator-npm-0.9.1-577e397aae-dbc6fa0656.zip b/.yarn/cache/optionator-npm-0.9.1-577e397aae-dbc6fa0656.zip deleted file mode 100644 index 6e6efe345b..0000000000 Binary files a/.yarn/cache/optionator-npm-0.9.1-577e397aae-dbc6fa0656.zip and /dev/null differ diff --git a/.yarn/cache/optionator-npm-0.9.3-56c3a4bf80-0928199944.zip b/.yarn/cache/optionator-npm-0.9.3-56c3a4bf80-0928199944.zip new file mode 100644 index 0000000000..06266323c5 Binary files /dev/null and b/.yarn/cache/optionator-npm-0.9.3-56c3a4bf80-0928199944.zip differ diff --git a/.yarn/cache/ordered-binary-npm-1.4.1-9ad6b7c6b5-274940b4ef.zip b/.yarn/cache/ordered-binary-npm-1.4.1-9ad6b7c6b5-274940b4ef.zip new file mode 100644 index 0000000000..35ea485c2b Binary files /dev/null and b/.yarn/cache/ordered-binary-npm-1.4.1-9ad6b7c6b5-274940b4ef.zip differ diff --git a/.yarn/cache/p-limit-npm-1.3.0-fdb471d864-281c1c0b8c.zip b/.yarn/cache/p-limit-npm-1.3.0-fdb471d864-281c1c0b8c.zip deleted file mode 100644 index 96906babdc..0000000000 Binary files a/.yarn/cache/p-limit-npm-1.3.0-fdb471d864-281c1c0b8c.zip and /dev/null differ diff --git a/.yarn/cache/p-locate-npm-2.0.0-3a2ee263dd-e2dceb9b49.zip b/.yarn/cache/p-locate-npm-2.0.0-3a2ee263dd-e2dceb9b49.zip deleted file mode 100644 index f6f9f09b9e..0000000000 Binary files a/.yarn/cache/p-locate-npm-2.0.0-3a2ee263dd-e2dceb9b49.zip and /dev/null differ diff --git a/.yarn/cache/p-try-npm-1.0.0-7373139e40-3b5303f77e.zip b/.yarn/cache/p-try-npm-1.0.0-7373139e40-3b5303f77e.zip deleted file mode 100644 index e12bd247e1..0000000000 Binary files a/.yarn/cache/p-try-npm-1.0.0-7373139e40-3b5303f77e.zip and /dev/null differ diff --git a/.yarn/cache/parcel-npm-2.12.0-96a4bb6cc3-d8e6cb690a.zip b/.yarn/cache/parcel-npm-2.12.0-96a4bb6cc3-d8e6cb690a.zip new file mode 100644 index 0000000000..965ad65ddc Binary files /dev/null and b/.yarn/cache/parcel-npm-2.12.0-96a4bb6cc3-d8e6cb690a.zip differ diff --git a/.yarn/cache/parcel-npm-2.8.2-7cad55fa52-b95ef40bad.zip b/.yarn/cache/parcel-npm-2.8.2-7cad55fa52-b95ef40bad.zip deleted file mode 100755 index 407cc84d40..0000000000 Binary files a/.yarn/cache/parcel-npm-2.8.2-7cad55fa52-b95ef40bad.zip and /dev/null differ diff --git a/.yarn/cache/path-exists-npm-3.0.0-e80371aa68-96e92643aa.zip b/.yarn/cache/path-exists-npm-3.0.0-e80371aa68-96e92643aa.zip deleted file mode 100644 index bdaa46fd30..0000000000 Binary files a/.yarn/cache/path-exists-npm-3.0.0-e80371aa68-96e92643aa.zip and /dev/null differ diff --git a/.yarn/cache/path-scurry-npm-1.9.1-b9d6b1c5bf-28caa788f1.zip b/.yarn/cache/path-scurry-npm-1.9.1-b9d6b1c5bf-28caa788f1.zip new file mode 100644 index 0000000000..3d6b3d39e4 Binary files /dev/null and b/.yarn/cache/path-scurry-npm-1.9.1-b9d6b1c5bf-28caa788f1.zip differ diff --git a/.yarn/cache/pinia-npm-2.0.28-9b0289223e-d515cd6220.zip b/.yarn/cache/pinia-npm-2.0.28-9b0289223e-d515cd6220.zip deleted file mode 100755 index acea97a136..0000000000 Binary files a/.yarn/cache/pinia-npm-2.0.28-9b0289223e-d515cd6220.zip and /dev/null differ diff --git a/.yarn/cache/pinia-npm-2.1.7-195409c154-1b7882aab2.zip b/.yarn/cache/pinia-npm-2.1.7-195409c154-1b7882aab2.zip new file mode 100644 index 0000000000..352e0a2f9f Binary files /dev/null and b/.yarn/cache/pinia-npm-2.1.7-195409c154-1b7882aab2.zip differ diff --git a/.yarn/cache/postcss-npm-8.4.12-e941d78a98-248e3d0f9b.zip b/.yarn/cache/postcss-npm-8.4.12-e941d78a98-248e3d0f9b.zip deleted file mode 100644 index 4f940728b9..0000000000 Binary files a/.yarn/cache/postcss-npm-8.4.12-e941d78a98-248e3d0f9b.zip and /dev/null differ diff --git a/.yarn/cache/postcss-npm-8.4.18-f1d73c0a84-9349fd9984.zip b/.yarn/cache/postcss-npm-8.4.18-f1d73c0a84-9349fd9984.zip deleted file mode 100644 index 2955faf8c2..0000000000 Binary files a/.yarn/cache/postcss-npm-8.4.18-f1d73c0a84-9349fd9984.zip and /dev/null differ diff --git a/.yarn/cache/postcss-npm-8.4.33-6ba8157009-6f98b2af4b.zip b/.yarn/cache/postcss-npm-8.4.33-6ba8157009-6f98b2af4b.zip new file mode 100644 index 0000000000..57638cbd81 Binary files /dev/null and b/.yarn/cache/postcss-npm-8.4.33-6ba8157009-6f98b2af4b.zip differ diff --git a/.yarn/cache/postcss-npm-8.4.35-6bc1848fff-cf3c3124d3.zip b/.yarn/cache/postcss-npm-8.4.35-6bc1848fff-cf3c3124d3.zip new file mode 100644 index 0000000000..888dccea0c Binary files /dev/null and b/.yarn/cache/postcss-npm-8.4.35-6bc1848fff-cf3c3124d3.zip differ diff --git a/.yarn/cache/postcss-selector-parser-npm-6.0.10-a4d7aaa270-46afaa60e3.zip b/.yarn/cache/postcss-selector-parser-npm-6.0.10-a4d7aaa270-46afaa60e3.zip deleted file mode 100644 index 496c72f70c..0000000000 Binary files a/.yarn/cache/postcss-selector-parser-npm-6.0.10-a4d7aaa270-46afaa60e3.zip and /dev/null differ diff --git a/.yarn/cache/postcss-selector-parser-npm-6.0.15-0ec4819b4e-57decb9415.zip b/.yarn/cache/postcss-selector-parser-npm-6.0.15-0ec4819b4e-57decb9415.zip new file mode 100644 index 0000000000..c6d454663e Binary files /dev/null and b/.yarn/cache/postcss-selector-parser-npm-6.0.15-0ec4819b4e-57decb9415.zip differ diff --git a/.yarn/cache/preact-npm-10.12.1-fdb903e9a5-0de99f4775.zip b/.yarn/cache/preact-npm-10.12.1-fdb903e9a5-0de99f4775.zip new file mode 100644 index 0000000000..86131908d7 Binary files /dev/null and b/.yarn/cache/preact-npm-10.12.1-fdb903e9a5-0de99f4775.zip differ diff --git a/.yarn/cache/preact-npm-10.7.2-dffb68bd4b-2f0655e043.zip b/.yarn/cache/preact-npm-10.7.2-dffb68bd4b-2f0655e043.zip deleted file mode 100644 index 40a53d7f1c..0000000000 Binary files a/.yarn/cache/preact-npm-10.7.2-dffb68bd4b-2f0655e043.zip and /dev/null differ diff --git a/.yarn/cache/pretty-format-npm-27.5.1-cd7d49696f-cf610cffcb.zip b/.yarn/cache/pretty-format-npm-27.5.1-cd7d49696f-cf610cffcb.zip deleted file mode 100644 index 8d28efe3e1..0000000000 Binary files a/.yarn/cache/pretty-format-npm-27.5.1-cd7d49696f-cf610cffcb.zip and /dev/null differ diff --git a/.yarn/cache/react-is-npm-17.0.2-091bbb8db6-9d6d111d89.zip b/.yarn/cache/react-is-npm-17.0.2-091bbb8db6-9d6d111d89.zip deleted file mode 100644 index 8b0c3e5460..0000000000 Binary files a/.yarn/cache/react-is-npm-17.0.2-091bbb8db6-9d6d111d89.zip and /dev/null differ diff --git a/.yarn/cache/regenerator-runtime-npm-0.14.0-e060897cf7-1c977ad82a.zip b/.yarn/cache/regenerator-runtime-npm-0.14.0-e060897cf7-1c977ad82a.zip new file mode 100644 index 0000000000..743dca6a4e Binary files /dev/null and b/.yarn/cache/regenerator-runtime-npm-0.14.0-e060897cf7-1c977ad82a.zip differ diff --git a/.yarn/cache/regexp.prototype.flags-npm-1.5.1-b8faeee306-869edff002.zip b/.yarn/cache/regexp.prototype.flags-npm-1.5.1-b8faeee306-869edff002.zip new file mode 100644 index 0000000000..d73fb5c3df Binary files /dev/null and b/.yarn/cache/regexp.prototype.flags-npm-1.5.1-b8faeee306-869edff002.zip differ diff --git a/.yarn/cache/resolve-npm-1.22.1-3980488690-07af5fc1e8.zip b/.yarn/cache/resolve-npm-1.22.1-3980488690-07af5fc1e8.zip deleted file mode 100644 index d41402c877..0000000000 Binary files a/.yarn/cache/resolve-npm-1.22.1-3980488690-07af5fc1e8.zip and /dev/null differ diff --git a/.yarn/cache/resolve-npm-1.22.3-f7dee15274-fb834b8134.zip b/.yarn/cache/resolve-npm-1.22.3-f7dee15274-fb834b8134.zip new file mode 100644 index 0000000000..f3daae8bc8 Binary files /dev/null and b/.yarn/cache/resolve-npm-1.22.3-f7dee15274-fb834b8134.zip differ diff --git a/.yarn/cache/resolve-npm-1.22.8-098f379dfe-f8a26958aa.zip b/.yarn/cache/resolve-npm-1.22.8-098f379dfe-f8a26958aa.zip new file mode 100644 index 0000000000..87b2b21978 Binary files /dev/null and b/.yarn/cache/resolve-npm-1.22.8-098f379dfe-f8a26958aa.zip differ diff --git a/.yarn/cache/resolve-patch-46f9469d0d-5656f4d0be.zip b/.yarn/cache/resolve-patch-46f9469d0d-5656f4d0be.zip deleted file mode 100644 index c3066c3608..0000000000 Binary files a/.yarn/cache/resolve-patch-46f9469d0d-5656f4d0be.zip and /dev/null differ diff --git a/.yarn/cache/resolve-patch-8df1eb26d0-ad59734723.zip b/.yarn/cache/resolve-patch-8df1eb26d0-ad59734723.zip new file mode 100644 index 0000000000..7d4960beb5 Binary files /dev/null and b/.yarn/cache/resolve-patch-8df1eb26d0-ad59734723.zip differ diff --git a/.yarn/cache/resolve-patch-f6b5304cab-5479b7d431.zip b/.yarn/cache/resolve-patch-f6b5304cab-5479b7d431.zip new file mode 100644 index 0000000000..84c63abe59 Binary files /dev/null and b/.yarn/cache/resolve-patch-f6b5304cab-5479b7d431.zip differ diff --git a/.yarn/cache/resolve-pkg-maps-npm-1.0.0-135b70c854-1012afc566.zip b/.yarn/cache/resolve-pkg-maps-npm-1.0.0-135b70c854-1012afc566.zip new file mode 100644 index 0000000000..53ff3fc69e Binary files /dev/null and b/.yarn/cache/resolve-pkg-maps-npm-1.0.0-135b70c854-1012afc566.zip differ diff --git a/.yarn/cache/rollup-npm-2.79.1-94e707a9a3-6a2bf167b3.zip b/.yarn/cache/rollup-npm-2.79.1-94e707a9a3-6a2bf167b3.zip deleted file mode 100644 index 84b0993034..0000000000 Binary files a/.yarn/cache/rollup-npm-2.79.1-94e707a9a3-6a2bf167b3.zip and /dev/null differ diff --git a/.yarn/cache/rollup-npm-3.29.4-5e5e5f2087-8bb20a39c8.zip b/.yarn/cache/rollup-npm-3.29.4-5e5e5f2087-8bb20a39c8.zip new file mode 100644 index 0000000000..9f6628aa42 Binary files /dev/null and b/.yarn/cache/rollup-npm-3.29.4-5e5e5f2087-8bb20a39c8.zip differ diff --git a/.yarn/cache/safe-array-concat-npm-1.0.1-8a42907bbf-001ecf1d8a.zip b/.yarn/cache/safe-array-concat-npm-1.0.1-8a42907bbf-001ecf1d8a.zip new file mode 100644 index 0000000000..6789308b81 Binary files /dev/null and b/.yarn/cache/safe-array-concat-npm-1.0.1-8a42907bbf-001ecf1d8a.zip differ diff --git a/.yarn/cache/safe-regex-test-npm-1.0.0-e94a09b84e-bc566d8beb.zip b/.yarn/cache/safe-regex-test-npm-1.0.0-e94a09b84e-bc566d8beb.zip new file mode 100644 index 0000000000..9e9dbfc637 Binary files /dev/null and b/.yarn/cache/safe-regex-test-npm-1.0.0-e94a09b84e-bc566d8beb.zip differ diff --git a/.yarn/cache/sass-npm-1.57.1-bafdba484f-734a08781b.zip b/.yarn/cache/sass-npm-1.57.1-bafdba484f-734a08781b.zip deleted file mode 100755 index 5979d6ab71..0000000000 Binary files a/.yarn/cache/sass-npm-1.57.1-bafdba484f-734a08781b.zip and /dev/null differ diff --git a/.yarn/cache/sass-npm-1.72.0-fb38bb530c-f420079c7d.zip b/.yarn/cache/sass-npm-1.72.0-fb38bb530c-f420079c7d.zip new file mode 100644 index 0000000000..a3aea4e668 Binary files /dev/null and b/.yarn/cache/sass-npm-1.72.0-fb38bb530c-f420079c7d.zip differ diff --git a/.yarn/cache/seemly-npm-0.3.3-1df3254399-b6445553f8.zip b/.yarn/cache/seemly-npm-0.3.3-1df3254399-b6445553f8.zip deleted file mode 100644 index 5265b57083..0000000000 Binary files a/.yarn/cache/seemly-npm-0.3.3-1df3254399-b6445553f8.zip and /dev/null differ diff --git a/.yarn/cache/seemly-npm-0.3.8-4940336497-98171fd4d9.zip b/.yarn/cache/seemly-npm-0.3.8-4940336497-98171fd4d9.zip new file mode 100644 index 0000000000..03ae0a8f50 Binary files /dev/null and b/.yarn/cache/seemly-npm-0.3.8-4940336497-98171fd4d9.zip differ diff --git a/.yarn/cache/semver-npm-6.3.1-bcba31fdbe-ae47d06de2.zip b/.yarn/cache/semver-npm-6.3.1-bcba31fdbe-ae47d06de2.zip new file mode 100644 index 0000000000..91f42cf845 Binary files /dev/null and b/.yarn/cache/semver-npm-6.3.1-bcba31fdbe-ae47d06de2.zip differ diff --git a/.yarn/cache/semver-npm-7.3.8-25a996cb4f-ba9c7cbbf2.zip b/.yarn/cache/semver-npm-7.3.8-25a996cb4f-ba9c7cbbf2.zip deleted file mode 100644 index c6d8940e6f..0000000000 Binary files a/.yarn/cache/semver-npm-7.3.8-25a996cb4f-ba9c7cbbf2.zip and /dev/null differ diff --git a/.yarn/cache/semver-npm-7.5.3-275095dbf3-9d58db1652.zip b/.yarn/cache/semver-npm-7.5.3-275095dbf3-9d58db1652.zip new file mode 100644 index 0000000000..79b7d4718c Binary files /dev/null and b/.yarn/cache/semver-npm-7.5.3-275095dbf3-9d58db1652.zip differ diff --git a/.yarn/cache/semver-npm-7.5.4-c4ad957fcd-12d8ad952f.zip b/.yarn/cache/semver-npm-7.5.4-c4ad957fcd-12d8ad952f.zip new file mode 100644 index 0000000000..f8689471f5 Binary files /dev/null and b/.yarn/cache/semver-npm-7.5.4-c4ad957fcd-12d8ad952f.zip differ diff --git a/.yarn/cache/semver-npm-7.6.0-f4630729f6-7427f05b70.zip b/.yarn/cache/semver-npm-7.6.0-f4630729f6-7427f05b70.zip new file mode 100644 index 0000000000..a5494e10ac Binary files /dev/null and b/.yarn/cache/semver-npm-7.6.0-f4630729f6-7427f05b70.zip differ diff --git a/.yarn/cache/set-function-length-npm-1.1.1-d362bf8221-c131d7569c.zip b/.yarn/cache/set-function-length-npm-1.1.1-d362bf8221-c131d7569c.zip new file mode 100644 index 0000000000..024add469c Binary files /dev/null and b/.yarn/cache/set-function-length-npm-1.1.1-d362bf8221-c131d7569c.zip differ diff --git a/.yarn/cache/set-function-name-npm-2.0.1-a9f970eea0-4975d17d90.zip b/.yarn/cache/set-function-name-npm-2.0.1-a9f970eea0-4975d17d90.zip new file mode 100644 index 0000000000..f18d53b599 Binary files /dev/null and b/.yarn/cache/set-function-name-npm-2.0.1-a9f970eea0-4975d17d90.zip differ diff --git a/.yarn/cache/shepherd.js-npm-10.0.1-64acc35968-be51f42734.zip b/.yarn/cache/shepherd.js-npm-10.0.1-64acc35968-be51f42734.zip deleted file mode 100755 index b2938d9d4e..0000000000 Binary files a/.yarn/cache/shepherd.js-npm-10.0.1-64acc35968-be51f42734.zip and /dev/null differ diff --git a/.yarn/cache/shepherd.js-npm-11.2.0-94b9af1487-0e71e63e51.zip b/.yarn/cache/shepherd.js-npm-11.2.0-94b9af1487-0e71e63e51.zip new file mode 100644 index 0000000000..6bd0d1e294 Binary files /dev/null and b/.yarn/cache/shepherd.js-npm-11.2.0-94b9af1487-0e71e63e51.zip differ diff --git a/.yarn/cache/signal-exit-npm-4.0.2-e3f0e8ed25-41f5928431.zip b/.yarn/cache/signal-exit-npm-4.0.2-e3f0e8ed25-41f5928431.zip new file mode 100644 index 0000000000..60c1f70c3a Binary files /dev/null and b/.yarn/cache/signal-exit-npm-4.0.2-e3f0e8ed25-41f5928431.zip differ diff --git a/.yarn/cache/slugify-npm-1.6.5-6db25d7016-a955a1b600.zip b/.yarn/cache/slugify-npm-1.6.5-6db25d7016-a955a1b600.zip deleted file mode 100644 index 3cea733dfa..0000000000 Binary files a/.yarn/cache/slugify-npm-1.6.5-6db25d7016-a955a1b600.zip and /dev/null differ diff --git a/.yarn/cache/slugify-npm-1.6.6-7ce458677d-04773c2d3b.zip b/.yarn/cache/slugify-npm-1.6.6-7ce458677d-04773c2d3b.zip new file mode 100644 index 0000000000..0630a8498b Binary files /dev/null and b/.yarn/cache/slugify-npm-1.6.6-7ce458677d-04773c2d3b.zip differ diff --git a/.yarn/cache/smoothscroll-polyfill-npm-0.4.4-69b5bb4bf7-b99ff7d916.zip b/.yarn/cache/smoothscroll-polyfill-npm-0.4.4-69b5bb4bf7-b99ff7d916.zip deleted file mode 100755 index 0943c81f3f..0000000000 Binary files a/.yarn/cache/smoothscroll-polyfill-npm-0.4.4-69b5bb4bf7-b99ff7d916.zip and /dev/null differ diff --git a/.yarn/cache/sortablejs-npm-1.15.0-f3a393abcc-bb82223a66.zip b/.yarn/cache/sortablejs-npm-1.15.0-f3a393abcc-bb82223a66.zip deleted file mode 100644 index 9028b71d1c..0000000000 Binary files a/.yarn/cache/sortablejs-npm-1.15.0-f3a393abcc-bb82223a66.zip and /dev/null differ diff --git a/.yarn/cache/sortablejs-npm-1.15.2-73347ae85a-36b20b144f.zip b/.yarn/cache/sortablejs-npm-1.15.2-73347ae85a-36b20b144f.zip new file mode 100644 index 0000000000..b303125761 Binary files /dev/null and b/.yarn/cache/sortablejs-npm-1.15.2-73347ae85a-36b20b144f.zip differ diff --git a/.yarn/cache/source-map-npm-0.8.0-beta.0-688a309e94-e94169be64.zip b/.yarn/cache/source-map-npm-0.8.0-beta.0-688a309e94-e94169be64.zip deleted file mode 100644 index 877220ab43..0000000000 Binary files a/.yarn/cache/source-map-npm-0.8.0-beta.0-688a309e94-e94169be64.zip and /dev/null differ diff --git a/.yarn/cache/source-map-support-npm-0.5.21-09ca99e250-43e98d700d.zip b/.yarn/cache/source-map-support-npm-0.5.21-09ca99e250-43e98d700d.zip deleted file mode 100644 index 5fc27c8438..0000000000 Binary files a/.yarn/cache/source-map-support-npm-0.5.21-09ca99e250-43e98d700d.zip and /dev/null differ diff --git a/.yarn/cache/sourcemap-codec-npm-1.4.8-3a1a9e60b1-b57981c056.zip b/.yarn/cache/sourcemap-codec-npm-1.4.8-3a1a9e60b1-b57981c056.zip deleted file mode 100644 index de84f79779..0000000000 Binary files a/.yarn/cache/sourcemap-codec-npm-1.4.8-3a1a9e60b1-b57981c056.zip and /dev/null differ diff --git a/.yarn/cache/srcset-npm-4.0.0-4e99d43236-aceb898c92.zip b/.yarn/cache/srcset-npm-4.0.0-4e99d43236-aceb898c92.zip new file mode 100644 index 0000000000..2c5170b494 Binary files /dev/null and b/.yarn/cache/srcset-npm-4.0.0-4e99d43236-aceb898c92.zip differ diff --git a/.yarn/cache/string-width-npm-5.1.2-bf60531341-7369deaa29.zip b/.yarn/cache/string-width-npm-5.1.2-bf60531341-7369deaa29.zip new file mode 100644 index 0000000000..bd88405658 Binary files /dev/null and b/.yarn/cache/string-width-npm-5.1.2-bf60531341-7369deaa29.zip differ diff --git a/.yarn/cache/string.prototype.trim-npm-1.2.8-7ed4517ce8-49eb1a862a.zip b/.yarn/cache/string.prototype.trim-npm-1.2.8-7ed4517ce8-49eb1a862a.zip new file mode 100644 index 0000000000..543f676ced Binary files /dev/null and b/.yarn/cache/string.prototype.trim-npm-1.2.8-7ed4517ce8-49eb1a862a.zip differ diff --git a/.yarn/cache/string.prototype.trimend-npm-1.0.4-a656b8fe24-17e5aa45c3.zip b/.yarn/cache/string.prototype.trimend-npm-1.0.4-a656b8fe24-17e5aa45c3.zip deleted file mode 100644 index 3a6cb8db61..0000000000 Binary files a/.yarn/cache/string.prototype.trimend-npm-1.0.4-a656b8fe24-17e5aa45c3.zip and /dev/null differ diff --git a/.yarn/cache/string.prototype.trimend-npm-1.0.7-159b9dcfbc-2375516272.zip b/.yarn/cache/string.prototype.trimend-npm-1.0.7-159b9dcfbc-2375516272.zip new file mode 100644 index 0000000000..93f30c147e Binary files /dev/null and b/.yarn/cache/string.prototype.trimend-npm-1.0.7-159b9dcfbc-2375516272.zip differ diff --git a/.yarn/cache/string.prototype.trimstart-npm-1.0.4-b31f5e7c85-3fb06818d3.zip b/.yarn/cache/string.prototype.trimstart-npm-1.0.4-b31f5e7c85-3fb06818d3.zip deleted file mode 100644 index 477439a720..0000000000 Binary files a/.yarn/cache/string.prototype.trimstart-npm-1.0.4-b31f5e7c85-3fb06818d3.zip and /dev/null differ diff --git a/.yarn/cache/string.prototype.trimstart-npm-1.0.7-ae2f803b78-13d0c2cb0d.zip b/.yarn/cache/string.prototype.trimstart-npm-1.0.7-ae2f803b78-13d0c2cb0d.zip new file mode 100644 index 0000000000..187509d052 Binary files /dev/null and b/.yarn/cache/string.prototype.trimstart-npm-1.0.7-ae2f803b78-13d0c2cb0d.zip differ diff --git a/.yarn/cache/strip-ansi-npm-7.0.1-668c121204-257f78fa43.zip b/.yarn/cache/strip-ansi-npm-7.0.1-668c121204-257f78fa43.zip new file mode 100644 index 0000000000..84c011395c Binary files /dev/null and b/.yarn/cache/strip-ansi-npm-7.0.1-668c121204-257f78fa43.zip differ diff --git a/.yarn/cache/terser-npm-5.13.1-c7df10bd07-0b1f5043cf.zip b/.yarn/cache/terser-npm-5.13.1-c7df10bd07-0b1f5043cf.zip deleted file mode 100644 index b8f30c7eff..0000000000 Binary files a/.yarn/cache/terser-npm-5.13.1-c7df10bd07-0b1f5043cf.zip and /dev/null differ diff --git a/.yarn/cache/tr46-npm-1.0.1-9547f343a4-96d4ed46bc.zip b/.yarn/cache/tr46-npm-1.0.1-9547f343a4-96d4ed46bc.zip deleted file mode 100644 index 3130815a0d..0000000000 Binary files a/.yarn/cache/tr46-npm-1.0.1-9547f343a4-96d4ed46bc.zip and /dev/null differ diff --git a/.yarn/cache/tsconfig-paths-npm-3.14.1-17a815b5c5-8afa01c673.zip b/.yarn/cache/tsconfig-paths-npm-3.14.1-17a815b5c5-8afa01c673.zip deleted file mode 100644 index 98a7ab1f87..0000000000 Binary files a/.yarn/cache/tsconfig-paths-npm-3.14.1-17a815b5c5-8afa01c673.zip and /dev/null differ diff --git a/.yarn/cache/tsconfig-paths-npm-3.15.0-ff68930e0e-59f35407a3.zip b/.yarn/cache/tsconfig-paths-npm-3.15.0-ff68930e0e-59f35407a3.zip new file mode 100644 index 0000000000..abfe8dd47e Binary files /dev/null and b/.yarn/cache/tsconfig-paths-npm-3.15.0-ff68930e0e-59f35407a3.zip differ diff --git a/.yarn/cache/typed-array-buffer-npm-1.0.0-95cb610310-3e0281c79b.zip b/.yarn/cache/typed-array-buffer-npm-1.0.0-95cb610310-3e0281c79b.zip new file mode 100644 index 0000000000..7e8dc8f1ed Binary files /dev/null and b/.yarn/cache/typed-array-buffer-npm-1.0.0-95cb610310-3e0281c79b.zip differ diff --git a/.yarn/cache/typed-array-byte-length-npm-1.0.0-94d79975ca-b03db16458.zip b/.yarn/cache/typed-array-byte-length-npm-1.0.0-94d79975ca-b03db16458.zip new file mode 100644 index 0000000000..9cd6f34788 Binary files /dev/null and b/.yarn/cache/typed-array-byte-length-npm-1.0.0-94d79975ca-b03db16458.zip differ diff --git a/.yarn/cache/typed-array-byte-offset-npm-1.0.0-8cbb911cf5-04f6f02d0e.zip b/.yarn/cache/typed-array-byte-offset-npm-1.0.0-8cbb911cf5-04f6f02d0e.zip new file mode 100644 index 0000000000..2318610bbc Binary files /dev/null and b/.yarn/cache/typed-array-byte-offset-npm-1.0.0-8cbb911cf5-04f6f02d0e.zip differ diff --git a/.yarn/cache/typed-array-length-npm-1.0.4-92771b81fc-2228febc93.zip b/.yarn/cache/typed-array-length-npm-1.0.4-92771b81fc-2228febc93.zip new file mode 100644 index 0000000000..f68a3c2c96 Binary files /dev/null and b/.yarn/cache/typed-array-length-npm-1.0.4-92771b81fc-2228febc93.zip differ diff --git a/.yarn/cache/v8-compile-cache-npm-2.3.0-961375f150-adb0a271ea.zip b/.yarn/cache/v8-compile-cache-npm-2.3.0-961375f150-adb0a271ea.zip deleted file mode 100644 index 0e04423cd8..0000000000 Binary files a/.yarn/cache/v8-compile-cache-npm-2.3.0-961375f150-adb0a271ea.zip and /dev/null differ diff --git a/.yarn/cache/vanillajs-datepicker-npm-1.3.4-bc86e15a9c-830958f8af.zip b/.yarn/cache/vanillajs-datepicker-npm-1.3.4-bc86e15a9c-830958f8af.zip new file mode 100644 index 0000000000..151d4c9230 Binary files /dev/null and b/.yarn/cache/vanillajs-datepicker-npm-1.3.4-bc86e15a9c-830958f8af.zip differ diff --git a/.yarn/cache/vite-npm-3.2.5-f23b9ecb5b-ad35b7008c.zip b/.yarn/cache/vite-npm-3.2.5-f23b9ecb5b-ad35b7008c.zip deleted file mode 100755 index f02dcb18e9..0000000000 Binary files a/.yarn/cache/vite-npm-3.2.5-f23b9ecb5b-ad35b7008c.zip and /dev/null differ diff --git a/.yarn/cache/vite-npm-4.5.3-5cedc7cb8f-fd3f512ce4.zip b/.yarn/cache/vite-npm-4.5.3-5cedc7cb8f-fd3f512ce4.zip new file mode 100644 index 0000000000..c6bb0e4ef7 Binary files /dev/null and b/.yarn/cache/vite-npm-4.5.3-5cedc7cb8f-fd3f512ce4.zip differ diff --git a/.yarn/cache/volar-service-html-npm-0.0.34-32b6d24136-83b50cd805.zip b/.yarn/cache/volar-service-html-npm-0.0.34-32b6d24136-83b50cd805.zip new file mode 100644 index 0000000000..0f1e9805f7 Binary files /dev/null and b/.yarn/cache/volar-service-html-npm-0.0.34-32b6d24136-83b50cd805.zip differ diff --git a/.yarn/cache/volar-service-pug-npm-0.0.34-6f5429e17c-4691aa1c8e.zip b/.yarn/cache/volar-service-pug-npm-0.0.34-6f5429e17c-4691aa1c8e.zip new file mode 100644 index 0000000000..d53f3521ee Binary files /dev/null and b/.yarn/cache/volar-service-pug-npm-0.0.34-6f5429e17c-4691aa1c8e.zip differ diff --git a/.yarn/cache/vscode-html-languageservice-npm-5.1.2-2ea2618bdd-3a2a5ee5ad.zip b/.yarn/cache/vscode-html-languageservice-npm-5.1.2-2ea2618bdd-3a2a5ee5ad.zip new file mode 100644 index 0000000000..d83607888b Binary files /dev/null and b/.yarn/cache/vscode-html-languageservice-npm-5.1.2-2ea2618bdd-3a2a5ee5ad.zip differ diff --git a/.yarn/cache/vscode-jsonrpc-npm-8.2.0-b7d2e5b553-f302a01e59.zip b/.yarn/cache/vscode-jsonrpc-npm-8.2.0-b7d2e5b553-f302a01e59.zip new file mode 100644 index 0000000000..75e2c086b6 Binary files /dev/null and b/.yarn/cache/vscode-jsonrpc-npm-8.2.0-b7d2e5b553-f302a01e59.zip differ diff --git a/.yarn/cache/vscode-languageserver-protocol-npm-3.17.5-2b07e16989-dfb42d276d.zip b/.yarn/cache/vscode-languageserver-protocol-npm-3.17.5-2b07e16989-dfb42d276d.zip new file mode 100644 index 0000000000..bcb5ae5b4e Binary files /dev/null and b/.yarn/cache/vscode-languageserver-protocol-npm-3.17.5-2b07e16989-dfb42d276d.zip differ diff --git a/.yarn/cache/vscode-languageserver-textdocument-npm-1.0.11-6fc94d2b7b-ea7cdc9d4f.zip b/.yarn/cache/vscode-languageserver-textdocument-npm-1.0.11-6fc94d2b7b-ea7cdc9d4f.zip new file mode 100644 index 0000000000..b1edfda12d Binary files /dev/null and b/.yarn/cache/vscode-languageserver-textdocument-npm-1.0.11-6fc94d2b7b-ea7cdc9d4f.zip differ diff --git a/.yarn/cache/vscode-languageserver-types-npm-3.17.5-aca3b71a5a-79b420e757.zip b/.yarn/cache/vscode-languageserver-types-npm-3.17.5-aca3b71a5a-79b420e757.zip new file mode 100644 index 0000000000..ec214b2903 Binary files /dev/null and b/.yarn/cache/vscode-languageserver-types-npm-3.17.5-aca3b71a5a-79b420e757.zip differ diff --git a/.yarn/cache/vscode-uri-npm-3.0.8-56f46b9d24-5142491268.zip b/.yarn/cache/vscode-uri-npm-3.0.8-56f46b9d24-5142491268.zip new file mode 100644 index 0000000000..6dadd110c9 Binary files /dev/null and b/.yarn/cache/vscode-uri-npm-3.0.8-56f46b9d24-5142491268.zip differ diff --git a/.yarn/cache/vue-demi-npm-0.13.1-a467bc3a9a-d26b060258.zip b/.yarn/cache/vue-demi-npm-0.13.1-a467bc3a9a-d26b060258.zip deleted file mode 100644 index 570ed5bd38..0000000000 Binary files a/.yarn/cache/vue-demi-npm-0.13.1-a467bc3a9a-d26b060258.zip and /dev/null differ diff --git a/.yarn/cache/vue-demi-npm-0.14.5-6e9e31189b-ff44b9372b.zip b/.yarn/cache/vue-demi-npm-0.14.5-6e9e31189b-ff44b9372b.zip new file mode 100644 index 0000000000..413aac531f Binary files /dev/null and b/.yarn/cache/vue-demi-npm-0.14.5-6e9e31189b-ff44b9372b.zip differ diff --git a/.yarn/cache/vue-eslint-parser-npm-9.0.3-1d52721799-61248eb504.zip b/.yarn/cache/vue-eslint-parser-npm-9.0.3-1d52721799-61248eb504.zip deleted file mode 100644 index 175f3e8846..0000000000 Binary files a/.yarn/cache/vue-eslint-parser-npm-9.0.3-1d52721799-61248eb504.zip and /dev/null differ diff --git a/.yarn/cache/vue-eslint-parser-npm-9.4.2-3e4e696025-67f14c8ea1.zip b/.yarn/cache/vue-eslint-parser-npm-9.4.2-3e4e696025-67f14c8ea1.zip new file mode 100644 index 0000000000..9ec85e189e Binary files /dev/null and b/.yarn/cache/vue-eslint-parser-npm-9.4.2-3e4e696025-67f14c8ea1.zip differ diff --git a/.yarn/cache/vue-npm-3.2.45-06b4b60efe-df60ca80cb.zip b/.yarn/cache/vue-npm-3.2.45-06b4b60efe-df60ca80cb.zip deleted file mode 100644 index aff8d8820e..0000000000 Binary files a/.yarn/cache/vue-npm-3.2.45-06b4b60efe-df60ca80cb.zip and /dev/null differ diff --git a/.yarn/cache/vue-npm-3.4.21-02110aa6d9-3c477982a0.zip b/.yarn/cache/vue-npm-3.4.21-02110aa6d9-3c477982a0.zip new file mode 100644 index 0000000000..c48b4e5dfe Binary files /dev/null and b/.yarn/cache/vue-npm-3.4.21-02110aa6d9-3c477982a0.zip differ diff --git a/.yarn/cache/vue-router-npm-4.1.6-ccab7109e1-c7f0156ac0.zip b/.yarn/cache/vue-router-npm-4.1.6-ccab7109e1-c7f0156ac0.zip deleted file mode 100644 index 2c00e0a7ce..0000000000 Binary files a/.yarn/cache/vue-router-npm-4.1.6-ccab7109e1-c7f0156ac0.zip and /dev/null differ diff --git a/.yarn/cache/vue-router-npm-4.3.0-b765d40138-0059261d39.zip b/.yarn/cache/vue-router-npm-4.3.0-b765d40138-0059261d39.zip new file mode 100644 index 0000000000..6b93953624 Binary files /dev/null and b/.yarn/cache/vue-router-npm-4.3.0-b765d40138-0059261d39.zip differ diff --git a/.yarn/cache/vueuc-npm-0.4.47-ad081ddd15-b82b77a882.zip b/.yarn/cache/vueuc-npm-0.4.47-ad081ddd15-b82b77a882.zip deleted file mode 100644 index 0a267b628a..0000000000 Binary files a/.yarn/cache/vueuc-npm-0.4.47-ad081ddd15-b82b77a882.zip and /dev/null differ diff --git a/.yarn/cache/vueuc-npm-0.4.58-be5584770c-fb0b9a69be.zip b/.yarn/cache/vueuc-npm-0.4.58-be5584770c-fb0b9a69be.zip new file mode 100644 index 0000000000..f62e5e32e8 Binary files /dev/null and b/.yarn/cache/vueuc-npm-0.4.58-be5584770c-fb0b9a69be.zip differ diff --git a/.yarn/cache/webidl-conversions-npm-4.0.2-1d159e6409-c93d8dfe90.zip b/.yarn/cache/webidl-conversions-npm-4.0.2-1d159e6409-c93d8dfe90.zip deleted file mode 100644 index a75f5ee65f..0000000000 Binary files a/.yarn/cache/webidl-conversions-npm-4.0.2-1d159e6409-c93d8dfe90.zip and /dev/null differ diff --git a/.yarn/cache/whatwg-url-npm-7.1.0-d6cae01571-fecb07c872.zip b/.yarn/cache/whatwg-url-npm-7.1.0-d6cae01571-fecb07c872.zip deleted file mode 100644 index 9f21814850..0000000000 Binary files a/.yarn/cache/whatwg-url-npm-7.1.0-d6cae01571-fecb07c872.zip and /dev/null differ diff --git a/.yarn/cache/which-typed-array-npm-1.1.13-92c18b4878-3828a0d5d7.zip b/.yarn/cache/which-typed-array-npm-1.1.13-92c18b4878-3828a0d5d7.zip new file mode 100644 index 0000000000..0d9d2479da Binary files /dev/null and b/.yarn/cache/which-typed-array-npm-1.1.13-92c18b4878-3828a0d5d7.zip differ diff --git a/.yarn/cache/word-wrap-npm-1.2.3-7fb15ab002-30b48f91fc.zip b/.yarn/cache/word-wrap-npm-1.2.3-7fb15ab002-30b48f91fc.zip deleted file mode 100644 index 518977eb88..0000000000 Binary files a/.yarn/cache/word-wrap-npm-1.2.3-7fb15ab002-30b48f91fc.zip and /dev/null differ diff --git a/.yarn/cache/wrap-ansi-npm-8.1.0-26a4e6ae28-371733296d.zip b/.yarn/cache/wrap-ansi-npm-8.1.0-26a4e6ae28-371733296d.zip new file mode 100644 index 0000000000..2ee78f31c8 Binary files /dev/null and b/.yarn/cache/wrap-ansi-npm-8.1.0-26a4e6ae28-371733296d.zip differ diff --git a/.yarn/cache/yargs-npm-16.2.0-547873d425-b14afbb51e.zip b/.yarn/cache/yargs-npm-16.2.0-547873d425-b14afbb51e.zip deleted file mode 100644 index d11c27d510..0000000000 Binary files a/.yarn/cache/yargs-npm-16.2.0-547873d425-b14afbb51e.zip and /dev/null differ diff --git a/.yarn/cache/yargs-npm-17.7.2-80b62638e1-73b572e863.zip b/.yarn/cache/yargs-npm-17.7.2-80b62638e1-73b572e863.zip new file mode 100644 index 0000000000..54c49dc9c6 Binary files /dev/null and b/.yarn/cache/yargs-npm-17.7.2-80b62638e1-73b572e863.zip differ diff --git a/.yarn/cache/yargs-parser-npm-20.2.9-a1d19e598d-8bb69015f2.zip b/.yarn/cache/yargs-parser-npm-20.2.9-a1d19e598d-8bb69015f2.zip deleted file mode 100644 index f230038cfc..0000000000 Binary files a/.yarn/cache/yargs-parser-npm-20.2.9-a1d19e598d-8bb69015f2.zip and /dev/null differ diff --git a/.yarn/cache/yargs-parser-npm-21.1.1-8fdc003314-ed2d96a616.zip b/.yarn/cache/yargs-parser-npm-21.1.1-8fdc003314-ed2d96a616.zip new file mode 100644 index 0000000000..d68ba748e7 Binary files /dev/null and b/.yarn/cache/yargs-parser-npm-21.1.1-8fdc003314-ed2d96a616.zip differ diff --git a/LICENSE b/LICENSE index aaed0ef57d..dc6e0c5663 100644 --- a/LICENSE +++ b/LICENSE @@ -1,6 +1,6 @@ BSD 3-Clause License -Copyright (c) 2008-2022, The IETF Trust +Copyright (c) 2008-2024, The IETF Trust All rights reserved. Redistribution and use in source and binary forms, with or without @@ -26,4 +26,4 @@ DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. \ No newline at end of file +OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. diff --git a/README.md b/README.md index 2acc75291c..baffc311e7 100644 --- a/README.md +++ b/README.md @@ -4,12 +4,11 @@ [![Release](https://img.shields.io/github/release/ietf-tools/datatracker.svg?style=flat&maxAge=300)](https://github.com/ietf-tools/datatracker/releases) [![License](https://img.shields.io/github/license/ietf-tools/datatracker)](https://github.com/ietf-tools/datatracker/blob/main/LICENSE) -[![Code Coverage](https://codecov.io/gh/ietf-tools/datatracker/branch/feat/bs5/graph/badge.svg?token=V4DXB0Q28C)](https://codecov.io/gh/ietf-tools/datatracker) -[![Nightly Dev DB Image](https://github.com/ietf-tools/datatracker/actions/workflows/dev-db-nightly.yml/badge.svg)](https://github.com/ietf-tools/datatracker/pkgs/container/datatracker-db) -[![Python Version](https://img.shields.io/badge/python-3.9-blue?logo=python&logoColor=white)](#prerequisites) -[![Django Version](https://img.shields.io/badge/django-2.x-51be95?logo=django&logoColor=white)](#prerequisites) +[![Code Coverage](https://codecov.io/gh/ietf-tools/datatracker/branch/feat/bs5/graph/badge.svg?token=V4DXB0Q28C)](https://codecov.io/gh/ietf-tools/datatracker) +[![Python Version](https://img.shields.io/badge/python-3.12-blue?logo=python&logoColor=white)](#prerequisites) +[![Django Version](https://img.shields.io/badge/django-4.x-51be95?logo=django&logoColor=white)](#prerequisites) [![Node Version](https://img.shields.io/badge/node.js-16.x-green?logo=node.js&logoColor=white)](#prerequisites) -[![MariaDB Version](https://img.shields.io/badge/mariadb-10-blue?logo=mariadb&logoColor=white)](#prerequisites) +[![MariaDB Version](https://img.shields.io/badge/postgres-17-blue?logo=postgresql&logoColor=white)](#prerequisites) ##### The day-to-day front-end to the IETF database for people who work on IETF standards. @@ -18,7 +17,8 @@ - [**Production Website**](https://datatracker.ietf.org) - [Changelog](https://github.com/ietf-tools/datatracker/releases) - [Contributing](https://github.com/ietf-tools/.github/blob/main/CONTRIBUTING.md) -- [Getting Started](#getting-started) +- [Getting Started](#getting-started) - *[ tl;dr ](#the-tldr-to-get-going)* + - [Creating a Fork](#creating-a-fork) - [Git Cloning Tips](#git-cloning-tips) - [Docker Dev Environment](docker/README.md) - [Database & Assets](#database--assets) @@ -44,14 +44,24 @@ This project is following the standard **Git Feature Workflow** development model. Learn about all the various steps of the development workflow, from creating a fork to submitting a pull request, in the [Contributing](https://github.com/ietf-tools/.github/blob/main/CONTRIBUTING.md) guide. +> [!TIP] > Make sure to read the [Styleguides](https://github.com/ietf-tools/.github/blob/main/CONTRIBUTING.md#styleguides) section to ensure a cohesive code format across the project. You can submit bug reports, enhancement and new feature requests in the [discussions](https://github.com/ietf-tools/datatracker/discussions) area. Accepted tickets will be converted to issues. +#### Creating a Fork + +Click the Fork button in the top-right corner of the repository to create a personal copy that you can work on. + +> [!NOTE] +> Some GitHub Actions might be enabled by default in your fork. You should disable them by going to **Settings** > **Actions** > **General** and selecting **Disable actions** (then Save). + #### Git Cloning Tips As outlined in the [Contributing](https://github.com/ietf-tools/.github/blob/main/CONTRIBUTING.md) guide, you will first want to create a fork of the datatracker project in your personal GitHub account before cloning it. +Windows developers: [Start with WSL2 from the beginning](https://github.com/ietf-tools/.github/blob/main/docs/windows-dev.md). + Because of the extensive history of this project, cloning the datatracker project locally can take a long time / disk space. You can speed up the cloning process by limiting the history depth, for example *(replace `USERNAME` with your GitHub username)*: - To fetch only up to the 10 latest commits: @@ -63,6 +73,25 @@ Because of the extensive history of this project, cloning the datatracker projec git clone --shallow-since=DATE https://github.com/USERNAME/datatracker.git ``` +#### The tl;dr to get going + +Note that you will have to have cloned the datatracker code locally - please read the above sections. + +Datatracker development is performed using Docker containers. You will need to be able to run docker (and docker-compose) on your machine to effectively develop. It is possible to get a purely native install working, but it is _very complicated_ and typically takes a first time datatracker developer a full day of setup, where the docker setup completes in a small number of minutes. + +Many developers are using [VS Code](https://code.visualstudio.com/) and taking advantage of VS Code's ability to start a project in a set of containers. If you are using VS Code, simply start VS Code in your clone and inside VS Code choose `Restart in container`. + +If VS Code is not available to you, in your clone, type `cd docker; ./run` + +Once the containers are started, run the tests to make sure your checkout is a good place to start from (all tests should pass - if any fail, ask for help at tools-help@). Inside the app container's shell type: +```sh +ietf/manage.py test --settings=settings_test +``` + +Note that we recently moved the datatracker onto PostgreSQL - you may still find older documentation that suggests testing with settings_sqlitetest. That will no longer work. + +For a more detailed description of getting going, see [docker/README.md](docker/README.md). + #### Overview of the datatracker models A beginning of a [walkthrough of the datatracker models](https://notes.ietf.org/iab-aid-datatracker-database-overview) was prepared for the IAB AID workshop. @@ -75,10 +104,27 @@ Read the [Docker Dev Environment](docker/README.md) guide to get started. ### Database & Assets -Nightly database dumps of the datatracker are available at -https://www.ietf.org/lib/dt/sprint/ietf_utf8.sql.gz +Nightly database dumps of the datatracker are available as Docker images: `ghcr.io/ietf-tools/datatracker-db:latest` + +> [!TIP] +> In order to update the database in your dev environment to the latest version, you should run the `docker/cleandb` script. + +### Blob storage for dev/test + +The dev and test environments use [minio](https://github.com/minio/minio) to provide local blob storage. See the settings files for how the app container communicates with the blobstore container. If you need to work with minio directly from outside the containers (to interact with its api or console), use `docker compose` from the top level directory of your clone to expose it at an ephemeral port. + +``` +$ docker compose port blobstore 9001 +0.0.0.0: + +$ curl -I http://localhost: +HTTP/1.1 200 OK +... +``` + + +The minio container exposes the minio api at port 9000 and the minio console at port 9001 -> Note that this link is provided as reference only. To update the database in your dev environment to the latest version, you should instead run the `docker/cleandb` script! ### Frontend Development @@ -96,7 +142,7 @@ Pages will gradually be updated to Vue 3 components. These components are locate Each Vue 3 app has its own sub-directory. For example, the agenda app is located under `/client/agenda`. -The datatracker makes use of the Django-Vite plugin to point to either the Vite.js server or the precompiled production files. The `DJANGO_VITE_DEV_MODE` flag, found in the `ietf/settings_local.py` file determines whether the Vite.js server is used or not. +The datatracker makes use of the Django-Vite plugin to point to either the Vite.js server or the precompiled production files. The `DJANGO_VITE["default"]["dev_mode"]` flag, found in the `ietf/settings_local.py` file determines whether the Vite.js server is used or not. In development mode, you must start the Vite.js development server, in addition to the usual Datatracker server: @@ -116,10 +162,10 @@ This will create packages under `ietf/static/dist-neue`, which are then served b #### Parcel *(Legacy/jQuery)* -The Datatracker includes these packages from the various Javascript and CSS files in `ietf/static/js` and `ietf/static/css`, respectively. +The Datatracker includes these packages from the various Javascript and CSS files in `ietf/static/js` and `ietf/static/css` respectively, bundled using Parcel. Static images are likewise in `ietf/static/images`. -Whenever changes are made to the files under `ietf/static`, you must re-run `parcel` to package them: +Whenever changes are made to the files under `ietf/static`, you must re-run the build command to package them: ``` shell yarn legacy:build @@ -202,9 +248,10 @@ before activating a new release. From a datatracker container, run the command: ```sh -./ietf/manage.py test --settings=settings_local_sqlitetest +./ietf/manage.py test --settings=settings_test ``` +> [!TIP] > You can limit the run to specific tests using the `--pattern` argument. ### Frontend Tests @@ -214,11 +261,13 @@ Frontend tests are done via Playwright. There're 2 different type of tests: - Tests that test Vue pages / components and run natively without any external dependency. - Tests that require a running datatracker instance to test against (usually legacy views). +> [!IMPORTANT] > Make sure you have Node.js 16.x or later installed on your machine. #### Run Vue Tests -> :warning: All commands below **MUST** be run from the `./playwright` directory, unless noted otherwise. +> [!WARNING] +> All commands below **MUST** be run from the `./playwright` directory, unless noted otherwise. 1. Run **once** to install dependencies on your system: ```sh @@ -251,7 +300,8 @@ Frontend tests are done via Playwright. There're 2 different type of tests: First, you need to start a datatracker instance (dev or prod), ideally from a docker container, exposing the 8000 port. -> :warning: All commands below **MUST** be run from the `./playwright` directory. +> [!WARNING] +> All commands below **MUST** be run from the `./playwright` directory. 1. Run **once** to install dependencies on your system: ```sh @@ -264,6 +314,7 @@ npm run install-deps npm run test:legacy ``` + ### Diff Tool To compare 2 different datatracker instances and look for diff, read the [diff tool instructions](dev/diff). diff --git a/bin/add-old-drafts-from-archive.py b/bin/add-old-drafts-from-archive.py deleted file mode 100755 index 239ba7837c..0000000000 --- a/bin/add-old-drafts-from-archive.py +++ /dev/null @@ -1,151 +0,0 @@ -#!/usr/bin/env python -# Copyright The IETF Trust 2017-2019, All Rights Reserved - -import datetime -import os -import sys -from pathlib import Path -from contextlib import closing - -os.environ["DJANGO_SETTINGS_MODULE"] = "ietf.settings" - -import django -django.setup() - -from django.conf import settings -from django.core.validators import validate_email, ValidationError -from ietf.utils.draft import PlaintextDraft -from ietf.submit.utils import update_authors -from ietf.utils.timezone import date_today - -import debug # pyflakes:ignore - -from ietf.doc.models import Document, NewRevisionDocEvent, DocEvent, State -from ietf.person.models import Person - -system = Person.objects.get(name="(System)") -expired = State.objects.get(type='draft',slug='expired') - -names = set() -print 'collecting draft names ...' -versions = 0 -for p in Path(settings.INTERNET_DRAFT_PATH).glob('draft*.txt'): - n = str(p).split('/')[-1].split('-') - if n[-1][:2].isdigit(): - name = '-'.join(n[:-1]) - if '--' in name or '.txt' in name or '[' in name or '=' in name or '&' in name: - continue - if name.startswith('draft-draft-'): - continue - if name == 'draft-ietf-trade-iotp-v1_0-dsig': - continue - if len(n[-1]) != 6: - continue - if name.startswith('draft-mlee-'): - continue - names.add('-'.join(n[:-1])) - -count=0 -print 'iterating through names ...' -for name in sorted(names): - if not Document.objects.filter(name=name).exists(): - paths = list(Path(settings.INTERNET_DRAFT_PATH).glob('%s-??.txt'%name)) - paths.sort() - doc = None - for p in paths: - n = str(p).split('/')[-1].split('-') - rev = n[-1][:2] - with open(str(p)) as txt_file: - raw = txt_file.read() - try: - text = raw.decode('utf8') - except UnicodeDecodeError: - text = raw.decode('latin1') - try: - draft = PlaintextDraft(text, txt_file.name, name_from_source=True) - except Exception as e: - print name, rev, "Can't parse", p,":",e - continue - if draft.errors and draft.errors.keys()!=['draftname',]: - print "Errors - could not process", name, rev, datetime.datetime.fromtimestamp(p.stat().st_mtime, datetime.timezone.utc), draft.errors, draft.get_title().encode('utf8') - else: - time = datetime.datetime.fromtimestamp(p.stat().st_mtime, datetime.timezone.utc) - if not doc: - doc = Document.objects.create(name=name, - time=time, - type_id='draft', - title=draft.get_title(), - abstract=draft.get_abstract(), - rev = rev, - pages=draft.get_pagecount(), - words=draft.get_wordcount(), - expires=time+datetime.timedelta(settings.INTERNET_DRAFT_DAYS_TO_EXPIRE), - ) - DocAlias.objects.create(name=doc.name).docs.add(doc) - doc.states.add(expired) - # update authors - authors = [] - for author in draft.get_author_list(): - full_name, first_name, middle_initial, last_name, name_suffix, email, country, company = author - - author_name = full_name.replace("\n", "").replace("\r", "").replace("<", "").replace(">", "").strip() - - if email: - try: - validate_email(email) - except ValidationError: - email = "" - - def turn_into_unicode(s): - if s is None: - return u"" - - if isinstance(s, unicode): - return s - else: - try: - return s.decode("utf-8") - except UnicodeDecodeError: - try: - return s.decode("latin-1") - except UnicodeDecodeError: - return "" - - author_name = turn_into_unicode(author_name) - email = turn_into_unicode(email) - company = turn_into_unicode(company) - - authors.append({ - "name": author_name, - "email": email, - "affiliation": company, - "country": country - }) - dummysubmission=type('', (), {})() #https://stackoverflow.com/questions/19476816/creating-an-empty-object-in-python - dummysubmission.authors = authors - update_authors(doc,dummysubmission) - - # add a docevent with words explaining where this came from - events = [] - e = NewRevisionDocEvent.objects.create( - type="new_revision", - doc=doc, - rev=rev, - by=system, - desc="New version available: %s-%s.txt" % (doc.name, doc.rev), - time=time, - ) - events.append(e) - e = DocEvent.objects.create( - type="comment", - doc = doc, - rev = rev, - by = system, - desc = "Revision added from id-archive on %s by %s"%(date_today(),sys.argv[0]), - time=time, - ) - events.append(e) - doc.time = time - doc.rev = rev - doc.save_with_history(events) - print "Added",name, rev diff --git a/bin/check-copyright b/bin/check-copyright deleted file mode 100755 index 13cbcd8582..0000000000 --- a/bin/check-copyright +++ /dev/null @@ -1,261 +0,0 @@ -#!/usr/bin/env python3.7 -# -*- mode: python; coding: utf-8 -*- -# Copyright The IETF Trust 2019, All Rights Reserved -""" -NAME - $program - Check for current copyright notice in given files - -SYNOPSIS - $program [OPTIONS] ARGS - -DESCRIPTION - Given a list of files or filename wildcard patterns, check all for - an IETF Trust copyright notice with the current year. Optionally - generate a diff on standard out which can be used by 'patch'. - - An invocation similar to the following can be particularly useful with - a set of changed version-controlled files, as it will fix up the - Copyright statements of any python files with pending changes: - - $ check-copyright -p $(svn st | cut -c 9- | grep '\.py$' ) | patch -p0 - - -%(options)s - -AUTHOR - Written by Henrik Levkowetz, - -COPYRIGHT - Copyright 2019 the IETF Trust - - This program is free software; you can redistribute it and/or modify - it under the terms of the Simplified BSD license as published by the - Open Source Initiative at http://opensource.org/licenses/BSD-2-Clause. - -""" - - -import datetime -import os -import sys -import time - -path = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) -if not path in sys.path: - sys.path.insert(0, path) - -import getopt -import re -import pytz -import tzparse -import debug - -version = "1.0.0" -program = os.path.basename(sys.argv[0]) -progdir = os.path.dirname(sys.argv[0]) - -debug.debug = True - -# ---------------------------------------------------------------------- -# Parse options - -options = "" -for line in re.findall("\n +(if|elif) +opt in \[(.+)\]:\s+#(.+)\n", open(sys.argv[0]).read()): - if not options: - options += "OPTIONS\n" - options += " %-16s %s\n" % (line[1].replace('"', ''), line[2]) -options = options.strip() - -# with ' < 1:' on the next line, this is a no-op: -if len(sys.argv) < 1: - print(__doc__ % locals()) - sys.exit(1) - -try: - opts, files = getopt.gnu_getopt(sys.argv[1:], "hC:pvV", ["help", "copyright=", "patch", "version", "verbose",]) -except Exception as e: - print( "%s: %s" % (program, e)) - sys.exit(1) - -# ---------------------------------------------------------------------- -# Handle options - -# set default values, if any -opt_verbose = 0 -opt_patch = False -opt_copyright = "Copyright The IETF Trust {years}, All Rights Reserved" - -# handle individual options -for opt, value in opts: - if opt in ["-h", "--help"]: # Output this help, then exit - print( __doc__ % locals() ) - sys.exit(1) - elif opt in ["-p", "--patch"]: # Generate patch output rather than error messages - opt_patch = True - elif opt in ["-C", "--copyright"]: # Copyright line pattern using {years} for years - opt_copyright = value - elif opt in ["-V", "--version"]: # Output version information, then exit - print( program, version ) - sys.exit(0) - elif opt in ["-v", "--verbose"]: # Be more verbose - opt_verbose += 1 - -# ---------------------------------------------------------------------- -def say(s): - sys.stderr.write("%s\n" % (s)) - -# ---------------------------------------------------------------------- -def note(s): - if opt_verbose: - sys.stderr.write("%s\n" % (s)) - -# ---------------------------------------------------------------------- -def die(s, error=1): - sys.stderr.write("\n%s: Error: %s\n\n" % (program, s)) - sys.exit(error) - -# ---------------------------------------------------------------------- - -def pipe(cmd, inp=None): - import shlex - from subprocess import Popen, PIPE - args = shlex.split(cmd) - bufsize = 4096 - stdin = PIPE if inp else None - pipe = Popen(args, stdin=stdin, stdout=PIPE, stderr=PIPE, bufsize=bufsize, encoding='utf-8', universal_newlines=True) - out, err = pipe.communicate(inp) - code = pipe.returncode - if code != 0: - raise OSError(err) - return out - -# ---------------------------------------------------------------------- -def split_loginfo(line): - try: - parts = line.split() - rev = parts[0][1:] - who = parts[2] - date = parts[4] - time = parts[5] - tz = parts[6] - when = tzparse.tzparse(" ".join(parts[4:7]), "%Y-%m-%d %H:%M:%S %Z") - when = when.astimezone(pytz.utc) - except ValueError as e: - sys.stderr.write("Bad log line format: %s\n %s\n" % (line, e)) - - return rev, who, when - -# ---------------------------------------------------------------------- -def get_first_commit(path): - note("Getting first commit for '%s'" % path) - cmd = 'svn log %s' % path - if opt_verbose > 1: - note("Running '%s' ..." % cmd) - try: - commit_log = pipe(cmd) - commit_log = commit_log.splitlines() - commit_log.reverse() - for line in commit_log: - if re.search(loginfo_format, line): - rev, who, when = split_loginfo(line) - break - else: - pass - except OSError: - rev, who, when = None, None, datetime.datetime.now(datetime.timezone.utc) - return { path: { 'rev': rev, 'who': who, 'date': when.strftime('%Y-%m-%d %H:%M:%S'), }, } - - -# ---------------------------------------------------------------------- -# The program itself - -import os -import json - -cwd = os.getcwd() - -# Get current initinfo from cache and svn -cachefn = os.path.join(os.environ.get('HOME', '.'), '.initinfo') - -if os.path.exists(cachefn): - note("Reading initinfo cache file %s" % cachefn) - with open(cachefn, "r") as file: - cache = json.load(file) -else: - sys.stderr.write("No initinfo cache file found -- will have to extract all information from SVN.\n"+ - "This may take some time.\n\n") - cache = {} -initinfo = cache - -merged_revs = {} -write_cache = False -loginfo_format = r'^r[0-9]+ \| [^@]+@[^@]+ \| \d\d\d\d-\d\d-\d\d ' - -year = time.strftime('%Y') -copyright_re = "(?i)"+opt_copyright.format(years=r"(\d+-)?\d+") -for path in files: - try: - if not os.path.exists(path): - note("File does not exist: %s" % path) - continue - note("Checking path %s" % path) - if not path in initinfo: - initinfo.update(get_first_commit(path)) - write_cache = True - date = initinfo[path]['date'] - init = date[:4] - - copyright_year_re = "(?i)"+opt_copyright.format(years=r"({init}-)?{year}".format(init=init, year=year)) - with open(path) as file: - try: - chunk = file.read(4000) - except UnicodeDecodeError as e: - sys.stderr.write(f'Error when reading {file.name}: {e}\n') - raise - if os.path.basename(path) == '__init__.py' and len(chunk)==0: - continue - if not re.search(copyright_year_re, chunk): - if year == init: - copyright = opt_copyright.format(years=year) - else: - copyright = opt_copyright.format(years=f"{init}-{year}") - if opt_patch: - print(f"--- {file.name}\t(original)") - print(f"+++ {file.name}\t(modified)") - if not re.search(copyright_re, chunk): - # Simple case, just insert copyright at the top - print( "@@ -1,3 +1,4 @@") - print(f"+# {copyright}") - for i, line in list(enumerate(chunk.splitlines()))[:3]: - print(f" {line}") - else: - # Find old copyright, then emit preceding lines, - # change, and following lines. - pos = None - for i, line in enumerate(chunk.splitlines(), start=1): - if re.search(copyright_re, line): - pos = i - break - if not pos: - raise RuntimeError("Unexpected state: Expected a copyright line, but found none") - print(f"@@ -1,{pos+3} +1,{pos+3} @@") - for i, line in list(enumerate(chunk.splitlines(), start=1))[:pos+3]: - if i == pos: - print(f"-{line}") - print(f"+# {copyright}") - else: - print(f" {line}") - else: - sys.stderr.write(f"{path}(1): Error: Missing or bad copyright. Expected: {copyright}") - except Exception: - if write_cache: - cache = initinfo - with open(cachefn, "w") as file: - json.dump(cache, file, indent=2, sort_keys=True) - raise - -if write_cache: - cache = initinfo - with open(cachefn, "w") as file: - json.dump(cache, file, indent=2, sort_keys=True) - diff --git a/bin/count.c b/bin/count.c deleted file mode 100644 index 786f15eb97..0000000000 --- a/bin/count.c +++ /dev/null @@ -1,26 +0,0 @@ -#include - -int main( void ) -{ - int c; - int count = 0; - - //turn off buffering - setvbuf(stdin, NULL, _IONBF, 0); - setvbuf(stdout, NULL, _IONBF, 0); - setvbuf(stderr, NULL, _IONBF, 0); - - c = fgetc(stdin); - while(c != EOF) - { - if (c=='.' || c=='E' || c=='F' || c=='s') count++; else count=0; - fputc(c, stdout); - fflush(stdout); - if (count && count % 76 == 0) { - fprintf(stderr, "%4d\n", count); - fflush(stderr); - } - c = fgetc(stdin); - } - return 0; -} diff --git a/bin/daily b/bin/daily deleted file mode 100755 index 40cf3fd2be..0000000000 --- a/bin/daily +++ /dev/null @@ -1,68 +0,0 @@ -#!/bin/bash - -# Nightly datatracker jobs. -# -# This script is expected to be triggered by cron from -# /etc/cron.d/datatracker -export LANG=en_US.UTF-8 -export PYTHONIOENCODING=utf-8 - -# Make sure we stop if something goes wrong: -program=${0##*/} -trap 'echo "$program($LINENO): Command failed with error code $? ([$$] $0 $*)"; exit 1' ERR - -# Datatracker directory -DTDIR=/a/www/ietf-datatracker/web -cd $DTDIR/ - -logger -p user.info -t cron "Running $DTDIR/bin/daily" - -# Run the hourly jobs first -$DTDIR/bin/hourly - -# Set up the virtual environment -source $DTDIR/env/bin/activate - - -# Update our information about the current version of some commands we use -$DTDIR/ietf/manage.py update_external_command_info - -# Get IANA-registered yang models -#YANG_IANA_DIR=$(python -c 'import ietf.settings; print ietf.settings.SUBMIT_YANG_IANA_MODEL_DIR') -# Hardcode the rsync target to avoid any unwanted deletes: -# rsync -avzq --delete rsync.ietf.org::iana/yang-parameters/ /a/www/ietf-ftp/yang/ianamod/ -rsync -avzq --delete /a/www/ietf-ftp/iana/yang-parameters/ /a/www/ietf-ftp/yang/ianamod/ - -# Get Yang models from Yangcatalog. -rsync -avzq rsync://yangcatalog.org:10873/yangdeps /a/www/ietf-ftp/yang/catalogmod/ - -# Populate the yang repositories -$DTDIR/ietf/manage.py populate_yang_model_dirs -v0 - -# Re-run yang checks on active documents -$DTDIR/ietf/manage.py run_yang_model_checks -v0 - -# Expire internet drafts -# Enable when removed from /a/www/ietf-datatracker/scripts/Cron-runner: -$DTDIR/ietf/bin/expire-ids - -# Send nomcom reminders about nomination acceptance and questionnaires -$DTDIR/ietf/manage.py send_reminders - -# Expire last calls -# Enable when removed from /a/www/ietf-datatracker/scripts/Cron-runner: -$DTDIR/ietf/bin/expire-last-calls - -# Run an extended version of the rfc editor update, to catch changes -# with backdated timestamps -# Enable when removed from /a/www/ietf-datatracker/scripts/Cron-runner: -$DTDIR/ietf/bin/rfc-editor-index-updates -d 1969-01-01 - -# Fetch meeting attendance data from ietf.org/registration/attendees -$DTDIR/ietf/manage.py fetch_meeting_attendance --latest 2 - -# Send reminders originating from the review app -$DTDIR/ietf/bin/send-review-reminders - -# Purge older PersonApiKeyEvents -$DTDIR/ietf/manage.py purge_old_personal_api_key_events 14 diff --git a/bin/drop-new-tables b/bin/drop-new-tables deleted file mode 100755 index ec1594ae26..0000000000 --- a/bin/drop-new-tables +++ /dev/null @@ -1,13 +0,0 @@ -#!/bin/bash - -# Drop tables which don't exist in the database dump. - -[ -n "$1" ] || { echo -e "\nUsage: $0 DUMPFILE\n\nError: No database dump file given"; exit 1; } - -zcat $1 | head | grep "Database: ietf_utf8" || { echo "Is this a database dump? Expected to see 'Database: ietf_utf8' "; exit 1; } - -echo -e "\nSQL commands:\n" - -diff <(zcat $1 | grep '^DROP TABLE IF EXISTS' | tr -d '`;' | field 5) <(ietf/manage.py dbshell <<< 'show tables;' | tail -n +2) | grep '^>' | awk '{print "drop table if exists", $2, ";";}' | tee /dev/stderr | ietf/manage.py dbshell - -echo -e "\nDone" diff --git a/bin/dump-to-names-json b/bin/dump-to-names-json deleted file mode 100644 index 9c7dfac07d..0000000000 --- a/bin/dump-to-names-json +++ /dev/null @@ -1,16 +0,0 @@ -#!/bin/bash - -# This script provides a limited selected dump of database content with the -# purpose of generating a test fixture that provides the test data needed -# by the test suite. -# -# The generated data fixture is sorted and normalized in order to produce -# minimal commit diffs which reflect only actual changes in the fixture data, -# without apparent changes resulting only from ordering changes. - -set -x -ietf/manage.py dumpdata --indent 1 doc.State doc.BallotType doc.StateType \ - mailtrigger.MailTrigger mailtrigger.Recipient name utils.VersionInfo \ - group.GroupFeatures stats.CountryAlias dbtemplate.DBTemplate \ - | jq --sort-keys "sort_by(.model, .pk)" \ - | jq '[.[] | select(.model!="dbtemplate.dbtemplate" or .pk==354)]' > ietf/name/fixtures/names.json diff --git a/bin/every15m b/bin/every15m deleted file mode 100755 index c0b0752f6a..0000000000 --- a/bin/every15m +++ /dev/null @@ -1,28 +0,0 @@ -#!/bin/bash - -# datatracker jobs to run every 15 minutes -# -# This script is expected to be triggered by cron from -# /etc/cron.d/datatracker - -export LANG=en_US.UTF-8 -export PYTHONIOENCODING=utf-8 - -DTDIR=/a/www/ietf-datatracker/web -cd $DTDIR/ - -# Set up the virtual environment -source $DTDIR/env/bin/activate - -logger -p user.info -t cron "Running $DTDIR/bin/every15m" - -# Send mail scheduled to go out at certain times -$DTDIR/ietf/bin/send-scheduled-mail all - -# Reparse the last _year_ of RFC index entries -# (which is the default if -d is not provided) -# until https://github.com/ietf-tools/datatracker/issues/3734 -# is addressed. -# This takes about 20s on production as of 2022-08-11 -$DTDIR/ietf/bin/rfc-editor-index-updates - diff --git a/bin/hourly b/bin/hourly deleted file mode 100755 index 77310302ce..0000000000 --- a/bin/hourly +++ /dev/null @@ -1,101 +0,0 @@ -#!/bin/bash - -# Hourly datatracker jobs -# -# This script is expected to be triggered by cron from -# /etc/cron.d/datatracker -export LANG=en_US.UTF-8 -export PYTHONIOENCODING=utf-8 - -# Make sure we stop if something goes wrong: -program=${0##*/} -trap 'echo "$program($LINENO): Command failed with error code $? ([$$] $0 $*)"; exit 1' ERR - -DTDIR=/a/www/ietf-datatracker/web -cd $DTDIR/ - -# Set up the virtual environment -source $DTDIR/env/bin/activate - -logger -p user.info -t cron "Running $DTDIR/bin/hourly" - -# *** Enable when removed from /a/www/ietf-datatracker/scripts/Cron-runner: *** - -# # Update community lists. Remove once the community rewrite (will be around 6.20.0 ) -# $DTDIR/ietf/manage.py update_community_lists -# -# # Polling backup for iana and rfc-editory post APIs -$DTDIR/ietf/bin/iana-changes-updates -$DTDIR/ietf/bin/iana-protocols-updates -# $DTDIR/ietf/bin/rfc-editor-index-updates -# $DTDIR/ietf/bin/rfc-editor-queue-updates -# -# # Generate alias and virtual files for draft email aliases -# $DTDIR/ietf/bin/generate-draft-aliases && \ -# ( cd /a/postfix; /usr/sbin/postalias -o draft-aliases; ) && \ -# ( cd /a/postfix; /usr/sbin/postmap -o draft-virtual; ) -# -# # Generate alias and virtual files for group email aliases -# $DTDIR/ietf/bin/generate-wg-aliases && \ -# ( cd /a/postfix; /usr/sbin/postalias -o group-aliases; ) && \ -# ( cd /a/postfix; /usr/sbin/postmap -o group-virtual; ) -# -# Generate some static files -ID=/a/ietfdata/doc/draft/repository -DERIVED=/a/ietfdata/derived -DOWNLOAD=/a/www/www6s/download - -export TMPDIR=/a/tmp - -TMPFILE1=`mktemp` || exit 1 -TMPFILE2=`mktemp` || exit 1 -TMPFILE3=`mktemp` || exit 1 -TMPFILE4=`mktemp` || exit 1 -TMPFILE5=`mktemp` || exit 1 -TMPFILE6=`mktemp` || exit 1 -TMPFILE7=`mktemp` || exit 1 -TMPFILE8=`mktemp` || exit 1 -TMPFILE9=`mktemp` || exit 1 -TMPFILEA=`mktemp` || exit 1 -TMPFILEB=`mktemp` || exit 1 - -chmod a+r $TMPFILE1 $TMPFILE2 $TMPFILE3 $TMPFILE4 $TMPFILE5 $TMPFILE6 $TMPFILE7 $TMPFILE8 $TMPFILE9 $TMPFILEA $TMPFILEB - -python -m ietf.idindex.generate_all_id_txt >> $TMPFILE1 -python -m ietf.idindex.generate_id_index_txt >> $TMPFILE2 -python -m ietf.idindex.generate_id_abstracts_txt >> $TMPFILE3 -cp $TMPFILE1 $TMPFILE4 -cp $TMPFILE2 $TMPFILE5 -cp $TMPFILE3 $TMPFILE6 -cp $TMPFILE1 $TMPFILE8 -cp $TMPFILE2 $TMPFILE9 -cp $TMPFILE3 $TMPFILEA -python -m ietf.idindex.generate_all_id2_txt >> $TMPFILE7 -cp $TMPFILE7 $TMPFILEB - -mv $TMPFILE1 $ID/all_id.txt -mv $TMPFILE2 $ID/1id-index.txt -mv $TMPFILE3 $ID/1id-abstracts.txt -mv $TMPFILE4 $DOWNLOAD/id-all.txt -mv $TMPFILE5 $DOWNLOAD/id-index.txt -mv $TMPFILE6 $DOWNLOAD/id-abstract.txt -mv $TMPFILE7 $ID/all_id2.txt -mv $TMPFILE8 $DERIVED/all_id.txt -mv $TMPFILE9 $DERIVED/1id-index.txt -mv $TMPFILEA $DERIVED/1id-abstracts.txt -mv $TMPFILEB $DERIVED/all_id2.txt - -$DTDIR/ietf/manage.py generate_idnits2_rfc_status -$DTDIR/ietf/manage.py generate_idnits2_rfcs_obsoleted - -CHARTER=/a/www/ietf-ftp/charter -wget -q https://datatracker.ietf.org/wg/1wg-charters-by-acronym.txt -O $CHARTER/1wg-charters-by-acronym.txt -wget -q https://datatracker.ietf.org/wg/1wg-charters.txt -O $CHARTER/1wg-charters.txt - -# Regenerate the last week of bibxml-ids -$DTDIR/ietf/manage.py generate_draft_bibxml_files - -# Create and update group wikis -#$DTDIR/ietf/manage.py create_group_wikis - -# exit 0 diff --git a/bin/mkdiagram b/bin/mkdiagram deleted file mode 100755 index 4f015c0abe..0000000000 --- a/bin/mkdiagram +++ /dev/null @@ -1,50 +0,0 @@ -#!/bin/bash - -# assume we're in bin/, sibling to ietf/ - -cd ${0%/*}/../ietf || { echo "CD to ietf directory failed, bailing out"; exit; } - -trap 'echo "$program($LINENO): Command failed with error code $? ($0 $*)"; exit 1' ERR - -if [ "$*" ]; then apps="$@"; graph="${1%.*}"; else apps=$(ls */models.py | sed 's!/models.py!!'); graph="models"; fi - -newapps="doc group meeting message person name" -legacyapps="announcements idindex idrfc idtracker iesg ietfauth ipr liaisons mailinglists proceedings redirects submit wgcharter wginfo" - -proxy="$(grep ^class */proxy.py | tr '()' ' ' | awk '{printf $2 ","}')" -names="$(grep ^class name/models.py | tr '()' ' ' | awk '{printf $2 ","}')" -legacy="$(for app in $legacyapps; do grep ^class $app/models.py | tr '()' ' '; done | grep -v ' Meeting\\(' | awk '{printf $2 ","}')" -events="$(egrep '^class .+DocEvent' doc/models.py | tr '()' ' ' | awk '{printf $2 ","}')" - -echo -e "proxy: $proxy\n" -echo -e "names: $names\n" -echo -e "legacy:$legacy\n" -echo -e "events:$events\n" - -exclude="--exclude=$proxy,$names,$legacy" - -export PYTHONPATH="$PWD/.." - -echo "Validating..." -./manage.py validate - -export PYTHONPATH=`dirname $PWD` -module=${PWD##*/} -export DJANGO_SETTINGS_MODULE=$module.settings -export graph -export title - -echo "Generate model graph" -graph="models-with-names-and-events" -title="New IETF Database schema" -${0%/*}/../ietf/manage.py graph_models --exclude="$proxy,$legacy" --title "$title" $apps > $graph.dot && dot -Tpng $graph.dot > $graph.png - -echo "Generate new model without names" -graph="models-with-names" -title="New IETF Database schema, without name tables" -modelviz.py --exclude="$proxy,$legacy,$names" --title "$title" $apps > $graph.dot && dot -Tpng $graph.dot > $graph.png - -echo "Generate new model without names and subevents" -graph="models" -title="New IETF Database schema, without name tables and subevents" -modelviz.py --exclude="$proxy,$legacy,$names,$events" --title "$title" $apps > $graph.dot && dot -Tpng $graph.dot > $graph.png diff --git a/bin/mm_hourly b/bin/mm_hourly deleted file mode 100755 index 0d1da2e572..0000000000 --- a/bin/mm_hourly +++ /dev/null @@ -1,24 +0,0 @@ -#!/bin/bash - -# Hourly datatracker jobs, ***run as mailman*** -# -# This script is expected to be triggered by cron from -# $DTDIR/etc/cron.d/datatracker which should be symlinked from -# /etc/cron.d/ - -export LANG=en_US.UTF-8 -export PYTHONIOENCODING=utf-8 - -# Make sure we stop if something goes wrong: -program=${0##*/} -trap 'echo "$program($LINENO): Command failed with error code $? ([$$] $0 $*)"; exit 1' ERR - -DTDIR=/a/www/ietf-datatracker/web -cd $DTDIR/ - -# Set up the virtual environment -source $DTDIR/env/bin/activate - -logger -p user.info -t cron "Running $DTDIR/bin/mm_hourly" - -$DTDIR/ietf/manage.py import_mailman_listinfo diff --git a/bin/monthly b/bin/monthly deleted file mode 100755 index 1d36abc210..0000000000 --- a/bin/monthly +++ /dev/null @@ -1,17 +0,0 @@ -#!/bin/bash - -# Weekly datatracker jobs. -# -# This script is expected to be triggered by cron from -# /etc/cron.d/datatracker -export LANG=en_US.UTF-8 -export PYTHONIOENCODING=utf-8 - -DTDIR=/a/www/ietf-datatracker/web -cd $DTDIR/ - -# Set up the virtual environment -source $DTDIR/env/bin/activate - -logger -p user.info -t cron "Running $DTDIR/bin/monthly" - diff --git a/bin/release-coverage b/bin/release-coverage deleted file mode 100755 index 22177c17a6..0000000000 --- a/bin/release-coverage +++ /dev/null @@ -1,2 +0,0 @@ -#!/bin/bash -zcat release-coverage.json.gz | jq 'to_entries[] | [.value.time, .key, .value.code.coverage, .value.template.coverage, .value.url.coverage] ' 2>/dev/null | tr "\n][" " \n" | tr -d ' "Z' | tr ",T" " " | sort -n | cut -c 2- | sed -n '/2015-03-10/,$p' diff --git a/bin/setupenv b/bin/setupenv deleted file mode 100755 index b9f0f72da0..0000000000 --- a/bin/setupenv +++ /dev/null @@ -1,45 +0,0 @@ -#!/usr/bin/env python - - -import os -import subprocess -import requests -import sys -import stat -import shutil - -basedir = os.path.dirname(os.path.dirname(__file__)) - -sys.path.append(basedir) - -shutil.copyfile(os.path.join(basedir, 'docker/settings_local.py'), os.path.join(basedir, 'settings_local.py')) - -from ietf.settings_sqlitetest import * # we don't import from django.conf here, on purpose - -for dir in [ AGENDA_PATH, IDSUBMIT_REPOSITORY_PATH, IDSUBMIT_STAGING_PATH, - INTERNET_DRAFT_ARCHIVE_DIR, os.path.dirname(DRAFT_ALIASES_PATH), PHOTOS_DIR, - os.path.dirname(os.path.abspath(TEST_GHOSTDRIVER_LOG_PATH)), ]: - if not os.path.exists(dir): - print("Creating %s" % dir) - os.makedirs(dir) - -for path in [ DRAFT_ALIASES_PATH, DRAFT_VIRTUAL_PATH, GROUP_ALIASES_PATH, GROUP_VIRTUAL_PATH, ]: - if not os.path.exists(path): - print("Setting up %s" % path) - dir, fn = os.path.split(path) - url = "https://zinfandel.tools.ietf.org/src/db/tmp/%s" % fn - r = requests.get(url) - if r.status_code == 200: - with open(path, "w") as of: - of.write(r.text) - else: - print("Error %s fetching '%s'" % (r.status_code, url)) - -path = IDSUBMIT_IDNITS_BINARY -if not os.path.exists(path): - print("Setting up %s" % path) - r = requests.get('https://tools.ietf.org/tools/idnits/idnits') - with open(path, 'w') as idnits: - idnits.write(r.text) - os.chmod(path, 0755) - \ No newline at end of file diff --git a/bin/update b/bin/update deleted file mode 100755 index bcb6e8b129..0000000000 --- a/bin/update +++ /dev/null @@ -1,229 +0,0 @@ -#!/bin/bash - -version="0.34" -program=$(basename $0) - -NEW="" # If there are more than $NEW % new lines, skip update -OLD="" # If there are more than $OLD % deleted lines, skip update -FILE="" -verbose="" -silent="" - -# ---------------------------------------------------------------------- -function usage() { -cat < -EOF -exit -} - - -# ---------------------------------------------------------------------- -function note() { - if [ -n "$verbose" ]; then - echo -e "$program: $*" - fi -} - -# ---------------------------------------------------------------------- -function warn() { - [ "$QUIET" ] || echo -e "$program: $*" -} - -# ---------------------------------------------------------------------- -function err() { - echo -e "$program: $*" > /dev/stderr -} - -# ----------------------------------------------------------------------------- -function leave() { - errcode=$1; shift - if [ "$errcode" -ge "2" ]; then warn "$*"; else note "$*"; fi - if [ -f "$tempfile" ]; then rm $tempfile; fi - if [ -f "$difffile" ]; then rm $difffile; fi - if [ "$errcode" = "1" -a "$RESULT" = "0" ]; then exit 0; else exit $errcode; fi -} - -# ---------------------------------------------------------------------- -# Set up error trap -trap 'leave 127 "$program($LINENO): Command failed with error code $? while processing '$origfile'."' ERR - -# exit with a message if a command fails -set -e - -# ---------------------------------------------------------------------- -# Get any options -# - -# Default values -PAT="\$path\$base.%Y-%m-%d_%H%M" -RESULT="0" -QUIET="" - -# Based on the sample code in /usr/share/doc/util-linux/examples/parse.bash.gz -if [ "$(uname)" = "Linux" ]; then - GETOPT_RESULT=$(getopt -o bc:ef:hn:o:p:qrvV --long backup,maxchg:,empty,file:,help,maxnew:,maxold:,prefix:,report,quiet,verbose,version -n "$program" -- "$@") -else - GETOPT_RESULT=$(getopt bc:ef:hn:o:p:qrvV "$@") -fi - -if [ $? != 0 ] ; then echo "Terminating..." >&2 ; exit 1 ; fi - -note "GETOPT_RESULT: $GETOPT_RESULT" -eval set -- "$GETOPT_RESULT" - -while true ; do - case "$1" in - -b|--backup) backup=1; shift ;; # Back up earlier versions by creating a backup file - -c|--maxchg) CHG="$2"; shift 2 ;; # Limit on percentage of changed lines - -e|--empty) empty=1; shift ;; # Permit the update to be empty (default: discard) - -f|--file) FILE="$2"; shift 2 ;; # Read input from FILE instead of standard input - -h|--help) usage; shift ;; # Show this text and exit - -n|--maxnew) NEW="$2"; shift 2 ;; # Limit on percentage of new (added) lines - -o|--maxold) OLD="$2"; shift 2 ;; # Limit on percentage of old (deleted) lines - -p|--pat*) PAT="$2"; shift 2 ;; # Backup name base ('$path$base.%Y%m%d_%H%M') - -q|--quiet) QUIET=1; shift;; # Be less verbose - -r|--result) RESULT=1; shift ;; # Return 1 if update not done - -v|--verbose) verbose=1; shift ;; # Be more verbose about what's happening - -V|--version) echo -e "$program\t$version"; exit;; # Show version and exit - --) shift ; break ;; - *) echo "$program: Internal error, inconsistent option specification." ; exit 1 ;; - esac -done - -if [ $CHG ]; then OLD=$CHG; NEW=$CHG; fi - -if [ $# -lt 1 ]; then echo -e "$program: Missing output filename\n"; usage; fi - -origfile=$1 -tempfile=$(mktemp) -difffile=$(mktemp) - -if [ -e "$origfile" ]; then - cp -p $origfile $tempfile # For ownership and permissions - cat $FILE > $tempfile - [ "$FILE" ] && touch -r $FILE $tempfile - # This won't work if we don't have sufficient privileges: - #chown --reference=$origfile $tempfile - #chmod --reference=$origfile $tempfile -else - cat $FILE > $origfile - [ "$FILE" ] && touch -r $FILE $tempfile - leave 0 "Created file '$origfile'" -fi - -origlen=$(wc -c < $origfile) -newlen=$(wc -c < $tempfile) - -if [ $origlen = 0 -a $newlen = 0 ]; then - rm $tempfile - leave 1 "New content is identical (and void) - not updating '$origfile'." -fi -if [ $newlen = 0 -a -z "$empty" ]; then - leave 1 "New content is void - not updating '$origfile'." -fi - -diff $origfile $tempfile > $difffile || [ $? -le 1 ] && true # suppress the '1' error code on differences -difflen=$(wc -l < $difffile) -if [ $difflen = 0 ]; then - leave 1 "New content is identical - not updating '$origfile'." -fi - -if [ "$OLD" -o "$NEW" ]; then - - if [ "$NEW" ]; then maxnew=$(( $origlen * $NEW / 100 )); fi - if [ "$OLD" ]; then maxdel=$(( $origlen * $OLD / 100 )); fi - - newcount=$(grep "^> " $difffile | wc -c) - outcount=$(grep "^< " $difffile | wc -c) - delcount=$(grep "^! " $difffile | wc -c) - delcount=$(( $outcount + $delcount )) - rm $difffile - - if [ "$OLD" ]; then - if [ "$delcount" -ge "$maxdel" ]; then - cp $tempfile $origfile.update - leave 2 "New content has too many removed lines ($delcount/$origlen)\n - not updating '$origfile'.\nNew content placed in '$origfile.update' instead" - fi - fi - if [ "$NEW" ]; then - if [ "$newcount" -ge "$maxnew" ]; then - cp $tempfile $origfile.update - leave 2 "New content has too many added lines ($newcount/$origlen)\n - not updating '$origfile'.\nNew content placed in '$origfile.update' instead" - fi - fi -fi - -if [ "$backup" ]; then - - path=${origfile%/*} - name=${origfile##*/} - base=${name%.*} - ext=${origfile##*.} - - if [ "$ext" = "$origfile" ]; then - ext="" - elif [ ! "${ext%/*}" = "$ext" ]; then - ext="" - else - ext=".$ext" - fi - - if [ "$path" = "$origfile" ]; then - path="" - else - path="$path/" - fi - - ver=1 - backfile=$(eval date +"$PAT") - backpath="${backfile%/*}" - if [ "$backpath" = "$backfile" ]; then - backpath="." - fi - if [ ! -d $backpath ]; then - if [ -e $backpath ]; then - leave 3 "The backup path '$backpath' exists but isn't a directory" - else - mkdir -p $backpath - fi - fi - while [ -e "$backfile,$ver$ext" ]; do - ver=$(( $ver+1 )) - done - note "Saving backup: $backfile,$ver$ext" - cp -p "$origfile" "$backfile,$ver$ext" - chmod -w "$backfile,$ver$ext" || true -fi - -if ! mv $tempfile $origfile; then cp -p $tempfile $origfile; fi -leave 0 "Updated file '$origfile'" diff --git a/bin/vnu.jar b/bin/vnu.jar index 776d5836ba..1766224071 100644 Binary files a/bin/vnu.jar and b/bin/vnu.jar differ diff --git a/bin/weekly b/bin/weekly deleted file mode 100755 index cca8403fd4..0000000000 --- a/bin/weekly +++ /dev/null @@ -1,25 +0,0 @@ -#!/bin/bash - -# Weekly datatracker jobs. -# -# This script is expected to be triggered by cron from -# /etc/cron.d/datatracker -export LANG=en_US.UTF-8 -export PYTHONIOENCODING=utf-8 - -DTDIR=/a/www/ietf-datatracker/web -cd $DTDIR/ - -# Set up the virtual environment -source $DTDIR/env/bin/activate - -logger -p user.info -t cron "Running $DTDIR/bin/weekly" - - -# Send out weekly summaries of apikey usage - -$DTDIR/ietf/manage.py send_apikey_usage_emails - -# Send notifications about coming expirations -$DTDIR/ietf/bin/notify-expirations - diff --git a/client/App.vue b/client/App.vue index 0664525ca4..7750674296 100644 --- a/client/App.vue +++ b/client/App.vue @@ -26,6 +26,28 @@ const siteStore = useSiteStore() const appContainer = ref(null) +// -------------------------------------------------------------------- +// Set user theme +// -------------------------------------------------------------------- + +function updateTheme() { + const desiredTheme = window.localStorage?.getItem('theme') + if (desiredTheme === 'dark') { + siteStore.theme = 'dark' + } else if (desiredTheme === 'light') { + siteStore.theme = 'light' + } else if (window.matchMedia("(prefers-color-scheme: dark)").matches) { + siteStore.theme = 'dark' + } else { + siteStore.theme = 'light' + } +} + +updateTheme() + +// this change event fires for either light or dark changes +window.matchMedia('(prefers-color-scheme: dark)').addEventListener('change', updateTheme) + // -------------------------------------------------------------------- // Handle browser resize // -------------------------------------------------------------------- diff --git a/client/Embedded.vue b/client/Embedded.vue index a0f0d2831e..80b105dc15 100644 --- a/client/Embedded.vue +++ b/client/Embedded.vue @@ -1,12 +1,13 @@ diff --git a/client/agenda/AgendaScheduleCalendar.vue b/client/agenda/AgendaScheduleCalendar.vue index 0dbea168ab..9863296341 100644 --- a/client/agenda/AgendaScheduleCalendar.vue +++ b/client/agenda/AgendaScheduleCalendar.vue @@ -11,14 +11,17 @@ n-drawer(v-model:show='isShown', placement='bottom', :height='state.drawerHeight n-button( :type='agendaStore.isTimezoneMeeting ? `primary` : `default`' @click='setTimezone(`meeting`)' + :text-color='agendaStore.isTimezoneMeeting ? `#FFF` : null' ) Meeting n-button( :type='agendaStore.isTimezoneLocal ? `primary` : `default`' @click='setTimezone(`local`)' + :text-color='agendaStore.isTimezoneLocal ? `#FFF` : null' ) Local n-button( :type='agendaStore.timezone === `UTC` ? `primary` : `default`' @click='setTimezone(`UTC`)' + :text-color='agendaStore.timezone === `UTC` ? `#FFF` : null' ) UTC n-divider(vertical) n-button.me-2( @@ -32,7 +35,7 @@ n-drawer(v-model:show='isShown', placement='bottom', :height='state.drawerHeight n-badge.ms-2(:value='agendaStore.selectedCatSubs.length', processing) n-button( ghost - color='gray' + :color='siteStore.theme === `dark` ? `#e35d6a` : `gray`' strong @click='close' ) @@ -81,11 +84,10 @@ import { NPopover } from 'naive-ui' -import '@fullcalendar/core/vdom' // solves problem with Vite import FullCalendar from '@fullcalendar/vue3' import timeGridPlugin from '@fullcalendar/timegrid' import interactionPlugin from '@fullcalendar/interaction' -import luxonPlugin from '@fullcalendar/luxon2' +import luxonPlugin from '@fullcalendar/luxon3' import bootstrap5Plugin from '@fullcalendar/bootstrap5' import AgendaDetailsModal from './AgendaDetailsModal.vue' @@ -185,6 +187,7 @@ function refreshData () { let earliestDate = DateTime.fromISO('2200-01-01') let latestDate = DateTime.fromISO('1990-01-01') let nowDate = DateTime.now() + let hasCrossDayEvents = false calendarOptions.events = agendaStore.scheduleAdjusted.map(ev => { // -> Determine boundaries @@ -200,6 +203,9 @@ function refreshData () { if (ev.adjustedEnd < latestDate) { latestDate = ev.adjustedEnd } + if (ev.adjustedStart.day !== ev.adjustedEnd.day) { + hasCrossDayEvents = true + } // -> Build event object return { id: ev.id, @@ -212,8 +218,8 @@ function refreshData () { }) // -> Display settings - calendarOptions.slotMinTime = `${earliestHour.toString().padStart(2, '0')}:00:00` - calendarOptions.slotMaxTime = `${latestHour.toString().padStart(2, '0')}:00:00` + calendarOptions.slotMinTime = hasCrossDayEvents ? '00:00:00' : `${earliestHour.toString().padStart(2, '0')}:00:00` + calendarOptions.slotMaxTime = hasCrossDayEvents ? '23:59:59' : `${latestHour.toString().padStart(2, '0')}:00:00` calendarOptions.validRange.start = earliestDate.minus({ days: 1 }).toISODate() calendarOptions.validRange.end = latestDate.plus({ days: 1 }).toISODate() // calendarOptions.scrollTime = `${earliestHour.toString().padStart(2, '0')}:00:00` @@ -324,7 +330,6 @@ function close () { } .badge { - width: 30px; font-size: .7em; border: 1px solid #CCC; text-transform: uppercase; diff --git a/client/agenda/AgendaScheduleList.vue b/client/agenda/AgendaScheduleList.vue index 87982ee626..bbe5dfee8b 100644 --- a/client/agenda/AgendaScheduleList.vue +++ b/client/agenda/AgendaScheduleList.vue @@ -15,6 +15,7 @@ td(:colspan='pickerModeActive ? 6 : 5') i.bi.bi-exclamation-triangle.me-2 span(v-if='agendaStore.searchVisible && agendaStore.searchText') No event matching your search query. + span(v-else-if='agendaStore.meeting.prelimAgendaDate') A preliminary agenda is expected to be released on {{ agendaStore.meeting.prelimAgendaDate }} span(v-else) Nothing to display tr( v-for='item of meetingEvents' @@ -24,7 +25,7 @@ ) //- ROW - DAY HEADING ----------------------- template(v-if='item.displayType === `day`') - td(:id='`agenda-day-` + item.id', :colspan='pickerModeActive ? 6 : 5') {{item.date}} + td(:id='item.slug', :colspan='pickerModeActive ? 6 : 5') {{item.date}} //- ROW - SESSION HEADING ------------------- template(v-else-if='item.displayType === `session-head`') td.agenda-table-cell-check(v-if='pickerModeActive')   @@ -83,6 +84,14 @@ template(#trigger) span.badge.is-bof BoF span #[a(href='https://www.ietf.org/how/bofs/', target='_blank') Birds of a Feather] sessions (BoFs) are initial discussions about a particular topic of interest to the IETF community. + n-popover( + v-if='item.isProposed' + trigger='hover' + :width='250' + ) + template(#trigger) + span.badge.is-proposed Proposed + span #[a(href='https://www.ietf.org/process/wgs/', target='_blank') Proposed WGs] are groups in the process of being chartered. If the charter is not approved by the IESG before the IETF meeting, the session may be canceled. .agenda-table-note(v-if='item.note') i.bi.bi-arrow-return-right.me-1 span {{item.note}} @@ -112,20 +121,12 @@ :options='item.links' key-field='id' :render-icon='renderLinkIcon' - :render-label='renderLinkLabel' + :render-label='renderLink' ) n-button(size='tiny') i.bi.bi-three-dots .agenda-table-cell-links-buttons(v-else-if='item.links && item.links.length > 0') - template(v-if='item.flags.agenda') - n-popover - template(#trigger) - i.bi.bi-collection( - :id='`btn-lnk-` + item.key + `-mat`' - @click='showMaterials(item.key)' - ) - span Show meeting materials - template(v-else-if='item.type === `regular`') + template(v-if='!item.flags.agenda && item.type === `regular`') n-popover template(#trigger) i.no-meeting-materials @@ -134,7 +135,16 @@ span No meeting materials yet. n-popover(v-for='lnk of item.links', :key='lnk.id') template(#trigger) + button( + v-if="lnk.click" + type="button" + :id='`btn-` + lnk.id' + @click='lnk.click' + :aria-label='lnk.label' + :class='`border-0 bg-transparent text-` + lnk.color' + ): i.bi(:class='`bi-` + lnk.icon') a( + v-else :id='`btn-` + lnk.id' :href='lnk.href' :aria-label='lnk.label' @@ -200,7 +210,7 @@ import { import AgendaDetailsModal from './AgendaDetailsModal.vue' -import { useAgendaStore } from './store' +import { useAgendaStore, daySlugPrefix, daySlug } from './store' import { useSiteStore } from '../shared/store' import { getUrl } from '../shared/urls' @@ -245,33 +255,55 @@ const meetingEvents = computed(() => { // -> Add date row const itemDate = DateTime.fromISO(item.adjustedStartDate) + let willRenderDateRow = false if (itemDate.toISODate() !== acc.lastDate) { acc.result.push({ id: item.id, + slug: daySlug(item), key: `day-${itemDate.toISODate()}`, displayType: 'day', date: itemDate.toLocaleString(DateTime.DATE_HUGE), cssClasses: 'agenda-table-display-day' }) + willRenderDateRow = true } acc.lastDate = itemDate.toISODate() // -> Add session header row - if (item.type === 'regular' && acc.lastTypeName !== `${item.type}-${item.name}`) { + const typeName = `${item.type}-${item.slotName}` + if (item.type === 'regular' && (acc.lastTypeName !== typeName || willRenderDateRow)) { acc.result.push({ key: `sesshd-${item.id}`, displayType: 'session-head', timeslot: itemTimeSlot, - name: `${item.adjustedStart.toFormat('cccc')} ${item.name}`, + name: `${item.adjustedStart.setZone(agendaStore.meeting.timezone).toFormat('cccc')} ${item.slotName}`, cssClasses: 'agenda-table-display-session-head' + (isLive ? ' agenda-table-live' : '') }) } - acc.lastTypeName = `${item.type}-${item.name}` - - // -> Populate event links + acc.lastTypeName = typeName + + // + /** + * -> Populate event menu items + * + * links is an array of either, + * 1. { href: "...", click: undefined, ...sharedProps } + * 2. { click: () => {...}, href: undefined, ...sharedProps } + */ const links = [] - if (item.flags.showAgenda || ['regular', 'plenary'].includes(item.type)) { + const typesWithLinks = ['regular', 'plenary', 'other'] + const purposesWithoutLinks = ['admin', 'closed_meeting', 'officehours', 'social'] + if (item.flags.showAgenda || (typesWithLinks.includes(item.type) && !purposesWithoutLinks.includes(item.purpose))) { if (item.flags.agenda) { + // -> Meeting Materials + links.push({ + id: `btn-${item.id}-mat`, + label: 'Show meeting materials', + icon: 'collection', + href: undefined, + click: () => showMaterials(item.id), + color: 'darkgray' + }) links.push({ id: `lnk-${item.id}-tar`, label: 'Download meeting materials as .tar archive', @@ -293,7 +325,18 @@ const meetingEvents = computed(() => { color: 'red' }) } - if (agendaStore.useNotes) { + // -> Point to Wiki for Hackathon sessions, HedgeDocs otherwise + if (item.groupAcronym === 'hackathon') { + links.push({ + id: `lnk-${item.id}-wiki`, + label: 'Wiki', + icon: 'book', + href: getUrl('hackathonWiki', { + meetingNumber: agendaStore.meeting.number + }), + color: 'blue' + }) + } else if (agendaStore.usesNotes) { links.push({ id: `lnk-${item.id}-note`, label: 'Notepad for note-takers', @@ -355,16 +398,6 @@ const meetingEvents = computed(() => { color: 'teal' }) } - // -> Calendar item - if (item.links.calendar) { - links.push({ - id: `lnk-${item.id}-calendar`, - label: isMobile.value ? `Calendar (.ics) entry for this session` : `Calendar (.ics) entry for ${item.acronym} session on ${item.adjustedStart.toFormat('fff')}`, - icon: 'calendar-check', - href: item.links.calendar, - color: 'pink' - }) - } } else { // -> Post event if (meetingNumberInt >= 60) { @@ -421,9 +454,36 @@ const meetingEvents = computed(() => { color: 'purple' }) } + // -> Keep showing video client / on-site tool for Plenary until end of day, in case it goes over the planned time range + if (item.type === 'plenary' && item.adjustedEnd.day === current.day) { + links.push({ + id: `lnk-${item.id}-video`, + label: 'Full Client with Video', + icon: 'camera-video', + href: item.links.videoStream, + color: 'purple' + }) + links.push({ + id: `lnk-${item.id}-onsitetool`, + label: 'Onsite tool', + icon: 'telephone-outbound', + href: item.links.onsiteTool, + color: 'teal' + }) + } } } } + // Add Calendar item for all events that has a calendar link + if (item.adjustedEnd > current && item.links.calendar) { + links.push({ + id: `lnk-${item.id}-calendar`, + label: 'Calendar (.ics) entry for this session', + icon: 'calendar-check', + href: item.links.calendar, + color: 'pink' + }) + } // Event icon let icon = null @@ -437,7 +497,7 @@ const meetingEvents = computed(() => { case 'other': if (item.name.toLowerCase().indexOf('office hours') >= 0) { icon = 'bi-building' - } else if (item.name.toLowerCase().indexOf('hackathon') >= 0) { + } else if (item.groupAcronym === 'hackathon') { icon = 'bi-command bi-pink' } break @@ -464,6 +524,7 @@ const meetingEvents = computed(() => { // groupParentName: item.groupParent?.name, icon, isBoF: item.isBoF, + isProposed: item.isProposed, isSessionEvent: item.type === 'regular', links, location: item.location, @@ -557,21 +618,53 @@ function renderLinkIcon (opt) { return h('i', { class: `bi bi-${opt.icon} text-${opt.color}` }) } -function renderLinkLabel (opt) { +function renderLink (opt) { + if (opt.click) { + return h('button', { type: 'button', class: 'overflow-button', onClick: opt.click }, opt.label) + } + return h('a', { href: opt.href, target: '_blank' }, opt.label) } function recalculateRedLine () { state.currentMinute = DateTime.local().minute - const lastEventId = agendaStore.findCurrentEventId() + const currentEventId = agendaStore.findCurrentEventId() - if (lastEventId) { - state.redhandOffset = document.getElementById(`agenda-rowid-${lastEventId}`)?.offsetTop || 0 + if (currentEventId) { + state.redhandOffset = document.getElementById(`agenda-rowid-${currentEventId}`)?.offsetTop || 0 } else { state.redhandOffset = 0 } } +/** + * On page load when browser location hash contains '#now' or '#agenda-day-*' then scroll accordingly + */ +;(function scrollToHashInit() { + if (!window.location.hash) { + return + } + if (!(window.location.hash === "#now" || window.location.hash.startsWith(`#${daySlugPrefix}`))) { + return + } + const unsubscribe = agendaStore.$subscribe((_mutation, agendaStoreState) => { + if (agendaStoreState.schedule.length === 0) { + return + } + unsubscribe() // we only need to scroll once, so unsubscribe from future updates + if (window.location.hash === "#now") { + const nowEventId = agendaStore.findNowEvent() + if (nowEventId) { + document.getElementById(`agenda-rowid-${nowEventId}`)?.scrollIntoView(true) + } else { + message.warning('There is no event happening right now or in the future.') + } + } else if(window.location.hash.startsWith(`#${daySlugPrefix}`)) { + document.getElementById(window.location.hash.substring(1))?.scrollIntoView(true) + } + }) +})() + // MOUNTED onMounted(() => { @@ -681,6 +774,10 @@ onBeforeUnmount(() => { border-radius: 5px; border-collapse: separate; border-spacing: 0; + + @at-root .theme-dark & { + border-color: #000; + } } // -> Table HEADER @@ -700,6 +797,11 @@ onBeforeUnmount(() => { font-weight: 600; border-right: 1px solid #FFF; + @at-root .theme-dark & { + border-bottom-color: #000; + border-right-color: #000; + } + @media screen and (max-width: $bs5-break-md) { font-size: .8em; padding: 0 6px; @@ -754,6 +856,10 @@ onBeforeUnmount(() => { tr:nth-child(odd) td { background-color: #F9F9F9; + + @at-root .theme-dark & { + background-color: darken($gray-900, 5%); + } } &-display-noresult > td { @@ -763,6 +869,12 @@ onBeforeUnmount(() => { color: $gray-800; text-shadow: 1px 1px 0 #FFF; font-weight: 600; + + @at-root .theme-dark & { + background: linear-gradient(to bottom, $gray-900, $gray-800); + color: #FFF; + text-shadow: 1px 1px 0 $gray-900; + } } &-display-day > td { @@ -774,6 +886,10 @@ onBeforeUnmount(() => { font-weight: 600; scroll-margin-top: 25px; + @at-root .theme-dark & { + border-bottom-color: #000; + } + @media screen and (max-width: $bs5-break-md) { font-size: .9em; } @@ -786,6 +902,11 @@ onBeforeUnmount(() => { padding: 0 12px; color: #333; + @at-root .theme-dark & { + background: linear-gradient(to top, lighten($blue-900, 8%), lighten($blue-900, 4%)) !important; + color: $blue-100; + } + @media screen and (max-width: $bs5-break-md) { padding: 0 6px; } @@ -793,12 +914,21 @@ onBeforeUnmount(() => { &.agenda-table-cell-ts { border-right: 1px solid $blue-200 !important; color: $blue-700; + + @at-root .theme-dark & { + border-right-color: $blue-700 !important; + color: $blue-200; + } } &.agenda-table-cell-name { color: $blue-700; font-weight: 600; + @at-root .theme-dark & { + color: $blue-200; + } + @media screen and (max-width: $bs5-break-md) { font-size: .9em; } @@ -810,6 +940,10 @@ onBeforeUnmount(() => { padding: 0 12px; color: #333; + @at-root .theme-dark & { + color: #FFF; + } + @media screen and (max-width: $bs5-break-md) { padding: 2px 6px; } @@ -818,6 +952,11 @@ onBeforeUnmount(() => { background-color: desaturate($blue-700, 50%) !important; border-bottom: 1px solid #FFF; padding-bottom: 2px; + + @at-root .theme-dark & { + background-color: $gray-800 !important; + border-bottom-color: #000; + } } &.agenda-table-cell-ts { @@ -826,6 +965,13 @@ onBeforeUnmount(() => { border-right: 1px solid $blue-200 !important; color: $blue-200; border-bottom: 1px solid #FFF; + + @at-root .theme-dark & { + background: linear-gradient(to right, rgba(lighten($blue-900, 8%), .1), lighten($blue-900, 5%)); + border-right-color: $blue-700 !important; + border-bottom-color: $blue-700; + color: $blue-700; + } } } @@ -834,6 +980,11 @@ onBeforeUnmount(() => { border-right: 1px solid $gray-300 !important; white-space: nowrap; + @at-root .theme-dark & { + color: $yellow-100; + border-right-color: $gray-700 !important; + } + @media screen and (max-width: 1300px) { font-size: .85rem; } @@ -877,6 +1028,11 @@ onBeforeUnmount(() => { border-right: 1px solid $gray-300 !important; white-space: nowrap; + @at-root .theme-dark & { + color: $gray-400; + border-right-color: $gray-700 !important; + } + @media screen and (max-width: $bs5-break-md) { font-size: .7rem; word-break: break-all; @@ -903,6 +1059,14 @@ onBeforeUnmount(() => { border-top-left-radius: 0; border-bottom-left-radius: 0; margin-right: 6px; + + @at-root .theme-dark & { + background-color: $gray-700; + border-bottom-color: $gray-600; + border-right-color: $gray-600; + color: $gray-200; + text-shadow: 1px 1px $gray-800; + } } } @@ -913,13 +1077,26 @@ onBeforeUnmount(() => { word-wrap: break-word; } - .badge.is-bof { - background-color: $teal-500; + .badge { margin: 0 8px; + &.is-bof { + background-color: $teal-500; + + @at-root .theme-dark & { + background-color: $teal-700; + } + } + + &.is-proposed { + background-color: $gray-500; + + @at-root .theme-dark & { + background-color: $gray-700; + } + } + @media screen and (max-width: $bs5-break-md) { - width: 30px; - display: block; margin: 2px 0 0 0; } } @@ -934,6 +1111,10 @@ onBeforeUnmount(() => { } &.bi-green { color: $green-500; + + @at-root .theme-dark & { + color: $green-300; + } } &.bi-pink { color: $pink-500; @@ -974,7 +1155,7 @@ onBeforeUnmount(() => { .agenda-table-cell-links-buttons { white-space: nowrap; - > a, > i { + > a, > i, > button { margin-left: 3px; color: #666; cursor: pointer; @@ -983,6 +1164,11 @@ onBeforeUnmount(() => { padding: 2px 3px; transition: background-color .6s ease; + @at-root .theme-dark & { + background-color: rgba(0, 0, 0, .2); + color: $gray-200; + } + &:hover, &:focus { color: $blue; } @@ -991,6 +1177,10 @@ onBeforeUnmount(() => { color: $red-500; background-color: rgba($red-500, .1); + @at-root .theme-dark & { + color: $red-400; + } + &:hover, &:focus { background-color: rgba($red-500, .3); } @@ -999,14 +1189,34 @@ onBeforeUnmount(() => { color: $orange-700; background-color: rgba($orange-500, .1); + @at-root .theme-dark & { + color: $orange-400; + } + &:hover, &:focus { background-color: rgba($orange-500, .3); } } + &.text-darkgray { + color: $gray-900; + background-color: rgba($gray-700, .1); + + @at-root .theme-dark & { + color: $gray-100; + } + + &:hover, &:focus { + background-color: rgba($gray-700, .3); + } + } &.text-blue { color: $blue-600; background-color: rgba($blue-300, .1); + @at-root .theme-dark & { + color: $blue-300; + } + &:hover, &:focus { background-color: rgba($blue-300, .3); } @@ -1015,6 +1225,10 @@ onBeforeUnmount(() => { color: $green-500; background-color: rgba($green-300, .1); + @at-root .theme-dark & { + color: $green-300; + } + &:hover, &:focus { background-color: rgba($green-300, .3); } @@ -1023,6 +1237,10 @@ onBeforeUnmount(() => { color: $purple-500; background-color: rgba($purple-400, .1); + @at-root .theme-dark & { + color: $purple-300; + } + &:hover, &:focus { background-color: rgba($purple-400, .3); } @@ -1031,6 +1249,10 @@ onBeforeUnmount(() => { color: $pink-500; background-color: rgba($pink-400, .1); + @at-root .theme-dark & { + color: $pink-400; + } + &:hover, &:focus { background-color: rgba($pink-400, .3); } @@ -1039,6 +1261,10 @@ onBeforeUnmount(() => { color: $teal-600; background-color: rgba($teal-400, .1); + @at-root .theme-dark & { + color: $teal-300; + } + &:hover, &:focus { background-color: rgba($teal-400, .3); } @@ -1058,13 +1284,17 @@ onBeforeUnmount(() => { &-cell-ts { border-right: 1px solid $gray-300 !important; - // -> Use system font instead of Montserrat so that all digits align vertically + // -> Use system font instead of Inter so that all digits align vertically font-family: system-ui, -apple-system, BlinkMacSystemFont, 'Segoe UI', Roboto, Oxygen, Ubuntu, Cantarell, 'Open Sans', 'Helvetica Neue', sans-serif; font-size: 1rem; font-weight: 700; text-align: right; white-space: nowrap; + @at-root .theme-dark & { + border-right-color: $gray-700 !important; + } + @media screen and (max-width: 1300px) { font-size: .9rem; } @@ -1090,8 +1320,22 @@ onBeforeUnmount(() => { border-bottom: none; } + &.agenda-table-cell-ts.is-session-event { + @at-root .theme-dark & { + background: transparent; + color: $red-300; + border-top: 1px solid darken($red-100, 5%); + border-bottom-color: darken($red-100, 5%); + } + } + &.agenda-table-cell-room { border-right: 1px solid darken($red-100, 5%) !important; + text-decoration: line-through; + } + + &.agenda-table-cell-name > a, &.agenda-table-cell-name > span { + text-decoration: line-through; } &:last-child { @@ -1108,8 +1352,22 @@ onBeforeUnmount(() => { border-bottom: none; } + &.agenda-table-cell-ts.is-session-event { + @at-root .theme-dark & { + background: transparent; + color: $orange-300; + border-top: 1px solid darken($orange-100, 5%); + border-bottom-color: darken($orange-100, 5%); + } + } + &.agenda-table-cell-room { border-right: 1px solid darken($orange-100, 5%) !important; + text-decoration: line-through; + } + + &.agenda-table-cell-name > a, &.agenda-table-cell-name > span { + text-decoration: line-through; } &:last-child { @@ -1121,10 +1379,21 @@ onBeforeUnmount(() => { border-top: 1px solid darken($indigo-100, 5%); border-bottom: 1px solid darken($indigo-100, 5%); + @at-root .theme-dark & { + color: $indigo-100; + // border-bottom-color: #000; + } + &.agenda-table-cell-ts { background: linear-gradient(to right, lighten($indigo-100, 8%), lighten($indigo-100, 5%)); color: $indigo-700; border-right: 1px solid $indigo-100 !important; + + @at-root .theme-dark & { + background: rgba($indigo, .1) !important; + color: $indigo-100; + border-right-color: $indigo-500 !important; + } } &.agenda-table-cell-room { @@ -1134,10 +1403,18 @@ onBeforeUnmount(() => { &.agenda-table-cell-name { color: $indigo-700; font-style: italic; + + @at-root .theme-dark & { + color: $indigo-200; + } } &.agenda-table-cell-links { background: linear-gradient(to right, lighten($indigo-100, 5%), lighten($indigo-100, 8%)); + + @at-root .theme-dark & { + background: rgba($indigo, .1) !important; + } } } &-type-plenary td { @@ -1146,9 +1423,19 @@ onBeforeUnmount(() => { border-top: 1px solid darken($teal-100, 5%); border-bottom: 1px solid darken($teal-100, 5%); + @at-root .theme-dark & { + background: rgba($teal, .15) !important; + color: $teal-100; + border-bottom: 1px solid darken($teal-600, 5%); + } + &.agenda-table-cell-ts { background: linear-gradient(to right, lighten($teal-100, 8%), lighten($teal-100, 2%)); border-right: 1px solid $teal-200 !important; + + @at-root .theme-dark & { + border-right-color: $teal-700 !important; + } } &.agenda-table-cell-room { @@ -1158,10 +1445,18 @@ onBeforeUnmount(() => { &.agenda-table-cell-name { font-weight: 600; color: $teal-700; + + @at-root .theme-dark & { + color: $teal-200; + } } &.agenda-table-cell-links { background: linear-gradient(to right, rgba(lighten($teal, 54%), 0), lighten($teal, 54%)); + + @at-root .theme-dark & { + background: rgba($teal, .15) !important; + } } } @@ -1326,6 +1621,22 @@ onBeforeUnmount(() => { } } +.overflow-button { + font-size: inherit; + padding: 0; + border: 0; + background: transparent; + + &:before { + content: ""; + position: absolute; + left: 0; + right: 0; + top: 0; + bottom: 0; + } +} + @keyframes fadeInAnim { 0% { opacity: 0; diff --git a/client/agenda/AgendaSettings.vue b/client/agenda/AgendaSettings.vue index 41cd226b6c..b074bc3247 100644 --- a/client/agenda/AgendaSettings.vue +++ b/client/agenda/AgendaSettings.vue @@ -457,6 +457,14 @@ onMounted(() => { font-size: .8rem; color: $gray-700; text-shadow: 1px 1px 0 #FFF; + + @at-root .theme-dark & { + background-color: $gray-900; + text-shadow: none; + border-bottom-color: $gray-700; + border-right-color: $gray-700; + color: #FFF; + } } &-calcoffset { diff --git a/client/agenda/AgendaShareModal.vue b/client/agenda/AgendaShareModal.vue index 8cdbb291af..a71938673b 100644 --- a/client/agenda/AgendaShareModal.vue +++ b/client/agenda/AgendaShareModal.vue @@ -20,7 +20,7 @@ n-modal(v-model:show='modalShown') i.bi.bi-share span Share this view .agenda-share-content - .text-muted.pb-2 Use the following URL for sharing the current view #[em (including any active filters)] with other users: + .text-body-secondary.pb-2 Use the following URL for sharing the current view #[em (including any active filters)] with other users: n-input-group n-input( ref='filteredUrlIpt' diff --git a/client/agenda/FloorPlan.vue b/client/agenda/FloorPlan.vue index 41d3f13ffe..eccd59aeaa 100644 --- a/client/agenda/FloorPlan.vue +++ b/client/agenda/FloorPlan.vue @@ -187,10 +187,10 @@ function handleDesiredRoom () { if (rm) { state.currentFloor = fl.id state.currentRoom = rm.id + state.desiredRoom = null break } } - state.desiredRoom = null } } @@ -258,6 +258,11 @@ onMounted(() => { border-radius: 5px; font-weight: 500; + @at-root .theme-dark & { + background-color: darken($gray-900, 5%); + border-color: $gray-700; + } + a { cursor: pointer; diff --git a/client/agenda/agenda.scss b/client/agenda/agenda.scss index 6d5cd634b5..e3e46d14f0 100644 --- a/client/agenda/agenda.scss +++ b/client/agenda/agenda.scss @@ -10,6 +10,10 @@ justify-content: space-between; align-items: center; + @at-root .theme-dark & { + color: $gray-300; + } + @media screen and (max-width: $bs5-break-sm) { justify-content: center; @@ -25,6 +29,10 @@ -webkit-background-clip: text; -webkit-text-fill-color: transparent; box-decoration-break: clone; + + @at-root .theme-dark & { + background-image: linear-gradient(220deg, $yellow-200 20%, $orange-400 70%); + } } } diff --git a/client/agenda/store.js b/client/agenda/store.js index 839d464c43..b5498303a6 100644 --- a/client/agenda/store.js +++ b/client/agenda/store.js @@ -50,7 +50,7 @@ export const useAgendaStore = defineStore('agenda', { selectedCatSubs: [], settingsShown: false, timezone: DateTime.local().zoneName, - useNotes: false, + usesNotes: false, visibleDays: [] }), getters: { @@ -121,7 +121,7 @@ export const useAgendaStore = defineStore('agenda', { meetingDays () { const siteStore = useSiteStore() return uniqBy(this.scheduleAdjusted, 'adjustedStartDate').sort().map(s => ({ - slug: s.id.toString(), + slug: daySlug(s), ts: s.adjustedStartDate, label: siteStore.viewport < 1350 ? DateTime.fromISO(s.adjustedStartDate).toFormat('ccc LLL d') : DateTime.fromISO(s.adjustedStartDate).toLocaleString(DateTime.DATE_HUGE) })) @@ -141,7 +141,7 @@ export const useAgendaStore = defineStore('agenda', { meetingNumber = meetingData.meetingNumber } - const resp = await fetch(`/api/meeting/${meetingNumber}/agenda-data`, { credentials: 'omit' }) + const resp = await fetch(`/api/meeting/${meetingNumber}/agenda-data`) if (!resp.ok) { throw new Error(resp.statusText) } @@ -160,7 +160,7 @@ export const useAgendaStore = defineStore('agenda', { this.isCurrentMeeting = agendaData.isCurrentMeeting this.meeting = agendaData.meeting this.schedule = agendaData.schedule - this.useNotes = agendaData.useNotes + this.usesNotes = agendaData.usesNotes // -> Compute current info note hash this.infoNoteHash = murmur(agendaData.meeting.infoNote, 0).toString() @@ -230,6 +230,28 @@ export const useAgendaStore = defineStore('agenda', { return lastEvent.id || null }, + findNowEventId () { + const currentEventId = this.findCurrentEventId() + + if (currentEventId) { + return currentEventId + } + + // if there isn't a current event then instead find the next event + + const current = (this.nowDebugDiff ? DateTime.local().minus(this.nowDebugDiff) : DateTime.local()).setZone(this.timezone) + + // -> Find next event after current time + let nextEventId = undefined + for(const sh of this.scheduleAdjusted) { + if (sh.adjustedStart > current) { + nextEventId = sh.id + break + } + } + + return nextEventId || null + }, hideLoadingScreen () { // -> Hide loading screen const loadingRef = document.querySelector('#app-loading') @@ -292,3 +314,8 @@ function findFirstConferenceUrl (txt) { } catch (err) { } return null } + +export const daySlugPrefix = 'agenda-day-' +export function daySlug(s) { + return `${daySlugPrefix}${s.adjustedStartDate}` // eg 'agenda-day-2024-08-13' +} diff --git a/client/components/ChatLog.vue b/client/components/ChatLog.vue index d393b18664..b3a4f7b40f 100644 --- a/client/components/ChatLog.vue +++ b/client/components/ChatLog.vue @@ -15,7 +15,7 @@ ) template(#default) div(v-html='item.text') - span.text-muted(v-else) + span.text-body-secondary(v-else) em No chat log available. @@ -159,4 +159,18 @@ onMounted(() => { } } } + +[data-bs-theme="dark"] .chatlog { + .n-timeline-item-content__title { + color: #d63384 !important; + } + + .n-timeline-item-content__content { + color: #fff !important; + } + + .n-timeline-item-content__meta { + color: #0569ffd9 !important; + } +} diff --git a/client/components/Polls.vue b/client/components/Polls.vue index 72c8e1c633..0846d4ed16 100644 --- a/client/components/Polls.vue +++ b/client/components/Polls.vue @@ -3,19 +3,20 @@ n-data-table( v-if='state.items.length > 0' :data='state.items' - :columns='columns' + :columns='state.columns' striped ) - span.text-muted(v-else) + span.text-danger(v-else-if='state.errMessage') + em {{ state.errMessage }} + span.text-body-secondary(v-else) em No polls available. + diff --git a/client/components/Status.vue b/client/components/Status.vue new file mode 100644 index 0000000000..4fded5bbe4 --- /dev/null +++ b/client/components/Status.vue @@ -0,0 +1,80 @@ + diff --git a/client/components/n-theme.vue b/client/components/n-theme.vue index 9fff81671e..b55c9772b9 100644 --- a/client/components/n-theme.vue +++ b/client/components/n-theme.vue @@ -1,24 +1,45 @@ - diff --git a/client/embedded.js b/client/embedded.js index f3b01f68f5..0509c0aecf 100644 --- a/client/embedded.js +++ b/client/embedded.js @@ -1,5 +1,12 @@ import { createApp } from 'vue' +import piniaPersist from 'pinia-plugin-persist' import Embedded from './Embedded.vue' +import { createPiniaSingleton } from './shared/create-pinia-singleton' + +// Initialize store (Pinia) + +const pinia = createPiniaSingleton() +pinia.use(piniaPersist) // Mount App @@ -9,5 +16,6 @@ for (const mnt of mountEls) { componentName: mnt.dataset.component, componentId: mnt.dataset.componentId }) + app.use(pinia) app.mount(mnt) } diff --git a/client/index.html b/client/index.html index e6c1648def..75d6f77727 100644 --- a/client/index.html +++ b/client/index.html @@ -8,8 +8,11 @@ + + +
@@ -18,5 +21,6 @@
+ diff --git a/client/main.js b/client/main.js index 0dc5cf32e0..3fbad907b1 100644 --- a/client/main.js +++ b/client/main.js @@ -1,14 +1,14 @@ import { createApp } from 'vue' -import { createPinia } from 'pinia' import piniaPersist from 'pinia-plugin-persist' import App from './App.vue' import router from './router' +import { createPiniaSingleton } from './shared/create-pinia-singleton' const app = createApp(App, {}) // Initialize store (Pinia) -const pinia = createPinia() +const pinia = createPiniaSingleton() pinia.use(piniaPersist) app.use(pinia) diff --git a/client/shared/create-pinia-singleton.js b/client/shared/create-pinia-singleton.js new file mode 100644 index 0000000000..f0013245a1 --- /dev/null +++ b/client/shared/create-pinia-singleton.js @@ -0,0 +1,6 @@ +import { createPinia } from 'pinia' + +export function createPiniaSingleton(){ + window.pinia = window.pinia ?? createPinia() + return window.pinia +} diff --git a/client/shared/json-wrapper.js b/client/shared/json-wrapper.js new file mode 100644 index 0000000000..e080b5a479 --- /dev/null +++ b/client/shared/json-wrapper.js @@ -0,0 +1,20 @@ +export const JSONWrapper = { + parse(jsonString, defaultValue) { + if(typeof jsonString !== "string") { + return defaultValue + } + try { + return JSON.parse(jsonString); + } catch (e) { + console.error(e); + } + return defaultValue + }, + stringify(data) { + try { + return JSON.stringify(data); + } catch (e) { + console.error(e) + } + }, +} diff --git a/client/shared/local-storage-wrapper.js b/client/shared/local-storage-wrapper.js new file mode 100644 index 0000000000..88cd3dc589 --- /dev/null +++ b/client/shared/local-storage-wrapper.js @@ -0,0 +1,42 @@ + +/* + * DEVELOPER NOTE + * + * Some browsers can block storage (localStorage, sessionStorage) + * access for privacy reasons, and all browsers can have storage + * that's full, and then they throw exceptions. + * + * See https://michalzalecki.com/why-using-localStorage-directly-is-a-bad-idea/ + * + * Exceptions can even be thrown when testing if localStorage + * even exists. This can throw: + * + * if (window.localStorage) + * + * Also localStorage/sessionStorage can be enabled after DOMContentLoaded + * so we handle it gracefully. + * + * 1) we need to wrap all usage in try/catch + * 2) we need to defer actual usage of these until + * necessary, + * + */ + +export const localStorageWrapper = { + getItem: (key) => { + try { + return localStorage.getItem(key) + } catch (e) { + console.error(e); + } + return null; + }, + setItem: (key, value) => { + try { + return localStorage.setItem(key, value) + } catch (e) { + console.error(e); + } + return; + }, +} diff --git a/client/shared/status-common.js b/client/shared/status-common.js new file mode 100644 index 0000000000..6503bfbf63 --- /dev/null +++ b/client/shared/status-common.js @@ -0,0 +1,5 @@ +// Used in Playwright Status and components + +export const STATUS_STORAGE_KEY = "status-dismissed" + +export const generateStatusTestId = (id) => `status-${id}` diff --git a/client/shared/store.js b/client/shared/store.js index 389d5e5a82..5dd57f2c81 100644 --- a/client/shared/store.js +++ b/client/shared/store.js @@ -6,6 +6,7 @@ export const useSiteStore = defineStore('site', { criticalErrorLink: null, criticalErrorLinkText: null, isMobile: /Mobi/i.test(navigator.userAgent), - viewport: Math.round(window.innerWidth) + viewport: Math.round(window.innerWidth), + theme: null }) }) diff --git a/client/shared/timezones.js b/client/shared/timezones.js index c3bc473000..8239c1f98e 100644 --- a/client/shared/timezones.js +++ b/client/shared/timezones.js @@ -1,252 +1,252 @@ export default [ - { label: '(GMT-11:00) Niue', value: 'Pacific/Niue' }, - { label: '(GMT-11:00) Pago Pago', value: 'Pacific/Pago_Pago' }, - { label: '(GMT-10:00) Hawaii Time', value: 'Pacific/Honolulu' }, - { label: '(GMT-10:00) Rarotonga', value: 'Pacific/Rarotonga' }, - { label: '(GMT-10:00) Tahiti', value: 'Pacific/Tahiti' }, - { label: '(GMT-09:30) Marquesas', value: 'Pacific/Marquesas' }, - { label: '(GMT-09:00) Alaska Time', value: 'America/Anchorage' }, - { label: '(GMT-09:00) Gambier', value: 'Pacific/Gambier' }, - { label: '(GMT-08:00) Pacific Time - Los Angeles', value: 'America/Los_Angeles' }, - { label: '(GMT-08:00) Pacific Time - Tijuana', value: 'America/Tijuana' }, - { label: '(GMT-08:00) Pacific Time - Vancouver', value: 'America/Vancouver' }, - { label: '(GMT-08:00) Pacific Time - Whitehorse', value: 'America/Whitehorse' }, - { label: '(GMT-08:00) Pitcairn', value: 'Pacific/Pitcairn' }, - { label: '(GMT-07:00) Mountain Time - Arizona', value: 'America/Phoenix' }, - { label: '(GMT-07:00) Mountain Time - Chihuahua, Mazatlan', value: 'America/Mazatlan' }, - { label: '(GMT-07:00) Mountain Time - Dawson Creek', value: 'America/Dawson_Creek' }, - { label: '(GMT-07:00) Mountain Time - Denver', value: 'America/Denver' }, - { label: '(GMT-07:00) Mountain Time - Edmonton', value: 'America/Edmonton' }, - { label: '(GMT-07:00) Mountain Time - Hermosillo', value: 'America/Hermosillo' }, - { label: '(GMT-07:00) Mountain Time - Yellowknife', value: 'America/Yellowknife' }, - { label: '(GMT-06:00) Belize', value: 'America/Belize' }, - { label: '(GMT-06:00) Central Time - Chicago', value: 'America/Chicago' }, - { label: '(GMT-06:00) Central Time - Mexico City', value: 'America/Mexico_City' }, - { label: '(GMT-06:00) Central Time - Regina', value: 'America/Regina' }, - { label: '(GMT-06:00) Central Time - Tegucigalpa', value: 'America/Tegucigalpa' }, - { label: '(GMT-06:00) Central Time - Winnipeg', value: 'America/Winnipeg' }, - { label: '(GMT-06:00) Costa Rica', value: 'America/Costa_Rica' }, - { label: '(GMT-06:00) El Salvador', value: 'America/El_Salvador' }, - { label: '(GMT-06:00) Galapagos', value: 'Pacific/Galapagos' }, - { label: '(GMT-06:00) Guatemala', value: 'America/Guatemala' }, - { label: '(GMT-06:00) Managua', value: 'America/Managua' }, - { label: '(GMT-05:00) America Cancun', value: 'America/Cancun' }, - { label: '(GMT-05:00) Bogota', value: 'America/Bogota' }, - { label: '(GMT-05:00) Easter Island', value: 'Pacific/Easter' }, - { label: '(GMT-05:00) Eastern Time - New York', value: 'America/New_York' }, - { label: '(GMT-05:00) Eastern Time - Iqaluit', value: 'America/Iqaluit' }, - { label: '(GMT-05:00) Eastern Time - Toronto', value: 'America/Toronto' }, - { label: '(GMT-05:00) Guayaquil', value: 'America/Guayaquil' }, - { label: '(GMT-05:00) Havana', value: 'America/Havana' }, - { label: '(GMT-05:00) Jamaica', value: 'America/Jamaica' }, - { label: '(GMT-05:00) Lima', value: 'America/Lima' }, - { label: '(GMT-05:00) Nassau', value: 'America/Nassau' }, - { label: '(GMT-05:00) Panama', value: 'America/Panama' }, - { label: '(GMT-05:00) Port-au-Prince', value: 'America/Port-au-Prince' }, - { label: '(GMT-05:00) Rio Branco', value: 'America/Rio_Branco' }, - { label: '(GMT-04:00) Atlantic Time - Halifax', value: 'America/Halifax' }, - { label: '(GMT-04:00) Barbados', value: 'America/Barbados' }, - { label: '(GMT-04:00) Bermuda', value: 'Atlantic/Bermuda' }, - { label: '(GMT-04:00) Boa Vista', value: 'America/Boa_Vista' }, - { label: '(GMT-04:00) Caracas', value: 'America/Caracas' }, - { label: '(GMT-04:00) Curacao', value: 'America/Curacao' }, - { label: '(GMT-04:00) Grand Turk', value: 'America/Grand_Turk' }, - { label: '(GMT-04:00) Guyana', value: 'America/Guyana' }, - { label: '(GMT-04:00) La Paz', value: 'America/La_Paz' }, - { label: '(GMT-04:00) Manaus', value: 'America/Manaus' }, - { label: '(GMT-04:00) Martinique', value: 'America/Martinique' }, - { label: '(GMT-04:00) Port of Spain', value: 'America/Port_of_Spain' }, - { label: '(GMT-04:00) Porto Velho', value: 'America/Porto_Velho' }, - { label: '(GMT-04:00) Puerto Rico', value: 'America/Puerto_Rico' }, - { label: '(GMT-04:00) Santo Domingo', value: 'America/Santo_Domingo' }, - { label: '(GMT-04:00) Thule', value: 'America/Thule' }, - { label: '(GMT-03:30) Newfoundland Time - St. Johns', value: 'America/St_Johns' }, - { label: '(GMT-03:00) Araguaina', value: 'America/Araguaina' }, - { label: '(GMT-03:00) Asuncion', value: 'America/Asuncion' }, - { label: '(GMT-03:00) Belem', value: 'America/Belem' }, - { label: '(GMT-03:00) Buenos Aires', value: 'America/Argentina/Buenos_Aires' }, - { label: '(GMT-03:00) Campo Grande', value: 'America/Campo_Grande' }, - { label: '(GMT-03:00) Cayenne', value: 'America/Cayenne' }, - { label: '(GMT-03:00) Cuiaba', value: 'America/Cuiaba' }, - { label: '(GMT-03:00) Fortaleza', value: 'America/Fortaleza' }, - { label: '(GMT-03:00) Godthab', value: 'America/Godthab' }, - { label: '(GMT-03:00) Maceio', value: 'America/Maceio' }, - { label: '(GMT-03:00) Miquelon', value: 'America/Miquelon' }, - { label: '(GMT-03:00) Montevideo', value: 'America/Montevideo' }, - { label: '(GMT-03:00) Palmer', value: 'Antarctica/Palmer' }, - { label: '(GMT-03:00) Paramaribo', value: 'America/Paramaribo' }, - { label: '(GMT-03:00) Punta Arenas', value: 'America/Punta_Arenas' }, - { label: '(GMT-03:00) Recife', value: 'America/Recife' }, - { label: '(GMT-03:00) Rothera', value: 'Antarctica/Rothera' }, - { label: '(GMT-03:00) Salvador', value: 'America/Bahia' }, - { label: '(GMT-03:00) Santiago', value: 'America/Santiago' }, - { label: '(GMT-03:00) Stanley', value: 'Atlantic/Stanley' }, - { label: '(GMT-02:00) Noronha', value: 'America/Noronha' }, - { label: '(GMT-02:00) Sao Paulo', value: 'America/Sao_Paulo' }, - { label: '(GMT-02:00) South Georgia', value: 'Atlantic/South_Georgia' }, - { label: '(GMT-01:00) Azores', value: 'Atlantic/Azores' }, - { label: '(GMT-01:00) Cape Verde', value: 'Atlantic/Cape_Verde' }, - { label: '(GMT-01:00) Scoresbysund', value: 'America/Scoresbysund' }, - { label: '(GMT+00:00) Abidjan', value: 'Africa/Abidjan' }, - { label: '(GMT+00:00) Accra', value: 'Africa/Accra' }, - { label: '(GMT+00:00) Bissau', value: 'Africa/Bissau' }, - { label: '(GMT+00:00) Canary Islands', value: 'Atlantic/Canary' }, - { label: '(GMT+00:00) Casablanca', value: 'Africa/Casablanca' }, - { label: '(GMT+00:00) Danmarkshavn', value: 'America/Danmarkshavn' }, - { label: '(GMT+00:00) Dublin', value: 'Europe/Dublin' }, - { label: '(GMT+00:00) El Aaiun', value: 'Africa/El_Aaiun' }, - { label: '(GMT+00:00) Faeroe', value: 'Atlantic/Faroe' }, - { label: '(GMT+00:00) UTC / GMT', value: 'UTC' }, - { label: '(GMT+00:00) Lisbon', value: 'Europe/Lisbon' }, - { label: '(GMT+00:00) London', value: 'Europe/London' }, - { label: '(GMT+00:00) Monrovia', value: 'Africa/Monrovia' }, - { label: '(GMT+00:00) Reykjavik', value: 'Atlantic/Reykjavik' }, - { label: '(GMT+01:00) Algiers', value: 'Africa/Algiers' }, - { label: '(GMT+01:00) Amsterdam', value: 'Europe/Amsterdam' }, - { label: '(GMT+01:00) Andorra', value: 'Europe/Andorra' }, - { label: '(GMT+01:00) Berlin', value: 'Europe/Berlin' }, - { label: '(GMT+01:00) Brussels', value: 'Europe/Brussels' }, - { label: '(GMT+01:00) Budapest', value: 'Europe/Budapest' }, - { label: '(GMT+01:00) Central European Time - Belgrade', value: 'Europe/Belgrade' }, - { label: '(GMT+01:00) Central European Time - Prague', value: 'Europe/Prague' }, - { label: '(GMT+01:00) Ceuta', value: 'Africa/Ceuta' }, - { label: '(GMT+01:00) Copenhagen', value: 'Europe/Copenhagen' }, - { label: '(GMT+01:00) Gibraltar', value: 'Europe/Gibraltar' }, - { label: '(GMT+01:00) Lagos', value: 'Africa/Lagos' }, - { label: '(GMT+01:00) Luxembourg', value: 'Europe/Luxembourg' }, - { label: '(GMT+01:00) Madrid', value: 'Europe/Madrid' }, - { label: '(GMT+01:00) Malta', value: 'Europe/Malta' }, - { label: '(GMT+01:00) Monaco', value: 'Europe/Monaco' }, - { label: '(GMT+01:00) Ndjamena', value: 'Africa/Ndjamena' }, - { label: '(GMT+01:00) Oslo', value: 'Europe/Oslo' }, - { label: '(GMT+01:00) Paris', value: 'Europe/Paris' }, - { label: '(GMT+01:00) Rome', value: 'Europe/Rome' }, - { label: '(GMT+01:00) Stockholm', value: 'Europe/Stockholm' }, - { label: '(GMT+01:00) Tirane', value: 'Europe/Tirane' }, - { label: '(GMT+01:00) Tunis', value: 'Africa/Tunis' }, - { label: '(GMT+01:00) Vienna', value: 'Europe/Vienna' }, - { label: '(GMT+01:00) Warsaw', value: 'Europe/Warsaw' }, - { label: '(GMT+01:00) Zurich', value: 'Europe/Zurich' }, - { label: '(GMT+02:00) Amman', value: 'Asia/Amman' }, - { label: '(GMT+02:00) Athens', value: 'Europe/Athens' }, - { label: '(GMT+02:00) Beirut', value: 'Asia/Beirut' }, - { label: '(GMT+02:00) Bucharest', value: 'Europe/Bucharest' }, - { label: '(GMT+02:00) Cairo', value: 'Africa/Cairo' }, - { label: '(GMT+02:00) Chisinau', value: 'Europe/Chisinau' }, - { label: '(GMT+02:00) Damascus', value: 'Asia/Damascus' }, - { label: '(GMT+02:00) Gaza', value: 'Asia/Gaza' }, - { label: '(GMT+02:00) Helsinki', value: 'Europe/Helsinki' }, - { label: '(GMT+02:00) Jerusalem', value: 'Asia/Jerusalem' }, - { label: '(GMT+02:00) Johannesburg', value: 'Africa/Johannesburg' }, - { label: '(GMT+02:00) Khartoum', value: 'Africa/Khartoum' }, - { label: '(GMT+02:00) Kiev', value: 'Europe/Kiev' }, - { label: '(GMT+02:00) Maputo', value: 'Africa/Maputo' }, - { label: '(GMT+02:00) Moscow-01 - Kaliningrad', value: 'Europe/Kaliningrad' }, - { label: '(GMT+02:00) Nicosia', value: 'Asia/Nicosia' }, - { label: '(GMT+02:00) Riga', value: 'Europe/Riga' }, - { label: '(GMT+02:00) Sofia', value: 'Europe/Sofia' }, - { label: '(GMT+02:00) Tallinn', value: 'Europe/Tallinn' }, - { label: '(GMT+02:00) Tripoli', value: 'Africa/Tripoli' }, - { label: '(GMT+02:00) Vilnius', value: 'Europe/Vilnius' }, - { label: '(GMT+02:00) Windhoek', value: 'Africa/Windhoek' }, - { label: '(GMT+03:00) Baghdad', value: 'Asia/Baghdad' }, - { label: '(GMT+03:00) Istanbul', value: 'Europe/Istanbul' }, - { label: '(GMT+03:00) Minsk', value: 'Europe/Minsk' }, - { label: '(GMT+03:00) Moscow+00 - Moscow', value: 'Europe/Moscow' }, - { label: '(GMT+03:00) Nairobi', value: 'Africa/Nairobi' }, - { label: '(GMT+03:00) Qatar', value: 'Asia/Qatar' }, - { label: '(GMT+03:00) Riyadh', value: 'Asia/Riyadh' }, - { label: '(GMT+03:00) Syowa', value: 'Antarctica/Syowa' }, - { label: '(GMT+03:30) Tehran', value: 'Asia/Tehran' }, - { label: '(GMT+04:00) Baku', value: 'Asia/Baku' }, - { label: '(GMT+04:00) Dubai', value: 'Asia/Dubai' }, - { label: '(GMT+04:00) Mahe', value: 'Indian/Mahe' }, - { label: '(GMT+04:00) Mauritius', value: 'Indian/Mauritius' }, - { label: '(GMT+04:00) Moscow+01 - Samara', value: 'Europe/Samara' }, - { label: '(GMT+04:00) Reunion', value: 'Indian/Reunion' }, - { label: '(GMT+04:00) Tbilisi', value: 'Asia/Tbilisi' }, - { label: '(GMT+04:00) Yerevan', value: 'Asia/Yerevan' }, - { label: '(GMT+04:30) Kabul', value: 'Asia/Kabul' }, - { label: '(GMT+05:00) Aqtau', value: 'Asia/Aqtau' }, - { label: '(GMT+05:00) Aqtobe', value: 'Asia/Aqtobe' }, - { label: '(GMT+05:00) Ashgabat', value: 'Asia/Ashgabat' }, - { label: '(GMT+05:00) Dushanbe', value: 'Asia/Dushanbe' }, - { label: '(GMT+05:00) Karachi', value: 'Asia/Karachi' }, - { label: '(GMT+05:00) Kerguelen', value: 'Indian/Kerguelen' }, - { label: '(GMT+05:00) Maldives', value: 'Indian/Maldives' }, - { label: '(GMT+05:00) Mawson', value: 'Antarctica/Mawson' }, - { label: '(GMT+05:00) Moscow+02 - Yekaterinburg', value: 'Asia/Yekaterinburg' }, - { label: '(GMT+05:00) Tashkent', value: 'Asia/Tashkent' }, - { label: '(GMT+05:30) Colombo', value: 'Asia/Colombo' }, - { label: '(GMT+05:30) India Standard Time', value: 'Asia/Kolkata' }, - { label: '(GMT+05:45) Kathmandu', value: 'Asia/Kathmandu' }, - { label: '(GMT+06:00) Almaty', value: 'Asia/Almaty' }, - { label: '(GMT+06:00) Bishkek', value: 'Asia/Bishkek' }, - { label: '(GMT+06:00) Chagos', value: 'Indian/Chagos' }, - { label: '(GMT+06:00) Dhaka', value: 'Asia/Dhaka' }, - { label: '(GMT+06:00) Moscow+03 - Omsk', value: 'Asia/Omsk' }, - { label: '(GMT+06:00) Thimphu', value: 'Asia/Thimphu' }, - { label: '(GMT+06:00) Vostok', value: 'Antarctica/Vostok' }, - { label: '(GMT+06:30) Cocos', value: 'Indian/Cocos' }, - { label: '(GMT+06:30) Rangoon', value: 'Asia/Yangon' }, - { label: '(GMT+07:00) Bangkok', value: 'Asia/Bangkok' }, - { label: '(GMT+07:00) Christmas', value: 'Indian/Christmas' }, - { label: '(GMT+07:00) Davis', value: 'Antarctica/Davis' }, - { label: '(GMT+07:00) Hanoi', value: 'Asia/Saigon' }, - { label: '(GMT+07:00) Hovd', value: 'Asia/Hovd' }, - { label: '(GMT+07:00) Jakarta', value: 'Asia/Jakarta' }, - { label: '(GMT+07:00) Moscow+04 - Krasnoyarsk', value: 'Asia/Krasnoyarsk' }, - { label: '(GMT+08:00) Brunei', value: 'Asia/Brunei' }, - { label: '(GMT+08:00) China Time - Beijing', value: 'Asia/Shanghai' }, - { label: '(GMT+08:00) Choibalsan', value: 'Asia/Choibalsan' }, - { label: '(GMT+08:00) Hong Kong', value: 'Asia/Hong_Kong' }, - { label: '(GMT+08:00) Kuala Lumpur', value: 'Asia/Kuala_Lumpur' }, - { label: '(GMT+08:00) Macau', value: 'Asia/Macau' }, - { label: '(GMT+08:00) Makassar', value: 'Asia/Makassar' }, - { label: '(GMT+08:00) Manila', value: 'Asia/Manila' }, - { label: '(GMT+08:00) Moscow+05 - Irkutsk', value: 'Asia/Irkutsk' }, - { label: '(GMT+08:00) Singapore', value: 'Asia/Singapore' }, - { label: '(GMT+08:00) Taipei', value: 'Asia/Taipei' }, - { label: '(GMT+08:00) Ulaanbaatar', value: 'Asia/Ulaanbaatar' }, - { label: '(GMT+08:00) Western Time - Perth', value: 'Australia/Perth' }, - { label: '(GMT+08:30) Pyongyang', value: 'Asia/Pyongyang' }, - { label: '(GMT+09:00) Dili', value: 'Asia/Dili' }, - { label: '(GMT+09:00) Jayapura', value: 'Asia/Jayapura' }, - { label: '(GMT+09:00) Moscow+06 - Yakutsk', value: 'Asia/Yakutsk' }, - { label: '(GMT+09:00) Palau', value: 'Pacific/Palau' }, - { label: '(GMT+09:00) Seoul', value: 'Asia/Seoul' }, - { label: '(GMT+09:00) Tokyo', value: 'Asia/Tokyo' }, - { label: '(GMT+09:30) Central Time - Darwin', value: 'Australia/Darwin' }, - { label: '(GMT+10:00) Dumont D\'Urville', value: 'Antarctica/DumontDUrville' }, - { label: '(GMT+10:00) Eastern Time - Brisbane', value: 'Australia/Brisbane' }, - { label: '(GMT+10:00) Guam', value: 'Pacific/Guam' }, - { label: '(GMT+10:00) Moscow+07 - Vladivostok', value: 'Asia/Vladivostok' }, - { label: '(GMT+10:00) Port Moresby', value: 'Pacific/Port_Moresby' }, - { label: '(GMT+10:00) Truk', value: 'Pacific/Chuuk' }, - { label: '(GMT+10:30) Central Time - Adelaide', value: 'Australia/Adelaide' }, - { label: '(GMT+11:00) Casey', value: 'Antarctica/Casey' }, - { label: '(GMT+11:00) Eastern Time - Hobart', value: 'Australia/Hobart' }, - { label: '(GMT+11:00) Eastern Time - Melbourne, Sydney', value: 'Australia/Sydney' }, - { label: '(GMT+11:00) Efate', value: 'Pacific/Efate' }, - { label: '(GMT+11:00) Guadalcanal', value: 'Pacific/Guadalcanal' }, - { label: '(GMT+11:00) Kosrae', value: 'Pacific/Kosrae' }, - { label: '(GMT+11:00) Moscow+08 - Magadan', value: 'Asia/Magadan' }, - { label: '(GMT+11:00) Norfolk', value: 'Pacific/Norfolk' }, - { label: '(GMT+11:00) Noumea', value: 'Pacific/Noumea' }, - { label: '(GMT+11:00) Ponape', value: 'Pacific/Pohnpei' }, - { label: '(GMT+12:00) Funafuti', value: 'Pacific/Funafuti' }, - { label: '(GMT+12:00) Kwajalein', value: 'Pacific/Kwajalein' }, - { label: '(GMT+12:00) Majuro', value: 'Pacific/Majuro' }, - { label: '(GMT+12:00) Moscow+09 - Petropavlovsk-Kamchatskiy', value: 'Asia/Kamchatka' }, - { label: '(GMT+12:00) Nauru', value: 'Pacific/Nauru' }, - { label: '(GMT+12:00) Tarawa', value: 'Pacific/Tarawa' }, - { label: '(GMT+12:00) Wake', value: 'Pacific/Wake' }, - { label: '(GMT+12:00) Wallis', value: 'Pacific/Wallis' }, - { label: '(GMT+13:00) Auckland', value: 'Pacific/Auckland' }, - { label: '(GMT+13:00) Enderbury', value: 'Pacific/Enderbury' }, - { label: '(GMT+13:00) Fakaofo', value: 'Pacific/Fakaofo' }, - { label: '(GMT+13:00) Fiji', value: 'Pacific/Fiji' }, - { label: '(GMT+13:00) Tongatapu', value: 'Pacific/Tongatapu' }, - { label: '(GMT+14:00) Apia', value: 'Pacific/Apia' }, - { label: '(GMT+14:00) Kiritimati', value: 'Pacific/Kiritimati' } + { label: 'Pacific - Niue', value: 'Pacific/Niue' }, + { label: 'Pacific - Pago Pago', value: 'Pacific/Pago_Pago' }, + { label: 'Pacific - Hawaii Time', value: 'Pacific/Honolulu' }, + { label: 'Pacific - Rarotonga', value: 'Pacific/Rarotonga' }, + { label: 'Pacific - Tahiti', value: 'Pacific/Tahiti' }, + { label: 'Pacific - Marquesas', value: 'Pacific/Marquesas' }, + { label: 'America - Alaska Time', value: 'America/Anchorage' }, + { label: 'Pacific - Gambier', value: 'Pacific/Gambier' }, + { label: 'America - Pacific Time - Los Angeles', value: 'America/Los_Angeles' }, + { label: 'America - Pacific Time - Tijuana', value: 'America/Tijuana' }, + { label: 'America - Pacific Time - Vancouver', value: 'America/Vancouver' }, + { label: 'America - Pacific Time - Whitehorse', value: 'America/Whitehorse' }, + { label: 'Pacific - Pitcairn', value: 'Pacific/Pitcairn' }, + { label: 'America - Mountain Time - Arizona', value: 'America/Phoenix' }, + { label: 'America - Mountain Time - Chihuahua, Mazatlan', value: 'America/Mazatlan' }, + { label: 'America - Mountain Time - Dawson Creek', value: 'America/Dawson_Creek' }, + { label: 'America - Mountain Time - Denver', value: 'America/Denver' }, + { label: 'America - Mountain Time - Edmonton', value: 'America/Edmonton' }, + { label: 'America - Mountain Time - Hermosillo', value: 'America/Hermosillo' }, + { label: 'America - Mountain Time - Yellowknife', value: 'America/Yellowknife' }, + { label: 'America - Belize', value: 'America/Belize' }, + { label: 'America - Central Time - Chicago', value: 'America/Chicago' }, + { label: 'America - Central Time - Mexico City', value: 'America/Mexico_City' }, + { label: 'America - Central Time - Regina', value: 'America/Regina' }, + { label: 'America - Central Time - Tegucigalpa', value: 'America/Tegucigalpa' }, + { label: 'America - Central Time - Winnipeg', value: 'America/Winnipeg' }, + { label: 'America - Costa Rica', value: 'America/Costa_Rica' }, + { label: 'America - El Salvador', value: 'America/El_Salvador' }, + { label: 'Pacific - Galapagos', value: 'Pacific/Galapagos' }, + { label: 'America - Guatemala', value: 'America/Guatemala' }, + { label: 'America - Managua', value: 'America/Managua' }, + { label: 'America - America Cancun', value: 'America/Cancun' }, + { label: 'America - Bogota', value: 'America/Bogota' }, + { label: 'Pacific - Easter Island', value: 'Pacific/Easter' }, + { label: 'America - Eastern Time - New York', value: 'America/New_York' }, + { label: 'America - Eastern Time - Iqaluit', value: 'America/Iqaluit' }, + { label: 'America - Eastern Time - Toronto', value: 'America/Toronto' }, + { label: 'America - Guayaquil', value: 'America/Guayaquil' }, + { label: 'America - Havana', value: 'America/Havana' }, + { label: 'America - Jamaica', value: 'America/Jamaica' }, + { label: 'America - Lima', value: 'America/Lima' }, + { label: 'America - Nassau', value: 'America/Nassau' }, + { label: 'America - Panama', value: 'America/Panama' }, + { label: 'America - Port-au-Prince', value: 'America/Port-au-Prince' }, + { label: 'America - Rio Branco', value: 'America/Rio_Branco' }, + { label: 'America - Atlantic Time - Halifax', value: 'America/Halifax' }, + { label: 'America - Barbados', value: 'America/Barbados' }, + { label: 'Atlantic - Bermuda', value: 'Atlantic/Bermuda' }, + { label: 'America - Boa Vista', value: 'America/Boa_Vista' }, + { label: 'America - Caracas', value: 'America/Caracas' }, + { label: 'America - Curacao', value: 'America/Curacao' }, + { label: 'America - Grand Turk', value: 'America/Grand_Turk' }, + { label: 'America - Guyana', value: 'America/Guyana' }, + { label: 'America - La Paz', value: 'America/La_Paz' }, + { label: 'America - Manaus', value: 'America/Manaus' }, + { label: 'America - Martinique', value: 'America/Martinique' }, + { label: 'America - Port of Spain', value: 'America/Port_of_Spain' }, + { label: 'America - Porto Velho', value: 'America/Porto_Velho' }, + { label: 'America - Puerto Rico', value: 'America/Puerto_Rico' }, + { label: 'America - Santo Domingo', value: 'America/Santo_Domingo' }, + { label: 'America - Thule', value: 'America/Thule' }, + { label: 'America - Newfoundland Time - St. Johns', value: 'America/St_Johns' }, + { label: 'America - Araguaina', value: 'America/Araguaina' }, + { label: 'America - Asuncion', value: 'America/Asuncion' }, + { label: 'America - Belem', value: 'America/Belem' }, + { label: 'America - Buenos Aires', value: 'America/Argentina/Buenos_Aires' }, + { label: 'America - Campo Grande', value: 'America/Campo_Grande' }, + { label: 'America - Cayenne', value: 'America/Cayenne' }, + { label: 'America - Cuiaba', value: 'America/Cuiaba' }, + { label: 'America - Fortaleza', value: 'America/Fortaleza' }, + { label: 'America - Godthab', value: 'America/Godthab' }, + { label: 'America - Maceio', value: 'America/Maceio' }, + { label: 'America - Miquelon', value: 'America/Miquelon' }, + { label: 'America - Montevideo', value: 'America/Montevideo' }, + { label: 'Antarctica - Palmer', value: 'Antarctica/Palmer' }, + { label: 'America - Paramaribo', value: 'America/Paramaribo' }, + { label: 'America - Punta Arenas', value: 'America/Punta_Arenas' }, + { label: 'America - Recife', value: 'America/Recife' }, + { label: 'Antarctica - Rothera', value: 'Antarctica/Rothera' }, + { label: 'America - Salvador', value: 'America/Bahia' }, + { label: 'America - Santiago', value: 'America/Santiago' }, + { label: 'Atlantic - Stanley', value: 'Atlantic/Stanley' }, + { label: 'America - Noronha', value: 'America/Noronha' }, + { label: 'America - Sao Paulo', value: 'America/Sao_Paulo' }, + { label: 'Atlantic - South Georgia', value: 'Atlantic/South_Georgia' }, + { label: 'Atlantic - Azores', value: 'Atlantic/Azores' }, + { label: 'Atlantic - Cape Verde', value: 'Atlantic/Cape_Verde' }, + { label: 'America - Scoresbysund', value: 'America/Scoresbysund' }, + { label: 'Africa - Abidjan', value: 'Africa/Abidjan' }, + { label: 'Africa - Accra', value: 'Africa/Accra' }, + { label: 'Africa - Bissau', value: 'Africa/Bissau' }, + { label: 'Atlantic - Canary Islands', value: 'Atlantic/Canary' }, + { label: 'Africa - Casablanca', value: 'Africa/Casablanca' }, + { label: 'America - Danmarkshavn', value: 'America/Danmarkshavn' }, + { label: 'Europe - Dublin', value: 'Europe/Dublin' }, + { label: 'Africa - El Aaiun', value: 'Africa/El_Aaiun' }, + { label: 'Atlantic - Faeroe', value: 'Atlantic/Faroe' }, + { label: 'UTC / GMT', value: 'UTC' }, + { label: 'Europe - Lisbon', value: 'Europe/Lisbon' }, + { label: 'Europe - London', value: 'Europe/London' }, + { label: 'Africa - Monrovia', value: 'Africa/Monrovia' }, + { label: 'Atlantic - Reykjavik', value: 'Atlantic/Reykjavik' }, + { label: 'Africa - Algiers', value: 'Africa/Algiers' }, + { label: 'Europe - Amsterdam', value: 'Europe/Amsterdam' }, + { label: 'Europe - Andorra', value: 'Europe/Andorra' }, + { label: 'Europe - Berlin', value: 'Europe/Berlin' }, + { label: 'Europe - Brussels', value: 'Europe/Brussels' }, + { label: 'Europe - Budapest', value: 'Europe/Budapest' }, + { label: 'Europe - Central European Time - Belgrade', value: 'Europe/Belgrade' }, + { label: 'Europe - Central European Time - Prague', value: 'Europe/Prague' }, + { label: 'Africa - Ceuta', value: 'Africa/Ceuta' }, + { label: 'Europe - Copenhagen', value: 'Europe/Copenhagen' }, + { label: 'Europe - Gibraltar', value: 'Europe/Gibraltar' }, + { label: 'Africa - Lagos', value: 'Africa/Lagos' }, + { label: 'Europe - Luxembourg', value: 'Europe/Luxembourg' }, + { label: 'Europe - Madrid', value: 'Europe/Madrid' }, + { label: 'Europe - Malta', value: 'Europe/Malta' }, + { label: 'Europe - Monaco', value: 'Europe/Monaco' }, + { label: 'Africa - Ndjamena', value: 'Africa/Ndjamena' }, + { label: 'Europe - Oslo', value: 'Europe/Oslo' }, + { label: 'Europe - Paris', value: 'Europe/Paris' }, + { label: 'Europe - Rome', value: 'Europe/Rome' }, + { label: 'Europe - Stockholm', value: 'Europe/Stockholm' }, + { label: 'Europe - Tirane', value: 'Europe/Tirane' }, + { label: 'Africa - Tunis', value: 'Africa/Tunis' }, + { label: 'Europe - Vienna', value: 'Europe/Vienna' }, + { label: 'Europe - Warsaw', value: 'Europe/Warsaw' }, + { label: 'Europe - Zurich', value: 'Europe/Zurich' }, + { label: 'Asia - Amman', value: 'Asia/Amman' }, + { label: 'Europe - Athens', value: 'Europe/Athens' }, + { label: 'Asia - Beirut', value: 'Asia/Beirut' }, + { label: 'Europe - Bucharest', value: 'Europe/Bucharest' }, + { label: 'Africa - Cairo', value: 'Africa/Cairo' }, + { label: 'Europe - Chisinau', value: 'Europe/Chisinau' }, + { label: 'Asia - Damascus', value: 'Asia/Damascus' }, + { label: 'Asia - Gaza', value: 'Asia/Gaza' }, + { label: 'Europe - Helsinki', value: 'Europe/Helsinki' }, + { label: 'Asia - Jerusalem', value: 'Asia/Jerusalem' }, + { label: 'Africa - Johannesburg', value: 'Africa/Johannesburg' }, + { label: 'Africa - Khartoum', value: 'Africa/Khartoum' }, + { label: 'Europe - Kiev', value: 'Europe/Kiev' }, + { label: 'Africa - Maputo', value: 'Africa/Maputo' }, + { label: 'Europe - Moscow-01 - Kaliningrad', value: 'Europe/Kaliningrad' }, + { label: 'Asia - Nicosia', value: 'Asia/Nicosia' }, + { label: 'Europe - Riga', value: 'Europe/Riga' }, + { label: 'Europe - Sofia', value: 'Europe/Sofia' }, + { label: 'Europe - Tallinn', value: 'Europe/Tallinn' }, + { label: 'Africa - Tripoli', value: 'Africa/Tripoli' }, + { label: 'Europe - Vilnius', value: 'Europe/Vilnius' }, + { label: 'Africa - Windhoek', value: 'Africa/Windhoek' }, + { label: 'Asia - Baghdad', value: 'Asia/Baghdad' }, + { label: 'Europe - Istanbul', value: 'Europe/Istanbul' }, + { label: 'Europe - Minsk', value: 'Europe/Minsk' }, + { label: 'Europe - Moscow+00 - Moscow', value: 'Europe/Moscow' }, + { label: 'Africa - Nairobi', value: 'Africa/Nairobi' }, + { label: 'Asia - Qatar', value: 'Asia/Qatar' }, + { label: 'Asia - Riyadh', value: 'Asia/Riyadh' }, + { label: 'Antarctica - Syowa', value: 'Antarctica/Syowa' }, + { label: 'Asia - Tehran', value: 'Asia/Tehran' }, + { label: 'Asia - Baku', value: 'Asia/Baku' }, + { label: 'Asia - Dubai', value: 'Asia/Dubai' }, + { label: 'Indian - Mahe', value: 'Indian/Mahe' }, + { label: 'Indian - Mauritius', value: 'Indian/Mauritius' }, + { label: 'Europe - Moscow+01 - Samara', value: 'Europe/Samara' }, + { label: 'Indian - Reunion', value: 'Indian/Reunion' }, + { label: 'Asia - Tbilisi', value: 'Asia/Tbilisi' }, + { label: 'Asia - Yerevan', value: 'Asia/Yerevan' }, + { label: 'Asia - Kabul', value: 'Asia/Kabul' }, + { label: 'Asia - Aqtau', value: 'Asia/Aqtau' }, + { label: 'Asia - Aqtobe', value: 'Asia/Aqtobe' }, + { label: 'Asia - Ashgabat', value: 'Asia/Ashgabat' }, + { label: 'Asia - Dushanbe', value: 'Asia/Dushanbe' }, + { label: 'Asia - Karachi', value: 'Asia/Karachi' }, + { label: 'Indian - Kerguelen', value: 'Indian/Kerguelen' }, + { label: 'Indian - Maldives', value: 'Indian/Maldives' }, + { label: 'Antarctica - Mawson', value: 'Antarctica/Mawson' }, + { label: 'Asia - Moscow+02 - Yekaterinburg', value: 'Asia/Yekaterinburg' }, + { label: 'Asia - Tashkent', value: 'Asia/Tashkent' }, + { label: 'Asia - Colombo', value: 'Asia/Colombo' }, + { label: 'Asia - India Standard Time', value: 'Asia/Kolkata' }, + { label: 'Asia - Kathmandu', value: 'Asia/Kathmandu' }, + { label: 'Asia - Almaty', value: 'Asia/Almaty' }, + { label: 'Asia - Bishkek', value: 'Asia/Bishkek' }, + { label: 'Indian - Chagos', value: 'Indian/Chagos' }, + { label: 'Asia - Dhaka', value: 'Asia/Dhaka' }, + { label: 'Asia - Moscow+03 - Omsk', value: 'Asia/Omsk' }, + { label: 'Asia - Thimphu', value: 'Asia/Thimphu' }, + { label: 'Antarctica - Vostok', value: 'Antarctica/Vostok' }, + { label: 'Indian - Cocos', value: 'Indian/Cocos' }, + { label: 'Asia - Rangoon', value: 'Asia/Yangon' }, + { label: 'Asia - Bangkok', value: 'Asia/Bangkok' }, + { label: 'Indian - Christmas', value: 'Indian/Christmas' }, + { label: 'Antarctica - Davis', value: 'Antarctica/Davis' }, + { label: 'Asia - Hanoi', value: 'Asia/Saigon' }, + { label: 'Asia - Hovd', value: 'Asia/Hovd' }, + { label: 'Asia - Jakarta', value: 'Asia/Jakarta' }, + { label: 'Asia - Moscow+04 - Krasnoyarsk', value: 'Asia/Krasnoyarsk' }, + { label: 'Asia - Brunei', value: 'Asia/Brunei' }, + { label: 'Asia - China Time - Beijing', value: 'Asia/Shanghai' }, + { label: 'Asia - Choibalsan', value: 'Asia/Choibalsan' }, + { label: 'Asia - Hong Kong', value: 'Asia/Hong_Kong' }, + { label: 'Asia - Kuala Lumpur', value: 'Asia/Kuala_Lumpur' }, + { label: 'Asia - Macau', value: 'Asia/Macau' }, + { label: 'Asia - Makassar', value: 'Asia/Makassar' }, + { label: 'Asia - Manila', value: 'Asia/Manila' }, + { label: 'Asia - Moscow+05 - Irkutsk', value: 'Asia/Irkutsk' }, + { label: 'Asia - Singapore', value: 'Asia/Singapore' }, + { label: 'Asia - Taipei', value: 'Asia/Taipei' }, + { label: 'Asia - Ulaanbaatar', value: 'Asia/Ulaanbaatar' }, + { label: 'Australia - Western Time - Perth', value: 'Australia/Perth' }, + { label: 'Asia - Pyongyang', value: 'Asia/Pyongyang' }, + { label: 'Asia - Dili', value: 'Asia/Dili' }, + { label: 'Asia - Jayapura', value: 'Asia/Jayapura' }, + { label: 'Asia - Moscow+06 - Yakutsk', value: 'Asia/Yakutsk' }, + { label: 'Pacific - Palau', value: 'Pacific/Palau' }, + { label: 'Asia - Seoul', value: 'Asia/Seoul' }, + { label: 'Asia - Tokyo', value: 'Asia/Tokyo' }, + { label: 'Australia - Central Time - Darwin', value: 'Australia/Darwin' }, + { label: 'Antarctica - Dumont D\'Urville', value: 'Antarctica/DumontDUrville' }, + { label: 'Australia - Eastern Time - Brisbane', value: 'Australia/Brisbane' }, + { label: 'Pacific - Guam', value: 'Pacific/Guam' }, + { label: 'Asia - Moscow+07 - Vladivostok', value: 'Asia/Vladivostok' }, + { label: 'Pacific - Port Moresby', value: 'Pacific/Port_Moresby' }, + { label: 'Pacific - Truk', value: 'Pacific/Chuuk' }, + { label: 'Australia - Central Time - Adelaide', value: 'Australia/Adelaide' }, + { label: 'Antarctica - Casey', value: 'Antarctica/Casey' }, + { label: 'Australia - Eastern Time - Hobart', value: 'Australia/Hobart' }, + { label: 'Australia - Eastern Time - Melbourne, Sydney', value: 'Australia/Sydney' }, + { label: 'Pacific - Efate', value: 'Pacific/Efate' }, + { label: 'Pacific - Guadalcanal', value: 'Pacific/Guadalcanal' }, + { label: 'Pacific - Kosrae', value: 'Pacific/Kosrae' }, + { label: 'Asia - Moscow+08 - Magadan', value: 'Asia/Magadan' }, + { label: 'Pacific - Norfolk', value: 'Pacific/Norfolk' }, + { label: 'Pacific - Noumea', value: 'Pacific/Noumea' }, + { label: 'Pacific - Ponape', value: 'Pacific/Pohnpei' }, + { label: 'Pacific - Funafuti', value: 'Pacific/Funafuti' }, + { label: 'Pacific - Kwajalein', value: 'Pacific/Kwajalein' }, + { label: 'Pacific - Majuro', value: 'Pacific/Majuro' }, + { label: 'Asia - Moscow+09 - Petropavlovsk-Kamchatskiy', value: 'Asia/Kamchatka' }, + { label: 'Pacific - Nauru', value: 'Pacific/Nauru' }, + { label: 'Pacific - Tarawa', value: 'Pacific/Tarawa' }, + { label: 'Pacific - Wake', value: 'Pacific/Wake' }, + { label: 'Pacific - Wallis', value: 'Pacific/Wallis' }, + { label: 'Pacific - Auckland', value: 'Pacific/Auckland' }, + { label: 'Pacific - Enderbury', value: 'Pacific/Enderbury' }, + { label: 'Pacific - Fakaofo', value: 'Pacific/Fakaofo' }, + { label: 'Pacific - Fiji', value: 'Pacific/Fiji' }, + { label: 'Pacific - Tongatapu', value: 'Pacific/Tongatapu' }, + { label: 'Pacific - Apia', value: 'Pacific/Apia' }, + { label: 'Pacific - Kiritimati', value: 'Pacific/Kiritimati' } ] diff --git a/client/shared/urls.json b/client/shared/urls.json index 285caa07d2..15410d68df 100644 --- a/client/shared/urls.json +++ b/client/shared/urls.json @@ -1,5 +1,6 @@ { "bofDefinition": "https://www.ietf.org/how/bofs/", + "hackathonWiki": "https://wiki.ietf.org/meeting/{meetingNumber}/hackathon", "meetingCalIcs": "/meeting/{meetingNumber}/agenda.ics", "meetingDetails": "/meeting/{meetingNumber}/session/{eventAcronym}/", "meetingMaterialsPdf": "/meeting/{meetingNumber}/agenda/{eventAcronym}-drafts.pdf", diff --git a/client/shared/xslugify.js b/client/shared/xslugify.js index daf0bdf2ba..e1ac556ddf 100644 --- a/client/shared/xslugify.js +++ b/client/shared/xslugify.js @@ -1,5 +1,5 @@ import slugify from 'slugify' export default (str) => { - return slugify(str.replace('/', '-'), { lower: true }) + return slugify(str.replaceAll('/', '-').replaceAll(/['&]/g, ''), { lower: true }) } diff --git a/debug.py b/debug.py index bf34367cce..4f0d64bae2 100644 --- a/debug.py +++ b/debug.py @@ -3,15 +3,7 @@ import sys import time as timeutils import inspect -from typing import Callable -try: - import syslog - logger = syslog.syslog # type: Callable -except ImportError: # import syslog will fail on Windows boxes - import logging - logging.basicConfig(filename='tracker.log',level=logging.INFO) - logger = logging.info try: from pprint import pformat @@ -55,7 +47,7 @@ def fix(s,n=64): if len(s) > n+3: s = s[:n]+"..." return s - def wrap(fn, *params,**kwargs): + def wrap(*params,**kwargs): call = wrap.callcount = wrap.callcount + 1 indent = ' ' * _report_indent[0] @@ -81,8 +73,8 @@ def wrap(fn, *params,**kwargs): return ret wrap.callcount = 0 if debug: - from decorator import decorator - return decorator(wrap, fn) + from functools import update_wrapper + return update_wrapper(wrap, fn) else: return fn @@ -119,7 +111,7 @@ def clock(s): def time(fn): """Decorator to print timing information about a function call. """ - def wrap(fn, *params,**kwargs): + def wrap(*params,**kwargs): indent = ' ' * _report_indent[0] fc = "%s.%s()" % (fn.__module__, fn.__name__,) @@ -132,8 +124,8 @@ def wrap(fn, *params,**kwargs): return ret wrap.callcount = 0 if debug: - from decorator import decorator - return decorator(wrap, fn) + from functools import update_wrapper + return update_wrapper(wrap, fn) else: return fn @@ -155,13 +147,6 @@ def showpos(name): indent = ' ' * (_report_indent[0]) sys.stderr.write("%s%s:%s: %s: '%s'\n" % (indent, fn, line, name, value)) -def log(name): - if debug: - frame = inspect.stack()[1][0] - value = eval(name, frame.f_globals, frame.f_locals) - indent = ' ' * (_report_indent[0]) - logger("%s%s: %s" % (indent, name, value)) - def pprint(name): if debug: frame = inspect.stack()[1][0] @@ -190,7 +175,7 @@ def type(name): value = eval(name, frame.f_globals, frame.f_locals) indent = ' ' * (_report_indent[0]) sys.stderr.write("%s%s: %s\n" % (indent, name, value)) - + def say(s): if debug: indent = ' ' * (_report_indent[0]) @@ -205,11 +190,11 @@ def wrapper(*args, **kwargs): prof.dump_stats(datafn) return retval if debug: - from decorator import decorator - return decorator(wrapper, fn) + from functools import update_wrapper + return update_wrapper(wrapper, fn) else: return fn - + def traceback(levels=None): if debug: indent = ' ' * (_report_indent[0]) diff --git a/dev/INSTALL b/dev/INSTALL deleted file mode 100644 index 15c7472972..0000000000 --- a/dev/INSTALL +++ /dev/null @@ -1,155 +0,0 @@ -============================================================================== - IETF Datatracker -============================================================================== - ------------------------------------------------------------------------------- - Installation Instructions ------------------------------------------------------------------------------- - -General Instructions for Deployment of a New Release -==================================================== - - 0. Prepare to hold different roles at different stages of the instructions below. - You will need to be root, wwwrun, and some user in group docker. - Consider using separate shells for the wwwrun and other roles. These instructions - are written assuming you will only use one shell. - - 1. Make a directory to hold the new release as wwwrun:: - sudo su - -s /bin/bash wwwrun - mkdir /a/www/ietf-datatracker/${releasenumber} - cd /a/www/ietf-datatracker/${releasenumber} - - 2. Fetch the release tarball from github - (see https://github.com/ietf-tools/datatracker/releases):: - - wget https://github.com/ietf-tools/datatracker/releases/download/${releasenumber}/release.tar.gz - tar xzvf release.tar.gz - - 3. Copy ietf/settings_local.py from previous release:: - - cp ../web/ietf/settings_local.py ietf/ - - 4. Setup a new virtual environment and install requirements:: - - python3.9 -mvenv env - source env/bin/activate - pip install -r requirements.txt - pip freeze > frozen-requirements.txt - - (The pip freeze command records the exact versions of the Python libraries that pip installed. - This is used by the celery docker container to ensure it uses the same library versions as - the datatracker service.) - - 5. Move static files into place for CDN (/a/www/www6s/lib/dt): - - ietf/manage.py collectstatic - - 6. Run system checks (which patches the just installed modules):: - - ietf/manage.py check - - 7. Switch to the docker directory and update images as a user in group docker: - - exit - cd /a/docker/datatracker - docker image tag ghcr.io/ietf-tools/datatracker-celery:latest datatracker-celery-fallback - docker image tag ghcr.io/ietf-tools/datatracker-mq:latest datatracker-mq-fallback - docker-compose pull - - 8. Stop and remove the async task containers: - Wait for this to finish cleanly. Usually this will only be a few seconds, but it may take up - to about 10 minutes for the 'down' command to complete if a long-running task is in progress. - - docker-compose down - - 9. Stop the datatracker - - sudo systemctl stop datatracker.socket datatracker.service - - 10. Return to the release directory and run migrations as wwwrun: - - sudo su - -s /bin/bash wwwrun - cd /a/www/ietf-datatracker/${releasenumber} - ietf/manage.py migrate - - Take note if any migrations were executed. - - 11. Back out one directory level, then re-point the 'web' symlink:: - - cd .. - rm ./web; ln -s ${releasenumber} web - - 12. Start the datatracker service (it is no longer necessary to restart apache) :: - - exit - sudo systemctl start datatracker.service datatracker.socket - - 13. Start async task worker and message broker: - - cd /a/docker/datatracker - bash startcommand - - 14. Verify operation: - - http://datatracker.ietf.org/ - - 15. If install failed and there were no migrations at step 9, revert web symlink and docker update and repeat the - restart in steps 11 and 12. To revert the docker update: - - cd /a/docker/datatracker - docker-compose down - docker image rm ghcr.io/ietf-tools/datatracker-celery:latest ghcr.io/ietf-tools/datatracker-mq:latest - docker image tag datatracker-celery-fallback ghcr.io/ietf-tools/datatracker-celery:latest - docker image tag datatracker-mq-fallback ghcr.io/ietf-tools/datatracker-mq:latest - cd - - - If there were migrations at step 10, they will need to be reversed before the restart at step 12. - If it's not obvious what to do to reverse the migrations, contact the dev team. - - -Patching a Production Release -============================= - -Sometimes it can prove necessary to patch an existing release. -The following process should be used: - - 1. Code and test the patch on an copy of the release with any - previously applied patches put in place. - - 2. Produce a patch file, named with date and subject:: - - $ git diff > 2013-03-25-ballot-calculation.patch - - 3. Move the patch file to the production server, and place it in - '/a/www/ietf-datatracker/patches/' - - 4. Make a recursive copy of the production code to a new directory, named with a patch number. - - /a/www/ietf-datatracker $ rsync -a web/ ${releasenumber}.p1/ - - 5. Apply the patch:: - - /a/www/ietf-datatracker $ cd ${releasenumber}.p1/ - /a/www/ietf-datatracker/${releasnumber}.p1 $ patch -p1 \ - < ../patches/2013-03-25-ballot-calculation.patch - - This must not produce any messages about failing to apply any chunks; - if it does, go back to 1. and figure out why. - - 6. Edit ``.../ietf/__init__.py`` in the new patched release to indicate the patch - version in the ``__patch__`` string. - - 7. Stop the async task container (this may take a few minutes if tasks are in progress): - - cd /a/docker/datatracker - docker-compose down - - 8. Change the 'web' symlink, reload etc. as described in - `General Instructions for Deployment of a New Release`_. - - 9. Start async task worker: - - cd /a/docker/datatracker - bash startcommand - - diff --git a/dev/build/Dockerfile b/dev/build/Dockerfile new file mode 100644 index 0000000000..e57fecd5f2 --- /dev/null +++ b/dev/build/Dockerfile @@ -0,0 +1,41 @@ +FROM ghcr.io/ietf-tools/datatracker-app-base:20260410T1557 +LABEL maintainer="IETF Tools Team " + +ENV DEBIAN_FRONTEND=noninteractive + +# uid 498 = wwwrun and gid 496 = www on ietfa +RUN groupadd -g 1000 datatracker && \ + useradd -c "Datatracker User" -u 1000 -g datatracker -m -s /bin/false datatracker + +RUN apt-get purge -y imagemagick imagemagick-6-common + +# Install libreoffice (needed via PPT2PDF_COMMAND) +RUN apt-get update && \ + apt-get -qy install libreoffice-nogui + +COPY . . +COPY ./dev/build/start.sh ./start.sh +COPY ./dev/build/datatracker-start.sh ./datatracker-start.sh +COPY ./dev/build/migration-start.sh ./migration-start.sh +COPY ./dev/build/celery-start.sh ./celery-start.sh +COPY ./dev/build/gunicorn.conf.py ./gunicorn.conf.py + +RUN pip3 --disable-pip-version-check --no-cache-dir install -r requirements.txt && \ + echo '# empty' > ietf/settings_local.py && \ + ietf/manage.py patch_libraries && \ + rm -f ietf/settings_local.py + +RUN chmod +x start.sh && \ + chmod +x datatracker-start.sh && \ + chmod +x migration-start.sh && \ + chmod +x celery-start.sh && \ + chmod +x docker/scripts/app-create-dirs.sh && \ + sh ./docker/scripts/app-create-dirs.sh + +RUN mkdir -p /a + +VOLUME [ "/a" ] + +EXPOSE 8000 + +CMD ["./start.sh"] diff --git a/dev/build/TARGET_BASE b/dev/build/TARGET_BASE new file mode 100644 index 0000000000..f430037c09 --- /dev/null +++ b/dev/build/TARGET_BASE @@ -0,0 +1 @@ +20260410T1557 diff --git a/dev/build/celery-start.sh b/dev/build/celery-start.sh new file mode 100644 index 0000000000..69dcd7bbda --- /dev/null +++ b/dev/build/celery-start.sh @@ -0,0 +1,52 @@ +#!/bin/bash -e +# +# Run a celery worker +# +echo "Running Datatracker checks..." +./ietf/manage.py check + +# Check whether the blobdb database exists - inspectdb will return a false +# status if not. +if ietf/manage.py inspectdb --database blobdb > /dev/null 2>&1; then + HAVE_BLOBDB="yes" +fi + +migrations_applied_for () { + local DATABASE=${1:-default} + ietf/manage.py migrate --check --database "$DATABASE" +} + +migrations_all_applied () { + if [[ "$HAVE_BLOBDB" == "yes" ]]; then + migrations_applied_for default && migrations_applied_for blobdb + else + migrations_applied_for default + fi +} + +if ! migrations_all_applied; then + echo "Unapplied migrations found, waiting to start..." + sleep 5 + while ! migrations_all_applied ; do + echo "... still waiting for migrations..." + sleep 5 + done +fi + +echo "Starting Celery..." + +cleanup () { + # Cleanly terminate the celery app by sending it a TERM, then waiting for it to exit. + if [[ -n "${celery_pid}" ]]; then + echo "Gracefully terminating celery worker. This may take a few minutes if tasks are in progress..." + kill -TERM "${celery_pid}" + wait "${celery_pid}" + fi +} + +trap 'trap "" TERM; cleanup' TERM + +# start celery in the background so we can trap the TERM signal +celery "$@" & +celery_pid=$! +wait "${celery_pid}" diff --git a/dev/build/collectstatics.sh b/dev/build/collectstatics.sh new file mode 100644 index 0000000000..44f1c608a9 --- /dev/null +++ b/dev/build/collectstatics.sh @@ -0,0 +1,13 @@ +#!/bin/bash + +# Copy temp local settings +cp dev/build/settings_local_collectstatics.py ietf/settings_local.py + +# Install Python dependencies +pip --disable-pip-version-check --no-cache-dir install -r requirements.txt + +# Collect statics +ietf/manage.py collectstatic + +# Delete temp local settings +rm ietf/settings_local.py \ No newline at end of file diff --git a/dev/build/datatracker-start.sh b/dev/build/datatracker-start.sh new file mode 100644 index 0000000000..a676415a26 --- /dev/null +++ b/dev/build/datatracker-start.sh @@ -0,0 +1,60 @@ +#!/bin/bash -e + +echo "Running Datatracker checks..." +./ietf/manage.py check + +# Check whether the blobdb database exists - inspectdb will return a false +# status if not. +if ietf/manage.py inspectdb --database blobdb > /dev/null 2>&1; then + HAVE_BLOBDB="yes" +fi + +migrations_applied_for () { + local DATABASE=${1:-default} + ietf/manage.py migrate --check --database "$DATABASE" +} + +migrations_all_applied () { + if [[ "$HAVE_BLOBDB" == "yes" ]]; then + migrations_applied_for default && migrations_applied_for blobdb + else + migrations_applied_for default + fi +} + +if ! migrations_all_applied; then + echo "Unapplied migrations found, waiting to start..." + sleep 5 + while ! migrations_all_applied ; do + echo "... still waiting for migrations..." + sleep 5 + done +fi + +echo "Starting Datatracker..." + +# trap TERM and shut down gunicorn +cleanup () { + if [[ -n "${gunicorn_pid}" ]]; then + echo "Terminating gunicorn..." + kill -TERM "${gunicorn_pid}" + wait "${gunicorn_pid}" + fi +} + +trap 'trap "" TERM; cleanup' TERM + +# start gunicorn in the background so we can trap the TERM signal +gunicorn \ + -c /workspace/gunicorn.conf.py \ + --workers "${DATATRACKER_GUNICORN_WORKERS:-9}" \ + --max-requests "${DATATRACKER_GUNICORN_MAX_REQUESTS:-32768}" \ + --timeout "${DATATRACKER_GUNICORN_TIMEOUT:-180}" \ + --bind :8000 \ + --log-level "${DATATRACKER_GUNICORN_LOG_LEVEL:-info}" \ + --capture-output \ + --access-logfile -\ + ${DATATRACKER_GUNICORN_EXTRA_ARGS} \ + ietf.wsgi:application & +gunicorn_pid=$! +wait "${gunicorn_pid}" diff --git a/dev/deploy/exclude-patterns.txt b/dev/build/exclude-patterns.txt similarity index 100% rename from dev/deploy/exclude-patterns.txt rename to dev/build/exclude-patterns.txt diff --git a/dev/build/gunicorn.conf.py b/dev/build/gunicorn.conf.py new file mode 100644 index 0000000000..9af4478685 --- /dev/null +++ b/dev/build/gunicorn.conf.py @@ -0,0 +1,164 @@ +# Copyright The IETF Trust 2024-2025, All Rights Reserved + +import os +import ietf +from opentelemetry import trace +from opentelemetry.exporter.otlp.proto.http.trace_exporter import OTLPSpanExporter +from opentelemetry.sdk.resources import Resource +from opentelemetry.sdk.trace import TracerProvider +from opentelemetry.sdk.trace.export import BatchSpanProcessor +from opentelemetry.instrumentation.django import DjangoInstrumentor +from opentelemetry.instrumentation.psycopg2 import Psycopg2Instrumentor +from opentelemetry.instrumentation.pymemcache import PymemcacheInstrumentor +from opentelemetry.instrumentation.requests import RequestsInstrumentor + +# Configure security scheme headers for forwarded requests. Cloudflare sets X-Forwarded-Proto +# for us. Don't trust any of the other similar headers. Only trust the header if it's coming +# from localhost, as all legitimate traffic will reach gunicorn via co-located nginx. +secure_scheme_headers = {"X-FORWARDED-PROTO": "https"} +forwarded_allow_ips = "127.0.0.1, ::1" # this is the default + +# Log as JSON on stdout (to distinguish from Django's logs on stderr) +# +# This is applied as an update to gunicorn's glogging.CONFIG_DEFAULTS. +logconfig_dict = { + "version": 1, + "disable_existing_loggers": False, + "root": {"level": "INFO", "handlers": ["console"]}, + "loggers": { + "gunicorn.error": { + "level": "INFO", + "handlers": ["console"], + "propagate": False, + "qualname": "gunicorn.error", + }, + "gunicorn.access": { + "level": "INFO", + "handlers": ["access_console"], + "propagate": False, + "qualname": "gunicorn.access", + }, + }, + "handlers": { + "console": { + "class": "logging.StreamHandler", + "formatter": "json", + "stream": "ext://sys.stdout", + }, + "access_console": { + "class": "logging.StreamHandler", + "formatter": "access_json", + "stream": "ext://sys.stdout", + }, + }, + "formatters": { + "json": { + "class": "ietf.utils.jsonlogger.DatatrackerJsonFormatter", + "style": "{", + "format": "{asctime}{levelname}{message}{name}{process}", + }, + "access_json": { + "class": "ietf.utils.jsonlogger.GunicornRequestJsonFormatter", + "style": "{", + "format": "{asctime}{levelname}{message}{name}{process}", + }, + }, +} + +# Track in-flight requests and emit a list of what was happeningwhen a worker is terminated. +# For the default sync worker, there will only be one request per PID, but allow for the +# possibility of multiple requests in case we switch to a different worker class. +# +# This dict is only visible within a single worker, but key by pid to guarantee no conflicts. +# +# Use a list rather than a set to allow for the possibility of overlapping identical requests. +in_flight_by_pid: dict[str, list[str]] = {} # pid -> list of in-flight requests + + +def _describe_request(req): + """Generate a consistent description of a request + + The return value is used identify in-flight requests, so it must not vary between the + start and end of handling a request. E.g., do not include a timestamp. + """ + client_ip = "-" + asn = "-" + cf_ray = "-" + for header, value in req.headers: + header = header.lower() + if header == "cf-connecting-ip": + client_ip = value + elif header == "x-ip-src-asnum": + asn = value + elif header == "cf-ray": + cf_ray = value + if req.query: + path = f"{req.path}?{req.query}" + else: + path = req.path + return f"{req.method} {path} (client_ip={client_ip}, asn={asn}, cf_ray={cf_ray})" + + +def pre_request(worker, req): + """Log the start of a request and add it to the in-flight list""" + request_description = _describe_request(req) + worker.log.info(f"gunicorn starting to process {request_description}") + in_flight = in_flight_by_pid.setdefault(worker.pid, []) + in_flight.append(request_description) + + +def worker_abort(worker): + """Emit an error log if any requests were in-flight""" + in_flight = in_flight_by_pid.get(worker.pid, []) + if len(in_flight) > 0: + worker.log.error( + f"Aborted worker {worker.pid} with in-flight requests: {', '.join(in_flight)}" + ) + + +def worker_int(worker): + """Emit an error log if any requests were in-flight""" + in_flight = in_flight_by_pid.get(worker.pid, []) + if len(in_flight) > 0: + worker.log.error( + f"Interrupted worker {worker.pid} with in-flight requests: {', '.join(in_flight)}" + ) + + +def post_request(worker, req, environ, resp): + """Remove request from in-flight list when we finish handling it""" + request_description = _describe_request(req) + in_flight = in_flight_by_pid.get(worker.pid, []) + if request_description in in_flight: + in_flight.remove(request_description) + +def post_fork(server, worker): + server.log.info("Worker spawned (pid: %s)", worker.pid) + + # Setting DATATRACKER_OPENTELEMETRY_ENABLE=all in the environment will enable all + # opentelemetry instrumentations. Individual instrumentations can be selected by + # using a space-separated list. See the code below for available instrumentations. + telemetry_env = os.environ.get("DATATRACKER_OPENTELEMETRY_ENABLE", "").strip() + if telemetry_env != "": + enabled_telemetry = [tok.strip().lower() for tok in telemetry_env.split()] + resource = Resource.create(attributes={ + "service.name": "datatracker", + "service.version": ietf.__version__, + "service.instance.id": worker.pid, + "service.namespace": "datatracker", + "deployment.environment.name": os.environ.get("DATATRACKER_SERVICE_ENV", "dev") + }) + trace.set_tracer_provider(TracerProvider(resource=resource)) + otlp_exporter = OTLPSpanExporter(endpoint="https://heimdall-otlp.ietf.org/v1/traces") + + trace.get_tracer_provider().add_span_processor(BatchSpanProcessor(otlp_exporter)) + + # Instrumentations + if "all" in enabled_telemetry or "django" in enabled_telemetry: + DjangoInstrumentor().instrument() + if "all" in enabled_telemetry or "psycopg2" in enabled_telemetry: + Psycopg2Instrumentor().instrument() + if "all" in enabled_telemetry or "pymemcache" in enabled_telemetry: + PymemcacheInstrumentor().instrument() + if "all" in enabled_telemetry or "requests" in enabled_telemetry: + RequestsInstrumentor().instrument() diff --git a/dev/build/migration-start.sh b/dev/build/migration-start.sh new file mode 100644 index 0000000000..578daf5cef --- /dev/null +++ b/dev/build/migration-start.sh @@ -0,0 +1,13 @@ +#!/bin/bash -e + +echo "Running Datatracker migrations..." +./ietf/manage.py migrate --settings=settings_local + +# Check whether the blobdb database exists - inspectdb will return a false +# status if not. +if ./ietf/manage.py inspectdb --database blobdb > /dev/null 2>&1; then + echo "Running Blobdb migrations ..." + ./ietf/manage.py migrate --settings=settings_local --database=blobdb +fi + +echo "Done!" diff --git a/dev/build/settings_local_collectstatics.py b/dev/build/settings_local_collectstatics.py new file mode 100644 index 0000000000..ccb4b33979 --- /dev/null +++ b/dev/build/settings_local_collectstatics.py @@ -0,0 +1,8 @@ +# Copyright The IETF Trust 2007-2019, All Rights Reserved +# -*- coding: utf-8 -*- + +from ietf import __version__ +from ietf.settings import * # pyflakes:ignore + +STATIC_URL = "https://static.ietf.org/dt/%s/"%__version__ +STATIC_ROOT = os.path.abspath(BASE_DIR + "/../static/") diff --git a/dev/build/start.sh b/dev/build/start.sh new file mode 100644 index 0000000000..3b03637068 --- /dev/null +++ b/dev/build/start.sh @@ -0,0 +1,29 @@ +#!/bin/bash +# +# Environment config: +# +# CONTAINER_ROLE - datatracker, celery, beat, migrations, or replicator (defaults to datatracker) +# +case "${CONTAINER_ROLE:-datatracker}" in + auth) + exec ./datatracker-start.sh + ;; + beat) + exec ./celery-start.sh --app=ietf beat + ;; + celery) + exec ./celery-start.sh --app=ietf worker + ;; + datatracker) + exec ./datatracker-start.sh + ;; + migrations) + exec ./migration-start.sh + ;; + replicator) + exec ./celery-start.sh --app=ietf worker --queues=blobdb --concurrency=1 + ;; + *) + echo "Unknown role '${CONTAINER_ROLE}'" + exit 255 +esac diff --git a/dev/celery/Dockerfile b/dev/celery/Dockerfile index 91f2949a6b..e69de29bb2 100644 --- a/dev/celery/Dockerfile +++ b/dev/celery/Dockerfile @@ -1,20 +0,0 @@ -# Dockerfile for celery worker -# -FROM ghcr.io/ietf-tools/datatracker-app-base:latest -LABEL maintainer="IETF Tools Team " - -ENV DEBIAN_FRONTEND=noninteractive - -RUN apt-get purge -y imagemagick imagemagick-6-common - -# Copy the startup file -COPY dev/celery/docker-init.sh /docker-init.sh -RUN sed -i 's/\r$//' /docker-init.sh && \ - chmod +x /docker-init.sh - -# Install current datatracker python dependencies -COPY requirements.txt /tmp/pip-tmp/ -RUN pip3 --disable-pip-version-check --no-cache-dir install -r /tmp/pip-tmp/requirements.txt -RUN rm -rf /tmp/pip-tmp - -ENTRYPOINT [ "/docker-init.sh" ] \ No newline at end of file diff --git a/dev/celery/docker-init.sh b/dev/celery/docker-init.sh deleted file mode 100755 index 79671bba81..0000000000 --- a/dev/celery/docker-init.sh +++ /dev/null @@ -1,90 +0,0 @@ -#!/bin/bash -e -# -# Environment parameters: -# -# CELERY_APP - name of application to pass to celery (defaults to ietf) -# -# CELERY_ROLE - 'worker' or 'beat' (defaults to 'worker') -# -# CELERY_UID - numeric uid for the celery worker process -# -# CELERY_GID - numeric gid for the celery worker process -# -# UPDATES_REQUIREMENTS_FROM - path, relative to /workspace mount, to a pip requirements -# file that should be installed at container startup. Default is no package install/update. -# -# DEBUG_TERM_TIMING - if non-empty, writes debug messages during shutdown after a TERM signal -# -WORKSPACEDIR="/workspace" -CELERY_ROLE="${CELERY_ROLE:-worker}" - -cd "$WORKSPACEDIR" || exit 255 - -if [[ -n "${UPDATE_REQUIREMENTS_FROM}" ]]; then - # Need to run as root in the container for this - reqs_file="${WORKSPACEDIR}/${UPDATE_REQUIREMENTS_FROM}" - echo "Updating requirements from ${reqs_file}..." - pip install --upgrade -r "${reqs_file}" -fi - -CELERY_OPTS=( "${CELERY_ROLE}" ) -if [[ -n "${CELERY_UID}" ]]; then - # ensure that a user with the necessary UID exists in container - if ! id "${CELERY_UID}" ; then - adduser --system --uid "${CELERY_UID}" --no-create-home --disabled-login "celery-user-${CELERY_UID}" - fi - CELERY_OPTS+=("--uid=${CELERY_UID}") - CELERY_USERNAME="$(id -nu ${CELERY_UID})" -fi - -if [[ -n "${CELERY_GID}" ]]; then - # ensure that some group with the necessary GID exists in container - if ! getent group "${CELERY_GID}" ; then - addgroup --gid "${CELERY_GID}" "celery-group-${CELERY_GID}" - fi - CELERY_OPTS+=("--gid=${CELERY_GID}") - CELERY_GROUP="$(getent group ${CELERY_GID} | awk -F: '{print $1}')" -fi - -run_as_celery_uid () { - SU_OPTS=() - if [[ -n "${CELERY_GROUP}" ]]; then - SU_OPTS+=("-g" "${CELERY_GROUP}") - fi - su "${SU_OPTS[@]}" "${CELERY_USERNAME:-root}" -s /bin/sh -c "$@" -} - -log_term_timing_msgs () { - # output periodic debug message - while true; do - echo "Waiting for celery worker shutdown ($(date --utc --iso-8601=ns))" - sleep 0.5s - done -} - -cleanup () { - # Cleanly terminate the celery app by sending it a TERM, then waiting for it to exit. - if [[ -n "${celery_pid}" ]]; then - echo "Gracefully terminating celery worker. This may take a few minutes if tasks are in progress..." - kill -TERM "${celery_pid}" - if [[ -n "${DEBUG_TERM_TIMING}" ]]; then - log_term_timing_msgs & - fi - wait "${celery_pid}" - fi -} - -echo "Running checks as root to apply patches..." -/usr/local/bin/python $WORKSPACEDIR/ietf/manage.py check - -if [[ "${CELERY_ROLE}" == "worker" ]]; then - echo "Running initial checks..." - # Run checks as celery worker if one was specified - run_as_celery_uid /usr/local/bin/python $WORKSPACEDIR/ietf/manage.py check -fi - -trap 'trap "" TERM; cleanup' TERM -# start celery in the background so we can trap the TERM signal -celery --app="${CELERY_APP:-ietf}" "${CELERY_OPTS[@]}" "$@" & -celery_pid=$! -wait "${celery_pid}" diff --git a/dev/codecov.yml b/dev/codecov.yml index 8e51bc54ea..7ec1def6c5 100644 --- a/dev/codecov.yml +++ b/dev/codecov.yml @@ -4,5 +4,8 @@ coverage: default: target: auto threshold: 1% + patch: + default: + informational: true github_checks: annotations: false diff --git a/dev/coverage-action/action.yml b/dev/coverage-action/action.yml index b8d732a534..60c8de2d92 100644 --- a/dev/coverage-action/action.yml +++ b/dev/coverage-action/action.yml @@ -35,7 +35,7 @@ outputs: changelog: description: Changelog with headers prepended and coverage stats + chart appended runs: - using: 'node16' + using: 'node20' main: 'index.js' branding: icon: layers diff --git a/dev/coverage-action/index.js b/dev/coverage-action/index.js index 57249bfdb1..5a1c690be3 100644 --- a/dev/coverage-action/index.js +++ b/dev/coverage-action/index.js @@ -5,20 +5,20 @@ const find = require('lodash/find') const round = require('lodash/round') const fs = require('fs/promises') const { DateTime } = require('luxon') -const isPlainObject = require('lodash/isPlainObject') +// const isPlainObject = require('lodash/isPlainObject') const dec = new TextDecoder() async function main () { const token = core.getInput('token') - const tokenCommon = core.getInput('tokenCommon') + // const tokenCommon = core.getInput('tokenCommon') const inputCovPath = core.getInput('coverageResultsPath') // 'data/coverage-raw.json' const outputCovPath = core.getInput('coverageResultsPath') // 'data/coverage.json' const outputHistPath = core.getInput('histCoveragePath') // 'data/historical-coverage.json' const relVersionRaw = core.getInput('version') // 'v7.47.0' const relVersion = relVersionRaw.indexOf('v') === 0 ? relVersionRaw.substring(1) : relVersionRaw const gh = github.getOctokit(token) - const ghCommon = github.getOctokit(tokenCommon) + // const ghCommon = github.getOctokit(tokenCommon) const owner = github.context.repo.owner // 'ietf-tools' const repo = github.context.repo.repo // 'datatracker' const sender = github.context.payload.sender.login // 'rjsparks' @@ -116,137 +116,137 @@ async function main () { } // -> Coverage Chart - if (chartsDirListing.some(c => c.name === `${newRelease.id}.svg`)) { - console.info(`Chart SVG already exists for ${newRelease.name}, skipping...`) - } else { - console.info(`Generating chart SVG for ${newRelease.name}...`) + // if (chartsDirListing.some(c => c.name === `${newRelease.id}.svg`)) { + // console.info(`Chart SVG already exists for ${newRelease.name}, skipping...`) + // } else { + // console.info(`Generating chart SVG for ${newRelease.name}...`) - const { ChartJSNodeCanvas } = require('chartjs-node-canvas') - const chartJSNodeCanvas = new ChartJSNodeCanvas({ type: 'svg', width: 850, height: 300, backgroundColour: '#FFFFFF' }) + // const { ChartJSNodeCanvas } = require('chartjs-node-canvas') + // const chartJSNodeCanvas = new ChartJSNodeCanvas({ type: 'svg', width: 850, height: 300, backgroundColour: '#FFFFFF' }) - // -> Reorder versions - const versions = [] - for (const [key, value] of Object.entries(covData)) { - if (isPlainObject(value)) { - const vRel = find(releases, r => r.tag_name === key || r.tag_name === `v${key}`) - if (!vRel) { - continue - } - versions.push({ - tag: key, - time: vRel.created_at, - stats: { - code: round(value.code * 100, 2), - template: round(value.template * 100, 2), - url: round(value.url * 100, 2) - } - }) - } - } - const roVersions = orderBy(versions, ['time', 'tag'], ['asc', 'asc']) + // // -> Reorder versions + // const versions = [] + // for (const [key, value] of Object.entries(covData)) { + // if (isPlainObject(value)) { + // const vRel = find(releases, r => r.tag_name === key || r.tag_name === `v${key}`) + // if (!vRel) { + // continue + // } + // versions.push({ + // tag: key, + // time: vRel.created_at, + // stats: { + // code: round(value.code * 100, 2), + // template: round(value.template * 100, 2), + // url: round(value.url * 100, 2) + // } + // }) + // } + // } + // const roVersions = orderBy(versions, ['time', 'tag'], ['asc', 'asc']) - // -> Fill axis + data points - const labels = [] - const datasetCode = [] - const datasetTemplate = [] - const datasetUrl = [] + // // -> Fill axis + data points + // const labels = [] + // const datasetCode = [] + // const datasetTemplate = [] + // const datasetUrl = [] - for (const ver of roVersions) { - labels.push(ver.tag) - datasetCode.push(ver.stats.code) - datasetTemplate.push(ver.stats.template) - datasetUrl.push(ver.stats.url) - } + // for (const ver of roVersions) { + // labels.push(ver.tag) + // datasetCode.push(ver.stats.code) + // datasetTemplate.push(ver.stats.template) + // datasetUrl.push(ver.stats.url) + // } - // -> Generate chart SVG - const outputStream = chartJSNodeCanvas.renderToBufferSync({ - type: 'line', - options: { - borderColor: '#CCC', - layout: { - padding: 20 - }, - plugins: { - legend: { - position: 'bottom', - labels: { - font: { - size: 11 - } - } - } - }, - scales: { - x: { - ticks: { - font: { - size: 10 - } - } - }, - y: { - ticks: { - callback: (value) => { - return `${value}%` - }, - font: { - size: 10 - } - } - } - } - }, - data: { - labels, - datasets: [ - { - label: 'Code', - data: datasetCode, - borderWidth: 2, - borderColor: '#E53935', - backgroundColor: '#C6282833', - fill: false, - cubicInterpolationMode: 'monotone', - tension: 0.4, - pointRadius: 0 - }, - { - label: 'Templates', - data: datasetTemplate, - borderWidth: 2, - borderColor: '#039BE5', - backgroundColor: '#0277BD33', - fill: false, - cubicInterpolationMode: 'monotone', - tension: 0.4, - pointRadius: 0 - }, - { - label: 'URLs', - data: datasetUrl, - borderWidth: 2, - borderColor: '#7CB342', - backgroundColor: '#558B2F33', - fill: false, - cubicInterpolationMode: 'monotone', - tension: 0.4, - pointRadius: 0 - } - ] - } - }, 'image/svg+xml') - const svg = Buffer.from(outputStream).toString('base64') + // // -> Generate chart SVG + // const outputStream = chartJSNodeCanvas.renderToBufferSync({ + // type: 'line', + // options: { + // borderColor: '#CCC', + // layout: { + // padding: 20 + // }, + // plugins: { + // legend: { + // position: 'bottom', + // labels: { + // font: { + // size: 11 + // } + // } + // } + // }, + // scales: { + // x: { + // ticks: { + // font: { + // size: 10 + // } + // } + // }, + // y: { + // ticks: { + // callback: (value) => { + // return `${value}%` + // }, + // font: { + // size: 10 + // } + // } + // } + // } + // }, + // data: { + // labels, + // datasets: [ + // { + // label: 'Code', + // data: datasetCode, + // borderWidth: 2, + // borderColor: '#E53935', + // backgroundColor: '#C6282833', + // fill: false, + // cubicInterpolationMode: 'monotone', + // tension: 0.4, + // pointRadius: 0 + // }, + // { + // label: 'Templates', + // data: datasetTemplate, + // borderWidth: 2, + // borderColor: '#039BE5', + // backgroundColor: '#0277BD33', + // fill: false, + // cubicInterpolationMode: 'monotone', + // tension: 0.4, + // pointRadius: 0 + // }, + // { + // label: 'URLs', + // data: datasetUrl, + // borderWidth: 2, + // borderColor: '#7CB342', + // backgroundColor: '#558B2F33', + // fill: false, + // cubicInterpolationMode: 'monotone', + // tension: 0.4, + // pointRadius: 0 + // } + // ] + // } + // }, 'image/svg+xml') + // const svg = Buffer.from(outputStream).toString('base64') - // -> Upload to common repo - console.info(`Uploading chart SVG for ${newRelease.name}...`) - await ghCommon.rest.repos.createOrUpdateFileContents({ - owner, - repo: repoCommon, - path: `assets/graphs/datatracker/${newRelease.id}.svg`, - message: `chore: update datatracker release chart for release ${newRelease.name}`, - content: svg - }) - } + // // -> Upload to common repo + // console.info(`Uploading chart SVG for ${newRelease.name}...`) + // await ghCommon.rest.repos.createOrUpdateFileContents({ + // owner, + // repo: repoCommon, + // path: `assets/graphs/datatracker/${newRelease.id}.svg`, + // message: `chore: update datatracker release chart for release ${newRelease.name}`, + // content: svg + // }) + // } // -> Add to changelog body let formattedBody = '' @@ -265,7 +265,7 @@ async function main () { formattedBody += `![](https://img.shields.io/badge/Code-${covInfo.code}%25-${getCoverageColor(covInfo.code)}?style=flat-square)` formattedBody += `![](https://img.shields.io/badge/Templates-${covInfo.template}%25-${getCoverageColor(covInfo.template)}?style=flat-square)` formattedBody += `![](https://img.shields.io/badge/URLs-${covInfo.url}%25-${getCoverageColor(covInfo.url)}?style=flat-square)\n\n` - formattedBody += `![chart](https://raw.githubusercontent.com/${owner}/${repoCommon}/main/assets/graphs/datatracker/${newRelease.id}.svg)` + // formattedBody += `![chart](https://raw.githubusercontent.com/${owner}/${repoCommon}/main/assets/graphs/datatracker/${newRelease.id}.svg)` core.setOutput('changelog', formattedBody) } diff --git a/dev/coverage-action/package-lock.json b/dev/coverage-action/package-lock.json index 8841a57fe2..09570ee0e4 100644 --- a/dev/coverage-action/package-lock.json +++ b/dev/coverage-action/package-lock.json @@ -9,4679 +9,510 @@ "version": "1.0.0", "license": "BSD-3-Clause", "dependencies": { - "@actions/core": "1.9.1", - "@actions/github": "5.0.0", - "chart.js": "3.7.1", - "chartjs-node-canvas": "4.1.6", + "@actions/core": "1.11.1", + "@actions/github": "6.0.1", "lodash": "4.17.21", - "luxon": "2.3.1" - }, - "devDependencies": { - "eslint": "7.32.0", - "eslint-config-standard": "16.0.3", - "eslint-plugin-import": "2.25.4", - "eslint-plugin-node": "11.1.0", - "eslint-plugin-promise": "5.2.0" + "luxon": "3.7.1" } }, "node_modules/@actions/core": { - "version": "1.9.1", - "resolved": "https://registry.npmjs.org/@actions/core/-/core-1.9.1.tgz", - "integrity": "sha512-5ad+U2YGrmmiw6du20AQW5XuWo7UKN2052FjSV7MX+Wfjf8sCqcsZe62NfgHys4QI4/Y+vQvLKYL8jWtA1ZBTA==", + "version": "1.11.1", + "resolved": "https://registry.npmjs.org/@actions/core/-/core-1.11.1.tgz", + "integrity": "sha512-hXJCSrkwfA46Vd9Z3q4cpEpHB1rL5NG04+/rbqW9d3+CSvtB1tYe8UTpAlixa1vj0m/ULglfEK2UKxMGxCxv5A==", "dependencies": { - "@actions/http-client": "^2.0.1", - "uuid": "^8.3.2" + "@actions/exec": "^1.1.1", + "@actions/http-client": "^2.0.1" } }, - "node_modules/@actions/core/node_modules/@actions/http-client": { - "version": "2.0.1", - "resolved": "https://registry.npmjs.org/@actions/http-client/-/http-client-2.0.1.tgz", - "integrity": "sha512-PIXiMVtz6VvyaRsGY268qvj57hXQEpsYogYOu2nrQhlf+XCGmZstmuZBbAybUl1nQGnvS1k1eEsQ69ZoD7xlSw==", + "node_modules/@actions/exec": { + "version": "1.1.1", + "resolved": "https://registry.npmjs.org/@actions/exec/-/exec-1.1.1.tgz", + "integrity": "sha512-+sCcHHbVdk93a0XT19ECtO/gIXoxvdsgQLzb2fE2/5sIZmWQuluYyjPQtrtTHdU1YzTZ7bAPN4sITq2xi1679w==", "dependencies": { - "tunnel": "^0.0.6" + "@actions/io": "^1.0.1" } }, "node_modules/@actions/github": { - "version": "5.0.0", - "resolved": "https://registry.npmjs.org/@actions/github/-/github-5.0.0.tgz", - "integrity": "sha512-QvE9eAAfEsS+yOOk0cylLBIO/d6WyWIOvsxxzdrPFaud39G6BOkUwScXZn1iBzQzHyu9SBkkLSWlohDWdsasAQ==", + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/@actions/github/-/github-6.0.1.tgz", + "integrity": "sha512-xbZVcaqD4XnQAe35qSQqskb3SqIAfRyLBrHMd/8TuL7hJSz2QtbDwnNM8zWx4zO5l2fnGtseNE3MbEvD7BxVMw==", + "license": "MIT", "dependencies": { - "@actions/http-client": "^1.0.11", - "@octokit/core": "^3.4.0", - "@octokit/plugin-paginate-rest": "^2.13.3", - "@octokit/plugin-rest-endpoint-methods": "^5.1.1" + "@actions/http-client": "^2.2.0", + "@octokit/core": "^5.0.1", + "@octokit/plugin-paginate-rest": "^9.2.2", + "@octokit/plugin-rest-endpoint-methods": "^10.4.0", + "@octokit/request": "^8.4.1", + "@octokit/request-error": "^5.1.1", + "undici": "^5.28.5" } }, "node_modules/@actions/http-client": { - "version": "1.0.11", - "resolved": "https://registry.npmjs.org/@actions/http-client/-/http-client-1.0.11.tgz", - "integrity": "sha512-VRYHGQV1rqnROJqdMvGUbY/Kn8vriQe/F9HR2AlYHzmKuM/p3kjNuXhmdBfcVgsvRWTz5C5XW5xvndZrVBuAYg==", + "version": "2.2.0", + "resolved": "https://registry.npmjs.org/@actions/http-client/-/http-client-2.2.0.tgz", + "integrity": "sha512-q+epW0trjVUUHboliPb4UF9g2msf+w61b32tAkFEwL/IwP0DQWgbCMM0Hbe3e3WXSKz5VcUXbzJQgy8Hkra/Lg==", "dependencies": { - "tunnel": "0.0.6" + "tunnel": "^0.0.6", + "undici": "^5.25.4" } }, - "node_modules/@babel/code-frame": { - "version": "7.12.11", - "resolved": "https://registry.npmjs.org/@babel/code-frame/-/code-frame-7.12.11.tgz", - "integrity": "sha512-Zt1yodBx1UcyiePMSkWnU4hPqhwq7hGi2nFL1LeA3EUl+q2LQx16MISgJ0+z7dnmgvP9QtIleuETGOiOH1RcIw==", - "dev": true, - "dependencies": { - "@babel/highlight": "^7.10.4" - } + "node_modules/@actions/io": { + "version": "1.1.3", + "resolved": "https://registry.npmjs.org/@actions/io/-/io-1.1.3.tgz", + "integrity": "sha512-wi9JjgKLYS7U/z8PPbco+PvTb/nRWjeoFlJ1Qer83k/3C5PHQi28hiVdeE2kHXmIL99mQFawx8qt/JPjZilJ8Q==" }, - "node_modules/@babel/helper-validator-identifier": { - "version": "7.16.7", - "resolved": "https://registry.npmjs.org/@babel/helper-validator-identifier/-/helper-validator-identifier-7.16.7.tgz", - "integrity": "sha512-hsEnFemeiW4D08A5gUAZxLBTXpZ39P+a+DGDsHw1yxqyQ/jzFEnxf5uTEGp+3bzAbNOxU1paTgYS4ECU/IgfDw==", - "dev": true, + "node_modules/@fastify/busboy": { + "version": "2.0.0", + "resolved": "https://registry.npmjs.org/@fastify/busboy/-/busboy-2.0.0.tgz", + "integrity": "sha512-JUFJad5lv7jxj926GPgymrWQxxjPYuJNiNjNMzqT+HiuP6Vl3dk5xzG+8sTX96np0ZAluvaMzPsjhHZ5rNuNQQ==", "engines": { - "node": ">=6.9.0" + "node": ">=14" } }, - "node_modules/@babel/highlight": { - "version": "7.16.10", - "resolved": "https://registry.npmjs.org/@babel/highlight/-/highlight-7.16.10.tgz", - "integrity": "sha512-5FnTQLSLswEj6IkgVw5KusNUUFY9ZGqe/TRFnP/BKYHYgfh7tc+C7mwiy95/yNP7Dh9x580Vv8r7u7ZfTBFxdw==", - "dev": true, - "dependencies": { - "@babel/helper-validator-identifier": "^7.16.7", - "chalk": "^2.0.0", - "js-tokens": "^4.0.0" - }, + "node_modules/@octokit/auth-token": { + "version": "4.0.0", + "resolved": "https://registry.npmjs.org/@octokit/auth-token/-/auth-token-4.0.0.tgz", + "integrity": "sha512-tY/msAuJo6ARbK6SPIxZrPBms3xPbfwBrulZe0Wtr/DIY9lje2HeV1uoebShn6mx7SjCHif6EjMvoREj+gZ+SA==", + "license": "MIT", "engines": { - "node": ">=6.9.0" + "node": ">= 18" } }, - "node_modules/@babel/highlight/node_modules/ansi-styles": { - "version": "3.2.1", - "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-3.2.1.tgz", - "integrity": "sha512-VT0ZI6kZRdTh8YyJw3SMbYm/u+NqfsAxEpWO0Pf9sq8/e94WxxOpPKx9FR1FlyCtOVDNOQ+8ntlqFxiRc+r5qA==", - "dev": true, - "dependencies": { - "color-convert": "^1.9.0" + "node_modules/@octokit/core": { + "version": "5.2.1", + "resolved": "https://registry.npmjs.org/@octokit/core/-/core-5.2.1.tgz", + "integrity": "sha512-dKYCMuPO1bmrpuogcjQ8z7ICCH3FP6WmxpwC03yjzGfZhj9fTJg6+bS1+UAplekbN2C+M61UNllGOOoAfGCrdQ==", + "license": "MIT", + "dependencies": { + "@octokit/auth-token": "^4.0.0", + "@octokit/graphql": "^7.1.0", + "@octokit/request": "^8.4.1", + "@octokit/request-error": "^5.1.1", + "@octokit/types": "^13.0.0", + "before-after-hook": "^2.2.0", + "universal-user-agent": "^6.0.0" }, "engines": { - "node": ">=4" + "node": ">= 18" } }, - "node_modules/@babel/highlight/node_modules/chalk": { - "version": "2.4.2", - "resolved": "https://registry.npmjs.org/chalk/-/chalk-2.4.2.tgz", - "integrity": "sha512-Mti+f9lpJNcwF4tWV8/OrTTtF1gZi+f8FqlyAdouralcFWFQWF2+NgCHShjkCb+IFBLq9buZwE1xckQU4peSuQ==", - "dev": true, + "node_modules/@octokit/endpoint": { + "version": "9.0.6", + "resolved": "https://registry.npmjs.org/@octokit/endpoint/-/endpoint-9.0.6.tgz", + "integrity": "sha512-H1fNTMA57HbkFESSt3Y9+FBICv+0jFceJFPWDePYlR/iMGrwM5ph+Dd4XRQs+8X+PUFURLQgX9ChPfhJ/1uNQw==", + "license": "MIT", "dependencies": { - "ansi-styles": "^3.2.1", - "escape-string-regexp": "^1.0.5", - "supports-color": "^5.3.0" + "@octokit/types": "^13.1.0", + "universal-user-agent": "^6.0.0" }, "engines": { - "node": ">=4" - } - }, - "node_modules/@babel/highlight/node_modules/color-convert": { - "version": "1.9.3", - "resolved": "https://registry.npmjs.org/color-convert/-/color-convert-1.9.3.tgz", - "integrity": "sha512-QfAUtd+vFdAtFQcC8CCyYt1fYWxSqAiK2cSD6zDB8N3cpsEBAvRxp9zOGg6G/SHHJYAT88/az/IuDGALsNVbGg==", - "dev": true, - "dependencies": { - "color-name": "1.1.3" - } - }, - "node_modules/@babel/highlight/node_modules/color-name": { - "version": "1.1.3", - "resolved": "https://registry.npmjs.org/color-name/-/color-name-1.1.3.tgz", - "integrity": "sha1-p9BVi9icQveV3UIyj3QIMcpTvCU=", - "dev": true - }, - "node_modules/@babel/highlight/node_modules/escape-string-regexp": { - "version": "1.0.5", - "resolved": "https://registry.npmjs.org/escape-string-regexp/-/escape-string-regexp-1.0.5.tgz", - "integrity": "sha1-G2HAViGQqN/2rjuyzwIAyhMLhtQ=", - "dev": true, - "engines": { - "node": ">=0.8.0" - } - }, - "node_modules/@babel/highlight/node_modules/has-flag": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/has-flag/-/has-flag-3.0.0.tgz", - "integrity": "sha1-tdRU3CGZriJWmfNGfloH87lVuv0=", - "dev": true, - "engines": { - "node": ">=4" + "node": ">= 18" } }, - "node_modules/@babel/highlight/node_modules/supports-color": { - "version": "5.5.0", - "resolved": "https://registry.npmjs.org/supports-color/-/supports-color-5.5.0.tgz", - "integrity": "sha512-QjVjwdXIt408MIiAqCX4oUKsgU2EqAGzs2Ppkm4aQYbjm+ZEWEcW4SfFNTr4uMNZma0ey4f5lgLrkB0aX0QMow==", - "dev": true, + "node_modules/@octokit/graphql": { + "version": "7.1.1", + "resolved": "https://registry.npmjs.org/@octokit/graphql/-/graphql-7.1.1.tgz", + "integrity": "sha512-3mkDltSfcDUoa176nlGoA32RGjeWjl3K7F/BwHwRMJUW/IteSa4bnSV8p2ThNkcIcZU2umkZWxwETSSCJf2Q7g==", + "license": "MIT", "dependencies": { - "has-flag": "^3.0.0" + "@octokit/request": "^8.4.1", + "@octokit/types": "^13.0.0", + "universal-user-agent": "^6.0.0" }, "engines": { - "node": ">=4" + "node": ">= 18" } }, - "node_modules/@eslint/eslintrc": { - "version": "0.4.3", - "resolved": "https://registry.npmjs.org/@eslint/eslintrc/-/eslintrc-0.4.3.tgz", - "integrity": "sha512-J6KFFz5QCYUJq3pf0mjEcCJVERbzv71PUIDczuh9JkwGEzced6CO5ADLHB1rbf/+oPBtoPfMYNOpGDzCANlbXw==", - "dev": true, - "dependencies": { - "ajv": "^6.12.4", - "debug": "^4.1.1", - "espree": "^7.3.0", - "globals": "^13.9.0", - "ignore": "^4.0.6", - "import-fresh": "^3.2.1", - "js-yaml": "^3.13.1", - "minimatch": "^3.0.4", - "strip-json-comments": "^3.1.1" - }, - "engines": { - "node": "^10.12.0 || >=12.0.0" - } + "node_modules/@octokit/openapi-types": { + "version": "24.2.0", + "resolved": "https://registry.npmjs.org/@octokit/openapi-types/-/openapi-types-24.2.0.tgz", + "integrity": "sha512-9sIH3nSUttelJSXUrmGzl7QUBFul0/mB8HRYl3fOlgHbIWG+WnYDXU3v/2zMtAvuzZ/ed00Ei6on975FhBfzrg==", + "license": "MIT" }, - "node_modules/@humanwhocodes/config-array": { - "version": "0.5.0", - "resolved": "https://registry.npmjs.org/@humanwhocodes/config-array/-/config-array-0.5.0.tgz", - "integrity": "sha512-FagtKFz74XrTl7y6HCzQpwDfXP0yhxe9lHLD1UZxjvZIcbyRz8zTFF/yYNfSfzU414eDwZ1SrO0Qvtyf+wFMQg==", - "dev": true, + "node_modules/@octokit/plugin-paginate-rest": { + "version": "9.2.2", + "resolved": "https://registry.npmjs.org/@octokit/plugin-paginate-rest/-/plugin-paginate-rest-9.2.2.tgz", + "integrity": "sha512-u3KYkGF7GcZnSD/3UP0S7K5XUFT2FkOQdcfXZGZQPGv3lm4F2Xbf71lvjldr8c1H3nNbF+33cLEkWYbokGWqiQ==", + "license": "MIT", "dependencies": { - "@humanwhocodes/object-schema": "^1.2.0", - "debug": "^4.1.1", - "minimatch": "^3.0.4" + "@octokit/types": "^12.6.0" }, "engines": { - "node": ">=10.10.0" - } - }, - "node_modules/@humanwhocodes/object-schema": { - "version": "1.2.1", - "resolved": "https://registry.npmjs.org/@humanwhocodes/object-schema/-/object-schema-1.2.1.tgz", - "integrity": "sha512-ZnQMnLV4e7hDlUvw8H+U8ASL02SS2Gn6+9Ac3wGGLIe7+je2AeAOxPY+izIPJDfFDb7eDjev0Us8MO1iFRN8hA==", - "dev": true - }, - "node_modules/@mapbox/node-pre-gyp": { - "version": "1.0.8", - "resolved": "https://registry.npmjs.org/@mapbox/node-pre-gyp/-/node-pre-gyp-1.0.8.tgz", - "integrity": "sha512-CMGKi28CF+qlbXh26hDe6NxCd7amqeAzEqnS6IHeO6LoaKyM/n+Xw3HT1COdq8cuioOdlKdqn/hCmqPUOMOywg==", - "dependencies": { - "detect-libc": "^1.0.3", - "https-proxy-agent": "^5.0.0", - "make-dir": "^3.1.0", - "node-fetch": "^2.6.5", - "nopt": "^5.0.0", - "npmlog": "^5.0.1", - "rimraf": "^3.0.2", - "semver": "^7.3.5", - "tar": "^6.1.11" + "node": ">= 18" }, - "bin": { - "node-pre-gyp": "bin/node-pre-gyp" - } - }, - "node_modules/@octokit/auth-token": { - "version": "2.5.0", - "resolved": "https://registry.npmjs.org/@octokit/auth-token/-/auth-token-2.5.0.tgz", - "integrity": "sha512-r5FVUJCOLl19AxiuZD2VRZ/ORjp/4IN98Of6YJoJOkY75CIBuYfmiNHGrDwXr+aLGG55igl9QrxX3hbiXlLb+g==", - "dependencies": { - "@octokit/types": "^6.0.3" - } - }, - "node_modules/@octokit/core": { - "version": "3.6.0", - "resolved": "https://registry.npmjs.org/@octokit/core/-/core-3.6.0.tgz", - "integrity": "sha512-7RKRKuA4xTjMhY+eG3jthb3hlZCsOwg3rztWh75Xc+ShDWOfDDATWbeZpAHBNRpm4Tv9WgBMOy1zEJYXG6NJ7Q==", - "dependencies": { - "@octokit/auth-token": "^2.4.4", - "@octokit/graphql": "^4.5.8", - "@octokit/request": "^5.6.3", - "@octokit/request-error": "^2.0.5", - "@octokit/types": "^6.0.3", - "before-after-hook": "^2.2.0", - "universal-user-agent": "^6.0.0" + "peerDependencies": { + "@octokit/core": "5" } }, - "node_modules/@octokit/endpoint": { - "version": "6.0.12", - "resolved": "https://registry.npmjs.org/@octokit/endpoint/-/endpoint-6.0.12.tgz", - "integrity": "sha512-lF3puPwkQWGfkMClXb4k/eUT/nZKQfxinRWJrdZaJO85Dqwo/G0yOC434Jr2ojwafWJMYqFGFa5ms4jJUgujdA==", - "dependencies": { - "@octokit/types": "^6.0.3", - "is-plain-object": "^5.0.0", - "universal-user-agent": "^6.0.0" - } + "node_modules/@octokit/plugin-paginate-rest/node_modules/@octokit/openapi-types": { + "version": "20.0.0", + "resolved": "https://registry.npmjs.org/@octokit/openapi-types/-/openapi-types-20.0.0.tgz", + "integrity": "sha512-EtqRBEjp1dL/15V7WiX5LJMIxxkdiGJnabzYx5Apx4FkQIFgAfKumXeYAqqJCj1s+BMX4cPFIFC4OLCR6stlnA==", + "license": "MIT" }, - "node_modules/@octokit/graphql": { - "version": "4.8.0", - "resolved": "https://registry.npmjs.org/@octokit/graphql/-/graphql-4.8.0.tgz", - "integrity": "sha512-0gv+qLSBLKF0z8TKaSKTsS39scVKF9dbMxJpj3U0vC7wjNWFuIpL/z76Qe2fiuCbDRcJSavkXsVtMS6/dtQQsg==", + "node_modules/@octokit/plugin-paginate-rest/node_modules/@octokit/types": { + "version": "12.6.0", + "resolved": "https://registry.npmjs.org/@octokit/types/-/types-12.6.0.tgz", + "integrity": "sha512-1rhSOfRa6H9w4YwK0yrf5faDaDTb+yLyBUKOCV4xtCDB5VmIPqd/v9yr9o6SAzOAlRxMiRiCic6JVM1/kunVkw==", + "license": "MIT", "dependencies": { - "@octokit/request": "^5.6.0", - "@octokit/types": "^6.0.3", - "universal-user-agent": "^6.0.0" + "@octokit/openapi-types": "^20.0.0" } }, - "node_modules/@octokit/openapi-types": { - "version": "11.2.0", - "resolved": "https://registry.npmjs.org/@octokit/openapi-types/-/openapi-types-11.2.0.tgz", - "integrity": "sha512-PBsVO+15KSlGmiI8QAzaqvsNlZlrDlyAJYcrXBCvVUxCp7VnXjkwPoFHgjEJXx3WF9BAwkA6nfCUA7i9sODzKA==" - }, - "node_modules/@octokit/plugin-paginate-rest": { - "version": "2.17.0", - "resolved": "https://registry.npmjs.org/@octokit/plugin-paginate-rest/-/plugin-paginate-rest-2.17.0.tgz", - "integrity": "sha512-tzMbrbnam2Mt4AhuyCHvpRkS0oZ5MvwwcQPYGtMv4tUa5kkzG58SVB0fcsLulOZQeRnOgdkZWkRUiyBlh0Bkyw==", + "node_modules/@octokit/plugin-rest-endpoint-methods": { + "version": "10.4.1", + "resolved": "https://registry.npmjs.org/@octokit/plugin-rest-endpoint-methods/-/plugin-rest-endpoint-methods-10.4.1.tgz", + "integrity": "sha512-xV1b+ceKV9KytQe3zCVqjg+8GTGfDYwaT1ATU5isiUyVtlVAO3HNdzpS4sr4GBx4hxQ46s7ITtZrAsxG22+rVg==", + "license": "MIT", "dependencies": { - "@octokit/types": "^6.34.0" + "@octokit/types": "^12.6.0" + }, + "engines": { + "node": ">= 18" }, "peerDependencies": { - "@octokit/core": ">=2" + "@octokit/core": "5" } }, - "node_modules/@octokit/plugin-rest-endpoint-methods": { - "version": "5.13.0", - "resolved": "https://registry.npmjs.org/@octokit/plugin-rest-endpoint-methods/-/plugin-rest-endpoint-methods-5.13.0.tgz", - "integrity": "sha512-uJjMTkN1KaOIgNtUPMtIXDOjx6dGYysdIFhgA52x4xSadQCz3b/zJexvITDVpANnfKPW/+E0xkOvLntqMYpviA==", + "node_modules/@octokit/plugin-rest-endpoint-methods/node_modules/@octokit/openapi-types": { + "version": "20.0.0", + "resolved": "https://registry.npmjs.org/@octokit/openapi-types/-/openapi-types-20.0.0.tgz", + "integrity": "sha512-EtqRBEjp1dL/15V7WiX5LJMIxxkdiGJnabzYx5Apx4FkQIFgAfKumXeYAqqJCj1s+BMX4cPFIFC4OLCR6stlnA==", + "license": "MIT" + }, + "node_modules/@octokit/plugin-rest-endpoint-methods/node_modules/@octokit/types": { + "version": "12.6.0", + "resolved": "https://registry.npmjs.org/@octokit/types/-/types-12.6.0.tgz", + "integrity": "sha512-1rhSOfRa6H9w4YwK0yrf5faDaDTb+yLyBUKOCV4xtCDB5VmIPqd/v9yr9o6SAzOAlRxMiRiCic6JVM1/kunVkw==", + "license": "MIT", "dependencies": { - "@octokit/types": "^6.34.0", - "deprecation": "^2.3.1" - }, - "peerDependencies": { - "@octokit/core": ">=3" + "@octokit/openapi-types": "^20.0.0" } }, "node_modules/@octokit/request": { - "version": "5.6.3", - "resolved": "https://registry.npmjs.org/@octokit/request/-/request-5.6.3.tgz", - "integrity": "sha512-bFJl0I1KVc9jYTe9tdGGpAMPy32dLBXXo1dS/YwSCTL/2nd9XeHsY616RE3HPXDVk+a+dBuzyz5YdlXwcDTr2A==", - "dependencies": { - "@octokit/endpoint": "^6.0.1", - "@octokit/request-error": "^2.1.0", - "@octokit/types": "^6.16.1", - "is-plain-object": "^5.0.0", - "node-fetch": "^2.6.7", + "version": "8.4.1", + "resolved": "https://registry.npmjs.org/@octokit/request/-/request-8.4.1.tgz", + "integrity": "sha512-qnB2+SY3hkCmBxZsR/MPCybNmbJe4KAlfWErXq+rBKkQJlbjdJeS85VI9r8UqeLYLvnAenU8Q1okM/0MBsAGXw==", + "license": "MIT", + "dependencies": { + "@octokit/endpoint": "^9.0.6", + "@octokit/request-error": "^5.1.1", + "@octokit/types": "^13.1.0", "universal-user-agent": "^6.0.0" + }, + "engines": { + "node": ">= 18" } }, "node_modules/@octokit/request-error": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/@octokit/request-error/-/request-error-2.1.0.tgz", - "integrity": "sha512-1VIvgXxs9WHSjicsRwq8PlR2LR2x6DwsJAaFgzdi0JfJoGSO8mYI/cHJQ+9FbN21aa+DrgNLnwObmyeSC8Rmpg==", + "version": "5.1.1", + "resolved": "https://registry.npmjs.org/@octokit/request-error/-/request-error-5.1.1.tgz", + "integrity": "sha512-v9iyEQJH6ZntoENr9/yXxjuezh4My67CBSu9r6Ve/05Iu5gNgnisNWOsoJHTP6k0Rr0+HQIpnH+kyammu90q/g==", + "license": "MIT", "dependencies": { - "@octokit/types": "^6.0.3", + "@octokit/types": "^13.1.0", "deprecation": "^2.0.0", "once": "^1.4.0" - } - }, - "node_modules/@octokit/types": { - "version": "6.34.0", - "resolved": "https://registry.npmjs.org/@octokit/types/-/types-6.34.0.tgz", - "integrity": "sha512-s1zLBjWhdEI2zwaoSgyOFoKSl109CUcVBCc7biPJ3aAf6LGLU6szDvi31JPU7bxfla2lqfhjbbg/5DdFNxOwHw==", - "dependencies": { - "@octokit/openapi-types": "^11.2.0" - } - }, - "node_modules/@types/json5": { - "version": "0.0.29", - "resolved": "https://registry.npmjs.org/@types/json5/-/json5-0.0.29.tgz", - "integrity": "sha1-7ihweulOEdK4J7y+UnC86n8+ce4=", - "dev": true - }, - "node_modules/abbrev": { - "version": "1.1.1", - "resolved": "https://registry.npmjs.org/abbrev/-/abbrev-1.1.1.tgz", - "integrity": "sha512-nne9/IiQ/hzIhY6pdDnbBtz7DjPTKrY00P/zvPSm5pOFkl6xuGrGnXn/VtTNNfNtAfZ9/1RtehkszU9qcTii0Q==" - }, - "node_modules/acorn": { - "version": "7.4.1", - "resolved": "https://registry.npmjs.org/acorn/-/acorn-7.4.1.tgz", - "integrity": "sha512-nQyp0o1/mNdbTO1PO6kHkwSrmgZ0MT/jCCpNiwbUjGoRN4dlBhqJtoQuCnEOKzgTVwg0ZWiCoQy6SxMebQVh8A==", - "dev": true, - "bin": { - "acorn": "bin/acorn" - }, - "engines": { - "node": ">=0.4.0" - } - }, - "node_modules/acorn-jsx": { - "version": "5.3.2", - "resolved": "https://registry.npmjs.org/acorn-jsx/-/acorn-jsx-5.3.2.tgz", - "integrity": "sha512-rq9s+JNhf0IChjtDXxllJ7g41oZk5SlXtp0LHwyA5cejwn7vKmKp4pPri6YEePv2PU65sAsegbXtIinmDFDXgQ==", - "dev": true, - "peerDependencies": { - "acorn": "^6.0.0 || ^7.0.0 || ^8.0.0" - } - }, - "node_modules/agent-base": { - "version": "6.0.2", - "resolved": "https://registry.npmjs.org/agent-base/-/agent-base-6.0.2.tgz", - "integrity": "sha512-RZNwNclF7+MS/8bDg70amg32dyeZGZxiDuQmZxKLAlQjr3jGyLx+4Kkk58UO7D2QdgFIQCovuSuZESne6RG6XQ==", - "dependencies": { - "debug": "4" }, "engines": { - "node": ">= 6.0.0" + "node": ">= 18" } }, - "node_modules/ajv": { - "version": "6.12.6", - "resolved": "https://registry.npmjs.org/ajv/-/ajv-6.12.6.tgz", - "integrity": "sha512-j3fVLgvTo527anyYyJOGTYJbG+vnnQYvE0m5mmkc1TK+nxAppkCLMIL0aZ4dblVCNoGShhm+kzE4ZUykBoMg4g==", - "dev": true, + "node_modules/@octokit/types": { + "version": "13.10.0", + "resolved": "https://registry.npmjs.org/@octokit/types/-/types-13.10.0.tgz", + "integrity": "sha512-ifLaO34EbbPj0Xgro4G5lP5asESjwHracYJvVaPIyXMuiuXLlhic3S47cBdTb+jfODkTE5YtGCLt3Ay3+J97sA==", + "license": "MIT", "dependencies": { - "fast-deep-equal": "^3.1.1", - "fast-json-stable-stringify": "^2.0.0", - "json-schema-traverse": "^0.4.1", - "uri-js": "^4.2.2" - }, - "funding": { - "type": "github", - "url": "https://github.com/sponsors/epoberezkin" - } - }, - "node_modules/ansi-colors": { - "version": "4.1.1", - "resolved": "https://registry.npmjs.org/ansi-colors/-/ansi-colors-4.1.1.tgz", - "integrity": "sha512-JoX0apGbHaUJBNl6yF+p6JAFYZ666/hhCGKN5t9QFjbJQKUU/g8MNbFDbvfrgKXvI1QpZplPOnwIo99lX/AAmA==", - "dev": true, - "engines": { - "node": ">=6" + "@octokit/openapi-types": "^24.2.0" } }, - "node_modules/ansi-regex": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", - "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==", - "engines": { - "node": ">=8" - } + "node_modules/before-after-hook": { + "version": "2.2.3", + "resolved": "https://registry.npmjs.org/before-after-hook/-/before-after-hook-2.2.3.tgz", + "integrity": "sha512-NzUnlZexiaH/46WDhANlyR2bXRopNg4F/zuSA3OpZnllCUgRaOF2znDioDWrmbNVsuZk6l9pMquQB38cfBZwkQ==", + "license": "Apache-2.0" }, - "node_modules/ansi-styles": { - "version": "4.3.0", - "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz", - "integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==", - "dev": true, - "dependencies": { - "color-convert": "^2.0.1" - }, - "engines": { - "node": ">=8" - }, - "funding": { - "url": "https://github.com/chalk/ansi-styles?sponsor=1" - } + "node_modules/deprecation": { + "version": "2.3.1", + "resolved": "https://registry.npmjs.org/deprecation/-/deprecation-2.3.1.tgz", + "integrity": "sha512-xmHIy4F3scKVwMsQ4WnVaS8bHOx0DmVwRywosKhaILI0ywMDWPtBSku2HNxRvF7jtwDRsoEwYQSfbxj8b7RlJQ==", + "license": "ISC" }, - "node_modules/aproba": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/aproba/-/aproba-2.0.0.tgz", - "integrity": "sha512-lYe4Gx7QT+MKGbDsA+Z+he/Wtef0BiwDOlK/XkBrdfsh9J/jPPXbX0tE9x9cl27Tmu5gg3QUbUrQYa/y+KOHPQ==" + "node_modules/lodash": { + "version": "4.17.21", + "resolved": "https://registry.npmjs.org/lodash/-/lodash-4.17.21.tgz", + "integrity": "sha512-v2kDEe57lecTulaDIuNTPy3Ry4gLGJ6Z1O3vE1krgXZNrsQ+LFTGHVxVjcXPs17LhbZVGedAJv8XZ1tvj5FvSg==" }, - "node_modules/are-we-there-yet": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/are-we-there-yet/-/are-we-there-yet-2.0.0.tgz", - "integrity": "sha512-Ci/qENmwHnsYo9xKIcUJN5LeDKdJ6R1Z1j9V/J5wyq8nh/mYPEpIKJbBZXtZjG04HiK7zV/p6Vs9952MrMeUIw==", - "dependencies": { - "delegates": "^1.0.0", - "readable-stream": "^3.6.0" - }, + "node_modules/luxon": { + "version": "3.7.1", + "resolved": "https://registry.npmjs.org/luxon/-/luxon-3.7.1.tgz", + "integrity": "sha512-RkRWjA926cTvz5rAb1BqyWkKbbjzCGchDUIKMCUvNi17j6f6j8uHGDV82Aqcqtzd+icoYpELmG3ksgGiFNNcNg==", "engines": { - "node": ">=10" + "node": ">=12" } }, - "node_modules/argparse": { - "version": "1.0.10", - "resolved": "https://registry.npmjs.org/argparse/-/argparse-1.0.10.tgz", - "integrity": "sha512-o5Roy6tNG4SL/FOkCAN6RzjiakZS25RLYFrcMttJqbdd8BWrnA+fGz57iN5Pb06pvBGvl5gQ0B48dJlslXvoTg==", - "dev": true, + "node_modules/once": { + "version": "1.4.0", + "resolved": "https://registry.npmjs.org/once/-/once-1.4.0.tgz", + "integrity": "sha512-lNaJgI+2Q5URQBkccEKHTQOPaXdUxnZZElQTZY0MFUAuaEqe1E+Nyvgdz/aIyNi6Z9MzO5dv1H8n58/GELp3+w==", + "license": "ISC", "dependencies": { - "sprintf-js": "~1.0.2" + "wrappy": "1" } }, - "node_modules/array-includes": { - "version": "3.1.4", - "resolved": "https://registry.npmjs.org/array-includes/-/array-includes-3.1.4.tgz", - "integrity": "sha512-ZTNSQkmWumEbiHO2GF4GmWxYVTiQyJy2XOTa15sdQSrvKn7l+180egQMqlrMOUMCyLMD7pmyQe4mMDUT6Behrw==", - "dev": true, - "dependencies": { - "call-bind": "^1.0.2", - "define-properties": "^1.1.3", - "es-abstract": "^1.19.1", - "get-intrinsic": "^1.1.1", - "is-string": "^1.0.7" - }, + "node_modules/tunnel": { + "version": "0.0.6", + "resolved": "https://registry.npmjs.org/tunnel/-/tunnel-0.0.6.tgz", + "integrity": "sha512-1h/Lnq9yajKY2PEbBadPXj3VxsDDu844OnaAo52UVmIzIvwwtBPIuNvkjuzBlTWpfJyUbG3ez0KSBibQkj4ojg==", "engines": { - "node": ">= 0.4" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" + "node": ">=0.6.11 <=0.7.0 || >=0.7.3" } }, - "node_modules/array.prototype.flat": { - "version": "1.2.5", - "resolved": "https://registry.npmjs.org/array.prototype.flat/-/array.prototype.flat-1.2.5.tgz", - "integrity": "sha512-KaYU+S+ndVqyUnignHftkwc58o3uVU1jzczILJ1tN2YaIZpFIKBiP/x/j97E5MVPsaCloPbqWLB/8qCTVvT2qg==", - "dev": true, + "node_modules/undici": { + "version": "5.29.0", + "resolved": "https://registry.npmjs.org/undici/-/undici-5.29.0.tgz", + "integrity": "sha512-raqeBD6NQK4SkWhQzeYKd1KmIG6dllBOTt55Rmkt4HtI9mwdWtJljnrXjAFUBLTSN67HWrOIZ3EPF4kjUw80Bg==", + "license": "MIT", "dependencies": { - "call-bind": "^1.0.2", - "define-properties": "^1.1.3", - "es-abstract": "^1.19.0" + "@fastify/busboy": "^2.0.0" }, "engines": { - "node": ">= 0.4" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" + "node": ">=14.0" } }, - "node_modules/astral-regex": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/astral-regex/-/astral-regex-2.0.0.tgz", - "integrity": "sha512-Z7tMw1ytTXt5jqMcOP+OQteU1VuNK9Y02uuJtKQ1Sv69jXQKKg5cibLwGJow8yzZP+eAc18EmLGPal0bp36rvQ==", - "dev": true, - "engines": { - "node": ">=8" - } + "node_modules/universal-user-agent": { + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-6.0.1.tgz", + "integrity": "sha512-yCzhz6FN2wU1NiiQRogkTQszlQSlpWaw8SvVegAc+bDxbzHgh1vX8uIe8OYyMH6DwH+sdTJsgMl36+mSMdRJIQ==", + "license": "ISC" }, - "node_modules/balanced-match": { + "node_modules/wrappy": { "version": "1.0.2", - "resolved": "https://registry.npmjs.org/balanced-match/-/balanced-match-1.0.2.tgz", - "integrity": "sha512-3oSeUO0TMV67hN1AmbXsK4yaqU7tjiHlbxRDZOpH0KW9+CeX4bRAaX0Anxt0tx2MrpRpWwQaPwIlISEJhYU5Pw==" - }, - "node_modules/before-after-hook": { - "version": "2.2.2", - "resolved": "https://registry.npmjs.org/before-after-hook/-/before-after-hook-2.2.2.tgz", - "integrity": "sha512-3pZEU3NT5BFUo/AD5ERPWOgQOCZITni6iavr5AUw5AUwQjMlI0kzu5btnyD39AF0gUEsDPwJT+oY1ORBJijPjQ==" - }, - "node_modules/brace-expansion": { - "version": "1.1.11", - "resolved": "https://registry.npmjs.org/brace-expansion/-/brace-expansion-1.1.11.tgz", - "integrity": "sha512-iCuPHDFgrHX7H2vEI/5xpz07zSHB00TpugqhmYtVmMO6518mCuRMoOYFldEBl0g187ufozdaHgWKcYFb61qGiA==", - "dependencies": { - "balanced-match": "^1.0.0", - "concat-map": "0.0.1" + "resolved": "https://registry.npmjs.org/wrappy/-/wrappy-1.0.2.tgz", + "integrity": "sha512-l4Sp/DRseor9wL6EvV2+TuQn63dMkPjZ/sp9XkghTEbV9KlPS1xUsZ3u7/IQO4wxtcFB4bgpQPRcR3QCvezPcQ==", + "license": "ISC" + } + }, + "dependencies": { + "@actions/core": { + "version": "1.11.1", + "resolved": "https://registry.npmjs.org/@actions/core/-/core-1.11.1.tgz", + "integrity": "sha512-hXJCSrkwfA46Vd9Z3q4cpEpHB1rL5NG04+/rbqW9d3+CSvtB1tYe8UTpAlixa1vj0m/ULglfEK2UKxMGxCxv5A==", + "requires": { + "@actions/exec": "^1.1.1", + "@actions/http-client": "^2.0.1" } }, - "node_modules/call-bind": { - "version": "1.0.2", - "resolved": "https://registry.npmjs.org/call-bind/-/call-bind-1.0.2.tgz", - "integrity": "sha512-7O+FbCihrB5WGbFYesctwmTKae6rOiIzmz1icreWJ+0aA7LJfuqhEso2T9ncpcFtzMQtzXf2QGGueWJGTYsqrA==", - "dev": true, - "dependencies": { - "function-bind": "^1.1.1", - "get-intrinsic": "^1.0.2" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" + "@actions/exec": { + "version": "1.1.1", + "resolved": "https://registry.npmjs.org/@actions/exec/-/exec-1.1.1.tgz", + "integrity": "sha512-+sCcHHbVdk93a0XT19ECtO/gIXoxvdsgQLzb2fE2/5sIZmWQuluYyjPQtrtTHdU1YzTZ7bAPN4sITq2xi1679w==", + "requires": { + "@actions/io": "^1.0.1" } }, - "node_modules/callsites": { - "version": "3.1.0", - "resolved": "https://registry.npmjs.org/callsites/-/callsites-3.1.0.tgz", - "integrity": "sha512-P8BjAsXvZS+VIDUI11hHCQEv74YT67YUi5JJFNWIqL235sBmjX4+qx9Muvls5ivyNENctx46xQLQ3aTuE7ssaQ==", - "dev": true, - "engines": { - "node": ">=6" + "@actions/github": { + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/@actions/github/-/github-6.0.1.tgz", + "integrity": "sha512-xbZVcaqD4XnQAe35qSQqskb3SqIAfRyLBrHMd/8TuL7hJSz2QtbDwnNM8zWx4zO5l2fnGtseNE3MbEvD7BxVMw==", + "requires": { + "@actions/http-client": "^2.2.0", + "@octokit/core": "^5.0.1", + "@octokit/plugin-paginate-rest": "^9.2.2", + "@octokit/plugin-rest-endpoint-methods": "^10.4.0", + "@octokit/request": "^8.4.1", + "@octokit/request-error": "^5.1.1", + "undici": "^5.28.5" } }, - "node_modules/canvas": { - "version": "2.9.1", - "resolved": "https://registry.npmjs.org/canvas/-/canvas-2.9.1.tgz", - "integrity": "sha512-vSQti1uG/2gjv3x6QLOZw7TctfufaerTWbVe+NSduHxxLGB+qf3kFgQ6n66DSnuoINtVUjrLLIK2R+lxrBG07A==", - "hasInstallScript": true, - "dependencies": { - "@mapbox/node-pre-gyp": "^1.0.0", - "nan": "^2.15.0", - "simple-get": "^3.0.3" - }, - "engines": { - "node": ">=6" + "@actions/http-client": { + "version": "2.2.0", + "resolved": "https://registry.npmjs.org/@actions/http-client/-/http-client-2.2.0.tgz", + "integrity": "sha512-q+epW0trjVUUHboliPb4UF9g2msf+w61b32tAkFEwL/IwP0DQWgbCMM0Hbe3e3WXSKz5VcUXbzJQgy8Hkra/Lg==", + "requires": { + "tunnel": "^0.0.6", + "undici": "^5.25.4" } }, - "node_modules/chalk": { - "version": "4.1.2", - "resolved": "https://registry.npmjs.org/chalk/-/chalk-4.1.2.tgz", - "integrity": "sha512-oKnbhFyRIXpUuez8iBMmyEa4nbj4IOQyuhc/wy9kY7/WVPcwIO9VA668Pu8RkO7+0G76SLROeyw9CpQ061i4mA==", - "dev": true, - "dependencies": { - "ansi-styles": "^4.1.0", - "supports-color": "^7.1.0" - }, - "engines": { - "node": ">=10" - }, - "funding": { - "url": "https://github.com/chalk/chalk?sponsor=1" - } + "@actions/io": { + "version": "1.1.3", + "resolved": "https://registry.npmjs.org/@actions/io/-/io-1.1.3.tgz", + "integrity": "sha512-wi9JjgKLYS7U/z8PPbco+PvTb/nRWjeoFlJ1Qer83k/3C5PHQi28hiVdeE2kHXmIL99mQFawx8qt/JPjZilJ8Q==" }, - "node_modules/chart.js": { - "version": "3.7.1", - "resolved": "https://registry.npmjs.org/chart.js/-/chart.js-3.7.1.tgz", - "integrity": "sha512-8knRegQLFnPQAheZV8MjxIXc5gQEfDFD897BJgv/klO/vtIyFFmgMXrNfgrXpbTr/XbTturxRgxIXx/Y+ASJBA==" + "@fastify/busboy": { + "version": "2.0.0", + "resolved": "https://registry.npmjs.org/@fastify/busboy/-/busboy-2.0.0.tgz", + "integrity": "sha512-JUFJad5lv7jxj926GPgymrWQxxjPYuJNiNjNMzqT+HiuP6Vl3dk5xzG+8sTX96np0ZAluvaMzPsjhHZ5rNuNQQ==" }, - "node_modules/chartjs-node-canvas": { - "version": "4.1.6", - "resolved": "https://registry.npmjs.org/chartjs-node-canvas/-/chartjs-node-canvas-4.1.6.tgz", - "integrity": "sha512-UQJbPWrvqB/FoLclGA9BaLQmZbzSYlujF4w8NZd6Xzb+sqgACBb2owDX6m7ifCXLjUW5Nz0Qx0qqrTtQkkSoYw==", - "dependencies": { - "canvas": "^2.8.0", - "tslib": "^2.3.1" - }, - "peerDependencies": { - "chart.js": "^3.5.1" - } + "@octokit/auth-token": { + "version": "4.0.0", + "resolved": "https://registry.npmjs.org/@octokit/auth-token/-/auth-token-4.0.0.tgz", + "integrity": "sha512-tY/msAuJo6ARbK6SPIxZrPBms3xPbfwBrulZe0Wtr/DIY9lje2HeV1uoebShn6mx7SjCHif6EjMvoREj+gZ+SA==" }, - "node_modules/chownr": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/chownr/-/chownr-2.0.0.tgz", - "integrity": "sha512-bIomtDF5KGpdogkLd9VspvFzk9KfpyyGlS8YFVZl7TGPBHL5snIOnxeshwVgPteQ9b4Eydl+pVbIyE1DcvCWgQ==", - "engines": { - "node": ">=10" + "@octokit/core": { + "version": "5.2.1", + "resolved": "https://registry.npmjs.org/@octokit/core/-/core-5.2.1.tgz", + "integrity": "sha512-dKYCMuPO1bmrpuogcjQ8z7ICCH3FP6WmxpwC03yjzGfZhj9fTJg6+bS1+UAplekbN2C+M61UNllGOOoAfGCrdQ==", + "requires": { + "@octokit/auth-token": "^4.0.0", + "@octokit/graphql": "^7.1.0", + "@octokit/request": "^8.4.1", + "@octokit/request-error": "^5.1.1", + "@octokit/types": "^13.0.0", + "before-after-hook": "^2.2.0", + "universal-user-agent": "^6.0.0" } }, - "node_modules/color-convert": { - "version": "2.0.1", - "resolved": "https://registry.npmjs.org/color-convert/-/color-convert-2.0.1.tgz", - "integrity": "sha512-RRECPsj7iu/xb5oKYcsFHSppFNnsj/52OVTRKb4zP5onXwVF3zVmmToNcOfGC+CRDpfK/U584fMg38ZHCaElKQ==", - "dev": true, - "dependencies": { - "color-name": "~1.1.4" - }, - "engines": { - "node": ">=7.0.0" + "@octokit/endpoint": { + "version": "9.0.6", + "resolved": "https://registry.npmjs.org/@octokit/endpoint/-/endpoint-9.0.6.tgz", + "integrity": "sha512-H1fNTMA57HbkFESSt3Y9+FBICv+0jFceJFPWDePYlR/iMGrwM5ph+Dd4XRQs+8X+PUFURLQgX9ChPfhJ/1uNQw==", + "requires": { + "@octokit/types": "^13.1.0", + "universal-user-agent": "^6.0.0" } }, - "node_modules/color-name": { - "version": "1.1.4", - "resolved": "https://registry.npmjs.org/color-name/-/color-name-1.1.4.tgz", - "integrity": "sha512-dOy+3AuW3a2wNbZHIuMZpTcgjGuLU/uBL/ubcZF9OXbDo8ff4O8yVp5Bf0efS8uEoYo5q4Fx7dY9OgQGXgAsQA==", - "dev": true - }, - "node_modules/color-support": { - "version": "1.1.3", - "resolved": "https://registry.npmjs.org/color-support/-/color-support-1.1.3.tgz", - "integrity": "sha512-qiBjkpbMLO/HL68y+lh4q0/O1MZFj2RX6X/KmMa3+gJD3z+WwI1ZzDHysvqHGS3mP6mznPckpXmw1nI9cJjyRg==", - "bin": { - "color-support": "bin.js" + "@octokit/graphql": { + "version": "7.1.1", + "resolved": "https://registry.npmjs.org/@octokit/graphql/-/graphql-7.1.1.tgz", + "integrity": "sha512-3mkDltSfcDUoa176nlGoA32RGjeWjl3K7F/BwHwRMJUW/IteSa4bnSV8p2ThNkcIcZU2umkZWxwETSSCJf2Q7g==", + "requires": { + "@octokit/request": "^8.4.1", + "@octokit/types": "^13.0.0", + "universal-user-agent": "^6.0.0" } }, - "node_modules/concat-map": { - "version": "0.0.1", - "resolved": "https://registry.npmjs.org/concat-map/-/concat-map-0.0.1.tgz", - "integrity": "sha1-2Klr13/Wjfd5OnMDajug1UBdR3s=" - }, - "node_modules/console-control-strings": { - "version": "1.1.0", - "resolved": "https://registry.npmjs.org/console-control-strings/-/console-control-strings-1.1.0.tgz", - "integrity": "sha1-PXz0Rk22RG6mRL9LOVB/mFEAjo4=" + "@octokit/openapi-types": { + "version": "24.2.0", + "resolved": "https://registry.npmjs.org/@octokit/openapi-types/-/openapi-types-24.2.0.tgz", + "integrity": "sha512-9sIH3nSUttelJSXUrmGzl7QUBFul0/mB8HRYl3fOlgHbIWG+WnYDXU3v/2zMtAvuzZ/ed00Ei6on975FhBfzrg==" }, - "node_modules/cross-spawn": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/cross-spawn/-/cross-spawn-7.0.3.tgz", - "integrity": "sha512-iRDPJKUPVEND7dHPO8rkbOnPpyDygcDFtWjpeWNCgy8WP2rXcxXL8TskReQl6OrB2G7+UJrags1q15Fudc7G6w==", - "dev": true, - "dependencies": { - "path-key": "^3.1.0", - "shebang-command": "^2.0.0", - "which": "^2.0.1" + "@octokit/plugin-paginate-rest": { + "version": "9.2.2", + "resolved": "https://registry.npmjs.org/@octokit/plugin-paginate-rest/-/plugin-paginate-rest-9.2.2.tgz", + "integrity": "sha512-u3KYkGF7GcZnSD/3UP0S7K5XUFT2FkOQdcfXZGZQPGv3lm4F2Xbf71lvjldr8c1H3nNbF+33cLEkWYbokGWqiQ==", + "requires": { + "@octokit/types": "^12.6.0" }, - "engines": { - "node": ">= 8" + "dependencies": { + "@octokit/openapi-types": { + "version": "20.0.0", + "resolved": "https://registry.npmjs.org/@octokit/openapi-types/-/openapi-types-20.0.0.tgz", + "integrity": "sha512-EtqRBEjp1dL/15V7WiX5LJMIxxkdiGJnabzYx5Apx4FkQIFgAfKumXeYAqqJCj1s+BMX4cPFIFC4OLCR6stlnA==" + }, + "@octokit/types": { + "version": "12.6.0", + "resolved": "https://registry.npmjs.org/@octokit/types/-/types-12.6.0.tgz", + "integrity": "sha512-1rhSOfRa6H9w4YwK0yrf5faDaDTb+yLyBUKOCV4xtCDB5VmIPqd/v9yr9o6SAzOAlRxMiRiCic6JVM1/kunVkw==", + "requires": { + "@octokit/openapi-types": "^20.0.0" + } + } } }, - "node_modules/debug": { - "version": "4.3.4", - "resolved": "https://registry.npmjs.org/debug/-/debug-4.3.4.tgz", - "integrity": "sha512-PRWFHuSU3eDtQJPvnNY7Jcket1j0t5OuOsFzPPzsekD52Zl8qUfFIPEiswXqIvHWGVHOgX+7G/vCNNhehwxfkQ==", - "dependencies": { - "ms": "2.1.2" - }, - "engines": { - "node": ">=6.0" + "@octokit/plugin-rest-endpoint-methods": { + "version": "10.4.1", + "resolved": "https://registry.npmjs.org/@octokit/plugin-rest-endpoint-methods/-/plugin-rest-endpoint-methods-10.4.1.tgz", + "integrity": "sha512-xV1b+ceKV9KytQe3zCVqjg+8GTGfDYwaT1ATU5isiUyVtlVAO3HNdzpS4sr4GBx4hxQ46s7ITtZrAsxG22+rVg==", + "requires": { + "@octokit/types": "^12.6.0" }, - "peerDependenciesMeta": { - "supports-color": { - "optional": true + "dependencies": { + "@octokit/openapi-types": { + "version": "20.0.0", + "resolved": "https://registry.npmjs.org/@octokit/openapi-types/-/openapi-types-20.0.0.tgz", + "integrity": "sha512-EtqRBEjp1dL/15V7WiX5LJMIxxkdiGJnabzYx5Apx4FkQIFgAfKumXeYAqqJCj1s+BMX4cPFIFC4OLCR6stlnA==" + }, + "@octokit/types": { + "version": "12.6.0", + "resolved": "https://registry.npmjs.org/@octokit/types/-/types-12.6.0.tgz", + "integrity": "sha512-1rhSOfRa6H9w4YwK0yrf5faDaDTb+yLyBUKOCV4xtCDB5VmIPqd/v9yr9o6SAzOAlRxMiRiCic6JVM1/kunVkw==", + "requires": { + "@octokit/openapi-types": "^20.0.0" + } } } }, - "node_modules/decompress-response": { - "version": "4.2.1", - "resolved": "https://registry.npmjs.org/decompress-response/-/decompress-response-4.2.1.tgz", - "integrity": "sha512-jOSne2qbyE+/r8G1VU+G/82LBs2Fs4LAsTiLSHOCOMZQl2OKZ6i8i4IyHemTe+/yIXOtTcRQMzPcgyhoFlqPkw==", - "dependencies": { - "mimic-response": "^2.0.0" - }, - "engines": { - "node": ">=8" + "@octokit/request": { + "version": "8.4.1", + "resolved": "https://registry.npmjs.org/@octokit/request/-/request-8.4.1.tgz", + "integrity": "sha512-qnB2+SY3hkCmBxZsR/MPCybNmbJe4KAlfWErXq+rBKkQJlbjdJeS85VI9r8UqeLYLvnAenU8Q1okM/0MBsAGXw==", + "requires": { + "@octokit/endpoint": "^9.0.6", + "@octokit/request-error": "^5.1.1", + "@octokit/types": "^13.1.0", + "universal-user-agent": "^6.0.0" } }, - "node_modules/deep-is": { - "version": "0.1.4", - "resolved": "https://registry.npmjs.org/deep-is/-/deep-is-0.1.4.tgz", - "integrity": "sha512-oIPzksmTg4/MriiaYGO+okXDT7ztn/w3Eptv/+gSIdMdKsJo0u4CfYNFJPy+4SKMuCqGw2wxnA+URMg3t8a/bQ==", - "dev": true + "@octokit/request-error": { + "version": "5.1.1", + "resolved": "https://registry.npmjs.org/@octokit/request-error/-/request-error-5.1.1.tgz", + "integrity": "sha512-v9iyEQJH6ZntoENr9/yXxjuezh4My67CBSu9r6Ve/05Iu5gNgnisNWOsoJHTP6k0Rr0+HQIpnH+kyammu90q/g==", + "requires": { + "@octokit/types": "^13.1.0", + "deprecation": "^2.0.0", + "once": "^1.4.0" + } }, - "node_modules/define-properties": { - "version": "1.1.3", - "resolved": "https://registry.npmjs.org/define-properties/-/define-properties-1.1.3.tgz", - "integrity": "sha512-3MqfYKj2lLzdMSf8ZIZE/V+Zuy+BgD6f164e8K2w7dgnpKArBDerGYpM46IYYcjnkdPNMjPk9A6VFB8+3SKlXQ==", - "dev": true, - "dependencies": { - "object-keys": "^1.0.12" - }, - "engines": { - "node": ">= 0.4" + "@octokit/types": { + "version": "13.10.0", + "resolved": "https://registry.npmjs.org/@octokit/types/-/types-13.10.0.tgz", + "integrity": "sha512-ifLaO34EbbPj0Xgro4G5lP5asESjwHracYJvVaPIyXMuiuXLlhic3S47cBdTb+jfODkTE5YtGCLt3Ay3+J97sA==", + "requires": { + "@octokit/openapi-types": "^24.2.0" } }, - "node_modules/delegates": { - "version": "1.0.0", - "resolved": "https://registry.npmjs.org/delegates/-/delegates-1.0.0.tgz", - "integrity": "sha1-hMbhWbgZBP3KWaDvRM2HDTElD5o=" + "before-after-hook": { + "version": "2.2.3", + "resolved": "https://registry.npmjs.org/before-after-hook/-/before-after-hook-2.2.3.tgz", + "integrity": "sha512-NzUnlZexiaH/46WDhANlyR2bXRopNg4F/zuSA3OpZnllCUgRaOF2znDioDWrmbNVsuZk6l9pMquQB38cfBZwkQ==" }, - "node_modules/deprecation": { + "deprecation": { "version": "2.3.1", "resolved": "https://registry.npmjs.org/deprecation/-/deprecation-2.3.1.tgz", "integrity": "sha512-xmHIy4F3scKVwMsQ4WnVaS8bHOx0DmVwRywosKhaILI0ywMDWPtBSku2HNxRvF7jtwDRsoEwYQSfbxj8b7RlJQ==" }, - "node_modules/detect-libc": { - "version": "1.0.3", - "resolved": "https://registry.npmjs.org/detect-libc/-/detect-libc-1.0.3.tgz", - "integrity": "sha1-+hN8S9aY7fVc1c0CrFWfkaTEups=", - "bin": { - "detect-libc": "bin/detect-libc.js" - }, - "engines": { - "node": ">=0.10" - } - }, - "node_modules/doctrine": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/doctrine/-/doctrine-3.0.0.tgz", - "integrity": "sha512-yS+Q5i3hBf7GBkd4KG8a7eBNNWNGLTaEwwYWUijIYM7zrlYDM0BFXHjjPWlWZ1Rg7UaddZeIDmi9jF3HmqiQ2w==", - "dev": true, - "dependencies": { - "esutils": "^2.0.2" - }, - "engines": { - "node": ">=6.0.0" - } + "lodash": { + "version": "4.17.21", + "resolved": "https://registry.npmjs.org/lodash/-/lodash-4.17.21.tgz", + "integrity": "sha512-v2kDEe57lecTulaDIuNTPy3Ry4gLGJ6Z1O3vE1krgXZNrsQ+LFTGHVxVjcXPs17LhbZVGedAJv8XZ1tvj5FvSg==" }, - "node_modules/emoji-regex": { - "version": "8.0.0", - "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", - "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" + "luxon": { + "version": "3.7.1", + "resolved": "https://registry.npmjs.org/luxon/-/luxon-3.7.1.tgz", + "integrity": "sha512-RkRWjA926cTvz5rAb1BqyWkKbbjzCGchDUIKMCUvNi17j6f6j8uHGDV82Aqcqtzd+icoYpELmG3ksgGiFNNcNg==" }, - "node_modules/enquirer": { - "version": "2.3.6", - "resolved": "https://registry.npmjs.org/enquirer/-/enquirer-2.3.6.tgz", - "integrity": "sha512-yjNnPr315/FjS4zIsUxYguYUPP2e1NK4d7E7ZOLiyYCcbFBiTMyID+2wvm2w6+pZ/odMA7cRkjhsPbltwBOrLg==", - "dev": true, - "dependencies": { - "ansi-colors": "^4.1.1" - }, - "engines": { - "node": ">=8.6" + "once": { + "version": "1.4.0", + "resolved": "https://registry.npmjs.org/once/-/once-1.4.0.tgz", + "integrity": "sha512-lNaJgI+2Q5URQBkccEKHTQOPaXdUxnZZElQTZY0MFUAuaEqe1E+Nyvgdz/aIyNi6Z9MzO5dv1H8n58/GELp3+w==", + "requires": { + "wrappy": "1" } }, - "node_modules/es-abstract": { - "version": "1.19.1", - "resolved": "https://registry.npmjs.org/es-abstract/-/es-abstract-1.19.1.tgz", - "integrity": "sha512-2vJ6tjA/UfqLm2MPs7jxVybLoB8i1t1Jd9R3kISld20sIxPcTbLuggQOUxeWeAvIUkduv/CfMjuh4WmiXr2v9w==", - "dev": true, - "dependencies": { - "call-bind": "^1.0.2", - "es-to-primitive": "^1.2.1", - "function-bind": "^1.1.1", - "get-intrinsic": "^1.1.1", - "get-symbol-description": "^1.0.0", - "has": "^1.0.3", - "has-symbols": "^1.0.2", - "internal-slot": "^1.0.3", - "is-callable": "^1.2.4", - "is-negative-zero": "^2.0.1", - "is-regex": "^1.1.4", - "is-shared-array-buffer": "^1.0.1", - "is-string": "^1.0.7", - "is-weakref": "^1.0.1", - "object-inspect": "^1.11.0", - "object-keys": "^1.1.1", - "object.assign": "^4.1.2", - "string.prototype.trimend": "^1.0.4", - "string.prototype.trimstart": "^1.0.4", - "unbox-primitive": "^1.0.1" - }, - "engines": { - "node": ">= 0.4" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/es-to-primitive": { - "version": "1.2.1", - "resolved": "https://registry.npmjs.org/es-to-primitive/-/es-to-primitive-1.2.1.tgz", - "integrity": "sha512-QCOllgZJtaUo9miYBcLChTUaHNjJF3PYs1VidD7AwiEj1kYxKeQTctLAezAOH5ZKRH0g2IgPn6KwB4IT8iRpvA==", - "dev": true, - "dependencies": { - "is-callable": "^1.1.4", - "is-date-object": "^1.0.1", - "is-symbol": "^1.0.2" - }, - "engines": { - "node": ">= 0.4" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/escape-string-regexp": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/escape-string-regexp/-/escape-string-regexp-4.0.0.tgz", - "integrity": "sha512-TtpcNJ3XAzx3Gq8sWRzJaVajRs0uVxA2YAkdb1jm2YkPz4G6egUFAyA3n5vtEIZefPk5Wa4UXbKuS5fKkJWdgA==", - "dev": true, - "engines": { - "node": ">=10" - }, - "funding": { - "url": "https://github.com/sponsors/sindresorhus" - } - }, - "node_modules/eslint": { - "version": "7.32.0", - "resolved": "https://registry.npmjs.org/eslint/-/eslint-7.32.0.tgz", - "integrity": "sha512-VHZ8gX+EDfz+97jGcgyGCyRia/dPOd6Xh9yPv8Bl1+SoaIwD+a/vlrOmGRUyOYu7MwUhc7CxqeaDZU13S4+EpA==", - "dev": true, - "dependencies": { - "@babel/code-frame": "7.12.11", - "@eslint/eslintrc": "^0.4.3", - "@humanwhocodes/config-array": "^0.5.0", - "ajv": "^6.10.0", - "chalk": "^4.0.0", - "cross-spawn": "^7.0.2", - "debug": "^4.0.1", - "doctrine": "^3.0.0", - "enquirer": "^2.3.5", - "escape-string-regexp": "^4.0.0", - "eslint-scope": "^5.1.1", - "eslint-utils": "^2.1.0", - "eslint-visitor-keys": "^2.0.0", - "espree": "^7.3.1", - "esquery": "^1.4.0", - "esutils": "^2.0.2", - "fast-deep-equal": "^3.1.3", - "file-entry-cache": "^6.0.1", - "functional-red-black-tree": "^1.0.1", - "glob-parent": "^5.1.2", - "globals": "^13.6.0", - "ignore": "^4.0.6", - "import-fresh": "^3.0.0", - "imurmurhash": "^0.1.4", - "is-glob": "^4.0.0", - "js-yaml": "^3.13.1", - "json-stable-stringify-without-jsonify": "^1.0.1", - "levn": "^0.4.1", - "lodash.merge": "^4.6.2", - "minimatch": "^3.0.4", - "natural-compare": "^1.4.0", - "optionator": "^0.9.1", - "progress": "^2.0.0", - "regexpp": "^3.1.0", - "semver": "^7.2.1", - "strip-ansi": "^6.0.0", - "strip-json-comments": "^3.1.0", - "table": "^6.0.9", - "text-table": "^0.2.0", - "v8-compile-cache": "^2.0.3" - }, - "bin": { - "eslint": "bin/eslint.js" - }, - "engines": { - "node": "^10.12.0 || >=12.0.0" - }, - "funding": { - "url": "https://opencollective.com/eslint" - } - }, - "node_modules/eslint-config-standard": { - "version": "16.0.3", - "resolved": "https://registry.npmjs.org/eslint-config-standard/-/eslint-config-standard-16.0.3.tgz", - "integrity": "sha512-x4fmJL5hGqNJKGHSjnLdgA6U6h1YW/G2dW9fA+cyVur4SK6lyue8+UgNKWlZtUDTXvgKDD/Oa3GQjmB5kjtVvg==", - "dev": true, - "funding": [ - { - "type": "github", - "url": "https://github.com/sponsors/feross" - }, - { - "type": "patreon", - "url": "https://www.patreon.com/feross" - }, - { - "type": "consulting", - "url": "https://feross.org/support" - } - ], - "peerDependencies": { - "eslint": "^7.12.1", - "eslint-plugin-import": "^2.22.1", - "eslint-plugin-node": "^11.1.0", - "eslint-plugin-promise": "^4.2.1 || ^5.0.0" - } - }, - "node_modules/eslint-import-resolver-node": { - "version": "0.3.6", - "resolved": "https://registry.npmjs.org/eslint-import-resolver-node/-/eslint-import-resolver-node-0.3.6.tgz", - "integrity": "sha512-0En0w03NRVMn9Uiyn8YRPDKvWjxCWkslUEhGNTdGx15RvPJYQ+lbOlqrlNI2vEAs4pDYK4f/HN2TbDmk5TP0iw==", - "dev": true, - "dependencies": { - "debug": "^3.2.7", - "resolve": "^1.20.0" - } - }, - "node_modules/eslint-import-resolver-node/node_modules/debug": { - "version": "3.2.7", - "resolved": "https://registry.npmjs.org/debug/-/debug-3.2.7.tgz", - "integrity": "sha512-CFjzYYAi4ThfiQvizrFQevTTXHtnCqWfe7x1AhgEscTz6ZbLbfoLRLPugTQyBth6f8ZERVUSyWHFD/7Wu4t1XQ==", - "dev": true, - "dependencies": { - "ms": "^2.1.1" - } - }, - "node_modules/eslint-module-utils": { - "version": "2.7.3", - "resolved": "https://registry.npmjs.org/eslint-module-utils/-/eslint-module-utils-2.7.3.tgz", - "integrity": "sha512-088JEC7O3lDZM9xGe0RerkOMd0EjFl+Yvd1jPWIkMT5u3H9+HC34mWWPnqPrN13gieT9pBOO+Qt07Nb/6TresQ==", - "dev": true, - "dependencies": { - "debug": "^3.2.7", - "find-up": "^2.1.0" - }, - "engines": { - "node": ">=4" - } - }, - "node_modules/eslint-module-utils/node_modules/debug": { - "version": "3.2.7", - "resolved": "https://registry.npmjs.org/debug/-/debug-3.2.7.tgz", - "integrity": "sha512-CFjzYYAi4ThfiQvizrFQevTTXHtnCqWfe7x1AhgEscTz6ZbLbfoLRLPugTQyBth6f8ZERVUSyWHFD/7Wu4t1XQ==", - "dev": true, - "dependencies": { - "ms": "^2.1.1" - } - }, - "node_modules/eslint-plugin-es": { - "version": "3.0.1", - "resolved": "https://registry.npmjs.org/eslint-plugin-es/-/eslint-plugin-es-3.0.1.tgz", - "integrity": "sha512-GUmAsJaN4Fc7Gbtl8uOBlayo2DqhwWvEzykMHSCZHU3XdJ+NSzzZcVhXh3VxX5icqQ+oQdIEawXX8xkR3mIFmQ==", - "dev": true, - "dependencies": { - "eslint-utils": "^2.0.0", - "regexpp": "^3.0.0" - }, - "engines": { - "node": ">=8.10.0" - }, - "funding": { - "url": "https://github.com/sponsors/mysticatea" - }, - "peerDependencies": { - "eslint": ">=4.19.1" - } - }, - "node_modules/eslint-plugin-import": { - "version": "2.25.4", - "resolved": "https://registry.npmjs.org/eslint-plugin-import/-/eslint-plugin-import-2.25.4.tgz", - "integrity": "sha512-/KJBASVFxpu0xg1kIBn9AUa8hQVnszpwgE7Ld0lKAlx7Ie87yzEzCgSkekt+le/YVhiaosO4Y14GDAOc41nfxA==", - "dev": true, - "dependencies": { - "array-includes": "^3.1.4", - "array.prototype.flat": "^1.2.5", - "debug": "^2.6.9", - "doctrine": "^2.1.0", - "eslint-import-resolver-node": "^0.3.6", - "eslint-module-utils": "^2.7.2", - "has": "^1.0.3", - "is-core-module": "^2.8.0", - "is-glob": "^4.0.3", - "minimatch": "^3.0.4", - "object.values": "^1.1.5", - "resolve": "^1.20.0", - "tsconfig-paths": "^3.12.0" - }, - "engines": { - "node": ">=4" - }, - "peerDependencies": { - "eslint": "^2 || ^3 || ^4 || ^5 || ^6 || ^7.2.0 || ^8" - } - }, - "node_modules/eslint-plugin-import/node_modules/debug": { - "version": "2.6.9", - "resolved": "https://registry.npmjs.org/debug/-/debug-2.6.9.tgz", - "integrity": "sha512-bC7ElrdJaJnPbAP+1EotYvqZsb3ecl5wi6Bfi6BJTUcNowp6cvspg0jXznRTKDjm/E7AdgFBVeAPVMNcKGsHMA==", - "dev": true, - "dependencies": { - "ms": "2.0.0" - } - }, - "node_modules/eslint-plugin-import/node_modules/doctrine": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/doctrine/-/doctrine-2.1.0.tgz", - "integrity": "sha512-35mSku4ZXK0vfCuHEDAwt55dg2jNajHZ1odvF+8SSr82EsZY4QmXfuWso8oEd8zRhVObSN18aM0CjSdoBX7zIw==", - "dev": true, - "dependencies": { - "esutils": "^2.0.2" - }, - "engines": { - "node": ">=0.10.0" - } - }, - "node_modules/eslint-plugin-import/node_modules/ms": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/ms/-/ms-2.0.0.tgz", - "integrity": "sha1-VgiurfwAvmwpAd9fmGF4jeDVl8g=", - "dev": true - }, - "node_modules/eslint-plugin-node": { - "version": "11.1.0", - "resolved": "https://registry.npmjs.org/eslint-plugin-node/-/eslint-plugin-node-11.1.0.tgz", - "integrity": "sha512-oUwtPJ1W0SKD0Tr+wqu92c5xuCeQqB3hSCHasn/ZgjFdA9iDGNkNf2Zi9ztY7X+hNuMib23LNGRm6+uN+KLE3g==", - "dev": true, - "dependencies": { - "eslint-plugin-es": "^3.0.0", - "eslint-utils": "^2.0.0", - "ignore": "^5.1.1", - "minimatch": "^3.0.4", - "resolve": "^1.10.1", - "semver": "^6.1.0" - }, - "engines": { - "node": ">=8.10.0" - }, - "peerDependencies": { - "eslint": ">=5.16.0" - } - }, - "node_modules/eslint-plugin-node/node_modules/ignore": { - "version": "5.2.0", - "resolved": "https://registry.npmjs.org/ignore/-/ignore-5.2.0.tgz", - "integrity": "sha512-CmxgYGiEPCLhfLnpPp1MoRmifwEIOgjcHXxOBjv7mY96c+eWScsOP9c112ZyLdWHi0FxHjI+4uVhKYp/gcdRmQ==", - "dev": true, - "engines": { - "node": ">= 4" - } - }, - "node_modules/eslint-plugin-node/node_modules/semver": { - "version": "6.3.0", - "resolved": "https://registry.npmjs.org/semver/-/semver-6.3.0.tgz", - "integrity": "sha512-b39TBaTSfV6yBrapU89p5fKekE2m/NwnDocOVruQFS1/veMgdzuPcnOM34M6CwxW8jH/lxEa5rBoDeUwu5HHTw==", - "dev": true, - "bin": { - "semver": "bin/semver.js" - } - }, - "node_modules/eslint-plugin-promise": { - "version": "5.2.0", - "resolved": "https://registry.npmjs.org/eslint-plugin-promise/-/eslint-plugin-promise-5.2.0.tgz", - "integrity": "sha512-SftLb1pUG01QYq2A/hGAWfDRXqYD82zE7j7TopDOyNdU+7SvvoXREls/+PRTY17vUXzXnZA/zfnyKgRH6x4JJw==", - "dev": true, - "engines": { - "node": "^10.12.0 || >=12.0.0" - }, - "peerDependencies": { - "eslint": "^7.0.0" - } - }, - "node_modules/eslint-scope": { - "version": "5.1.1", - "resolved": "https://registry.npmjs.org/eslint-scope/-/eslint-scope-5.1.1.tgz", - "integrity": "sha512-2NxwbF/hZ0KpepYN0cNbo+FN6XoK7GaHlQhgx/hIZl6Va0bF45RQOOwhLIy8lQDbuCiadSLCBnH2CFYquit5bw==", - "dev": true, - "dependencies": { - "esrecurse": "^4.3.0", - "estraverse": "^4.1.1" - }, - "engines": { - "node": ">=8.0.0" - } - }, - "node_modules/eslint-utils": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/eslint-utils/-/eslint-utils-2.1.0.tgz", - "integrity": "sha512-w94dQYoauyvlDc43XnGB8lU3Zt713vNChgt4EWwhXAP2XkBvndfxF0AgIqKOOasjPIPzj9JqgwkwbCYD0/V3Zg==", - "dev": true, - "dependencies": { - "eslint-visitor-keys": "^1.1.0" - }, - "engines": { - "node": ">=6" - }, - "funding": { - "url": "https://github.com/sponsors/mysticatea" - } - }, - "node_modules/eslint-utils/node_modules/eslint-visitor-keys": { - "version": "1.3.0", - "resolved": "https://registry.npmjs.org/eslint-visitor-keys/-/eslint-visitor-keys-1.3.0.tgz", - "integrity": "sha512-6J72N8UNa462wa/KFODt/PJ3IU60SDpC3QXC1Hjc1BXXpfL2C9R5+AU7jhe0F6GREqVMh4Juu+NY7xn+6dipUQ==", - "dev": true, - "engines": { - "node": ">=4" - } - }, - "node_modules/eslint-visitor-keys": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/eslint-visitor-keys/-/eslint-visitor-keys-2.1.0.tgz", - "integrity": "sha512-0rSmRBzXgDzIsD6mGdJgevzgezI534Cer5L/vyMX0kHzT/jiB43jRhd9YUlMGYLQy2zprNmoT8qasCGtY+QaKw==", - "dev": true, - "engines": { - "node": ">=10" - } - }, - "node_modules/espree": { - "version": "7.3.1", - "resolved": "https://registry.npmjs.org/espree/-/espree-7.3.1.tgz", - "integrity": "sha512-v3JCNCE64umkFpmkFGqzVKsOT0tN1Zr+ueqLZfpV1Ob8e+CEgPWa+OxCoGH3tnhimMKIaBm4m/vaRpJ/krRz2g==", - "dev": true, - "dependencies": { - "acorn": "^7.4.0", - "acorn-jsx": "^5.3.1", - "eslint-visitor-keys": "^1.3.0" - }, - "engines": { - "node": "^10.12.0 || >=12.0.0" - } - }, - "node_modules/espree/node_modules/eslint-visitor-keys": { - "version": "1.3.0", - "resolved": "https://registry.npmjs.org/eslint-visitor-keys/-/eslint-visitor-keys-1.3.0.tgz", - "integrity": "sha512-6J72N8UNa462wa/KFODt/PJ3IU60SDpC3QXC1Hjc1BXXpfL2C9R5+AU7jhe0F6GREqVMh4Juu+NY7xn+6dipUQ==", - "dev": true, - "engines": { - "node": ">=4" - } - }, - "node_modules/esprima": { - "version": "4.0.1", - "resolved": "https://registry.npmjs.org/esprima/-/esprima-4.0.1.tgz", - "integrity": "sha512-eGuFFw7Upda+g4p+QHvnW0RyTX/SVeJBDM/gCtMARO0cLuT2HcEKnTPvhjV6aGeqrCB/sbNop0Kszm0jsaWU4A==", - "dev": true, - "bin": { - "esparse": "bin/esparse.js", - "esvalidate": "bin/esvalidate.js" - }, - "engines": { - "node": ">=4" - } - }, - "node_modules/esquery": { - "version": "1.4.0", - "resolved": "https://registry.npmjs.org/esquery/-/esquery-1.4.0.tgz", - "integrity": "sha512-cCDispWt5vHHtwMY2YrAQ4ibFkAL8RbH5YGBnZBc90MolvvfkkQcJro/aZiAQUlQ3qgrYS6D6v8Gc5G5CQsc9w==", - "dev": true, - "dependencies": { - "estraverse": "^5.1.0" - }, - "engines": { - "node": ">=0.10" - } - }, - "node_modules/esquery/node_modules/estraverse": { - "version": "5.3.0", - "resolved": "https://registry.npmjs.org/estraverse/-/estraverse-5.3.0.tgz", - "integrity": "sha512-MMdARuVEQziNTeJD8DgMqmhwR11BRQ/cBP+pLtYdSTnf3MIO8fFeiINEbX36ZdNlfU/7A9f3gUw49B3oQsvwBA==", - "dev": true, - "engines": { - "node": ">=4.0" - } - }, - "node_modules/esrecurse": { - "version": "4.3.0", - "resolved": "https://registry.npmjs.org/esrecurse/-/esrecurse-4.3.0.tgz", - "integrity": "sha512-KmfKL3b6G+RXvP8N1vr3Tq1kL/oCFgn2NYXEtqP8/L3pKapUA4G8cFVaoF3SU323CD4XypR/ffioHmkti6/Tag==", - "dev": true, - "dependencies": { - "estraverse": "^5.2.0" - }, - "engines": { - "node": ">=4.0" - } - }, - "node_modules/esrecurse/node_modules/estraverse": { - "version": "5.3.0", - "resolved": "https://registry.npmjs.org/estraverse/-/estraverse-5.3.0.tgz", - "integrity": "sha512-MMdARuVEQziNTeJD8DgMqmhwR11BRQ/cBP+pLtYdSTnf3MIO8fFeiINEbX36ZdNlfU/7A9f3gUw49B3oQsvwBA==", - "dev": true, - "engines": { - "node": ">=4.0" - } - }, - "node_modules/estraverse": { - "version": "4.3.0", - "resolved": "https://registry.npmjs.org/estraverse/-/estraverse-4.3.0.tgz", - "integrity": "sha512-39nnKffWz8xN1BU/2c79n9nB9HDzo0niYUqx6xyqUnyoAnQyyWpOTdZEeiCch8BBu515t4wp9ZmgVfVhn9EBpw==", - "dev": true, - "engines": { - "node": ">=4.0" - } - }, - "node_modules/esutils": { - "version": "2.0.3", - "resolved": "https://registry.npmjs.org/esutils/-/esutils-2.0.3.tgz", - "integrity": "sha512-kVscqXk4OCp68SZ0dkgEKVi6/8ij300KBWTJq32P/dYeWTSwK41WyTxalN1eRmA5Z9UU/LX9D7FWSmV9SAYx6g==", - "dev": true, - "engines": { - "node": ">=0.10.0" - } - }, - "node_modules/fast-deep-equal": { - "version": "3.1.3", - "resolved": "https://registry.npmjs.org/fast-deep-equal/-/fast-deep-equal-3.1.3.tgz", - "integrity": "sha512-f3qQ9oQy9j2AhBe/H9VC91wLmKBCCU/gDOnKNAYG5hswO7BLKj09Hc5HYNz9cGI++xlpDCIgDaitVs03ATR84Q==", - "dev": true - }, - "node_modules/fast-json-stable-stringify": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/fast-json-stable-stringify/-/fast-json-stable-stringify-2.1.0.tgz", - "integrity": "sha512-lhd/wF+Lk98HZoTCtlVraHtfh5XYijIjalXck7saUtuanSDyLMxnHhSXEDJqHxD7msR8D0uCmqlkwjCV8xvwHw==", - "dev": true - }, - "node_modules/fast-levenshtein": { - "version": "2.0.6", - "resolved": "https://registry.npmjs.org/fast-levenshtein/-/fast-levenshtein-2.0.6.tgz", - "integrity": "sha1-PYpcZog6FqMMqGQ+hR8Zuqd5eRc=", - "dev": true - }, - "node_modules/file-entry-cache": { - "version": "6.0.1", - "resolved": "https://registry.npmjs.org/file-entry-cache/-/file-entry-cache-6.0.1.tgz", - "integrity": "sha512-7Gps/XWymbLk2QLYK4NzpMOrYjMhdIxXuIvy2QBsLE6ljuodKvdkWs/cpyJJ3CVIVpH0Oi1Hvg1ovbMzLdFBBg==", - "dev": true, - "dependencies": { - "flat-cache": "^3.0.4" - }, - "engines": { - "node": "^10.12.0 || >=12.0.0" - } - }, - "node_modules/find-up": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/find-up/-/find-up-2.1.0.tgz", - "integrity": "sha1-RdG35QbHF93UgndaK3eSCjwMV6c=", - "dev": true, - "dependencies": { - "locate-path": "^2.0.0" - }, - "engines": { - "node": ">=4" - } - }, - "node_modules/flat-cache": { - "version": "3.0.4", - "resolved": "https://registry.npmjs.org/flat-cache/-/flat-cache-3.0.4.tgz", - "integrity": "sha512-dm9s5Pw7Jc0GvMYbshN6zchCA9RgQlzzEZX3vylR9IqFfS8XciblUXOKfW6SiuJ0e13eDYZoZV5wdrev7P3Nwg==", - "dev": true, - "dependencies": { - "flatted": "^3.1.0", - "rimraf": "^3.0.2" - }, - "engines": { - "node": "^10.12.0 || >=12.0.0" - } - }, - "node_modules/flatted": { - "version": "3.2.5", - "resolved": "https://registry.npmjs.org/flatted/-/flatted-3.2.5.tgz", - "integrity": "sha512-WIWGi2L3DyTUvUrwRKgGi9TwxQMUEqPOPQBVi71R96jZXJdFskXEmf54BoZaS1kknGODoIGASGEzBUYdyMCBJg==", - "dev": true - }, - "node_modules/fs-minipass": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/fs-minipass/-/fs-minipass-2.1.0.tgz", - "integrity": "sha512-V/JgOLFCS+R6Vcq0slCuaeWEdNC3ouDlJMNIsacH2VtALiu9mV4LPrHc5cDl8k5aw6J8jwgWWpiTo5RYhmIzvg==", - "dependencies": { - "minipass": "^3.0.0" - }, - "engines": { - "node": ">= 8" - } - }, - "node_modules/fs.realpath": { - "version": "1.0.0", - "resolved": "https://registry.npmjs.org/fs.realpath/-/fs.realpath-1.0.0.tgz", - "integrity": "sha1-FQStJSMVjKpA20onh8sBQRmU6k8=" - }, - "node_modules/function-bind": { - "version": "1.1.1", - "resolved": "https://registry.npmjs.org/function-bind/-/function-bind-1.1.1.tgz", - "integrity": "sha512-yIovAzMX49sF8Yl58fSCWJ5svSLuaibPxXQJFLmBObTuCr0Mf1KiPopGM9NiFjiYBCbfaa2Fh6breQ6ANVTI0A==", - "dev": true - }, - "node_modules/functional-red-black-tree": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/functional-red-black-tree/-/functional-red-black-tree-1.0.1.tgz", - "integrity": "sha1-GwqzvVU7Kg1jmdKcDj6gslIHgyc=", - "dev": true - }, - "node_modules/gauge": { - "version": "3.0.2", - "resolved": "https://registry.npmjs.org/gauge/-/gauge-3.0.2.tgz", - "integrity": "sha512-+5J6MS/5XksCuXq++uFRsnUd7Ovu1XenbeuIuNRJxYWjgQbPuFhT14lAvsWfqfAmnwluf1OwMjz39HjfLPci0Q==", - "dependencies": { - "aproba": "^1.0.3 || ^2.0.0", - "color-support": "^1.1.2", - "console-control-strings": "^1.0.0", - "has-unicode": "^2.0.1", - "object-assign": "^4.1.1", - "signal-exit": "^3.0.0", - "string-width": "^4.2.3", - "strip-ansi": "^6.0.1", - "wide-align": "^1.1.2" - }, - "engines": { - "node": ">=10" - } - }, - "node_modules/get-intrinsic": { - "version": "1.1.1", - "resolved": "https://registry.npmjs.org/get-intrinsic/-/get-intrinsic-1.1.1.tgz", - "integrity": "sha512-kWZrnVM42QCiEA2Ig1bG8zjoIMOgxWwYCEeNdwY6Tv/cOSeGpcoX4pXHfKUxNKVoArnrEr2e9srnAxxGIraS9Q==", - "dev": true, - "dependencies": { - "function-bind": "^1.1.1", - "has": "^1.0.3", - "has-symbols": "^1.0.1" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/get-symbol-description": { - "version": "1.0.0", - "resolved": "https://registry.npmjs.org/get-symbol-description/-/get-symbol-description-1.0.0.tgz", - "integrity": "sha512-2EmdH1YvIQiZpltCNgkuiUnyukzxM/R6NDJX31Ke3BG1Nq5b0S2PhX59UKi9vZpPDQVdqn+1IcaAwnzTT5vCjw==", - "dev": true, - "dependencies": { - "call-bind": "^1.0.2", - "get-intrinsic": "^1.1.1" - }, - "engines": { - "node": ">= 0.4" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/glob": { - "version": "7.2.0", - "resolved": "https://registry.npmjs.org/glob/-/glob-7.2.0.tgz", - "integrity": "sha512-lmLf6gtyrPq8tTjSmrO94wBeQbFR3HbLHbuyD69wuyQkImp2hWqMGB47OX65FBkPffO641IP9jWa1z4ivqG26Q==", - "dependencies": { - "fs.realpath": "^1.0.0", - "inflight": "^1.0.4", - "inherits": "2", - "minimatch": "^3.0.4", - "once": "^1.3.0", - "path-is-absolute": "^1.0.0" - }, - "engines": { - "node": "*" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - } - }, - "node_modules/glob-parent": { - "version": "5.1.2", - "resolved": "https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz", - "integrity": "sha512-AOIgSQCepiJYwP3ARnGx+5VnTu2HBYdzbGP45eLw1vr3zB3vZLeyed1sC9hnbcOc9/SrMyM5RPQrkGz4aS9Zow==", - "dev": true, - "dependencies": { - "is-glob": "^4.0.1" - }, - "engines": { - "node": ">= 6" - } - }, - "node_modules/globals": { - "version": "13.13.0", - "resolved": "https://registry.npmjs.org/globals/-/globals-13.13.0.tgz", - "integrity": "sha512-EQ7Q18AJlPwp3vUDL4mKA0KXrXyNIQyWon6T6XQiBQF0XHvRsiCSrWmmeATpUzdJN2HhWZU6Pdl0a9zdep5p6A==", - "dev": true, - "dependencies": { - "type-fest": "^0.20.2" - }, - "engines": { - "node": ">=8" - }, - "funding": { - "url": "https://github.com/sponsors/sindresorhus" - } - }, - "node_modules/has": { - "version": "1.0.3", - "resolved": "https://registry.npmjs.org/has/-/has-1.0.3.tgz", - "integrity": "sha512-f2dvO0VU6Oej7RkWJGrehjbzMAjFp5/VKPp5tTpWIV4JHHZK1/BxbFRtf/siA2SWTe09caDmVtYYzWEIbBS4zw==", - "dev": true, - "dependencies": { - "function-bind": "^1.1.1" - }, - "engines": { - "node": ">= 0.4.0" - } - }, - "node_modules/has-bigints": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/has-bigints/-/has-bigints-1.0.1.tgz", - "integrity": "sha512-LSBS2LjbNBTf6287JEbEzvJgftkF5qFkmCo9hDRpAzKhUOlJ+hx8dd4USs00SgsUNwc4617J9ki5YtEClM2ffA==", - "dev": true, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/has-flag": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/has-flag/-/has-flag-4.0.0.tgz", - "integrity": "sha512-EykJT/Q1KjTWctppgIAgfSO0tKVuZUjhgMr17kqTumMl6Afv3EISleU7qZUzoXDFTAHTDC4NOoG/ZxU3EvlMPQ==", - "dev": true, - "engines": { - "node": ">=8" - } - }, - "node_modules/has-symbols": { - "version": "1.0.3", - "resolved": "https://registry.npmjs.org/has-symbols/-/has-symbols-1.0.3.tgz", - "integrity": "sha512-l3LCuF6MgDNwTDKkdYGEihYjt5pRPbEg46rtlmnSPlUbgmB8LOIrKJbYYFBSbnPaJexMKtiPO8hmeRjRz2Td+A==", - "dev": true, - "engines": { - "node": ">= 0.4" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/has-tostringtag": { - "version": "1.0.0", - "resolved": "https://registry.npmjs.org/has-tostringtag/-/has-tostringtag-1.0.0.tgz", - "integrity": "sha512-kFjcSNhnlGV1kyoGk7OXKSawH5JOb/LzUc5w9B02hOTO0dfFRjbHQKvg1d6cf3HbeUmtU9VbbV3qzZ2Teh97WQ==", - "dev": true, - "dependencies": { - "has-symbols": "^1.0.2" - }, - "engines": { - "node": ">= 0.4" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/has-unicode": { - "version": "2.0.1", - "resolved": "https://registry.npmjs.org/has-unicode/-/has-unicode-2.0.1.tgz", - "integrity": "sha1-4Ob+aijPUROIVeCG0Wkedx3iqLk=" - }, - "node_modules/https-proxy-agent": { - "version": "5.0.0", - "resolved": "https://registry.npmjs.org/https-proxy-agent/-/https-proxy-agent-5.0.0.tgz", - "integrity": "sha512-EkYm5BcKUGiduxzSt3Eppko+PiNWNEpa4ySk9vTC6wDsQJW9rHSa+UhGNJoRYp7bz6Ht1eaRIa6QaJqO5rCFbA==", - "dependencies": { - "agent-base": "6", - "debug": "4" - }, - "engines": { - "node": ">= 6" - } - }, - "node_modules/ignore": { - "version": "4.0.6", - "resolved": "https://registry.npmjs.org/ignore/-/ignore-4.0.6.tgz", - "integrity": "sha512-cyFDKrqc/YdcWFniJhzI42+AzS+gNwmUzOSFcRCQYwySuBBBy/KjuxWLZ/FHEH6Moq1NizMOBWyTcv8O4OZIMg==", - "dev": true, - "engines": { - "node": ">= 4" - } - }, - "node_modules/import-fresh": { - "version": "3.3.0", - "resolved": "https://registry.npmjs.org/import-fresh/-/import-fresh-3.3.0.tgz", - "integrity": "sha512-veYYhQa+D1QBKznvhUHxb8faxlrwUnxseDAbAp457E0wLNio2bOSKnjYDhMj+YiAq61xrMGhQk9iXVk5FzgQMw==", - "dev": true, - "dependencies": { - "parent-module": "^1.0.0", - "resolve-from": "^4.0.0" - }, - "engines": { - "node": ">=6" - }, - "funding": { - "url": "https://github.com/sponsors/sindresorhus" - } - }, - "node_modules/imurmurhash": { - "version": "0.1.4", - "resolved": "https://registry.npmjs.org/imurmurhash/-/imurmurhash-0.1.4.tgz", - "integrity": "sha1-khi5srkoojixPcT7a21XbyMUU+o=", - "dev": true, - "engines": { - "node": ">=0.8.19" - } - }, - "node_modules/inflight": { - "version": "1.0.6", - "resolved": "https://registry.npmjs.org/inflight/-/inflight-1.0.6.tgz", - "integrity": "sha1-Sb1jMdfQLQwJvJEKEHW6gWW1bfk=", - "dependencies": { - "once": "^1.3.0", - "wrappy": "1" - } - }, - "node_modules/inherits": { - "version": "2.0.4", - "resolved": "https://registry.npmjs.org/inherits/-/inherits-2.0.4.tgz", - "integrity": "sha512-k/vGaX4/Yla3WzyMCvTQOXYeIHvqOKtnqBduzTHpzpQZzAskKMhZ2K+EnBiSM9zGSoIFeMpXKxa4dYeZIQqewQ==" - }, - "node_modules/internal-slot": { - "version": "1.0.3", - "resolved": "https://registry.npmjs.org/internal-slot/-/internal-slot-1.0.3.tgz", - "integrity": "sha512-O0DB1JC/sPyZl7cIo78n5dR7eUSwwpYPiXRhTzNxZVAMUuB8vlnRFyLxdrVToks6XPLVnFfbzaVd5WLjhgg+vA==", - "dev": true, - "dependencies": { - "get-intrinsic": "^1.1.0", - "has": "^1.0.3", - "side-channel": "^1.0.4" - }, - "engines": { - "node": ">= 0.4" - } - }, - "node_modules/is-bigint": { - "version": "1.0.4", - "resolved": "https://registry.npmjs.org/is-bigint/-/is-bigint-1.0.4.tgz", - "integrity": "sha512-zB9CruMamjym81i2JZ3UMn54PKGsQzsJeo6xvN3HJJ4CAsQNB6iRutp2To77OfCNuoxspsIhzaPoO1zyCEhFOg==", - "dev": true, - "dependencies": { - "has-bigints": "^1.0.1" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/is-boolean-object": { - "version": "1.1.2", - "resolved": "https://registry.npmjs.org/is-boolean-object/-/is-boolean-object-1.1.2.tgz", - "integrity": "sha512-gDYaKHJmnj4aWxyj6YHyXVpdQawtVLHU5cb+eztPGczf6cjuTdwve5ZIEfgXqH4e57An1D1AKf8CZ3kYrQRqYA==", - "dev": true, - "dependencies": { - "call-bind": "^1.0.2", - "has-tostringtag": "^1.0.0" - }, - "engines": { - "node": ">= 0.4" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/is-callable": { - "version": "1.2.4", - "resolved": "https://registry.npmjs.org/is-callable/-/is-callable-1.2.4.tgz", - "integrity": "sha512-nsuwtxZfMX67Oryl9LCQ+upnC0Z0BgpwntpS89m1H/TLF0zNfzfLMV/9Wa/6MZsj0acpEjAO0KF1xT6ZdLl95w==", - "dev": true, - "engines": { - "node": ">= 0.4" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/is-core-module": { - "version": "2.8.1", - "resolved": "https://registry.npmjs.org/is-core-module/-/is-core-module-2.8.1.tgz", - "integrity": "sha512-SdNCUs284hr40hFTFP6l0IfZ/RSrMXF3qgoRHd3/79unUTvrFO/JoXwkGm+5J/Oe3E/b5GsnG330uUNgRpu1PA==", - "dev": true, - "dependencies": { - "has": "^1.0.3" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/is-date-object": { - "version": "1.0.5", - "resolved": "https://registry.npmjs.org/is-date-object/-/is-date-object-1.0.5.tgz", - "integrity": "sha512-9YQaSxsAiSwcvS33MBk3wTCVnWK+HhF8VZR2jRxehM16QcVOdHqPn4VPHmRK4lSr38n9JriurInLcP90xsYNfQ==", - "dev": true, - "dependencies": { - "has-tostringtag": "^1.0.0" - }, - "engines": { - "node": ">= 0.4" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/is-extglob": { - "version": "2.1.1", - "resolved": "https://registry.npmjs.org/is-extglob/-/is-extglob-2.1.1.tgz", - "integrity": "sha1-qIwCU1eR8C7TfHahueqXc8gz+MI=", - "dev": true, - "engines": { - "node": ">=0.10.0" - } - }, - "node_modules/is-fullwidth-code-point": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/is-fullwidth-code-point/-/is-fullwidth-code-point-3.0.0.tgz", - "integrity": "sha512-zymm5+u+sCsSWyD9qNaejV3DFvhCKclKdizYaJUuHA83RLjb7nSuGnddCHGv0hk+KY7BMAlsWeK4Ueg6EV6XQg==", - "engines": { - "node": ">=8" - } - }, - "node_modules/is-glob": { - "version": "4.0.3", - "resolved": "https://registry.npmjs.org/is-glob/-/is-glob-4.0.3.tgz", - "integrity": "sha512-xelSayHH36ZgE7ZWhli7pW34hNbNl8Ojv5KVmkJD4hBdD3th8Tfk9vYasLM+mXWOZhFkgZfxhLSnrwRr4elSSg==", - "dev": true, - "dependencies": { - "is-extglob": "^2.1.1" - }, - "engines": { - "node": ">=0.10.0" - } - }, - "node_modules/is-negative-zero": { - "version": "2.0.2", - "resolved": "https://registry.npmjs.org/is-negative-zero/-/is-negative-zero-2.0.2.tgz", - "integrity": "sha512-dqJvarLawXsFbNDeJW7zAz8ItJ9cd28YufuuFzh0G8pNHjJMnY08Dv7sYX2uF5UpQOwieAeOExEYAWWfu7ZZUA==", - "dev": true, - "engines": { - "node": ">= 0.4" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/is-number-object": { - "version": "1.0.6", - "resolved": "https://registry.npmjs.org/is-number-object/-/is-number-object-1.0.6.tgz", - "integrity": "sha512-bEVOqiRcvo3zO1+G2lVMy+gkkEm9Yh7cDMRusKKu5ZJKPUYSJwICTKZrNKHA2EbSP0Tu0+6B/emsYNHZyn6K8g==", - "dev": true, - "dependencies": { - "has-tostringtag": "^1.0.0" - }, - "engines": { - "node": ">= 0.4" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/is-plain-object": { - "version": "5.0.0", - "resolved": "https://registry.npmjs.org/is-plain-object/-/is-plain-object-5.0.0.tgz", - "integrity": "sha512-VRSzKkbMm5jMDoKLbltAkFQ5Qr7VDiTFGXxYFXXowVj387GeGNOCsOH6Msy00SGZ3Fp84b1Naa1psqgcCIEP5Q==", - "engines": { - "node": ">=0.10.0" - } - }, - "node_modules/is-regex": { - "version": "1.1.4", - "resolved": "https://registry.npmjs.org/is-regex/-/is-regex-1.1.4.tgz", - "integrity": "sha512-kvRdxDsxZjhzUX07ZnLydzS1TU/TJlTUHHY4YLL87e37oUA49DfkLqgy+VjFocowy29cKvcSiu+kIv728jTTVg==", - "dev": true, - "dependencies": { - "call-bind": "^1.0.2", - "has-tostringtag": "^1.0.0" - }, - "engines": { - "node": ">= 0.4" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/is-shared-array-buffer": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/is-shared-array-buffer/-/is-shared-array-buffer-1.0.1.tgz", - "integrity": "sha512-IU0NmyknYZN0rChcKhRO1X8LYz5Isj/Fsqh8NJOSf+N/hCOTwy29F32Ik7a+QszE63IdvmwdTPDd6cZ5pg4cwA==", - "dev": true, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/is-string": { - "version": "1.0.7", - "resolved": "https://registry.npmjs.org/is-string/-/is-string-1.0.7.tgz", - "integrity": "sha512-tE2UXzivje6ofPW7l23cjDOMa09gb7xlAqG6jG5ej6uPV32TlWP3NKPigtaGeHNu9fohccRYvIiZMfOOnOYUtg==", - "dev": true, - "dependencies": { - "has-tostringtag": "^1.0.0" - }, - "engines": { - "node": ">= 0.4" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/is-symbol": { - "version": "1.0.4", - "resolved": "https://registry.npmjs.org/is-symbol/-/is-symbol-1.0.4.tgz", - "integrity": "sha512-C/CPBqKWnvdcxqIARxyOh4v1UUEOCHpgDa0WYgpKDFMszcrPcffg5uhwSgPCLD2WWxmq6isisz87tzT01tuGhg==", - "dev": true, - "dependencies": { - "has-symbols": "^1.0.2" - }, - "engines": { - "node": ">= 0.4" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/is-weakref": { - "version": "1.0.2", - "resolved": "https://registry.npmjs.org/is-weakref/-/is-weakref-1.0.2.tgz", - "integrity": "sha512-qctsuLZmIQ0+vSSMfoVvyFe2+GSEvnmZ2ezTup1SBse9+twCCeial6EEi3Nc2KFcf6+qz2FBPnjXsk8xhKSaPQ==", - "dev": true, - "dependencies": { - "call-bind": "^1.0.2" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/isexe": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/isexe/-/isexe-2.0.0.tgz", - "integrity": "sha1-6PvzdNxVb/iUehDcsFctYz8s+hA=", - "dev": true - }, - "node_modules/js-tokens": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/js-tokens/-/js-tokens-4.0.0.tgz", - "integrity": "sha512-RdJUflcE3cUzKiMqQgsCu06FPu9UdIJO0beYbPhHN4k6apgJtifcoCtT9bcxOpYBtpD2kCM6Sbzg4CausW/PKQ==", - "dev": true - }, - "node_modules/js-yaml": { - "version": "3.14.1", - "resolved": "https://registry.npmjs.org/js-yaml/-/js-yaml-3.14.1.tgz", - "integrity": "sha512-okMH7OXXJ7YrN9Ok3/SXrnu4iX9yOk+25nqX4imS2npuvTYDmo/QEZoqwZkYaIDk3jVvBOTOIEgEhaLOynBS9g==", - "dev": true, - "dependencies": { - "argparse": "^1.0.7", - "esprima": "^4.0.0" - }, - "bin": { - "js-yaml": "bin/js-yaml.js" - } - }, - "node_modules/json-schema-traverse": { - "version": "0.4.1", - "resolved": "https://registry.npmjs.org/json-schema-traverse/-/json-schema-traverse-0.4.1.tgz", - "integrity": "sha512-xbbCH5dCYU5T8LcEhhuh7HJ88HXuW3qsI3Y0zOZFKfZEHcpWiHU/Jxzk629Brsab/mMiHQti9wMP+845RPe3Vg==", - "dev": true - }, - "node_modules/json-stable-stringify-without-jsonify": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/json-stable-stringify-without-jsonify/-/json-stable-stringify-without-jsonify-1.0.1.tgz", - "integrity": "sha1-nbe1lJatPzz+8wp1FC0tkwrXJlE=", - "dev": true - }, - "node_modules/json5": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/json5/-/json5-1.0.1.tgz", - "integrity": "sha512-aKS4WQjPenRxiQsC93MNfjx+nbF4PAdYzmd/1JIj8HYzqfbu86beTuNgXDzPknWk0n0uARlyewZo4s++ES36Ow==", - "dev": true, - "dependencies": { - "minimist": "^1.2.0" - }, - "bin": { - "json5": "lib/cli.js" - } - }, - "node_modules/levn": { - "version": "0.4.1", - "resolved": "https://registry.npmjs.org/levn/-/levn-0.4.1.tgz", - "integrity": "sha512-+bT2uH4E5LGE7h/n3evcS/sQlJXCpIp6ym8OWJ5eV6+67Dsql/LaaT7qJBAt2rzfoa/5QBGBhxDix1dMt2kQKQ==", - "dev": true, - "dependencies": { - "prelude-ls": "^1.2.1", - "type-check": "~0.4.0" - }, - "engines": { - "node": ">= 0.8.0" - } - }, - "node_modules/locate-path": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/locate-path/-/locate-path-2.0.0.tgz", - "integrity": "sha1-K1aLJl7slExtnA3pw9u7ygNUzY4=", - "dev": true, - "dependencies": { - "p-locate": "^2.0.0", - "path-exists": "^3.0.0" - }, - "engines": { - "node": ">=4" - } - }, - "node_modules/lodash": { - "version": "4.17.21", - "resolved": "https://registry.npmjs.org/lodash/-/lodash-4.17.21.tgz", - "integrity": "sha512-v2kDEe57lecTulaDIuNTPy3Ry4gLGJ6Z1O3vE1krgXZNrsQ+LFTGHVxVjcXPs17LhbZVGedAJv8XZ1tvj5FvSg==" - }, - "node_modules/lodash.merge": { - "version": "4.6.2", - "resolved": "https://registry.npmjs.org/lodash.merge/-/lodash.merge-4.6.2.tgz", - "integrity": "sha512-0KpjqXRVvrYyCsX1swR/XTK0va6VQkQM6MNo7PqW77ByjAhoARA8EfrP1N4+KlKj8YS0ZUCtRT/YUuhyYDujIQ==", - "dev": true - }, - "node_modules/lodash.truncate": { - "version": "4.4.2", - "resolved": "https://registry.npmjs.org/lodash.truncate/-/lodash.truncate-4.4.2.tgz", - "integrity": "sha1-WjUNoLERO4N+z//VgSy+WNbq4ZM=", - "dev": true - }, - "node_modules/lru-cache": { - "version": "6.0.0", - "resolved": "https://registry.npmjs.org/lru-cache/-/lru-cache-6.0.0.tgz", - "integrity": "sha512-Jo6dJ04CmSjuznwJSS3pUeWmd/H0ffTlkXXgwZi+eq1UCmqQwCh+eLsYOYCwY991i2Fah4h1BEMCx4qThGbsiA==", - "dependencies": { - "yallist": "^4.0.0" - }, - "engines": { - "node": ">=10" - } - }, - "node_modules/luxon": { - "version": "2.3.1", - "resolved": "https://registry.npmjs.org/luxon/-/luxon-2.3.1.tgz", - "integrity": "sha512-I8vnjOmhXsMSlNMZlMkSOvgrxKJl0uOsEzdGgGNZuZPaS9KlefpE9KV95QFftlJSC+1UyCC9/I69R02cz/zcCA==", - "engines": { - "node": ">=12" - } - }, - "node_modules/make-dir": { - "version": "3.1.0", - "resolved": "https://registry.npmjs.org/make-dir/-/make-dir-3.1.0.tgz", - "integrity": "sha512-g3FeP20LNwhALb/6Cz6Dd4F2ngze0jz7tbzrD2wAV+o9FeNHe4rL+yK2md0J/fiSf1sa1ADhXqi5+oVwOM/eGw==", - "dependencies": { - "semver": "^6.0.0" - }, - "engines": { - "node": ">=8" - }, - "funding": { - "url": "https://github.com/sponsors/sindresorhus" - } - }, - "node_modules/make-dir/node_modules/semver": { - "version": "6.3.0", - "resolved": "https://registry.npmjs.org/semver/-/semver-6.3.0.tgz", - "integrity": "sha512-b39TBaTSfV6yBrapU89p5fKekE2m/NwnDocOVruQFS1/veMgdzuPcnOM34M6CwxW8jH/lxEa5rBoDeUwu5HHTw==", - "bin": { - "semver": "bin/semver.js" - } - }, - "node_modules/mimic-response": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/mimic-response/-/mimic-response-2.1.0.tgz", - "integrity": "sha512-wXqjST+SLt7R009ySCglWBCFpjUygmCIfD790/kVbiGmUgfYGuB14PiTd5DwVxSV4NcYHjzMkoj5LjQZwTQLEA==", - "engines": { - "node": ">=8" - }, - "funding": { - "url": "https://github.com/sponsors/sindresorhus" - } - }, - "node_modules/minimatch": { - "version": "3.1.2", - "resolved": "https://registry.npmjs.org/minimatch/-/minimatch-3.1.2.tgz", - "integrity": "sha512-J7p63hRiAjw1NDEww1W7i37+ByIrOWO5XQQAzZ3VOcL0PNybwpfmV/N05zFAzwQ9USyEcX6t3UO+K5aqBQOIHw==", - "dependencies": { - "brace-expansion": "^1.1.7" - }, - "engines": { - "node": "*" - } - }, - "node_modules/minimist": { - "version": "1.2.6", - "resolved": "https://registry.npmjs.org/minimist/-/minimist-1.2.6.tgz", - "integrity": "sha512-Jsjnk4bw3YJqYzbdyBiNsPWHPfO++UGG749Cxs6peCu5Xg4nrena6OVxOYxrQTqww0Jmwt+Ref8rggumkTLz9Q==", - "dev": true - }, - "node_modules/minipass": { - "version": "3.1.6", - "resolved": "https://registry.npmjs.org/minipass/-/minipass-3.1.6.tgz", - "integrity": "sha512-rty5kpw9/z8SX9dmxblFA6edItUmwJgMeYDZRrwlIVN27i8gysGbznJwUggw2V/FVqFSDdWy040ZPS811DYAqQ==", - "dependencies": { - "yallist": "^4.0.0" - }, - "engines": { - "node": ">=8" - } - }, - "node_modules/minizlib": { - "version": "2.1.2", - "resolved": "https://registry.npmjs.org/minizlib/-/minizlib-2.1.2.tgz", - "integrity": "sha512-bAxsR8BVfj60DWXHE3u30oHzfl4G7khkSuPW+qvpd7jFRHm7dLxOjUk1EHACJ/hxLY8phGJ0YhYHZo7jil7Qdg==", - "dependencies": { - "minipass": "^3.0.0", - "yallist": "^4.0.0" - }, - "engines": { - "node": ">= 8" - } - }, - "node_modules/mkdirp": { - "version": "1.0.4", - "resolved": "https://registry.npmjs.org/mkdirp/-/mkdirp-1.0.4.tgz", - "integrity": "sha512-vVqVZQyf3WLx2Shd0qJ9xuvqgAyKPLAiqITEtqW0oIUjzo3PePDd6fW9iFz30ef7Ysp/oiWqbhszeGWW2T6Gzw==", - "bin": { - "mkdirp": "bin/cmd.js" - }, - "engines": { - "node": ">=10" - } - }, - "node_modules/ms": { - "version": "2.1.2", - "resolved": "https://registry.npmjs.org/ms/-/ms-2.1.2.tgz", - "integrity": "sha512-sGkPx+VjMtmA6MX27oA4FBFELFCZZ4S4XqeGOXCv68tT+jb3vk/RyaKWP0PTKyWtmLSM0b+adUTEvbs1PEaH2w==" - }, - "node_modules/nan": { - "version": "2.15.0", - "resolved": "https://registry.npmjs.org/nan/-/nan-2.15.0.tgz", - "integrity": "sha512-8ZtvEnA2c5aYCZYd1cvgdnU6cqwixRoYg70xPLWUws5ORTa/lnw+u4amixRS/Ac5U5mQVgp9pnlSUnbNWFaWZQ==" - }, - "node_modules/natural-compare": { - "version": "1.4.0", - "resolved": "https://registry.npmjs.org/natural-compare/-/natural-compare-1.4.0.tgz", - "integrity": "sha1-Sr6/7tdUHywnrPspvbvRXI1bpPc=", - "dev": true - }, - "node_modules/node-fetch": { - "version": "2.6.7", - "resolved": "https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.7.tgz", - "integrity": "sha512-ZjMPFEfVx5j+y2yF35Kzx5sF7kDzxuDj6ziH4FFbOp87zKDZNx8yExJIb05OGF4Nlt9IHFIMBkRl41VdvcNdbQ==", - "dependencies": { - "whatwg-url": "^5.0.0" - }, - "engines": { - "node": "4.x || >=6.0.0" - }, - "peerDependencies": { - "encoding": "^0.1.0" - }, - "peerDependenciesMeta": { - "encoding": { - "optional": true - } - } - }, - "node_modules/nopt": { - "version": "5.0.0", - "resolved": "https://registry.npmjs.org/nopt/-/nopt-5.0.0.tgz", - "integrity": "sha512-Tbj67rffqceeLpcRXrT7vKAN8CwfPeIBgM7E6iBkmKLV7bEMwpGgYLGv0jACUsECaa/vuxP0IjEont6umdMgtQ==", - "dependencies": { - "abbrev": "1" - }, - "bin": { - "nopt": "bin/nopt.js" - }, - "engines": { - "node": ">=6" - } - }, - "node_modules/npmlog": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/npmlog/-/npmlog-5.0.1.tgz", - "integrity": "sha512-AqZtDUWOMKs1G/8lwylVjrdYgqA4d9nu8hc+0gzRxlDb1I10+FHBGMXs6aiQHFdCUUlqH99MUMuLfzWDNDtfxw==", - "dependencies": { - "are-we-there-yet": "^2.0.0", - "console-control-strings": "^1.1.0", - "gauge": "^3.0.0", - "set-blocking": "^2.0.0" - } - }, - "node_modules/object-assign": { - "version": "4.1.1", - "resolved": "https://registry.npmjs.org/object-assign/-/object-assign-4.1.1.tgz", - "integrity": "sha1-IQmtx5ZYh8/AXLvUQsrIv7s2CGM=", - "engines": { - "node": ">=0.10.0" - } - }, - "node_modules/object-inspect": { - "version": "1.12.0", - "resolved": "https://registry.npmjs.org/object-inspect/-/object-inspect-1.12.0.tgz", - "integrity": "sha512-Ho2z80bVIvJloH+YzRmpZVQe87+qASmBUKZDWgx9cu+KDrX2ZDH/3tMy+gXbZETVGs2M8YdxObOh7XAtim9Y0g==", - "dev": true, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/object-keys": { - "version": "1.1.1", - "resolved": "https://registry.npmjs.org/object-keys/-/object-keys-1.1.1.tgz", - "integrity": "sha512-NuAESUOUMrlIXOfHKzD6bpPu3tYt3xvjNdRIQ+FeT0lNb4K8WR70CaDxhuNguS2XG+GjkyMwOzsN5ZktImfhLA==", - "dev": true, - "engines": { - "node": ">= 0.4" - } - }, - "node_modules/object.assign": { - "version": "4.1.2", - "resolved": "https://registry.npmjs.org/object.assign/-/object.assign-4.1.2.tgz", - "integrity": "sha512-ixT2L5THXsApyiUPYKmW+2EHpXXe5Ii3M+f4e+aJFAHao5amFRW6J0OO6c/LU8Be47utCx2GL89hxGB6XSmKuQ==", - "dev": true, - "dependencies": { - "call-bind": "^1.0.0", - "define-properties": "^1.1.3", - "has-symbols": "^1.0.1", - "object-keys": "^1.1.1" - }, - "engines": { - "node": ">= 0.4" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/object.values": { - "version": "1.1.5", - "resolved": "https://registry.npmjs.org/object.values/-/object.values-1.1.5.tgz", - "integrity": "sha512-QUZRW0ilQ3PnPpbNtgdNV1PDbEqLIiSFB3l+EnGtBQ/8SUTLj1PZwtQHABZtLgwpJZTSZhuGLOGk57Drx2IvYg==", - "dev": true, - "dependencies": { - "call-bind": "^1.0.2", - "define-properties": "^1.1.3", - "es-abstract": "^1.19.1" - }, - "engines": { - "node": ">= 0.4" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/once": { - "version": "1.4.0", - "resolved": "https://registry.npmjs.org/once/-/once-1.4.0.tgz", - "integrity": "sha1-WDsap3WWHUsROsF9nFC6753Xa9E=", - "dependencies": { - "wrappy": "1" - } - }, - "node_modules/optionator": { - "version": "0.9.1", - "resolved": "https://registry.npmjs.org/optionator/-/optionator-0.9.1.tgz", - "integrity": "sha512-74RlY5FCnhq4jRxVUPKDaRwrVNXMqsGsiW6AJw4XK8hmtm10wC0ypZBLw5IIp85NZMr91+qd1RvvENwg7jjRFw==", - "dev": true, - "dependencies": { - "deep-is": "^0.1.3", - "fast-levenshtein": "^2.0.6", - "levn": "^0.4.1", - "prelude-ls": "^1.2.1", - "type-check": "^0.4.0", - "word-wrap": "^1.2.3" - }, - "engines": { - "node": ">= 0.8.0" - } - }, - "node_modules/p-limit": { - "version": "1.3.0", - "resolved": "https://registry.npmjs.org/p-limit/-/p-limit-1.3.0.tgz", - "integrity": "sha512-vvcXsLAJ9Dr5rQOPk7toZQZJApBl2K4J6dANSsEuh6QI41JYcsS/qhTGa9ErIUUgK3WNQoJYvylxvjqmiqEA9Q==", - "dev": true, - "dependencies": { - "p-try": "^1.0.0" - }, - "engines": { - "node": ">=4" - } - }, - "node_modules/p-locate": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/p-locate/-/p-locate-2.0.0.tgz", - "integrity": "sha1-IKAQOyIqcMj9OcwuWAaA893l7EM=", - "dev": true, - "dependencies": { - "p-limit": "^1.1.0" - }, - "engines": { - "node": ">=4" - } - }, - "node_modules/p-try": { - "version": "1.0.0", - "resolved": "https://registry.npmjs.org/p-try/-/p-try-1.0.0.tgz", - "integrity": "sha1-y8ec26+P1CKOE/Yh8rGiN8GyB7M=", - "dev": true, - "engines": { - "node": ">=4" - } - }, - "node_modules/parent-module": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/parent-module/-/parent-module-1.0.1.tgz", - "integrity": "sha512-GQ2EWRpQV8/o+Aw8YqtfZZPfNRWZYkbidE9k5rpl/hC3vtHHBfGm2Ifi6qWV+coDGkrUKZAxE3Lot5kcsRlh+g==", - "dev": true, - "dependencies": { - "callsites": "^3.0.0" - }, - "engines": { - "node": ">=6" - } - }, - "node_modules/path-exists": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/path-exists/-/path-exists-3.0.0.tgz", - "integrity": "sha1-zg6+ql94yxiSXqfYENe1mwEP1RU=", - "dev": true, - "engines": { - "node": ">=4" - } - }, - "node_modules/path-is-absolute": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/path-is-absolute/-/path-is-absolute-1.0.1.tgz", - "integrity": "sha1-F0uSaHNVNP+8es5r9TpanhtcX18=", - "engines": { - "node": ">=0.10.0" - } - }, - "node_modules/path-key": { - "version": "3.1.1", - "resolved": "https://registry.npmjs.org/path-key/-/path-key-3.1.1.tgz", - "integrity": "sha512-ojmeN0qd+y0jszEtoY48r0Peq5dwMEkIlCOu6Q5f41lfkswXuKtYrhgoTpLnyIcHm24Uhqx+5Tqm2InSwLhE6Q==", - "dev": true, - "engines": { - "node": ">=8" - } - }, - "node_modules/path-parse": { - "version": "1.0.7", - "resolved": "https://registry.npmjs.org/path-parse/-/path-parse-1.0.7.tgz", - "integrity": "sha512-LDJzPVEEEPR+y48z93A0Ed0yXb8pAByGWo/k5YYdYgpY2/2EsOsksJrq7lOHxryrVOn1ejG6oAp8ahvOIQD8sw==", - "dev": true - }, - "node_modules/prelude-ls": { - "version": "1.2.1", - "resolved": "https://registry.npmjs.org/prelude-ls/-/prelude-ls-1.2.1.tgz", - "integrity": "sha512-vkcDPrRZo1QZLbn5RLGPpg/WmIQ65qoWWhcGKf/b5eplkkarX0m9z8ppCat4mlOqUsWpyNuYgO3VRyrYHSzX5g==", - "dev": true, - "engines": { - "node": ">= 0.8.0" - } - }, - "node_modules/progress": { - "version": "2.0.3", - "resolved": "https://registry.npmjs.org/progress/-/progress-2.0.3.tgz", - "integrity": "sha512-7PiHtLll5LdnKIMw100I+8xJXR5gW2QwWYkT6iJva0bXitZKa/XMrSbdmg3r2Xnaidz9Qumd0VPaMrZlF9V9sA==", - "dev": true, - "engines": { - "node": ">=0.4.0" - } - }, - "node_modules/punycode": { - "version": "2.1.1", - "resolved": "https://registry.npmjs.org/punycode/-/punycode-2.1.1.tgz", - "integrity": "sha512-XRsRjdf+j5ml+y/6GKHPZbrF/8p2Yga0JPtdqTIY2Xe5ohJPD9saDJJLPvp9+NSBprVvevdXZybnj2cv8OEd0A==", - "dev": true, - "engines": { - "node": ">=6" - } - }, - "node_modules/readable-stream": { - "version": "3.6.0", - "resolved": "https://registry.npmjs.org/readable-stream/-/readable-stream-3.6.0.tgz", - "integrity": "sha512-BViHy7LKeTz4oNnkcLJ+lVSL6vpiFeX6/d3oSH8zCW7UxP2onchk+vTGB143xuFjHS3deTgkKoXXymXqymiIdA==", - "dependencies": { - "inherits": "^2.0.3", - "string_decoder": "^1.1.1", - "util-deprecate": "^1.0.1" - }, - "engines": { - "node": ">= 6" - } - }, - "node_modules/regexpp": { - "version": "3.2.0", - "resolved": "https://registry.npmjs.org/regexpp/-/regexpp-3.2.0.tgz", - "integrity": "sha512-pq2bWo9mVD43nbts2wGv17XLiNLya+GklZ8kaDLV2Z08gDCsGpnKn9BFMepvWuHCbyVvY7J5o5+BVvoQbmlJLg==", - "dev": true, - "engines": { - "node": ">=8" - }, - "funding": { - "url": "https://github.com/sponsors/mysticatea" - } - }, - "node_modules/require-from-string": { - "version": "2.0.2", - "resolved": "https://registry.npmjs.org/require-from-string/-/require-from-string-2.0.2.tgz", - "integrity": "sha512-Xf0nWe6RseziFMu+Ap9biiUbmplq6S9/p+7w7YXP/JBHhrUDDUhwa+vANyubuqfZWTveU//DYVGsDG7RKL/vEw==", - "dev": true, - "engines": { - "node": ">=0.10.0" - } - }, - "node_modules/resolve": { - "version": "1.22.0", - "resolved": "https://registry.npmjs.org/resolve/-/resolve-1.22.0.tgz", - "integrity": "sha512-Hhtrw0nLeSrFQ7phPp4OOcVjLPIeMnRlr5mcnVuMe7M/7eBn98A3hmFRLoFo3DLZkivSYwhRUJTyPyWAk56WLw==", - "dev": true, - "dependencies": { - "is-core-module": "^2.8.1", - "path-parse": "^1.0.7", - "supports-preserve-symlinks-flag": "^1.0.0" - }, - "bin": { - "resolve": "bin/resolve" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/resolve-from": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/resolve-from/-/resolve-from-4.0.0.tgz", - "integrity": "sha512-pb/MYmXstAkysRFx8piNI1tGFNQIFA3vkE3Gq4EuA1dF6gHp/+vgZqsCGJapvy8N3Q+4o7FwvquPJcnZ7RYy4g==", - "dev": true, - "engines": { - "node": ">=4" - } - }, - "node_modules/rimraf": { - "version": "3.0.2", - "resolved": "https://registry.npmjs.org/rimraf/-/rimraf-3.0.2.tgz", - "integrity": "sha512-JZkJMZkAGFFPP2YqXZXPbMlMBgsxzE8ILs4lMIX/2o0L9UBw9O/Y3o6wFw/i9YLapcUJWwqbi3kdxIPdC62TIA==", - "dependencies": { - "glob": "^7.1.3" - }, - "bin": { - "rimraf": "bin.js" - }, - "funding": { - "url": "https://github.com/sponsors/isaacs" - } - }, - "node_modules/safe-buffer": { - "version": "5.2.1", - "resolved": "https://registry.npmjs.org/safe-buffer/-/safe-buffer-5.2.1.tgz", - "integrity": "sha512-rp3So07KcdmmKbGvgaNxQSJr7bGVSVk5S9Eq1F+ppbRo70+YeaDxkw5Dd8NPN+GD6bjnYm2VuPuCXmpuYvmCXQ==", - "funding": [ - { - "type": "github", - "url": "https://github.com/sponsors/feross" - }, - { - "type": "patreon", - "url": "https://www.patreon.com/feross" - }, - { - "type": "consulting", - "url": "https://feross.org/support" - } - ] - }, - "node_modules/semver": { - "version": "7.3.5", - "resolved": "https://registry.npmjs.org/semver/-/semver-7.3.5.tgz", - "integrity": "sha512-PoeGJYh8HK4BTO/a9Tf6ZG3veo/A7ZVsYrSA6J8ny9nb3B1VrpkuN+z9OE5wfE5p6H4LchYZsegiQgbJD94ZFQ==", - "dependencies": { - "lru-cache": "^6.0.0" - }, - "bin": { - "semver": "bin/semver.js" - }, - "engines": { - "node": ">=10" - } - }, - "node_modules/set-blocking": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/set-blocking/-/set-blocking-2.0.0.tgz", - "integrity": "sha1-BF+XgtARrppoA93TgrJDkrPYkPc=" - }, - "node_modules/shebang-command": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/shebang-command/-/shebang-command-2.0.0.tgz", - "integrity": "sha512-kHxr2zZpYtdmrN1qDjrrX/Z1rR1kG8Dx+gkpK1G4eXmvXswmcE1hTWBWYUzlraYw1/yZp6YuDY77YtvbN0dmDA==", - "dev": true, - "dependencies": { - "shebang-regex": "^3.0.0" - }, - "engines": { - "node": ">=8" - } - }, - "node_modules/shebang-regex": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/shebang-regex/-/shebang-regex-3.0.0.tgz", - "integrity": "sha512-7++dFhtcx3353uBaq8DDR4NuxBetBzC7ZQOhmTQInHEd6bSrXdiEyzCvG07Z44UYdLShWUyXt5M/yhz8ekcb1A==", - "dev": true, - "engines": { - "node": ">=8" - } - }, - "node_modules/side-channel": { - "version": "1.0.4", - "resolved": "https://registry.npmjs.org/side-channel/-/side-channel-1.0.4.tgz", - "integrity": "sha512-q5XPytqFEIKHkGdiMIrY10mvLRvnQh42/+GoBlFW3b2LXLE2xxJpZFdm94we0BaoV3RwJyGqg5wS7epxTv0Zvw==", - "dev": true, - "dependencies": { - "call-bind": "^1.0.0", - "get-intrinsic": "^1.0.2", - "object-inspect": "^1.9.0" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/signal-exit": { - "version": "3.0.7", - "resolved": "https://registry.npmjs.org/signal-exit/-/signal-exit-3.0.7.tgz", - "integrity": "sha512-wnD2ZE+l+SPC/uoS0vXeE9L1+0wuaMqKlfz9AMUo38JsyLSBWSFcHR1Rri62LZc12vLr1gb3jl7iwQhgwpAbGQ==" - }, - "node_modules/simple-concat": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/simple-concat/-/simple-concat-1.0.1.tgz", - "integrity": "sha512-cSFtAPtRhljv69IK0hTVZQ+OfE9nePi/rtJmw5UjHeVyVroEqJXP1sFztKUy1qU+xvz3u/sfYJLa947b7nAN2Q==", - "funding": [ - { - "type": "github", - "url": "https://github.com/sponsors/feross" - }, - { - "type": "patreon", - "url": "https://www.patreon.com/feross" - }, - { - "type": "consulting", - "url": "https://feross.org/support" - } - ] - }, - "node_modules/simple-get": { - "version": "3.1.1", - "resolved": "https://registry.npmjs.org/simple-get/-/simple-get-3.1.1.tgz", - "integrity": "sha512-CQ5LTKGfCpvE1K0n2us+kuMPbk/q0EKl82s4aheV9oXjFEz6W/Y7oQFVJuU6QG77hRT4Ghb5RURteF5vnWjupA==", - "dependencies": { - "decompress-response": "^4.2.0", - "once": "^1.3.1", - "simple-concat": "^1.0.0" - } - }, - "node_modules/slice-ansi": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/slice-ansi/-/slice-ansi-4.0.0.tgz", - "integrity": "sha512-qMCMfhY040cVHT43K9BFygqYbUPFZKHOg7K73mtTWJRb8pyP3fzf4Ixd5SzdEJQ6MRUg/WBnOLxghZtKKurENQ==", - "dev": true, - "dependencies": { - "ansi-styles": "^4.0.0", - "astral-regex": "^2.0.0", - "is-fullwidth-code-point": "^3.0.0" - }, - "engines": { - "node": ">=10" - }, - "funding": { - "url": "https://github.com/chalk/slice-ansi?sponsor=1" - } - }, - "node_modules/sprintf-js": { - "version": "1.0.3", - "resolved": "https://registry.npmjs.org/sprintf-js/-/sprintf-js-1.0.3.tgz", - "integrity": "sha1-BOaSb2YolTVPPdAVIDYzuFcpfiw=", - "dev": true - }, - "node_modules/string_decoder": { - "version": "1.3.0", - "resolved": "https://registry.npmjs.org/string_decoder/-/string_decoder-1.3.0.tgz", - "integrity": "sha512-hkRX8U1WjJFd8LsDJ2yQ/wWWxaopEsABU1XfkM8A+j0+85JAGppt16cr1Whg6KIbb4okU6Mql6BOj+uup/wKeA==", - "dependencies": { - "safe-buffer": "~5.2.0" - } - }, - "node_modules/string-width": { - "version": "4.2.3", - "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", - "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", - "dependencies": { - "emoji-regex": "^8.0.0", - "is-fullwidth-code-point": "^3.0.0", - "strip-ansi": "^6.0.1" - }, - "engines": { - "node": ">=8" - } - }, - "node_modules/string.prototype.trimend": { - "version": "1.0.4", - "resolved": "https://registry.npmjs.org/string.prototype.trimend/-/string.prototype.trimend-1.0.4.tgz", - "integrity": "sha512-y9xCjw1P23Awk8EvTpcyL2NIr1j7wJ39f+k6lvRnSMz+mz9CGz9NYPelDk42kOz6+ql8xjfK8oYzy3jAP5QU5A==", - "dev": true, - "dependencies": { - "call-bind": "^1.0.2", - "define-properties": "^1.1.3" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/string.prototype.trimstart": { - "version": "1.0.4", - "resolved": "https://registry.npmjs.org/string.prototype.trimstart/-/string.prototype.trimstart-1.0.4.tgz", - "integrity": "sha512-jh6e984OBfvxS50tdY2nRZnoC5/mLFKOREQfw8t5yytkoUsJRNxvI/E39qu1sD0OtWI3OC0XgKSmcWwziwYuZw==", - "dev": true, - "dependencies": { - "call-bind": "^1.0.2", - "define-properties": "^1.1.3" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/strip-ansi": { - "version": "6.0.1", - "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", - "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", - "dependencies": { - "ansi-regex": "^5.0.1" - }, - "engines": { - "node": ">=8" - } - }, - "node_modules/strip-bom": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/strip-bom/-/strip-bom-3.0.0.tgz", - "integrity": "sha1-IzTBjpx1n3vdVv3vfprj1YjmjtM=", - "dev": true, - "engines": { - "node": ">=4" - } - }, - "node_modules/strip-json-comments": { - "version": "3.1.1", - "resolved": "https://registry.npmjs.org/strip-json-comments/-/strip-json-comments-3.1.1.tgz", - "integrity": "sha512-6fPc+R4ihwqP6N/aIv2f1gMH8lOVtWQHoqC4yK6oSDVVocumAsfCqjkXnqiYMhmMwS/mEHLp7Vehlt3ql6lEig==", - "dev": true, - "engines": { - "node": ">=8" - }, - "funding": { - "url": "https://github.com/sponsors/sindresorhus" - } - }, - "node_modules/supports-color": { - "version": "7.2.0", - "resolved": "https://registry.npmjs.org/supports-color/-/supports-color-7.2.0.tgz", - "integrity": "sha512-qpCAvRl9stuOHveKsn7HncJRvv501qIacKzQlO/+Lwxc9+0q2wLyv4Dfvt80/DPn2pqOBsJdDiogXGR9+OvwRw==", - "dev": true, - "dependencies": { - "has-flag": "^4.0.0" - }, - "engines": { - "node": ">=8" - } - }, - "node_modules/supports-preserve-symlinks-flag": { - "version": "1.0.0", - "resolved": "https://registry.npmjs.org/supports-preserve-symlinks-flag/-/supports-preserve-symlinks-flag-1.0.0.tgz", - "integrity": "sha512-ot0WnXS9fgdkgIcePe6RHNk1WA8+muPa6cSjeR3V8K27q9BB1rTE3R1p7Hv0z1ZyAc8s6Vvv8DIyWf681MAt0w==", - "dev": true, - "engines": { - "node": ">= 0.4" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/table": { - "version": "6.8.0", - "resolved": "https://registry.npmjs.org/table/-/table-6.8.0.tgz", - "integrity": "sha512-s/fitrbVeEyHKFa7mFdkuQMWlH1Wgw/yEXMt5xACT4ZpzWFluehAxRtUUQKPuWhaLAWhFcVx6w3oC8VKaUfPGA==", - "dev": true, - "dependencies": { - "ajv": "^8.0.1", - "lodash.truncate": "^4.4.2", - "slice-ansi": "^4.0.0", - "string-width": "^4.2.3", - "strip-ansi": "^6.0.1" - }, - "engines": { - "node": ">=10.0.0" - } - }, - "node_modules/table/node_modules/ajv": { - "version": "8.11.0", - "resolved": "https://registry.npmjs.org/ajv/-/ajv-8.11.0.tgz", - "integrity": "sha512-wGgprdCvMalC0BztXvitD2hC04YffAvtsUn93JbGXYLAtCUO4xd17mCCZQxUOItiBwZvJScWo8NIvQMQ71rdpg==", - "dev": true, - "dependencies": { - "fast-deep-equal": "^3.1.1", - "json-schema-traverse": "^1.0.0", - "require-from-string": "^2.0.2", - "uri-js": "^4.2.2" - }, - "funding": { - "type": "github", - "url": "https://github.com/sponsors/epoberezkin" - } - }, - "node_modules/table/node_modules/json-schema-traverse": { - "version": "1.0.0", - "resolved": "https://registry.npmjs.org/json-schema-traverse/-/json-schema-traverse-1.0.0.tgz", - "integrity": "sha512-NM8/P9n3XjXhIZn1lLhkFaACTOURQXjWhV4BA/RnOv8xvgqtqpAX9IO4mRQxSx1Rlo4tqzeqb0sOlruaOy3dug==", - "dev": true - }, - "node_modules/tar": { - "version": "6.1.11", - "resolved": "https://registry.npmjs.org/tar/-/tar-6.1.11.tgz", - "integrity": "sha512-an/KZQzQUkZCkuoAA64hM92X0Urb6VpRhAFllDzz44U2mcD5scmT3zBc4VgVpkugF580+DQn8eAFSyoQt0tznA==", - "dependencies": { - "chownr": "^2.0.0", - "fs-minipass": "^2.0.0", - "minipass": "^3.0.0", - "minizlib": "^2.1.1", - "mkdirp": "^1.0.3", - "yallist": "^4.0.0" - }, - "engines": { - "node": ">= 10" - } - }, - "node_modules/text-table": { - "version": "0.2.0", - "resolved": "https://registry.npmjs.org/text-table/-/text-table-0.2.0.tgz", - "integrity": "sha1-f17oI66AUgfACvLfSoTsP8+lcLQ=", - "dev": true - }, - "node_modules/tr46": { - "version": "0.0.3", - "resolved": "https://registry.npmjs.org/tr46/-/tr46-0.0.3.tgz", - "integrity": "sha1-gYT9NH2snNwYWZLzpmIuFLnZq2o=" - }, - "node_modules/tsconfig-paths": { - "version": "3.14.1", - "resolved": "https://registry.npmjs.org/tsconfig-paths/-/tsconfig-paths-3.14.1.tgz", - "integrity": "sha512-fxDhWnFSLt3VuTwtvJt5fpwxBHg5AdKWMsgcPOOIilyjymcYVZoCQF8fvFRezCNfblEXmi+PcM1eYHeOAgXCOQ==", - "dev": true, - "dependencies": { - "@types/json5": "^0.0.29", - "json5": "^1.0.1", - "minimist": "^1.2.6", - "strip-bom": "^3.0.0" - } - }, - "node_modules/tslib": { - "version": "2.3.1", - "resolved": "https://registry.npmjs.org/tslib/-/tslib-2.3.1.tgz", - "integrity": "sha512-77EbyPPpMz+FRFRuAFlWMtmgUWGe9UOG2Z25NqCwiIjRhOf5iKGuzSe5P2w1laq+FkRy4p+PCuVkJSGkzTEKVw==" - }, - "node_modules/tunnel": { - "version": "0.0.6", - "resolved": "https://registry.npmjs.org/tunnel/-/tunnel-0.0.6.tgz", - "integrity": "sha512-1h/Lnq9yajKY2PEbBadPXj3VxsDDu844OnaAo52UVmIzIvwwtBPIuNvkjuzBlTWpfJyUbG3ez0KSBibQkj4ojg==", - "engines": { - "node": ">=0.6.11 <=0.7.0 || >=0.7.3" - } - }, - "node_modules/type-check": { - "version": "0.4.0", - "resolved": "https://registry.npmjs.org/type-check/-/type-check-0.4.0.tgz", - "integrity": "sha512-XleUoc9uwGXqjWwXaUTZAmzMcFZ5858QA2vvx1Ur5xIcixXIP+8LnFDgRplU30us6teqdlskFfu+ae4K79Ooew==", - "dev": true, - "dependencies": { - "prelude-ls": "^1.2.1" - }, - "engines": { - "node": ">= 0.8.0" - } - }, - "node_modules/type-fest": { - "version": "0.20.2", - "resolved": "https://registry.npmjs.org/type-fest/-/type-fest-0.20.2.tgz", - "integrity": "sha512-Ne+eE4r0/iWnpAxD852z3A+N0Bt5RN//NjJwRd2VFHEmrywxf5vsZlh4R6lixl6B+wz/8d+maTSAkN1FIkI3LQ==", - "dev": true, - "engines": { - "node": ">=10" - }, - "funding": { - "url": "https://github.com/sponsors/sindresorhus" - } - }, - "node_modules/unbox-primitive": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/unbox-primitive/-/unbox-primitive-1.0.1.tgz", - "integrity": "sha512-tZU/3NqK3dA5gpE1KtyiJUrEB0lxnGkMFHptJ7q6ewdZ8s12QrODwNbhIJStmJkd1QDXa1NRA8aF2A1zk/Ypyw==", - "dev": true, - "dependencies": { - "function-bind": "^1.1.1", - "has-bigints": "^1.0.1", - "has-symbols": "^1.0.2", - "which-boxed-primitive": "^1.0.2" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/universal-user-agent": { - "version": "6.0.0", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-6.0.0.tgz", - "integrity": "sha512-isyNax3wXoKaulPDZWHQqbmIx1k2tb9fb3GGDBRxCscfYV2Ch7WxPArBsFEG8s/safwXTT7H4QGhaIkTp9447w==" - }, - "node_modules/uri-js": { - "version": "4.4.1", - "resolved": "https://registry.npmjs.org/uri-js/-/uri-js-4.4.1.tgz", - "integrity": "sha512-7rKUyy33Q1yc98pQ1DAmLtwX109F7TIfWlW1Ydo8Wl1ii1SeHieeh0HHfPeL2fMXK6z0s8ecKs9frCuLJvndBg==", - "dev": true, - "dependencies": { - "punycode": "^2.1.0" - } - }, - "node_modules/util-deprecate": { - "version": "1.0.2", - "resolved": "https://registry.npmjs.org/util-deprecate/-/util-deprecate-1.0.2.tgz", - "integrity": "sha1-RQ1Nyfpw3nMnYvvS1KKJgUGaDM8=" - }, - "node_modules/uuid": { - "version": "8.3.2", - "resolved": "https://registry.npmjs.org/uuid/-/uuid-8.3.2.tgz", - "integrity": "sha512-+NYs2QeMWy+GWFOEm9xnn6HCDp0l7QBD7ml8zLUmJ+93Q5NF0NocErnwkTkXVFNiX3/fpC6afS8Dhb/gz7R7eg==", - "bin": { - "uuid": "dist/bin/uuid" - } - }, - "node_modules/v8-compile-cache": { - "version": "2.3.0", - "resolved": "https://registry.npmjs.org/v8-compile-cache/-/v8-compile-cache-2.3.0.tgz", - "integrity": "sha512-l8lCEmLcLYZh4nbunNZvQCJc5pv7+RCwa8q/LdUx8u7lsWvPDKmpodJAJNwkAhJC//dFY48KuIEmjtd4RViDrA==", - "dev": true - }, - "node_modules/webidl-conversions": { - "version": "3.0.1", - "resolved": "https://registry.npmjs.org/webidl-conversions/-/webidl-conversions-3.0.1.tgz", - "integrity": "sha1-JFNCdeKnvGvnvIZhHMFq4KVlSHE=" - }, - "node_modules/whatwg-url": { - "version": "5.0.0", - "resolved": "https://registry.npmjs.org/whatwg-url/-/whatwg-url-5.0.0.tgz", - "integrity": "sha1-lmRU6HZUYuN2RNNib2dCzotwll0=", - "dependencies": { - "tr46": "~0.0.3", - "webidl-conversions": "^3.0.0" - } - }, - "node_modules/which": { - "version": "2.0.2", - "resolved": "https://registry.npmjs.org/which/-/which-2.0.2.tgz", - "integrity": "sha512-BLI3Tl1TW3Pvl70l3yq3Y64i+awpwXqsGBYWkkqMtnbXgrMD+yj7rhW0kuEDxzJaYXGjEW5ogapKNMEKNMjibA==", - "dev": true, - "dependencies": { - "isexe": "^2.0.0" - }, - "bin": { - "node-which": "bin/node-which" - }, - "engines": { - "node": ">= 8" - } - }, - "node_modules/which-boxed-primitive": { - "version": "1.0.2", - "resolved": "https://registry.npmjs.org/which-boxed-primitive/-/which-boxed-primitive-1.0.2.tgz", - "integrity": "sha512-bwZdv0AKLpplFY2KZRX6TvyuN7ojjr7lwkg6ml0roIy9YeuSr7JS372qlNW18UQYzgYK9ziGcerWqZOmEn9VNg==", - "dev": true, - "dependencies": { - "is-bigint": "^1.0.1", - "is-boolean-object": "^1.1.0", - "is-number-object": "^1.0.4", - "is-string": "^1.0.5", - "is-symbol": "^1.0.3" - }, - "funding": { - "url": "https://github.com/sponsors/ljharb" - } - }, - "node_modules/wide-align": { - "version": "1.1.5", - "resolved": "https://registry.npmjs.org/wide-align/-/wide-align-1.1.5.tgz", - "integrity": "sha512-eDMORYaPNZ4sQIuuYPDHdQvf4gyCF9rEEV/yPxGfwPkRodwEgiMUUXTx/dex+Me0wxx53S+NgUHaP7y3MGlDmg==", - "dependencies": { - "string-width": "^1.0.2 || 2 || 3 || 4" - } - }, - "node_modules/word-wrap": { - "version": "1.2.3", - "resolved": "https://registry.npmjs.org/word-wrap/-/word-wrap-1.2.3.tgz", - "integrity": "sha512-Hz/mrNwitNRh/HUAtM/VT/5VH+ygD6DV7mYKZAtHOrbs8U7lvPS6xf7EJKMF0uW1KJCl0H701g3ZGus+muE5vQ==", - "dev": true, - "engines": { - "node": ">=0.10.0" - } - }, - "node_modules/wrappy": { - "version": "1.0.2", - "resolved": "https://registry.npmjs.org/wrappy/-/wrappy-1.0.2.tgz", - "integrity": "sha1-tSQ9jz7BqjXxNkYFvA0QNuMKtp8=" - }, - "node_modules/yallist": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/yallist/-/yallist-4.0.0.tgz", - "integrity": "sha512-3wdGidZyq5PB084XLES5TpOSRA3wjXAlIWMhum2kRcv/41Sn2emQ0dycQW4uZXLejwKvg6EsvbdlVL+FYEct7A==" - } - }, - "dependencies": { - "@actions/core": { - "version": "1.9.1", - "resolved": "https://registry.npmjs.org/@actions/core/-/core-1.9.1.tgz", - "integrity": "sha512-5ad+U2YGrmmiw6du20AQW5XuWo7UKN2052FjSV7MX+Wfjf8sCqcsZe62NfgHys4QI4/Y+vQvLKYL8jWtA1ZBTA==", - "requires": { - "@actions/http-client": "^2.0.1", - "uuid": "^8.3.2" - }, - "dependencies": { - "@actions/http-client": { - "version": "2.0.1", - "resolved": "https://registry.npmjs.org/@actions/http-client/-/http-client-2.0.1.tgz", - "integrity": "sha512-PIXiMVtz6VvyaRsGY268qvj57hXQEpsYogYOu2nrQhlf+XCGmZstmuZBbAybUl1nQGnvS1k1eEsQ69ZoD7xlSw==", - "requires": { - "tunnel": "^0.0.6" - } - } - } - }, - "@actions/github": { - "version": "5.0.0", - "resolved": "https://registry.npmjs.org/@actions/github/-/github-5.0.0.tgz", - "integrity": "sha512-QvE9eAAfEsS+yOOk0cylLBIO/d6WyWIOvsxxzdrPFaud39G6BOkUwScXZn1iBzQzHyu9SBkkLSWlohDWdsasAQ==", - "requires": { - "@actions/http-client": "^1.0.11", - "@octokit/core": "^3.4.0", - "@octokit/plugin-paginate-rest": "^2.13.3", - "@octokit/plugin-rest-endpoint-methods": "^5.1.1" - } - }, - "@actions/http-client": { - "version": "1.0.11", - "resolved": "https://registry.npmjs.org/@actions/http-client/-/http-client-1.0.11.tgz", - "integrity": "sha512-VRYHGQV1rqnROJqdMvGUbY/Kn8vriQe/F9HR2AlYHzmKuM/p3kjNuXhmdBfcVgsvRWTz5C5XW5xvndZrVBuAYg==", - "requires": { - "tunnel": "0.0.6" - } - }, - "@babel/code-frame": { - "version": "7.12.11", - "resolved": "https://registry.npmjs.org/@babel/code-frame/-/code-frame-7.12.11.tgz", - "integrity": "sha512-Zt1yodBx1UcyiePMSkWnU4hPqhwq7hGi2nFL1LeA3EUl+q2LQx16MISgJ0+z7dnmgvP9QtIleuETGOiOH1RcIw==", - "dev": true, - "requires": { - "@babel/highlight": "^7.10.4" - } - }, - "@babel/helper-validator-identifier": { - "version": "7.16.7", - "resolved": "https://registry.npmjs.org/@babel/helper-validator-identifier/-/helper-validator-identifier-7.16.7.tgz", - "integrity": "sha512-hsEnFemeiW4D08A5gUAZxLBTXpZ39P+a+DGDsHw1yxqyQ/jzFEnxf5uTEGp+3bzAbNOxU1paTgYS4ECU/IgfDw==", - "dev": true - }, - "@babel/highlight": { - "version": "7.16.10", - "resolved": "https://registry.npmjs.org/@babel/highlight/-/highlight-7.16.10.tgz", - "integrity": "sha512-5FnTQLSLswEj6IkgVw5KusNUUFY9ZGqe/TRFnP/BKYHYgfh7tc+C7mwiy95/yNP7Dh9x580Vv8r7u7ZfTBFxdw==", - "dev": true, - "requires": { - "@babel/helper-validator-identifier": "^7.16.7", - "chalk": "^2.0.0", - "js-tokens": "^4.0.0" - }, - "dependencies": { - "ansi-styles": { - "version": "3.2.1", - "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-3.2.1.tgz", - "integrity": "sha512-VT0ZI6kZRdTh8YyJw3SMbYm/u+NqfsAxEpWO0Pf9sq8/e94WxxOpPKx9FR1FlyCtOVDNOQ+8ntlqFxiRc+r5qA==", - "dev": true, - "requires": { - "color-convert": "^1.9.0" - } - }, - "chalk": { - "version": "2.4.2", - "resolved": "https://registry.npmjs.org/chalk/-/chalk-2.4.2.tgz", - "integrity": "sha512-Mti+f9lpJNcwF4tWV8/OrTTtF1gZi+f8FqlyAdouralcFWFQWF2+NgCHShjkCb+IFBLq9buZwE1xckQU4peSuQ==", - "dev": true, - "requires": { - "ansi-styles": "^3.2.1", - "escape-string-regexp": "^1.0.5", - "supports-color": "^5.3.0" - } - }, - "color-convert": { - "version": "1.9.3", - "resolved": "https://registry.npmjs.org/color-convert/-/color-convert-1.9.3.tgz", - "integrity": "sha512-QfAUtd+vFdAtFQcC8CCyYt1fYWxSqAiK2cSD6zDB8N3cpsEBAvRxp9zOGg6G/SHHJYAT88/az/IuDGALsNVbGg==", - "dev": true, - "requires": { - "color-name": "1.1.3" - } - }, - "color-name": { - "version": "1.1.3", - "resolved": "https://registry.npmjs.org/color-name/-/color-name-1.1.3.tgz", - "integrity": "sha1-p9BVi9icQveV3UIyj3QIMcpTvCU=", - "dev": true - }, - "escape-string-regexp": { - "version": "1.0.5", - "resolved": "https://registry.npmjs.org/escape-string-regexp/-/escape-string-regexp-1.0.5.tgz", - "integrity": "sha1-G2HAViGQqN/2rjuyzwIAyhMLhtQ=", - "dev": true - }, - "has-flag": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/has-flag/-/has-flag-3.0.0.tgz", - "integrity": "sha1-tdRU3CGZriJWmfNGfloH87lVuv0=", - "dev": true - }, - "supports-color": { - "version": "5.5.0", - "resolved": "https://registry.npmjs.org/supports-color/-/supports-color-5.5.0.tgz", - "integrity": "sha512-QjVjwdXIt408MIiAqCX4oUKsgU2EqAGzs2Ppkm4aQYbjm+ZEWEcW4SfFNTr4uMNZma0ey4f5lgLrkB0aX0QMow==", - "dev": true, - "requires": { - "has-flag": "^3.0.0" - } - } - } - }, - "@eslint/eslintrc": { - "version": "0.4.3", - "resolved": "https://registry.npmjs.org/@eslint/eslintrc/-/eslintrc-0.4.3.tgz", - "integrity": "sha512-J6KFFz5QCYUJq3pf0mjEcCJVERbzv71PUIDczuh9JkwGEzced6CO5ADLHB1rbf/+oPBtoPfMYNOpGDzCANlbXw==", - "dev": true, - "requires": { - "ajv": "^6.12.4", - "debug": "^4.1.1", - "espree": "^7.3.0", - "globals": "^13.9.0", - "ignore": "^4.0.6", - "import-fresh": "^3.2.1", - "js-yaml": "^3.13.1", - "minimatch": "^3.0.4", - "strip-json-comments": "^3.1.1" - } - }, - "@humanwhocodes/config-array": { - "version": "0.5.0", - "resolved": "https://registry.npmjs.org/@humanwhocodes/config-array/-/config-array-0.5.0.tgz", - "integrity": "sha512-FagtKFz74XrTl7y6HCzQpwDfXP0yhxe9lHLD1UZxjvZIcbyRz8zTFF/yYNfSfzU414eDwZ1SrO0Qvtyf+wFMQg==", - "dev": true, - "requires": { - "@humanwhocodes/object-schema": "^1.2.0", - "debug": "^4.1.1", - "minimatch": "^3.0.4" - } - }, - "@humanwhocodes/object-schema": { - "version": "1.2.1", - "resolved": "https://registry.npmjs.org/@humanwhocodes/object-schema/-/object-schema-1.2.1.tgz", - "integrity": "sha512-ZnQMnLV4e7hDlUvw8H+U8ASL02SS2Gn6+9Ac3wGGLIe7+je2AeAOxPY+izIPJDfFDb7eDjev0Us8MO1iFRN8hA==", - "dev": true - }, - "@mapbox/node-pre-gyp": { - "version": "1.0.8", - "resolved": "https://registry.npmjs.org/@mapbox/node-pre-gyp/-/node-pre-gyp-1.0.8.tgz", - "integrity": "sha512-CMGKi28CF+qlbXh26hDe6NxCd7amqeAzEqnS6IHeO6LoaKyM/n+Xw3HT1COdq8cuioOdlKdqn/hCmqPUOMOywg==", - "requires": { - "detect-libc": "^1.0.3", - "https-proxy-agent": "^5.0.0", - "make-dir": "^3.1.0", - "node-fetch": "^2.6.5", - "nopt": "^5.0.0", - "npmlog": "^5.0.1", - "rimraf": "^3.0.2", - "semver": "^7.3.5", - "tar": "^6.1.11" - } - }, - "@octokit/auth-token": { - "version": "2.5.0", - "resolved": "https://registry.npmjs.org/@octokit/auth-token/-/auth-token-2.5.0.tgz", - "integrity": "sha512-r5FVUJCOLl19AxiuZD2VRZ/ORjp/4IN98Of6YJoJOkY75CIBuYfmiNHGrDwXr+aLGG55igl9QrxX3hbiXlLb+g==", - "requires": { - "@octokit/types": "^6.0.3" - } - }, - "@octokit/core": { - "version": "3.6.0", - "resolved": "https://registry.npmjs.org/@octokit/core/-/core-3.6.0.tgz", - "integrity": "sha512-7RKRKuA4xTjMhY+eG3jthb3hlZCsOwg3rztWh75Xc+ShDWOfDDATWbeZpAHBNRpm4Tv9WgBMOy1zEJYXG6NJ7Q==", - "requires": { - "@octokit/auth-token": "^2.4.4", - "@octokit/graphql": "^4.5.8", - "@octokit/request": "^5.6.3", - "@octokit/request-error": "^2.0.5", - "@octokit/types": "^6.0.3", - "before-after-hook": "^2.2.0", - "universal-user-agent": "^6.0.0" - } - }, - "@octokit/endpoint": { - "version": "6.0.12", - "resolved": "https://registry.npmjs.org/@octokit/endpoint/-/endpoint-6.0.12.tgz", - "integrity": "sha512-lF3puPwkQWGfkMClXb4k/eUT/nZKQfxinRWJrdZaJO85Dqwo/G0yOC434Jr2ojwafWJMYqFGFa5ms4jJUgujdA==", - "requires": { - "@octokit/types": "^6.0.3", - "is-plain-object": "^5.0.0", - "universal-user-agent": "^6.0.0" - } - }, - "@octokit/graphql": { - "version": "4.8.0", - "resolved": "https://registry.npmjs.org/@octokit/graphql/-/graphql-4.8.0.tgz", - "integrity": "sha512-0gv+qLSBLKF0z8TKaSKTsS39scVKF9dbMxJpj3U0vC7wjNWFuIpL/z76Qe2fiuCbDRcJSavkXsVtMS6/dtQQsg==", - "requires": { - "@octokit/request": "^5.6.0", - "@octokit/types": "^6.0.3", - "universal-user-agent": "^6.0.0" - } - }, - "@octokit/openapi-types": { - "version": "11.2.0", - "resolved": "https://registry.npmjs.org/@octokit/openapi-types/-/openapi-types-11.2.0.tgz", - "integrity": "sha512-PBsVO+15KSlGmiI8QAzaqvsNlZlrDlyAJYcrXBCvVUxCp7VnXjkwPoFHgjEJXx3WF9BAwkA6nfCUA7i9sODzKA==" - }, - "@octokit/plugin-paginate-rest": { - "version": "2.17.0", - "resolved": "https://registry.npmjs.org/@octokit/plugin-paginate-rest/-/plugin-paginate-rest-2.17.0.tgz", - "integrity": "sha512-tzMbrbnam2Mt4AhuyCHvpRkS0oZ5MvwwcQPYGtMv4tUa5kkzG58SVB0fcsLulOZQeRnOgdkZWkRUiyBlh0Bkyw==", - "requires": { - "@octokit/types": "^6.34.0" - } - }, - "@octokit/plugin-rest-endpoint-methods": { - "version": "5.13.0", - "resolved": "https://registry.npmjs.org/@octokit/plugin-rest-endpoint-methods/-/plugin-rest-endpoint-methods-5.13.0.tgz", - "integrity": "sha512-uJjMTkN1KaOIgNtUPMtIXDOjx6dGYysdIFhgA52x4xSadQCz3b/zJexvITDVpANnfKPW/+E0xkOvLntqMYpviA==", - "requires": { - "@octokit/types": "^6.34.0", - "deprecation": "^2.3.1" - } - }, - "@octokit/request": { - "version": "5.6.3", - "resolved": "https://registry.npmjs.org/@octokit/request/-/request-5.6.3.tgz", - "integrity": "sha512-bFJl0I1KVc9jYTe9tdGGpAMPy32dLBXXo1dS/YwSCTL/2nd9XeHsY616RE3HPXDVk+a+dBuzyz5YdlXwcDTr2A==", - "requires": { - "@octokit/endpoint": "^6.0.1", - "@octokit/request-error": "^2.1.0", - "@octokit/types": "^6.16.1", - "is-plain-object": "^5.0.0", - "node-fetch": "^2.6.7", - "universal-user-agent": "^6.0.0" - } - }, - "@octokit/request-error": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/@octokit/request-error/-/request-error-2.1.0.tgz", - "integrity": "sha512-1VIvgXxs9WHSjicsRwq8PlR2LR2x6DwsJAaFgzdi0JfJoGSO8mYI/cHJQ+9FbN21aa+DrgNLnwObmyeSC8Rmpg==", - "requires": { - "@octokit/types": "^6.0.3", - "deprecation": "^2.0.0", - "once": "^1.4.0" - } - }, - "@octokit/types": { - "version": "6.34.0", - "resolved": "https://registry.npmjs.org/@octokit/types/-/types-6.34.0.tgz", - "integrity": "sha512-s1zLBjWhdEI2zwaoSgyOFoKSl109CUcVBCc7biPJ3aAf6LGLU6szDvi31JPU7bxfla2lqfhjbbg/5DdFNxOwHw==", - "requires": { - "@octokit/openapi-types": "^11.2.0" - } - }, - "@types/json5": { - "version": "0.0.29", - "resolved": "https://registry.npmjs.org/@types/json5/-/json5-0.0.29.tgz", - "integrity": "sha1-7ihweulOEdK4J7y+UnC86n8+ce4=", - "dev": true - }, - "abbrev": { - "version": "1.1.1", - "resolved": "https://registry.npmjs.org/abbrev/-/abbrev-1.1.1.tgz", - "integrity": "sha512-nne9/IiQ/hzIhY6pdDnbBtz7DjPTKrY00P/zvPSm5pOFkl6xuGrGnXn/VtTNNfNtAfZ9/1RtehkszU9qcTii0Q==" - }, - "acorn": { - "version": "7.4.1", - "resolved": "https://registry.npmjs.org/acorn/-/acorn-7.4.1.tgz", - "integrity": "sha512-nQyp0o1/mNdbTO1PO6kHkwSrmgZ0MT/jCCpNiwbUjGoRN4dlBhqJtoQuCnEOKzgTVwg0ZWiCoQy6SxMebQVh8A==", - "dev": true - }, - "acorn-jsx": { - "version": "5.3.2", - "resolved": "https://registry.npmjs.org/acorn-jsx/-/acorn-jsx-5.3.2.tgz", - "integrity": "sha512-rq9s+JNhf0IChjtDXxllJ7g41oZk5SlXtp0LHwyA5cejwn7vKmKp4pPri6YEePv2PU65sAsegbXtIinmDFDXgQ==", - "dev": true, - "requires": {} - }, - "agent-base": { - "version": "6.0.2", - "resolved": "https://registry.npmjs.org/agent-base/-/agent-base-6.0.2.tgz", - "integrity": "sha512-RZNwNclF7+MS/8bDg70amg32dyeZGZxiDuQmZxKLAlQjr3jGyLx+4Kkk58UO7D2QdgFIQCovuSuZESne6RG6XQ==", - "requires": { - "debug": "4" - } - }, - "ajv": { - "version": "6.12.6", - "resolved": "https://registry.npmjs.org/ajv/-/ajv-6.12.6.tgz", - "integrity": "sha512-j3fVLgvTo527anyYyJOGTYJbG+vnnQYvE0m5mmkc1TK+nxAppkCLMIL0aZ4dblVCNoGShhm+kzE4ZUykBoMg4g==", - "dev": true, - "requires": { - "fast-deep-equal": "^3.1.1", - "fast-json-stable-stringify": "^2.0.0", - "json-schema-traverse": "^0.4.1", - "uri-js": "^4.2.2" - } - }, - "ansi-colors": { - "version": "4.1.1", - "resolved": "https://registry.npmjs.org/ansi-colors/-/ansi-colors-4.1.1.tgz", - "integrity": "sha512-JoX0apGbHaUJBNl6yF+p6JAFYZ666/hhCGKN5t9QFjbJQKUU/g8MNbFDbvfrgKXvI1QpZplPOnwIo99lX/AAmA==", - "dev": true - }, - "ansi-regex": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", - "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==" - }, - "ansi-styles": { - "version": "4.3.0", - "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz", - "integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==", - "dev": true, - "requires": { - "color-convert": "^2.0.1" - } - }, - "aproba": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/aproba/-/aproba-2.0.0.tgz", - "integrity": "sha512-lYe4Gx7QT+MKGbDsA+Z+he/Wtef0BiwDOlK/XkBrdfsh9J/jPPXbX0tE9x9cl27Tmu5gg3QUbUrQYa/y+KOHPQ==" - }, - "are-we-there-yet": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/are-we-there-yet/-/are-we-there-yet-2.0.0.tgz", - "integrity": "sha512-Ci/qENmwHnsYo9xKIcUJN5LeDKdJ6R1Z1j9V/J5wyq8nh/mYPEpIKJbBZXtZjG04HiK7zV/p6Vs9952MrMeUIw==", - "requires": { - "delegates": "^1.0.0", - "readable-stream": "^3.6.0" - } - }, - "argparse": { - "version": "1.0.10", - "resolved": "https://registry.npmjs.org/argparse/-/argparse-1.0.10.tgz", - "integrity": "sha512-o5Roy6tNG4SL/FOkCAN6RzjiakZS25RLYFrcMttJqbdd8BWrnA+fGz57iN5Pb06pvBGvl5gQ0B48dJlslXvoTg==", - "dev": true, - "requires": { - "sprintf-js": "~1.0.2" - } - }, - "array-includes": { - "version": "3.1.4", - "resolved": "https://registry.npmjs.org/array-includes/-/array-includes-3.1.4.tgz", - "integrity": "sha512-ZTNSQkmWumEbiHO2GF4GmWxYVTiQyJy2XOTa15sdQSrvKn7l+180egQMqlrMOUMCyLMD7pmyQe4mMDUT6Behrw==", - "dev": true, - "requires": { - "call-bind": "^1.0.2", - "define-properties": "^1.1.3", - "es-abstract": "^1.19.1", - "get-intrinsic": "^1.1.1", - "is-string": "^1.0.7" - } - }, - "array.prototype.flat": { - "version": "1.2.5", - "resolved": "https://registry.npmjs.org/array.prototype.flat/-/array.prototype.flat-1.2.5.tgz", - "integrity": "sha512-KaYU+S+ndVqyUnignHftkwc58o3uVU1jzczILJ1tN2YaIZpFIKBiP/x/j97E5MVPsaCloPbqWLB/8qCTVvT2qg==", - "dev": true, - "requires": { - "call-bind": "^1.0.2", - "define-properties": "^1.1.3", - "es-abstract": "^1.19.0" - } - }, - "astral-regex": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/astral-regex/-/astral-regex-2.0.0.tgz", - "integrity": "sha512-Z7tMw1ytTXt5jqMcOP+OQteU1VuNK9Y02uuJtKQ1Sv69jXQKKg5cibLwGJow8yzZP+eAc18EmLGPal0bp36rvQ==", - "dev": true - }, - "balanced-match": { - "version": "1.0.2", - "resolved": "https://registry.npmjs.org/balanced-match/-/balanced-match-1.0.2.tgz", - "integrity": "sha512-3oSeUO0TMV67hN1AmbXsK4yaqU7tjiHlbxRDZOpH0KW9+CeX4bRAaX0Anxt0tx2MrpRpWwQaPwIlISEJhYU5Pw==" - }, - "before-after-hook": { - "version": "2.2.2", - "resolved": "https://registry.npmjs.org/before-after-hook/-/before-after-hook-2.2.2.tgz", - "integrity": "sha512-3pZEU3NT5BFUo/AD5ERPWOgQOCZITni6iavr5AUw5AUwQjMlI0kzu5btnyD39AF0gUEsDPwJT+oY1ORBJijPjQ==" - }, - "brace-expansion": { - "version": "1.1.11", - "resolved": "https://registry.npmjs.org/brace-expansion/-/brace-expansion-1.1.11.tgz", - "integrity": "sha512-iCuPHDFgrHX7H2vEI/5xpz07zSHB00TpugqhmYtVmMO6518mCuRMoOYFldEBl0g187ufozdaHgWKcYFb61qGiA==", - "requires": { - "balanced-match": "^1.0.0", - "concat-map": "0.0.1" - } - }, - "call-bind": { - "version": "1.0.2", - "resolved": "https://registry.npmjs.org/call-bind/-/call-bind-1.0.2.tgz", - "integrity": "sha512-7O+FbCihrB5WGbFYesctwmTKae6rOiIzmz1icreWJ+0aA7LJfuqhEso2T9ncpcFtzMQtzXf2QGGueWJGTYsqrA==", - "dev": true, - "requires": { - "function-bind": "^1.1.1", - "get-intrinsic": "^1.0.2" - } - }, - "callsites": { - "version": "3.1.0", - "resolved": "https://registry.npmjs.org/callsites/-/callsites-3.1.0.tgz", - "integrity": "sha512-P8BjAsXvZS+VIDUI11hHCQEv74YT67YUi5JJFNWIqL235sBmjX4+qx9Muvls5ivyNENctx46xQLQ3aTuE7ssaQ==", - "dev": true - }, - "canvas": { - "version": "2.9.1", - "resolved": "https://registry.npmjs.org/canvas/-/canvas-2.9.1.tgz", - "integrity": "sha512-vSQti1uG/2gjv3x6QLOZw7TctfufaerTWbVe+NSduHxxLGB+qf3kFgQ6n66DSnuoINtVUjrLLIK2R+lxrBG07A==", - "requires": { - "@mapbox/node-pre-gyp": "^1.0.0", - "nan": "^2.15.0", - "simple-get": "^3.0.3" - } - }, - "chalk": { - "version": "4.1.2", - "resolved": "https://registry.npmjs.org/chalk/-/chalk-4.1.2.tgz", - "integrity": "sha512-oKnbhFyRIXpUuez8iBMmyEa4nbj4IOQyuhc/wy9kY7/WVPcwIO9VA668Pu8RkO7+0G76SLROeyw9CpQ061i4mA==", - "dev": true, - "requires": { - "ansi-styles": "^4.1.0", - "supports-color": "^7.1.0" - } - }, - "chart.js": { - "version": "3.7.1", - "resolved": "https://registry.npmjs.org/chart.js/-/chart.js-3.7.1.tgz", - "integrity": "sha512-8knRegQLFnPQAheZV8MjxIXc5gQEfDFD897BJgv/klO/vtIyFFmgMXrNfgrXpbTr/XbTturxRgxIXx/Y+ASJBA==" - }, - "chartjs-node-canvas": { - "version": "4.1.6", - "resolved": "https://registry.npmjs.org/chartjs-node-canvas/-/chartjs-node-canvas-4.1.6.tgz", - "integrity": "sha512-UQJbPWrvqB/FoLclGA9BaLQmZbzSYlujF4w8NZd6Xzb+sqgACBb2owDX6m7ifCXLjUW5Nz0Qx0qqrTtQkkSoYw==", - "requires": { - "canvas": "^2.8.0", - "tslib": "^2.3.1" - } - }, - "chownr": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/chownr/-/chownr-2.0.0.tgz", - "integrity": "sha512-bIomtDF5KGpdogkLd9VspvFzk9KfpyyGlS8YFVZl7TGPBHL5snIOnxeshwVgPteQ9b4Eydl+pVbIyE1DcvCWgQ==" - }, - "color-convert": { - "version": "2.0.1", - "resolved": "https://registry.npmjs.org/color-convert/-/color-convert-2.0.1.tgz", - "integrity": "sha512-RRECPsj7iu/xb5oKYcsFHSppFNnsj/52OVTRKb4zP5onXwVF3zVmmToNcOfGC+CRDpfK/U584fMg38ZHCaElKQ==", - "dev": true, - "requires": { - "color-name": "~1.1.4" - } - }, - "color-name": { - "version": "1.1.4", - "resolved": "https://registry.npmjs.org/color-name/-/color-name-1.1.4.tgz", - "integrity": "sha512-dOy+3AuW3a2wNbZHIuMZpTcgjGuLU/uBL/ubcZF9OXbDo8ff4O8yVp5Bf0efS8uEoYo5q4Fx7dY9OgQGXgAsQA==", - "dev": true - }, - "color-support": { - "version": "1.1.3", - "resolved": "https://registry.npmjs.org/color-support/-/color-support-1.1.3.tgz", - "integrity": "sha512-qiBjkpbMLO/HL68y+lh4q0/O1MZFj2RX6X/KmMa3+gJD3z+WwI1ZzDHysvqHGS3mP6mznPckpXmw1nI9cJjyRg==" - }, - "concat-map": { - "version": "0.0.1", - "resolved": "https://registry.npmjs.org/concat-map/-/concat-map-0.0.1.tgz", - "integrity": "sha1-2Klr13/Wjfd5OnMDajug1UBdR3s=" - }, - "console-control-strings": { - "version": "1.1.0", - "resolved": "https://registry.npmjs.org/console-control-strings/-/console-control-strings-1.1.0.tgz", - "integrity": "sha1-PXz0Rk22RG6mRL9LOVB/mFEAjo4=" - }, - "cross-spawn": { - "version": "7.0.3", - "resolved": "https://registry.npmjs.org/cross-spawn/-/cross-spawn-7.0.3.tgz", - "integrity": "sha512-iRDPJKUPVEND7dHPO8rkbOnPpyDygcDFtWjpeWNCgy8WP2rXcxXL8TskReQl6OrB2G7+UJrags1q15Fudc7G6w==", - "dev": true, - "requires": { - "path-key": "^3.1.0", - "shebang-command": "^2.0.0", - "which": "^2.0.1" - } - }, - "debug": { - "version": "4.3.4", - "resolved": "https://registry.npmjs.org/debug/-/debug-4.3.4.tgz", - "integrity": "sha512-PRWFHuSU3eDtQJPvnNY7Jcket1j0t5OuOsFzPPzsekD52Zl8qUfFIPEiswXqIvHWGVHOgX+7G/vCNNhehwxfkQ==", - "requires": { - "ms": "2.1.2" - } - }, - "decompress-response": { - "version": "4.2.1", - "resolved": "https://registry.npmjs.org/decompress-response/-/decompress-response-4.2.1.tgz", - "integrity": "sha512-jOSne2qbyE+/r8G1VU+G/82LBs2Fs4LAsTiLSHOCOMZQl2OKZ6i8i4IyHemTe+/yIXOtTcRQMzPcgyhoFlqPkw==", - "requires": { - "mimic-response": "^2.0.0" - } - }, - "deep-is": { - "version": "0.1.4", - "resolved": "https://registry.npmjs.org/deep-is/-/deep-is-0.1.4.tgz", - "integrity": "sha512-oIPzksmTg4/MriiaYGO+okXDT7ztn/w3Eptv/+gSIdMdKsJo0u4CfYNFJPy+4SKMuCqGw2wxnA+URMg3t8a/bQ==", - "dev": true - }, - "define-properties": { - "version": "1.1.3", - "resolved": "https://registry.npmjs.org/define-properties/-/define-properties-1.1.3.tgz", - "integrity": "sha512-3MqfYKj2lLzdMSf8ZIZE/V+Zuy+BgD6f164e8K2w7dgnpKArBDerGYpM46IYYcjnkdPNMjPk9A6VFB8+3SKlXQ==", - "dev": true, - "requires": { - "object-keys": "^1.0.12" - } - }, - "delegates": { - "version": "1.0.0", - "resolved": "https://registry.npmjs.org/delegates/-/delegates-1.0.0.tgz", - "integrity": "sha1-hMbhWbgZBP3KWaDvRM2HDTElD5o=" - }, - "deprecation": { - "version": "2.3.1", - "resolved": "https://registry.npmjs.org/deprecation/-/deprecation-2.3.1.tgz", - "integrity": "sha512-xmHIy4F3scKVwMsQ4WnVaS8bHOx0DmVwRywosKhaILI0ywMDWPtBSku2HNxRvF7jtwDRsoEwYQSfbxj8b7RlJQ==" - }, - "detect-libc": { - "version": "1.0.3", - "resolved": "https://registry.npmjs.org/detect-libc/-/detect-libc-1.0.3.tgz", - "integrity": "sha1-+hN8S9aY7fVc1c0CrFWfkaTEups=" - }, - "doctrine": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/doctrine/-/doctrine-3.0.0.tgz", - "integrity": "sha512-yS+Q5i3hBf7GBkd4KG8a7eBNNWNGLTaEwwYWUijIYM7zrlYDM0BFXHjjPWlWZ1Rg7UaddZeIDmi9jF3HmqiQ2w==", - "dev": true, - "requires": { - "esutils": "^2.0.2" - } - }, - "emoji-regex": { - "version": "8.0.0", - "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", - "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" - }, - "enquirer": { - "version": "2.3.6", - "resolved": "https://registry.npmjs.org/enquirer/-/enquirer-2.3.6.tgz", - "integrity": "sha512-yjNnPr315/FjS4zIsUxYguYUPP2e1NK4d7E7ZOLiyYCcbFBiTMyID+2wvm2w6+pZ/odMA7cRkjhsPbltwBOrLg==", - "dev": true, - "requires": { - "ansi-colors": "^4.1.1" - } - }, - "es-abstract": { - "version": "1.19.1", - "resolved": "https://registry.npmjs.org/es-abstract/-/es-abstract-1.19.1.tgz", - "integrity": "sha512-2vJ6tjA/UfqLm2MPs7jxVybLoB8i1t1Jd9R3kISld20sIxPcTbLuggQOUxeWeAvIUkduv/CfMjuh4WmiXr2v9w==", - "dev": true, - "requires": { - "call-bind": "^1.0.2", - "es-to-primitive": "^1.2.1", - "function-bind": "^1.1.1", - "get-intrinsic": "^1.1.1", - "get-symbol-description": "^1.0.0", - "has": "^1.0.3", - "has-symbols": "^1.0.2", - "internal-slot": "^1.0.3", - "is-callable": "^1.2.4", - "is-negative-zero": "^2.0.1", - "is-regex": "^1.1.4", - "is-shared-array-buffer": "^1.0.1", - "is-string": "^1.0.7", - "is-weakref": "^1.0.1", - "object-inspect": "^1.11.0", - "object-keys": "^1.1.1", - "object.assign": "^4.1.2", - "string.prototype.trimend": "^1.0.4", - "string.prototype.trimstart": "^1.0.4", - "unbox-primitive": "^1.0.1" - } - }, - "es-to-primitive": { - "version": "1.2.1", - "resolved": "https://registry.npmjs.org/es-to-primitive/-/es-to-primitive-1.2.1.tgz", - "integrity": "sha512-QCOllgZJtaUo9miYBcLChTUaHNjJF3PYs1VidD7AwiEj1kYxKeQTctLAezAOH5ZKRH0g2IgPn6KwB4IT8iRpvA==", - "dev": true, - "requires": { - "is-callable": "^1.1.4", - "is-date-object": "^1.0.1", - "is-symbol": "^1.0.2" - } - }, - "escape-string-regexp": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/escape-string-regexp/-/escape-string-regexp-4.0.0.tgz", - "integrity": "sha512-TtpcNJ3XAzx3Gq8sWRzJaVajRs0uVxA2YAkdb1jm2YkPz4G6egUFAyA3n5vtEIZefPk5Wa4UXbKuS5fKkJWdgA==", - "dev": true - }, - "eslint": { - "version": "7.32.0", - "resolved": "https://registry.npmjs.org/eslint/-/eslint-7.32.0.tgz", - "integrity": "sha512-VHZ8gX+EDfz+97jGcgyGCyRia/dPOd6Xh9yPv8Bl1+SoaIwD+a/vlrOmGRUyOYu7MwUhc7CxqeaDZU13S4+EpA==", - "dev": true, - "requires": { - "@babel/code-frame": "7.12.11", - "@eslint/eslintrc": "^0.4.3", - "@humanwhocodes/config-array": "^0.5.0", - "ajv": "^6.10.0", - "chalk": "^4.0.0", - "cross-spawn": "^7.0.2", - "debug": "^4.0.1", - "doctrine": "^3.0.0", - "enquirer": "^2.3.5", - "escape-string-regexp": "^4.0.0", - "eslint-scope": "^5.1.1", - "eslint-utils": "^2.1.0", - "eslint-visitor-keys": "^2.0.0", - "espree": "^7.3.1", - "esquery": "^1.4.0", - "esutils": "^2.0.2", - "fast-deep-equal": "^3.1.3", - "file-entry-cache": "^6.0.1", - "functional-red-black-tree": "^1.0.1", - "glob-parent": "^5.1.2", - "globals": "^13.6.0", - "ignore": "^4.0.6", - "import-fresh": "^3.0.0", - "imurmurhash": "^0.1.4", - "is-glob": "^4.0.0", - "js-yaml": "^3.13.1", - "json-stable-stringify-without-jsonify": "^1.0.1", - "levn": "^0.4.1", - "lodash.merge": "^4.6.2", - "minimatch": "^3.0.4", - "natural-compare": "^1.4.0", - "optionator": "^0.9.1", - "progress": "^2.0.0", - "regexpp": "^3.1.0", - "semver": "^7.2.1", - "strip-ansi": "^6.0.0", - "strip-json-comments": "^3.1.0", - "table": "^6.0.9", - "text-table": "^0.2.0", - "v8-compile-cache": "^2.0.3" - } - }, - "eslint-config-standard": { - "version": "16.0.3", - "resolved": "https://registry.npmjs.org/eslint-config-standard/-/eslint-config-standard-16.0.3.tgz", - "integrity": "sha512-x4fmJL5hGqNJKGHSjnLdgA6U6h1YW/G2dW9fA+cyVur4SK6lyue8+UgNKWlZtUDTXvgKDD/Oa3GQjmB5kjtVvg==", - "dev": true, - "requires": {} - }, - "eslint-import-resolver-node": { - "version": "0.3.6", - "resolved": "https://registry.npmjs.org/eslint-import-resolver-node/-/eslint-import-resolver-node-0.3.6.tgz", - "integrity": "sha512-0En0w03NRVMn9Uiyn8YRPDKvWjxCWkslUEhGNTdGx15RvPJYQ+lbOlqrlNI2vEAs4pDYK4f/HN2TbDmk5TP0iw==", - "dev": true, - "requires": { - "debug": "^3.2.7", - "resolve": "^1.20.0" - }, - "dependencies": { - "debug": { - "version": "3.2.7", - "resolved": "https://registry.npmjs.org/debug/-/debug-3.2.7.tgz", - "integrity": "sha512-CFjzYYAi4ThfiQvizrFQevTTXHtnCqWfe7x1AhgEscTz6ZbLbfoLRLPugTQyBth6f8ZERVUSyWHFD/7Wu4t1XQ==", - "dev": true, - "requires": { - "ms": "^2.1.1" - } - } - } - }, - "eslint-module-utils": { - "version": "2.7.3", - "resolved": "https://registry.npmjs.org/eslint-module-utils/-/eslint-module-utils-2.7.3.tgz", - "integrity": "sha512-088JEC7O3lDZM9xGe0RerkOMd0EjFl+Yvd1jPWIkMT5u3H9+HC34mWWPnqPrN13gieT9pBOO+Qt07Nb/6TresQ==", - "dev": true, - "requires": { - "debug": "^3.2.7", - "find-up": "^2.1.0" - }, - "dependencies": { - "debug": { - "version": "3.2.7", - "resolved": "https://registry.npmjs.org/debug/-/debug-3.2.7.tgz", - "integrity": "sha512-CFjzYYAi4ThfiQvizrFQevTTXHtnCqWfe7x1AhgEscTz6ZbLbfoLRLPugTQyBth6f8ZERVUSyWHFD/7Wu4t1XQ==", - "dev": true, - "requires": { - "ms": "^2.1.1" - } - } - } - }, - "eslint-plugin-es": { - "version": "3.0.1", - "resolved": "https://registry.npmjs.org/eslint-plugin-es/-/eslint-plugin-es-3.0.1.tgz", - "integrity": "sha512-GUmAsJaN4Fc7Gbtl8uOBlayo2DqhwWvEzykMHSCZHU3XdJ+NSzzZcVhXh3VxX5icqQ+oQdIEawXX8xkR3mIFmQ==", - "dev": true, - "requires": { - "eslint-utils": "^2.0.0", - "regexpp": "^3.0.0" - } - }, - "eslint-plugin-import": { - "version": "2.25.4", - "resolved": "https://registry.npmjs.org/eslint-plugin-import/-/eslint-plugin-import-2.25.4.tgz", - "integrity": "sha512-/KJBASVFxpu0xg1kIBn9AUa8hQVnszpwgE7Ld0lKAlx7Ie87yzEzCgSkekt+le/YVhiaosO4Y14GDAOc41nfxA==", - "dev": true, - "requires": { - "array-includes": "^3.1.4", - "array.prototype.flat": "^1.2.5", - "debug": "^2.6.9", - "doctrine": "^2.1.0", - "eslint-import-resolver-node": "^0.3.6", - "eslint-module-utils": "^2.7.2", - "has": "^1.0.3", - "is-core-module": "^2.8.0", - "is-glob": "^4.0.3", - "minimatch": "^3.0.4", - "object.values": "^1.1.5", - "resolve": "^1.20.0", - "tsconfig-paths": "^3.12.0" - }, - "dependencies": { - "debug": { - "version": "2.6.9", - "resolved": "https://registry.npmjs.org/debug/-/debug-2.6.9.tgz", - "integrity": "sha512-bC7ElrdJaJnPbAP+1EotYvqZsb3ecl5wi6Bfi6BJTUcNowp6cvspg0jXznRTKDjm/E7AdgFBVeAPVMNcKGsHMA==", - "dev": true, - "requires": { - "ms": "2.0.0" - } - }, - "doctrine": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/doctrine/-/doctrine-2.1.0.tgz", - "integrity": "sha512-35mSku4ZXK0vfCuHEDAwt55dg2jNajHZ1odvF+8SSr82EsZY4QmXfuWso8oEd8zRhVObSN18aM0CjSdoBX7zIw==", - "dev": true, - "requires": { - "esutils": "^2.0.2" - } - }, - "ms": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/ms/-/ms-2.0.0.tgz", - "integrity": "sha1-VgiurfwAvmwpAd9fmGF4jeDVl8g=", - "dev": true - } - } - }, - "eslint-plugin-node": { - "version": "11.1.0", - "resolved": "https://registry.npmjs.org/eslint-plugin-node/-/eslint-plugin-node-11.1.0.tgz", - "integrity": "sha512-oUwtPJ1W0SKD0Tr+wqu92c5xuCeQqB3hSCHasn/ZgjFdA9iDGNkNf2Zi9ztY7X+hNuMib23LNGRm6+uN+KLE3g==", - "dev": true, - "requires": { - "eslint-plugin-es": "^3.0.0", - "eslint-utils": "^2.0.0", - "ignore": "^5.1.1", - "minimatch": "^3.0.4", - "resolve": "^1.10.1", - "semver": "^6.1.0" - }, - "dependencies": { - "ignore": { - "version": "5.2.0", - "resolved": "https://registry.npmjs.org/ignore/-/ignore-5.2.0.tgz", - "integrity": "sha512-CmxgYGiEPCLhfLnpPp1MoRmifwEIOgjcHXxOBjv7mY96c+eWScsOP9c112ZyLdWHi0FxHjI+4uVhKYp/gcdRmQ==", - "dev": true - }, - "semver": { - "version": "6.3.0", - "resolved": "https://registry.npmjs.org/semver/-/semver-6.3.0.tgz", - "integrity": "sha512-b39TBaTSfV6yBrapU89p5fKekE2m/NwnDocOVruQFS1/veMgdzuPcnOM34M6CwxW8jH/lxEa5rBoDeUwu5HHTw==", - "dev": true - } - } - }, - "eslint-plugin-promise": { - "version": "5.2.0", - "resolved": "https://registry.npmjs.org/eslint-plugin-promise/-/eslint-plugin-promise-5.2.0.tgz", - "integrity": "sha512-SftLb1pUG01QYq2A/hGAWfDRXqYD82zE7j7TopDOyNdU+7SvvoXREls/+PRTY17vUXzXnZA/zfnyKgRH6x4JJw==", - "dev": true, - "requires": {} - }, - "eslint-scope": { - "version": "5.1.1", - "resolved": "https://registry.npmjs.org/eslint-scope/-/eslint-scope-5.1.1.tgz", - "integrity": "sha512-2NxwbF/hZ0KpepYN0cNbo+FN6XoK7GaHlQhgx/hIZl6Va0bF45RQOOwhLIy8lQDbuCiadSLCBnH2CFYquit5bw==", - "dev": true, - "requires": { - "esrecurse": "^4.3.0", - "estraverse": "^4.1.1" - } - }, - "eslint-utils": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/eslint-utils/-/eslint-utils-2.1.0.tgz", - "integrity": "sha512-w94dQYoauyvlDc43XnGB8lU3Zt713vNChgt4EWwhXAP2XkBvndfxF0AgIqKOOasjPIPzj9JqgwkwbCYD0/V3Zg==", - "dev": true, - "requires": { - "eslint-visitor-keys": "^1.1.0" - }, - "dependencies": { - "eslint-visitor-keys": { - "version": "1.3.0", - "resolved": "https://registry.npmjs.org/eslint-visitor-keys/-/eslint-visitor-keys-1.3.0.tgz", - "integrity": "sha512-6J72N8UNa462wa/KFODt/PJ3IU60SDpC3QXC1Hjc1BXXpfL2C9R5+AU7jhe0F6GREqVMh4Juu+NY7xn+6dipUQ==", - "dev": true - } - } - }, - "eslint-visitor-keys": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/eslint-visitor-keys/-/eslint-visitor-keys-2.1.0.tgz", - "integrity": "sha512-0rSmRBzXgDzIsD6mGdJgevzgezI534Cer5L/vyMX0kHzT/jiB43jRhd9YUlMGYLQy2zprNmoT8qasCGtY+QaKw==", - "dev": true - }, - "espree": { - "version": "7.3.1", - "resolved": "https://registry.npmjs.org/espree/-/espree-7.3.1.tgz", - "integrity": "sha512-v3JCNCE64umkFpmkFGqzVKsOT0tN1Zr+ueqLZfpV1Ob8e+CEgPWa+OxCoGH3tnhimMKIaBm4m/vaRpJ/krRz2g==", - "dev": true, - "requires": { - "acorn": "^7.4.0", - "acorn-jsx": "^5.3.1", - "eslint-visitor-keys": "^1.3.0" - }, - "dependencies": { - "eslint-visitor-keys": { - "version": "1.3.0", - "resolved": "https://registry.npmjs.org/eslint-visitor-keys/-/eslint-visitor-keys-1.3.0.tgz", - "integrity": "sha512-6J72N8UNa462wa/KFODt/PJ3IU60SDpC3QXC1Hjc1BXXpfL2C9R5+AU7jhe0F6GREqVMh4Juu+NY7xn+6dipUQ==", - "dev": true - } - } - }, - "esprima": { - "version": "4.0.1", - "resolved": "https://registry.npmjs.org/esprima/-/esprima-4.0.1.tgz", - "integrity": "sha512-eGuFFw7Upda+g4p+QHvnW0RyTX/SVeJBDM/gCtMARO0cLuT2HcEKnTPvhjV6aGeqrCB/sbNop0Kszm0jsaWU4A==", - "dev": true - }, - "esquery": { - "version": "1.4.0", - "resolved": "https://registry.npmjs.org/esquery/-/esquery-1.4.0.tgz", - "integrity": "sha512-cCDispWt5vHHtwMY2YrAQ4ibFkAL8RbH5YGBnZBc90MolvvfkkQcJro/aZiAQUlQ3qgrYS6D6v8Gc5G5CQsc9w==", - "dev": true, - "requires": { - "estraverse": "^5.1.0" - }, - "dependencies": { - "estraverse": { - "version": "5.3.0", - "resolved": "https://registry.npmjs.org/estraverse/-/estraverse-5.3.0.tgz", - "integrity": "sha512-MMdARuVEQziNTeJD8DgMqmhwR11BRQ/cBP+pLtYdSTnf3MIO8fFeiINEbX36ZdNlfU/7A9f3gUw49B3oQsvwBA==", - "dev": true - } - } - }, - "esrecurse": { - "version": "4.3.0", - "resolved": "https://registry.npmjs.org/esrecurse/-/esrecurse-4.3.0.tgz", - "integrity": "sha512-KmfKL3b6G+RXvP8N1vr3Tq1kL/oCFgn2NYXEtqP8/L3pKapUA4G8cFVaoF3SU323CD4XypR/ffioHmkti6/Tag==", - "dev": true, - "requires": { - "estraverse": "^5.2.0" - }, - "dependencies": { - "estraverse": { - "version": "5.3.0", - "resolved": "https://registry.npmjs.org/estraverse/-/estraverse-5.3.0.tgz", - "integrity": "sha512-MMdARuVEQziNTeJD8DgMqmhwR11BRQ/cBP+pLtYdSTnf3MIO8fFeiINEbX36ZdNlfU/7A9f3gUw49B3oQsvwBA==", - "dev": true - } - } - }, - "estraverse": { - "version": "4.3.0", - "resolved": "https://registry.npmjs.org/estraverse/-/estraverse-4.3.0.tgz", - "integrity": "sha512-39nnKffWz8xN1BU/2c79n9nB9HDzo0niYUqx6xyqUnyoAnQyyWpOTdZEeiCch8BBu515t4wp9ZmgVfVhn9EBpw==", - "dev": true - }, - "esutils": { - "version": "2.0.3", - "resolved": "https://registry.npmjs.org/esutils/-/esutils-2.0.3.tgz", - "integrity": "sha512-kVscqXk4OCp68SZ0dkgEKVi6/8ij300KBWTJq32P/dYeWTSwK41WyTxalN1eRmA5Z9UU/LX9D7FWSmV9SAYx6g==", - "dev": true - }, - "fast-deep-equal": { - "version": "3.1.3", - "resolved": "https://registry.npmjs.org/fast-deep-equal/-/fast-deep-equal-3.1.3.tgz", - "integrity": "sha512-f3qQ9oQy9j2AhBe/H9VC91wLmKBCCU/gDOnKNAYG5hswO7BLKj09Hc5HYNz9cGI++xlpDCIgDaitVs03ATR84Q==", - "dev": true - }, - "fast-json-stable-stringify": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/fast-json-stable-stringify/-/fast-json-stable-stringify-2.1.0.tgz", - "integrity": "sha512-lhd/wF+Lk98HZoTCtlVraHtfh5XYijIjalXck7saUtuanSDyLMxnHhSXEDJqHxD7msR8D0uCmqlkwjCV8xvwHw==", - "dev": true - }, - "fast-levenshtein": { - "version": "2.0.6", - "resolved": "https://registry.npmjs.org/fast-levenshtein/-/fast-levenshtein-2.0.6.tgz", - "integrity": "sha1-PYpcZog6FqMMqGQ+hR8Zuqd5eRc=", - "dev": true - }, - "file-entry-cache": { - "version": "6.0.1", - "resolved": "https://registry.npmjs.org/file-entry-cache/-/file-entry-cache-6.0.1.tgz", - "integrity": "sha512-7Gps/XWymbLk2QLYK4NzpMOrYjMhdIxXuIvy2QBsLE6ljuodKvdkWs/cpyJJ3CVIVpH0Oi1Hvg1ovbMzLdFBBg==", - "dev": true, - "requires": { - "flat-cache": "^3.0.4" - } - }, - "find-up": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/find-up/-/find-up-2.1.0.tgz", - "integrity": "sha1-RdG35QbHF93UgndaK3eSCjwMV6c=", - "dev": true, - "requires": { - "locate-path": "^2.0.0" - } - }, - "flat-cache": { - "version": "3.0.4", - "resolved": "https://registry.npmjs.org/flat-cache/-/flat-cache-3.0.4.tgz", - "integrity": "sha512-dm9s5Pw7Jc0GvMYbshN6zchCA9RgQlzzEZX3vylR9IqFfS8XciblUXOKfW6SiuJ0e13eDYZoZV5wdrev7P3Nwg==", - "dev": true, - "requires": { - "flatted": "^3.1.0", - "rimraf": "^3.0.2" - } - }, - "flatted": { - "version": "3.2.5", - "resolved": "https://registry.npmjs.org/flatted/-/flatted-3.2.5.tgz", - "integrity": "sha512-WIWGi2L3DyTUvUrwRKgGi9TwxQMUEqPOPQBVi71R96jZXJdFskXEmf54BoZaS1kknGODoIGASGEzBUYdyMCBJg==", - "dev": true - }, - "fs-minipass": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/fs-minipass/-/fs-minipass-2.1.0.tgz", - "integrity": "sha512-V/JgOLFCS+R6Vcq0slCuaeWEdNC3ouDlJMNIsacH2VtALiu9mV4LPrHc5cDl8k5aw6J8jwgWWpiTo5RYhmIzvg==", - "requires": { - "minipass": "^3.0.0" - } - }, - "fs.realpath": { - "version": "1.0.0", - "resolved": "https://registry.npmjs.org/fs.realpath/-/fs.realpath-1.0.0.tgz", - "integrity": "sha1-FQStJSMVjKpA20onh8sBQRmU6k8=" - }, - "function-bind": { - "version": "1.1.1", - "resolved": "https://registry.npmjs.org/function-bind/-/function-bind-1.1.1.tgz", - "integrity": "sha512-yIovAzMX49sF8Yl58fSCWJ5svSLuaibPxXQJFLmBObTuCr0Mf1KiPopGM9NiFjiYBCbfaa2Fh6breQ6ANVTI0A==", - "dev": true - }, - "functional-red-black-tree": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/functional-red-black-tree/-/functional-red-black-tree-1.0.1.tgz", - "integrity": "sha1-GwqzvVU7Kg1jmdKcDj6gslIHgyc=", - "dev": true - }, - "gauge": { - "version": "3.0.2", - "resolved": "https://registry.npmjs.org/gauge/-/gauge-3.0.2.tgz", - "integrity": "sha512-+5J6MS/5XksCuXq++uFRsnUd7Ovu1XenbeuIuNRJxYWjgQbPuFhT14lAvsWfqfAmnwluf1OwMjz39HjfLPci0Q==", - "requires": { - "aproba": "^1.0.3 || ^2.0.0", - "color-support": "^1.1.2", - "console-control-strings": "^1.0.0", - "has-unicode": "^2.0.1", - "object-assign": "^4.1.1", - "signal-exit": "^3.0.0", - "string-width": "^4.2.3", - "strip-ansi": "^6.0.1", - "wide-align": "^1.1.2" - } - }, - "get-intrinsic": { - "version": "1.1.1", - "resolved": "https://registry.npmjs.org/get-intrinsic/-/get-intrinsic-1.1.1.tgz", - "integrity": "sha512-kWZrnVM42QCiEA2Ig1bG8zjoIMOgxWwYCEeNdwY6Tv/cOSeGpcoX4pXHfKUxNKVoArnrEr2e9srnAxxGIraS9Q==", - "dev": true, - "requires": { - "function-bind": "^1.1.1", - "has": "^1.0.3", - "has-symbols": "^1.0.1" - } - }, - "get-symbol-description": { - "version": "1.0.0", - "resolved": "https://registry.npmjs.org/get-symbol-description/-/get-symbol-description-1.0.0.tgz", - "integrity": "sha512-2EmdH1YvIQiZpltCNgkuiUnyukzxM/R6NDJX31Ke3BG1Nq5b0S2PhX59UKi9vZpPDQVdqn+1IcaAwnzTT5vCjw==", - "dev": true, - "requires": { - "call-bind": "^1.0.2", - "get-intrinsic": "^1.1.1" - } - }, - "glob": { - "version": "7.2.0", - "resolved": "https://registry.npmjs.org/glob/-/glob-7.2.0.tgz", - "integrity": "sha512-lmLf6gtyrPq8tTjSmrO94wBeQbFR3HbLHbuyD69wuyQkImp2hWqMGB47OX65FBkPffO641IP9jWa1z4ivqG26Q==", - "requires": { - "fs.realpath": "^1.0.0", - "inflight": "^1.0.4", - "inherits": "2", - "minimatch": "^3.0.4", - "once": "^1.3.0", - "path-is-absolute": "^1.0.0" - } - }, - "glob-parent": { - "version": "5.1.2", - "resolved": "https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz", - "integrity": "sha512-AOIgSQCepiJYwP3ARnGx+5VnTu2HBYdzbGP45eLw1vr3zB3vZLeyed1sC9hnbcOc9/SrMyM5RPQrkGz4aS9Zow==", - "dev": true, - "requires": { - "is-glob": "^4.0.1" - } - }, - "globals": { - "version": "13.13.0", - "resolved": "https://registry.npmjs.org/globals/-/globals-13.13.0.tgz", - "integrity": "sha512-EQ7Q18AJlPwp3vUDL4mKA0KXrXyNIQyWon6T6XQiBQF0XHvRsiCSrWmmeATpUzdJN2HhWZU6Pdl0a9zdep5p6A==", - "dev": true, - "requires": { - "type-fest": "^0.20.2" - } - }, - "has": { - "version": "1.0.3", - "resolved": "https://registry.npmjs.org/has/-/has-1.0.3.tgz", - "integrity": "sha512-f2dvO0VU6Oej7RkWJGrehjbzMAjFp5/VKPp5tTpWIV4JHHZK1/BxbFRtf/siA2SWTe09caDmVtYYzWEIbBS4zw==", - "dev": true, - "requires": { - "function-bind": "^1.1.1" - } - }, - "has-bigints": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/has-bigints/-/has-bigints-1.0.1.tgz", - "integrity": "sha512-LSBS2LjbNBTf6287JEbEzvJgftkF5qFkmCo9hDRpAzKhUOlJ+hx8dd4USs00SgsUNwc4617J9ki5YtEClM2ffA==", - "dev": true - }, - "has-flag": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/has-flag/-/has-flag-4.0.0.tgz", - "integrity": "sha512-EykJT/Q1KjTWctppgIAgfSO0tKVuZUjhgMr17kqTumMl6Afv3EISleU7qZUzoXDFTAHTDC4NOoG/ZxU3EvlMPQ==", - "dev": true - }, - "has-symbols": { - "version": "1.0.3", - "resolved": "https://registry.npmjs.org/has-symbols/-/has-symbols-1.0.3.tgz", - "integrity": "sha512-l3LCuF6MgDNwTDKkdYGEihYjt5pRPbEg46rtlmnSPlUbgmB8LOIrKJbYYFBSbnPaJexMKtiPO8hmeRjRz2Td+A==", - "dev": true - }, - "has-tostringtag": { - "version": "1.0.0", - "resolved": "https://registry.npmjs.org/has-tostringtag/-/has-tostringtag-1.0.0.tgz", - "integrity": "sha512-kFjcSNhnlGV1kyoGk7OXKSawH5JOb/LzUc5w9B02hOTO0dfFRjbHQKvg1d6cf3HbeUmtU9VbbV3qzZ2Teh97WQ==", - "dev": true, - "requires": { - "has-symbols": "^1.0.2" - } - }, - "has-unicode": { - "version": "2.0.1", - "resolved": "https://registry.npmjs.org/has-unicode/-/has-unicode-2.0.1.tgz", - "integrity": "sha1-4Ob+aijPUROIVeCG0Wkedx3iqLk=" - }, - "https-proxy-agent": { - "version": "5.0.0", - "resolved": "https://registry.npmjs.org/https-proxy-agent/-/https-proxy-agent-5.0.0.tgz", - "integrity": "sha512-EkYm5BcKUGiduxzSt3Eppko+PiNWNEpa4ySk9vTC6wDsQJW9rHSa+UhGNJoRYp7bz6Ht1eaRIa6QaJqO5rCFbA==", - "requires": { - "agent-base": "6", - "debug": "4" - } - }, - "ignore": { - "version": "4.0.6", - "resolved": "https://registry.npmjs.org/ignore/-/ignore-4.0.6.tgz", - "integrity": "sha512-cyFDKrqc/YdcWFniJhzI42+AzS+gNwmUzOSFcRCQYwySuBBBy/KjuxWLZ/FHEH6Moq1NizMOBWyTcv8O4OZIMg==", - "dev": true - }, - "import-fresh": { - "version": "3.3.0", - "resolved": "https://registry.npmjs.org/import-fresh/-/import-fresh-3.3.0.tgz", - "integrity": "sha512-veYYhQa+D1QBKznvhUHxb8faxlrwUnxseDAbAp457E0wLNio2bOSKnjYDhMj+YiAq61xrMGhQk9iXVk5FzgQMw==", - "dev": true, - "requires": { - "parent-module": "^1.0.0", - "resolve-from": "^4.0.0" - } - }, - "imurmurhash": { - "version": "0.1.4", - "resolved": "https://registry.npmjs.org/imurmurhash/-/imurmurhash-0.1.4.tgz", - "integrity": "sha1-khi5srkoojixPcT7a21XbyMUU+o=", - "dev": true - }, - "inflight": { - "version": "1.0.6", - "resolved": "https://registry.npmjs.org/inflight/-/inflight-1.0.6.tgz", - "integrity": "sha1-Sb1jMdfQLQwJvJEKEHW6gWW1bfk=", - "requires": { - "once": "^1.3.0", - "wrappy": "1" - } - }, - "inherits": { - "version": "2.0.4", - "resolved": "https://registry.npmjs.org/inherits/-/inherits-2.0.4.tgz", - "integrity": "sha512-k/vGaX4/Yla3WzyMCvTQOXYeIHvqOKtnqBduzTHpzpQZzAskKMhZ2K+EnBiSM9zGSoIFeMpXKxa4dYeZIQqewQ==" - }, - "internal-slot": { - "version": "1.0.3", - "resolved": "https://registry.npmjs.org/internal-slot/-/internal-slot-1.0.3.tgz", - "integrity": "sha512-O0DB1JC/sPyZl7cIo78n5dR7eUSwwpYPiXRhTzNxZVAMUuB8vlnRFyLxdrVToks6XPLVnFfbzaVd5WLjhgg+vA==", - "dev": true, - "requires": { - "get-intrinsic": "^1.1.0", - "has": "^1.0.3", - "side-channel": "^1.0.4" - } - }, - "is-bigint": { - "version": "1.0.4", - "resolved": "https://registry.npmjs.org/is-bigint/-/is-bigint-1.0.4.tgz", - "integrity": "sha512-zB9CruMamjym81i2JZ3UMn54PKGsQzsJeo6xvN3HJJ4CAsQNB6iRutp2To77OfCNuoxspsIhzaPoO1zyCEhFOg==", - "dev": true, - "requires": { - "has-bigints": "^1.0.1" - } - }, - "is-boolean-object": { - "version": "1.1.2", - "resolved": "https://registry.npmjs.org/is-boolean-object/-/is-boolean-object-1.1.2.tgz", - "integrity": "sha512-gDYaKHJmnj4aWxyj6YHyXVpdQawtVLHU5cb+eztPGczf6cjuTdwve5ZIEfgXqH4e57An1D1AKf8CZ3kYrQRqYA==", - "dev": true, - "requires": { - "call-bind": "^1.0.2", - "has-tostringtag": "^1.0.0" - } - }, - "is-callable": { - "version": "1.2.4", - "resolved": "https://registry.npmjs.org/is-callable/-/is-callable-1.2.4.tgz", - "integrity": "sha512-nsuwtxZfMX67Oryl9LCQ+upnC0Z0BgpwntpS89m1H/TLF0zNfzfLMV/9Wa/6MZsj0acpEjAO0KF1xT6ZdLl95w==", - "dev": true - }, - "is-core-module": { - "version": "2.8.1", - "resolved": "https://registry.npmjs.org/is-core-module/-/is-core-module-2.8.1.tgz", - "integrity": "sha512-SdNCUs284hr40hFTFP6l0IfZ/RSrMXF3qgoRHd3/79unUTvrFO/JoXwkGm+5J/Oe3E/b5GsnG330uUNgRpu1PA==", - "dev": true, - "requires": { - "has": "^1.0.3" - } - }, - "is-date-object": { - "version": "1.0.5", - "resolved": "https://registry.npmjs.org/is-date-object/-/is-date-object-1.0.5.tgz", - "integrity": "sha512-9YQaSxsAiSwcvS33MBk3wTCVnWK+HhF8VZR2jRxehM16QcVOdHqPn4VPHmRK4lSr38n9JriurInLcP90xsYNfQ==", - "dev": true, - "requires": { - "has-tostringtag": "^1.0.0" - } - }, - "is-extglob": { - "version": "2.1.1", - "resolved": "https://registry.npmjs.org/is-extglob/-/is-extglob-2.1.1.tgz", - "integrity": "sha1-qIwCU1eR8C7TfHahueqXc8gz+MI=", - "dev": true - }, - "is-fullwidth-code-point": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/is-fullwidth-code-point/-/is-fullwidth-code-point-3.0.0.tgz", - "integrity": "sha512-zymm5+u+sCsSWyD9qNaejV3DFvhCKclKdizYaJUuHA83RLjb7nSuGnddCHGv0hk+KY7BMAlsWeK4Ueg6EV6XQg==" - }, - "is-glob": { - "version": "4.0.3", - "resolved": "https://registry.npmjs.org/is-glob/-/is-glob-4.0.3.tgz", - "integrity": "sha512-xelSayHH36ZgE7ZWhli7pW34hNbNl8Ojv5KVmkJD4hBdD3th8Tfk9vYasLM+mXWOZhFkgZfxhLSnrwRr4elSSg==", - "dev": true, - "requires": { - "is-extglob": "^2.1.1" - } - }, - "is-negative-zero": { - "version": "2.0.2", - "resolved": "https://registry.npmjs.org/is-negative-zero/-/is-negative-zero-2.0.2.tgz", - "integrity": "sha512-dqJvarLawXsFbNDeJW7zAz8ItJ9cd28YufuuFzh0G8pNHjJMnY08Dv7sYX2uF5UpQOwieAeOExEYAWWfu7ZZUA==", - "dev": true - }, - "is-number-object": { - "version": "1.0.6", - "resolved": "https://registry.npmjs.org/is-number-object/-/is-number-object-1.0.6.tgz", - "integrity": "sha512-bEVOqiRcvo3zO1+G2lVMy+gkkEm9Yh7cDMRusKKu5ZJKPUYSJwICTKZrNKHA2EbSP0Tu0+6B/emsYNHZyn6K8g==", - "dev": true, - "requires": { - "has-tostringtag": "^1.0.0" - } - }, - "is-plain-object": { - "version": "5.0.0", - "resolved": "https://registry.npmjs.org/is-plain-object/-/is-plain-object-5.0.0.tgz", - "integrity": "sha512-VRSzKkbMm5jMDoKLbltAkFQ5Qr7VDiTFGXxYFXXowVj387GeGNOCsOH6Msy00SGZ3Fp84b1Naa1psqgcCIEP5Q==" - }, - "is-regex": { - "version": "1.1.4", - "resolved": "https://registry.npmjs.org/is-regex/-/is-regex-1.1.4.tgz", - "integrity": "sha512-kvRdxDsxZjhzUX07ZnLydzS1TU/TJlTUHHY4YLL87e37oUA49DfkLqgy+VjFocowy29cKvcSiu+kIv728jTTVg==", - "dev": true, - "requires": { - "call-bind": "^1.0.2", - "has-tostringtag": "^1.0.0" - } - }, - "is-shared-array-buffer": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/is-shared-array-buffer/-/is-shared-array-buffer-1.0.1.tgz", - "integrity": "sha512-IU0NmyknYZN0rChcKhRO1X8LYz5Isj/Fsqh8NJOSf+N/hCOTwy29F32Ik7a+QszE63IdvmwdTPDd6cZ5pg4cwA==", - "dev": true - }, - "is-string": { - "version": "1.0.7", - "resolved": "https://registry.npmjs.org/is-string/-/is-string-1.0.7.tgz", - "integrity": "sha512-tE2UXzivje6ofPW7l23cjDOMa09gb7xlAqG6jG5ej6uPV32TlWP3NKPigtaGeHNu9fohccRYvIiZMfOOnOYUtg==", - "dev": true, - "requires": { - "has-tostringtag": "^1.0.0" - } - }, - "is-symbol": { - "version": "1.0.4", - "resolved": "https://registry.npmjs.org/is-symbol/-/is-symbol-1.0.4.tgz", - "integrity": "sha512-C/CPBqKWnvdcxqIARxyOh4v1UUEOCHpgDa0WYgpKDFMszcrPcffg5uhwSgPCLD2WWxmq6isisz87tzT01tuGhg==", - "dev": true, - "requires": { - "has-symbols": "^1.0.2" - } - }, - "is-weakref": { - "version": "1.0.2", - "resolved": "https://registry.npmjs.org/is-weakref/-/is-weakref-1.0.2.tgz", - "integrity": "sha512-qctsuLZmIQ0+vSSMfoVvyFe2+GSEvnmZ2ezTup1SBse9+twCCeial6EEi3Nc2KFcf6+qz2FBPnjXsk8xhKSaPQ==", - "dev": true, - "requires": { - "call-bind": "^1.0.2" - } - }, - "isexe": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/isexe/-/isexe-2.0.0.tgz", - "integrity": "sha1-6PvzdNxVb/iUehDcsFctYz8s+hA=", - "dev": true - }, - "js-tokens": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/js-tokens/-/js-tokens-4.0.0.tgz", - "integrity": "sha512-RdJUflcE3cUzKiMqQgsCu06FPu9UdIJO0beYbPhHN4k6apgJtifcoCtT9bcxOpYBtpD2kCM6Sbzg4CausW/PKQ==", - "dev": true - }, - "js-yaml": { - "version": "3.14.1", - "resolved": "https://registry.npmjs.org/js-yaml/-/js-yaml-3.14.1.tgz", - "integrity": "sha512-okMH7OXXJ7YrN9Ok3/SXrnu4iX9yOk+25nqX4imS2npuvTYDmo/QEZoqwZkYaIDk3jVvBOTOIEgEhaLOynBS9g==", - "dev": true, - "requires": { - "argparse": "^1.0.7", - "esprima": "^4.0.0" - } - }, - "json-schema-traverse": { - "version": "0.4.1", - "resolved": "https://registry.npmjs.org/json-schema-traverse/-/json-schema-traverse-0.4.1.tgz", - "integrity": "sha512-xbbCH5dCYU5T8LcEhhuh7HJ88HXuW3qsI3Y0zOZFKfZEHcpWiHU/Jxzk629Brsab/mMiHQti9wMP+845RPe3Vg==", - "dev": true - }, - "json-stable-stringify-without-jsonify": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/json-stable-stringify-without-jsonify/-/json-stable-stringify-without-jsonify-1.0.1.tgz", - "integrity": "sha1-nbe1lJatPzz+8wp1FC0tkwrXJlE=", - "dev": true - }, - "json5": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/json5/-/json5-1.0.1.tgz", - "integrity": "sha512-aKS4WQjPenRxiQsC93MNfjx+nbF4PAdYzmd/1JIj8HYzqfbu86beTuNgXDzPknWk0n0uARlyewZo4s++ES36Ow==", - "dev": true, - "requires": { - "minimist": "^1.2.0" - } - }, - "levn": { - "version": "0.4.1", - "resolved": "https://registry.npmjs.org/levn/-/levn-0.4.1.tgz", - "integrity": "sha512-+bT2uH4E5LGE7h/n3evcS/sQlJXCpIp6ym8OWJ5eV6+67Dsql/LaaT7qJBAt2rzfoa/5QBGBhxDix1dMt2kQKQ==", - "dev": true, - "requires": { - "prelude-ls": "^1.2.1", - "type-check": "~0.4.0" - } - }, - "locate-path": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/locate-path/-/locate-path-2.0.0.tgz", - "integrity": "sha1-K1aLJl7slExtnA3pw9u7ygNUzY4=", - "dev": true, - "requires": { - "p-locate": "^2.0.0", - "path-exists": "^3.0.0" - } - }, - "lodash": { - "version": "4.17.21", - "resolved": "https://registry.npmjs.org/lodash/-/lodash-4.17.21.tgz", - "integrity": "sha512-v2kDEe57lecTulaDIuNTPy3Ry4gLGJ6Z1O3vE1krgXZNrsQ+LFTGHVxVjcXPs17LhbZVGedAJv8XZ1tvj5FvSg==" - }, - "lodash.merge": { - "version": "4.6.2", - "resolved": "https://registry.npmjs.org/lodash.merge/-/lodash.merge-4.6.2.tgz", - "integrity": "sha512-0KpjqXRVvrYyCsX1swR/XTK0va6VQkQM6MNo7PqW77ByjAhoARA8EfrP1N4+KlKj8YS0ZUCtRT/YUuhyYDujIQ==", - "dev": true - }, - "lodash.truncate": { - "version": "4.4.2", - "resolved": "https://registry.npmjs.org/lodash.truncate/-/lodash.truncate-4.4.2.tgz", - "integrity": "sha1-WjUNoLERO4N+z//VgSy+WNbq4ZM=", - "dev": true - }, - "lru-cache": { - "version": "6.0.0", - "resolved": "https://registry.npmjs.org/lru-cache/-/lru-cache-6.0.0.tgz", - "integrity": "sha512-Jo6dJ04CmSjuznwJSS3pUeWmd/H0ffTlkXXgwZi+eq1UCmqQwCh+eLsYOYCwY991i2Fah4h1BEMCx4qThGbsiA==", - "requires": { - "yallist": "^4.0.0" - } - }, - "luxon": { - "version": "2.3.1", - "resolved": "https://registry.npmjs.org/luxon/-/luxon-2.3.1.tgz", - "integrity": "sha512-I8vnjOmhXsMSlNMZlMkSOvgrxKJl0uOsEzdGgGNZuZPaS9KlefpE9KV95QFftlJSC+1UyCC9/I69R02cz/zcCA==" - }, - "make-dir": { - "version": "3.1.0", - "resolved": "https://registry.npmjs.org/make-dir/-/make-dir-3.1.0.tgz", - "integrity": "sha512-g3FeP20LNwhALb/6Cz6Dd4F2ngze0jz7tbzrD2wAV+o9FeNHe4rL+yK2md0J/fiSf1sa1ADhXqi5+oVwOM/eGw==", - "requires": { - "semver": "^6.0.0" - }, - "dependencies": { - "semver": { - "version": "6.3.0", - "resolved": "https://registry.npmjs.org/semver/-/semver-6.3.0.tgz", - "integrity": "sha512-b39TBaTSfV6yBrapU89p5fKekE2m/NwnDocOVruQFS1/veMgdzuPcnOM34M6CwxW8jH/lxEa5rBoDeUwu5HHTw==" - } - } - }, - "mimic-response": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/mimic-response/-/mimic-response-2.1.0.tgz", - "integrity": "sha512-wXqjST+SLt7R009ySCglWBCFpjUygmCIfD790/kVbiGmUgfYGuB14PiTd5DwVxSV4NcYHjzMkoj5LjQZwTQLEA==" - }, - "minimatch": { - "version": "3.1.2", - "resolved": "https://registry.npmjs.org/minimatch/-/minimatch-3.1.2.tgz", - "integrity": "sha512-J7p63hRiAjw1NDEww1W7i37+ByIrOWO5XQQAzZ3VOcL0PNybwpfmV/N05zFAzwQ9USyEcX6t3UO+K5aqBQOIHw==", - "requires": { - "brace-expansion": "^1.1.7" - } - }, - "minimist": { - "version": "1.2.6", - "resolved": "https://registry.npmjs.org/minimist/-/minimist-1.2.6.tgz", - "integrity": "sha512-Jsjnk4bw3YJqYzbdyBiNsPWHPfO++UGG749Cxs6peCu5Xg4nrena6OVxOYxrQTqww0Jmwt+Ref8rggumkTLz9Q==", - "dev": true - }, - "minipass": { - "version": "3.1.6", - "resolved": "https://registry.npmjs.org/minipass/-/minipass-3.1.6.tgz", - "integrity": "sha512-rty5kpw9/z8SX9dmxblFA6edItUmwJgMeYDZRrwlIVN27i8gysGbznJwUggw2V/FVqFSDdWy040ZPS811DYAqQ==", - "requires": { - "yallist": "^4.0.0" - } - }, - "minizlib": { - "version": "2.1.2", - "resolved": "https://registry.npmjs.org/minizlib/-/minizlib-2.1.2.tgz", - "integrity": "sha512-bAxsR8BVfj60DWXHE3u30oHzfl4G7khkSuPW+qvpd7jFRHm7dLxOjUk1EHACJ/hxLY8phGJ0YhYHZo7jil7Qdg==", - "requires": { - "minipass": "^3.0.0", - "yallist": "^4.0.0" - } - }, - "mkdirp": { - "version": "1.0.4", - "resolved": "https://registry.npmjs.org/mkdirp/-/mkdirp-1.0.4.tgz", - "integrity": "sha512-vVqVZQyf3WLx2Shd0qJ9xuvqgAyKPLAiqITEtqW0oIUjzo3PePDd6fW9iFz30ef7Ysp/oiWqbhszeGWW2T6Gzw==" - }, - "ms": { - "version": "2.1.2", - "resolved": "https://registry.npmjs.org/ms/-/ms-2.1.2.tgz", - "integrity": "sha512-sGkPx+VjMtmA6MX27oA4FBFELFCZZ4S4XqeGOXCv68tT+jb3vk/RyaKWP0PTKyWtmLSM0b+adUTEvbs1PEaH2w==" - }, - "nan": { - "version": "2.15.0", - "resolved": "https://registry.npmjs.org/nan/-/nan-2.15.0.tgz", - "integrity": "sha512-8ZtvEnA2c5aYCZYd1cvgdnU6cqwixRoYg70xPLWUws5ORTa/lnw+u4amixRS/Ac5U5mQVgp9pnlSUnbNWFaWZQ==" - }, - "natural-compare": { - "version": "1.4.0", - "resolved": "https://registry.npmjs.org/natural-compare/-/natural-compare-1.4.0.tgz", - "integrity": "sha1-Sr6/7tdUHywnrPspvbvRXI1bpPc=", - "dev": true - }, - "node-fetch": { - "version": "2.6.7", - "resolved": "https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.7.tgz", - "integrity": "sha512-ZjMPFEfVx5j+y2yF35Kzx5sF7kDzxuDj6ziH4FFbOp87zKDZNx8yExJIb05OGF4Nlt9IHFIMBkRl41VdvcNdbQ==", - "requires": { - "whatwg-url": "^5.0.0" - } - }, - "nopt": { - "version": "5.0.0", - "resolved": "https://registry.npmjs.org/nopt/-/nopt-5.0.0.tgz", - "integrity": "sha512-Tbj67rffqceeLpcRXrT7vKAN8CwfPeIBgM7E6iBkmKLV7bEMwpGgYLGv0jACUsECaa/vuxP0IjEont6umdMgtQ==", - "requires": { - "abbrev": "1" - } - }, - "npmlog": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/npmlog/-/npmlog-5.0.1.tgz", - "integrity": "sha512-AqZtDUWOMKs1G/8lwylVjrdYgqA4d9nu8hc+0gzRxlDb1I10+FHBGMXs6aiQHFdCUUlqH99MUMuLfzWDNDtfxw==", - "requires": { - "are-we-there-yet": "^2.0.0", - "console-control-strings": "^1.1.0", - "gauge": "^3.0.0", - "set-blocking": "^2.0.0" - } - }, - "object-assign": { - "version": "4.1.1", - "resolved": "https://registry.npmjs.org/object-assign/-/object-assign-4.1.1.tgz", - "integrity": "sha1-IQmtx5ZYh8/AXLvUQsrIv7s2CGM=" - }, - "object-inspect": { - "version": "1.12.0", - "resolved": "https://registry.npmjs.org/object-inspect/-/object-inspect-1.12.0.tgz", - "integrity": "sha512-Ho2z80bVIvJloH+YzRmpZVQe87+qASmBUKZDWgx9cu+KDrX2ZDH/3tMy+gXbZETVGs2M8YdxObOh7XAtim9Y0g==", - "dev": true - }, - "object-keys": { - "version": "1.1.1", - "resolved": "https://registry.npmjs.org/object-keys/-/object-keys-1.1.1.tgz", - "integrity": "sha512-NuAESUOUMrlIXOfHKzD6bpPu3tYt3xvjNdRIQ+FeT0lNb4K8WR70CaDxhuNguS2XG+GjkyMwOzsN5ZktImfhLA==", - "dev": true - }, - "object.assign": { - "version": "4.1.2", - "resolved": "https://registry.npmjs.org/object.assign/-/object.assign-4.1.2.tgz", - "integrity": "sha512-ixT2L5THXsApyiUPYKmW+2EHpXXe5Ii3M+f4e+aJFAHao5amFRW6J0OO6c/LU8Be47utCx2GL89hxGB6XSmKuQ==", - "dev": true, - "requires": { - "call-bind": "^1.0.0", - "define-properties": "^1.1.3", - "has-symbols": "^1.0.1", - "object-keys": "^1.1.1" - } - }, - "object.values": { - "version": "1.1.5", - "resolved": "https://registry.npmjs.org/object.values/-/object.values-1.1.5.tgz", - "integrity": "sha512-QUZRW0ilQ3PnPpbNtgdNV1PDbEqLIiSFB3l+EnGtBQ/8SUTLj1PZwtQHABZtLgwpJZTSZhuGLOGk57Drx2IvYg==", - "dev": true, - "requires": { - "call-bind": "^1.0.2", - "define-properties": "^1.1.3", - "es-abstract": "^1.19.1" - } - }, - "once": { - "version": "1.4.0", - "resolved": "https://registry.npmjs.org/once/-/once-1.4.0.tgz", - "integrity": "sha1-WDsap3WWHUsROsF9nFC6753Xa9E=", - "requires": { - "wrappy": "1" - } - }, - "optionator": { - "version": "0.9.1", - "resolved": "https://registry.npmjs.org/optionator/-/optionator-0.9.1.tgz", - "integrity": "sha512-74RlY5FCnhq4jRxVUPKDaRwrVNXMqsGsiW6AJw4XK8hmtm10wC0ypZBLw5IIp85NZMr91+qd1RvvENwg7jjRFw==", - "dev": true, - "requires": { - "deep-is": "^0.1.3", - "fast-levenshtein": "^2.0.6", - "levn": "^0.4.1", - "prelude-ls": "^1.2.1", - "type-check": "^0.4.0", - "word-wrap": "^1.2.3" - } - }, - "p-limit": { - "version": "1.3.0", - "resolved": "https://registry.npmjs.org/p-limit/-/p-limit-1.3.0.tgz", - "integrity": "sha512-vvcXsLAJ9Dr5rQOPk7toZQZJApBl2K4J6dANSsEuh6QI41JYcsS/qhTGa9ErIUUgK3WNQoJYvylxvjqmiqEA9Q==", - "dev": true, - "requires": { - "p-try": "^1.0.0" - } - }, - "p-locate": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/p-locate/-/p-locate-2.0.0.tgz", - "integrity": "sha1-IKAQOyIqcMj9OcwuWAaA893l7EM=", - "dev": true, - "requires": { - "p-limit": "^1.1.0" - } - }, - "p-try": { - "version": "1.0.0", - "resolved": "https://registry.npmjs.org/p-try/-/p-try-1.0.0.tgz", - "integrity": "sha1-y8ec26+P1CKOE/Yh8rGiN8GyB7M=", - "dev": true - }, - "parent-module": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/parent-module/-/parent-module-1.0.1.tgz", - "integrity": "sha512-GQ2EWRpQV8/o+Aw8YqtfZZPfNRWZYkbidE9k5rpl/hC3vtHHBfGm2Ifi6qWV+coDGkrUKZAxE3Lot5kcsRlh+g==", - "dev": true, - "requires": { - "callsites": "^3.0.0" - } - }, - "path-exists": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/path-exists/-/path-exists-3.0.0.tgz", - "integrity": "sha1-zg6+ql94yxiSXqfYENe1mwEP1RU=", - "dev": true - }, - "path-is-absolute": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/path-is-absolute/-/path-is-absolute-1.0.1.tgz", - "integrity": "sha1-F0uSaHNVNP+8es5r9TpanhtcX18=" - }, - "path-key": { - "version": "3.1.1", - "resolved": "https://registry.npmjs.org/path-key/-/path-key-3.1.1.tgz", - "integrity": "sha512-ojmeN0qd+y0jszEtoY48r0Peq5dwMEkIlCOu6Q5f41lfkswXuKtYrhgoTpLnyIcHm24Uhqx+5Tqm2InSwLhE6Q==", - "dev": true - }, - "path-parse": { - "version": "1.0.7", - "resolved": "https://registry.npmjs.org/path-parse/-/path-parse-1.0.7.tgz", - "integrity": "sha512-LDJzPVEEEPR+y48z93A0Ed0yXb8pAByGWo/k5YYdYgpY2/2EsOsksJrq7lOHxryrVOn1ejG6oAp8ahvOIQD8sw==", - "dev": true - }, - "prelude-ls": { - "version": "1.2.1", - "resolved": "https://registry.npmjs.org/prelude-ls/-/prelude-ls-1.2.1.tgz", - "integrity": "sha512-vkcDPrRZo1QZLbn5RLGPpg/WmIQ65qoWWhcGKf/b5eplkkarX0m9z8ppCat4mlOqUsWpyNuYgO3VRyrYHSzX5g==", - "dev": true - }, - "progress": { - "version": "2.0.3", - "resolved": "https://registry.npmjs.org/progress/-/progress-2.0.3.tgz", - "integrity": "sha512-7PiHtLll5LdnKIMw100I+8xJXR5gW2QwWYkT6iJva0bXitZKa/XMrSbdmg3r2Xnaidz9Qumd0VPaMrZlF9V9sA==", - "dev": true - }, - "punycode": { - "version": "2.1.1", - "resolved": "https://registry.npmjs.org/punycode/-/punycode-2.1.1.tgz", - "integrity": "sha512-XRsRjdf+j5ml+y/6GKHPZbrF/8p2Yga0JPtdqTIY2Xe5ohJPD9saDJJLPvp9+NSBprVvevdXZybnj2cv8OEd0A==", - "dev": true - }, - "readable-stream": { - "version": "3.6.0", - "resolved": "https://registry.npmjs.org/readable-stream/-/readable-stream-3.6.0.tgz", - "integrity": "sha512-BViHy7LKeTz4oNnkcLJ+lVSL6vpiFeX6/d3oSH8zCW7UxP2onchk+vTGB143xuFjHS3deTgkKoXXymXqymiIdA==", - "requires": { - "inherits": "^2.0.3", - "string_decoder": "^1.1.1", - "util-deprecate": "^1.0.1" - } - }, - "regexpp": { - "version": "3.2.0", - "resolved": "https://registry.npmjs.org/regexpp/-/regexpp-3.2.0.tgz", - "integrity": "sha512-pq2bWo9mVD43nbts2wGv17XLiNLya+GklZ8kaDLV2Z08gDCsGpnKn9BFMepvWuHCbyVvY7J5o5+BVvoQbmlJLg==", - "dev": true - }, - "require-from-string": { - "version": "2.0.2", - "resolved": "https://registry.npmjs.org/require-from-string/-/require-from-string-2.0.2.tgz", - "integrity": "sha512-Xf0nWe6RseziFMu+Ap9biiUbmplq6S9/p+7w7YXP/JBHhrUDDUhwa+vANyubuqfZWTveU//DYVGsDG7RKL/vEw==", - "dev": true - }, - "resolve": { - "version": "1.22.0", - "resolved": "https://registry.npmjs.org/resolve/-/resolve-1.22.0.tgz", - "integrity": "sha512-Hhtrw0nLeSrFQ7phPp4OOcVjLPIeMnRlr5mcnVuMe7M/7eBn98A3hmFRLoFo3DLZkivSYwhRUJTyPyWAk56WLw==", - "dev": true, - "requires": { - "is-core-module": "^2.8.1", - "path-parse": "^1.0.7", - "supports-preserve-symlinks-flag": "^1.0.0" - } - }, - "resolve-from": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/resolve-from/-/resolve-from-4.0.0.tgz", - "integrity": "sha512-pb/MYmXstAkysRFx8piNI1tGFNQIFA3vkE3Gq4EuA1dF6gHp/+vgZqsCGJapvy8N3Q+4o7FwvquPJcnZ7RYy4g==", - "dev": true - }, - "rimraf": { - "version": "3.0.2", - "resolved": "https://registry.npmjs.org/rimraf/-/rimraf-3.0.2.tgz", - "integrity": "sha512-JZkJMZkAGFFPP2YqXZXPbMlMBgsxzE8ILs4lMIX/2o0L9UBw9O/Y3o6wFw/i9YLapcUJWwqbi3kdxIPdC62TIA==", - "requires": { - "glob": "^7.1.3" - } - }, - "safe-buffer": { - "version": "5.2.1", - "resolved": "https://registry.npmjs.org/safe-buffer/-/safe-buffer-5.2.1.tgz", - "integrity": "sha512-rp3So07KcdmmKbGvgaNxQSJr7bGVSVk5S9Eq1F+ppbRo70+YeaDxkw5Dd8NPN+GD6bjnYm2VuPuCXmpuYvmCXQ==" - }, - "semver": { - "version": "7.3.5", - "resolved": "https://registry.npmjs.org/semver/-/semver-7.3.5.tgz", - "integrity": "sha512-PoeGJYh8HK4BTO/a9Tf6ZG3veo/A7ZVsYrSA6J8ny9nb3B1VrpkuN+z9OE5wfE5p6H4LchYZsegiQgbJD94ZFQ==", - "requires": { - "lru-cache": "^6.0.0" - } - }, - "set-blocking": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/set-blocking/-/set-blocking-2.0.0.tgz", - "integrity": "sha1-BF+XgtARrppoA93TgrJDkrPYkPc=" - }, - "shebang-command": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/shebang-command/-/shebang-command-2.0.0.tgz", - "integrity": "sha512-kHxr2zZpYtdmrN1qDjrrX/Z1rR1kG8Dx+gkpK1G4eXmvXswmcE1hTWBWYUzlraYw1/yZp6YuDY77YtvbN0dmDA==", - "dev": true, - "requires": { - "shebang-regex": "^3.0.0" - } - }, - "shebang-regex": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/shebang-regex/-/shebang-regex-3.0.0.tgz", - "integrity": "sha512-7++dFhtcx3353uBaq8DDR4NuxBetBzC7ZQOhmTQInHEd6bSrXdiEyzCvG07Z44UYdLShWUyXt5M/yhz8ekcb1A==", - "dev": true - }, - "side-channel": { - "version": "1.0.4", - "resolved": "https://registry.npmjs.org/side-channel/-/side-channel-1.0.4.tgz", - "integrity": "sha512-q5XPytqFEIKHkGdiMIrY10mvLRvnQh42/+GoBlFW3b2LXLE2xxJpZFdm94we0BaoV3RwJyGqg5wS7epxTv0Zvw==", - "dev": true, - "requires": { - "call-bind": "^1.0.0", - "get-intrinsic": "^1.0.2", - "object-inspect": "^1.9.0" - } - }, - "signal-exit": { - "version": "3.0.7", - "resolved": "https://registry.npmjs.org/signal-exit/-/signal-exit-3.0.7.tgz", - "integrity": "sha512-wnD2ZE+l+SPC/uoS0vXeE9L1+0wuaMqKlfz9AMUo38JsyLSBWSFcHR1Rri62LZc12vLr1gb3jl7iwQhgwpAbGQ==" - }, - "simple-concat": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/simple-concat/-/simple-concat-1.0.1.tgz", - "integrity": "sha512-cSFtAPtRhljv69IK0hTVZQ+OfE9nePi/rtJmw5UjHeVyVroEqJXP1sFztKUy1qU+xvz3u/sfYJLa947b7nAN2Q==" - }, - "simple-get": { - "version": "3.1.1", - "resolved": "https://registry.npmjs.org/simple-get/-/simple-get-3.1.1.tgz", - "integrity": "sha512-CQ5LTKGfCpvE1K0n2us+kuMPbk/q0EKl82s4aheV9oXjFEz6W/Y7oQFVJuU6QG77hRT4Ghb5RURteF5vnWjupA==", - "requires": { - "decompress-response": "^4.2.0", - "once": "^1.3.1", - "simple-concat": "^1.0.0" - } - }, - "slice-ansi": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/slice-ansi/-/slice-ansi-4.0.0.tgz", - "integrity": "sha512-qMCMfhY040cVHT43K9BFygqYbUPFZKHOg7K73mtTWJRb8pyP3fzf4Ixd5SzdEJQ6MRUg/WBnOLxghZtKKurENQ==", - "dev": true, - "requires": { - "ansi-styles": "^4.0.0", - "astral-regex": "^2.0.0", - "is-fullwidth-code-point": "^3.0.0" - } - }, - "sprintf-js": { - "version": "1.0.3", - "resolved": "https://registry.npmjs.org/sprintf-js/-/sprintf-js-1.0.3.tgz", - "integrity": "sha1-BOaSb2YolTVPPdAVIDYzuFcpfiw=", - "dev": true - }, - "string_decoder": { - "version": "1.3.0", - "resolved": "https://registry.npmjs.org/string_decoder/-/string_decoder-1.3.0.tgz", - "integrity": "sha512-hkRX8U1WjJFd8LsDJ2yQ/wWWxaopEsABU1XfkM8A+j0+85JAGppt16cr1Whg6KIbb4okU6Mql6BOj+uup/wKeA==", - "requires": { - "safe-buffer": "~5.2.0" - } - }, - "string-width": { - "version": "4.2.3", - "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", - "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", - "requires": { - "emoji-regex": "^8.0.0", - "is-fullwidth-code-point": "^3.0.0", - "strip-ansi": "^6.0.1" - } - }, - "string.prototype.trimend": { - "version": "1.0.4", - "resolved": "https://registry.npmjs.org/string.prototype.trimend/-/string.prototype.trimend-1.0.4.tgz", - "integrity": "sha512-y9xCjw1P23Awk8EvTpcyL2NIr1j7wJ39f+k6lvRnSMz+mz9CGz9NYPelDk42kOz6+ql8xjfK8oYzy3jAP5QU5A==", - "dev": true, - "requires": { - "call-bind": "^1.0.2", - "define-properties": "^1.1.3" - } - }, - "string.prototype.trimstart": { - "version": "1.0.4", - "resolved": "https://registry.npmjs.org/string.prototype.trimstart/-/string.prototype.trimstart-1.0.4.tgz", - "integrity": "sha512-jh6e984OBfvxS50tdY2nRZnoC5/mLFKOREQfw8t5yytkoUsJRNxvI/E39qu1sD0OtWI3OC0XgKSmcWwziwYuZw==", - "dev": true, - "requires": { - "call-bind": "^1.0.2", - "define-properties": "^1.1.3" - } - }, - "strip-ansi": { - "version": "6.0.1", - "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", - "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", - "requires": { - "ansi-regex": "^5.0.1" - } - }, - "strip-bom": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/strip-bom/-/strip-bom-3.0.0.tgz", - "integrity": "sha1-IzTBjpx1n3vdVv3vfprj1YjmjtM=", - "dev": true - }, - "strip-json-comments": { - "version": "3.1.1", - "resolved": "https://registry.npmjs.org/strip-json-comments/-/strip-json-comments-3.1.1.tgz", - "integrity": "sha512-6fPc+R4ihwqP6N/aIv2f1gMH8lOVtWQHoqC4yK6oSDVVocumAsfCqjkXnqiYMhmMwS/mEHLp7Vehlt3ql6lEig==", - "dev": true - }, - "supports-color": { - "version": "7.2.0", - "resolved": "https://registry.npmjs.org/supports-color/-/supports-color-7.2.0.tgz", - "integrity": "sha512-qpCAvRl9stuOHveKsn7HncJRvv501qIacKzQlO/+Lwxc9+0q2wLyv4Dfvt80/DPn2pqOBsJdDiogXGR9+OvwRw==", - "dev": true, - "requires": { - "has-flag": "^4.0.0" - } - }, - "supports-preserve-symlinks-flag": { - "version": "1.0.0", - "resolved": "https://registry.npmjs.org/supports-preserve-symlinks-flag/-/supports-preserve-symlinks-flag-1.0.0.tgz", - "integrity": "sha512-ot0WnXS9fgdkgIcePe6RHNk1WA8+muPa6cSjeR3V8K27q9BB1rTE3R1p7Hv0z1ZyAc8s6Vvv8DIyWf681MAt0w==", - "dev": true - }, - "table": { - "version": "6.8.0", - "resolved": "https://registry.npmjs.org/table/-/table-6.8.0.tgz", - "integrity": "sha512-s/fitrbVeEyHKFa7mFdkuQMWlH1Wgw/yEXMt5xACT4ZpzWFluehAxRtUUQKPuWhaLAWhFcVx6w3oC8VKaUfPGA==", - "dev": true, - "requires": { - "ajv": "^8.0.1", - "lodash.truncate": "^4.4.2", - "slice-ansi": "^4.0.0", - "string-width": "^4.2.3", - "strip-ansi": "^6.0.1" - }, - "dependencies": { - "ajv": { - "version": "8.11.0", - "resolved": "https://registry.npmjs.org/ajv/-/ajv-8.11.0.tgz", - "integrity": "sha512-wGgprdCvMalC0BztXvitD2hC04YffAvtsUn93JbGXYLAtCUO4xd17mCCZQxUOItiBwZvJScWo8NIvQMQ71rdpg==", - "dev": true, - "requires": { - "fast-deep-equal": "^3.1.1", - "json-schema-traverse": "^1.0.0", - "require-from-string": "^2.0.2", - "uri-js": "^4.2.2" - } - }, - "json-schema-traverse": { - "version": "1.0.0", - "resolved": "https://registry.npmjs.org/json-schema-traverse/-/json-schema-traverse-1.0.0.tgz", - "integrity": "sha512-NM8/P9n3XjXhIZn1lLhkFaACTOURQXjWhV4BA/RnOv8xvgqtqpAX9IO4mRQxSx1Rlo4tqzeqb0sOlruaOy3dug==", - "dev": true - } - } - }, - "tar": { - "version": "6.1.11", - "resolved": "https://registry.npmjs.org/tar/-/tar-6.1.11.tgz", - "integrity": "sha512-an/KZQzQUkZCkuoAA64hM92X0Urb6VpRhAFllDzz44U2mcD5scmT3zBc4VgVpkugF580+DQn8eAFSyoQt0tznA==", - "requires": { - "chownr": "^2.0.0", - "fs-minipass": "^2.0.0", - "minipass": "^3.0.0", - "minizlib": "^2.1.1", - "mkdirp": "^1.0.3", - "yallist": "^4.0.0" - } - }, - "text-table": { - "version": "0.2.0", - "resolved": "https://registry.npmjs.org/text-table/-/text-table-0.2.0.tgz", - "integrity": "sha1-f17oI66AUgfACvLfSoTsP8+lcLQ=", - "dev": true - }, - "tr46": { - "version": "0.0.3", - "resolved": "https://registry.npmjs.org/tr46/-/tr46-0.0.3.tgz", - "integrity": "sha1-gYT9NH2snNwYWZLzpmIuFLnZq2o=" - }, - "tsconfig-paths": { - "version": "3.14.1", - "resolved": "https://registry.npmjs.org/tsconfig-paths/-/tsconfig-paths-3.14.1.tgz", - "integrity": "sha512-fxDhWnFSLt3VuTwtvJt5fpwxBHg5AdKWMsgcPOOIilyjymcYVZoCQF8fvFRezCNfblEXmi+PcM1eYHeOAgXCOQ==", - "dev": true, - "requires": { - "@types/json5": "^0.0.29", - "json5": "^1.0.1", - "minimist": "^1.2.6", - "strip-bom": "^3.0.0" - } - }, - "tslib": { - "version": "2.3.1", - "resolved": "https://registry.npmjs.org/tslib/-/tslib-2.3.1.tgz", - "integrity": "sha512-77EbyPPpMz+FRFRuAFlWMtmgUWGe9UOG2Z25NqCwiIjRhOf5iKGuzSe5P2w1laq+FkRy4p+PCuVkJSGkzTEKVw==" - }, "tunnel": { "version": "0.0.6", "resolved": "https://registry.npmjs.org/tunnel/-/tunnel-0.0.6.tgz", "integrity": "sha512-1h/Lnq9yajKY2PEbBadPXj3VxsDDu844OnaAo52UVmIzIvwwtBPIuNvkjuzBlTWpfJyUbG3ez0KSBibQkj4ojg==" }, - "type-check": { - "version": "0.4.0", - "resolved": "https://registry.npmjs.org/type-check/-/type-check-0.4.0.tgz", - "integrity": "sha512-XleUoc9uwGXqjWwXaUTZAmzMcFZ5858QA2vvx1Ur5xIcixXIP+8LnFDgRplU30us6teqdlskFfu+ae4K79Ooew==", - "dev": true, - "requires": { - "prelude-ls": "^1.2.1" - } - }, - "type-fest": { - "version": "0.20.2", - "resolved": "https://registry.npmjs.org/type-fest/-/type-fest-0.20.2.tgz", - "integrity": "sha512-Ne+eE4r0/iWnpAxD852z3A+N0Bt5RN//NjJwRd2VFHEmrywxf5vsZlh4R6lixl6B+wz/8d+maTSAkN1FIkI3LQ==", - "dev": true - }, - "unbox-primitive": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/unbox-primitive/-/unbox-primitive-1.0.1.tgz", - "integrity": "sha512-tZU/3NqK3dA5gpE1KtyiJUrEB0lxnGkMFHptJ7q6ewdZ8s12QrODwNbhIJStmJkd1QDXa1NRA8aF2A1zk/Ypyw==", - "dev": true, + "undici": { + "version": "5.29.0", + "resolved": "https://registry.npmjs.org/undici/-/undici-5.29.0.tgz", + "integrity": "sha512-raqeBD6NQK4SkWhQzeYKd1KmIG6dllBOTt55Rmkt4HtI9mwdWtJljnrXjAFUBLTSN67HWrOIZ3EPF4kjUw80Bg==", "requires": { - "function-bind": "^1.1.1", - "has-bigints": "^1.0.1", - "has-symbols": "^1.0.2", - "which-boxed-primitive": "^1.0.2" + "@fastify/busboy": "^2.0.0" } }, "universal-user-agent": { - "version": "6.0.0", - "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-6.0.0.tgz", - "integrity": "sha512-isyNax3wXoKaulPDZWHQqbmIx1k2tb9fb3GGDBRxCscfYV2Ch7WxPArBsFEG8s/safwXTT7H4QGhaIkTp9447w==" - }, - "uri-js": { - "version": "4.4.1", - "resolved": "https://registry.npmjs.org/uri-js/-/uri-js-4.4.1.tgz", - "integrity": "sha512-7rKUyy33Q1yc98pQ1DAmLtwX109F7TIfWlW1Ydo8Wl1ii1SeHieeh0HHfPeL2fMXK6z0s8ecKs9frCuLJvndBg==", - "dev": true, - "requires": { - "punycode": "^2.1.0" - } - }, - "util-deprecate": { - "version": "1.0.2", - "resolved": "https://registry.npmjs.org/util-deprecate/-/util-deprecate-1.0.2.tgz", - "integrity": "sha1-RQ1Nyfpw3nMnYvvS1KKJgUGaDM8=" - }, - "uuid": { - "version": "8.3.2", - "resolved": "https://registry.npmjs.org/uuid/-/uuid-8.3.2.tgz", - "integrity": "sha512-+NYs2QeMWy+GWFOEm9xnn6HCDp0l7QBD7ml8zLUmJ+93Q5NF0NocErnwkTkXVFNiX3/fpC6afS8Dhb/gz7R7eg==" - }, - "v8-compile-cache": { - "version": "2.3.0", - "resolved": "https://registry.npmjs.org/v8-compile-cache/-/v8-compile-cache-2.3.0.tgz", - "integrity": "sha512-l8lCEmLcLYZh4nbunNZvQCJc5pv7+RCwa8q/LdUx8u7lsWvPDKmpodJAJNwkAhJC//dFY48KuIEmjtd4RViDrA==", - "dev": true - }, - "webidl-conversions": { - "version": "3.0.1", - "resolved": "https://registry.npmjs.org/webidl-conversions/-/webidl-conversions-3.0.1.tgz", - "integrity": "sha1-JFNCdeKnvGvnvIZhHMFq4KVlSHE=" - }, - "whatwg-url": { - "version": "5.0.0", - "resolved": "https://registry.npmjs.org/whatwg-url/-/whatwg-url-5.0.0.tgz", - "integrity": "sha1-lmRU6HZUYuN2RNNib2dCzotwll0=", - "requires": { - "tr46": "~0.0.3", - "webidl-conversions": "^3.0.0" - } - }, - "which": { - "version": "2.0.2", - "resolved": "https://registry.npmjs.org/which/-/which-2.0.2.tgz", - "integrity": "sha512-BLI3Tl1TW3Pvl70l3yq3Y64i+awpwXqsGBYWkkqMtnbXgrMD+yj7rhW0kuEDxzJaYXGjEW5ogapKNMEKNMjibA==", - "dev": true, - "requires": { - "isexe": "^2.0.0" - } - }, - "which-boxed-primitive": { - "version": "1.0.2", - "resolved": "https://registry.npmjs.org/which-boxed-primitive/-/which-boxed-primitive-1.0.2.tgz", - "integrity": "sha512-bwZdv0AKLpplFY2KZRX6TvyuN7ojjr7lwkg6ml0roIy9YeuSr7JS372qlNW18UQYzgYK9ziGcerWqZOmEn9VNg==", - "dev": true, - "requires": { - "is-bigint": "^1.0.1", - "is-boolean-object": "^1.1.0", - "is-number-object": "^1.0.4", - "is-string": "^1.0.5", - "is-symbol": "^1.0.3" - } - }, - "wide-align": { - "version": "1.1.5", - "resolved": "https://registry.npmjs.org/wide-align/-/wide-align-1.1.5.tgz", - "integrity": "sha512-eDMORYaPNZ4sQIuuYPDHdQvf4gyCF9rEEV/yPxGfwPkRodwEgiMUUXTx/dex+Me0wxx53S+NgUHaP7y3MGlDmg==", - "requires": { - "string-width": "^1.0.2 || 2 || 3 || 4" - } - }, - "word-wrap": { - "version": "1.2.3", - "resolved": "https://registry.npmjs.org/word-wrap/-/word-wrap-1.2.3.tgz", - "integrity": "sha512-Hz/mrNwitNRh/HUAtM/VT/5VH+ygD6DV7mYKZAtHOrbs8U7lvPS6xf7EJKMF0uW1KJCl0H701g3ZGus+muE5vQ==", - "dev": true + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-6.0.1.tgz", + "integrity": "sha512-yCzhz6FN2wU1NiiQRogkTQszlQSlpWaw8SvVegAc+bDxbzHgh1vX8uIe8OYyMH6DwH+sdTJsgMl36+mSMdRJIQ==" }, "wrappy": { "version": "1.0.2", "resolved": "https://registry.npmjs.org/wrappy/-/wrappy-1.0.2.tgz", - "integrity": "sha1-tSQ9jz7BqjXxNkYFvA0QNuMKtp8=" - }, - "yallist": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/yallist/-/yallist-4.0.0.tgz", - "integrity": "sha512-3wdGidZyq5PB084XLES5TpOSRA3wjXAlIWMhum2kRcv/41Sn2emQ0dycQW4uZXLejwKvg6EsvbdlVL+FYEct7A==" + "integrity": "sha512-l4Sp/DRseor9wL6EvV2+TuQn63dMkPjZ/sp9XkghTEbV9KlPS1xUsZ3u7/IQO4wxtcFB4bgpQPRcR3QCvezPcQ==" } } } diff --git a/dev/coverage-action/package.json b/dev/coverage-action/package.json index 9e7417fe65..3f72b78028 100644 --- a/dev/coverage-action/package.json +++ b/dev/coverage-action/package.json @@ -6,18 +6,9 @@ "author": "IETF Trust", "license": "BSD-3-Clause", "dependencies": { - "@actions/core": "1.9.1", - "@actions/github": "5.0.0", - "chart.js": "3.7.1", - "chartjs-node-canvas": "4.1.6", + "@actions/core": "1.11.1", + "@actions/github": "6.0.1", "lodash": "4.17.21", - "luxon": "2.3.1" - }, - "devDependencies": { - "eslint": "7.32.0", - "eslint-config-standard": "16.0.3", - "eslint-plugin-import": "2.25.4", - "eslint-plugin-node": "11.1.0", - "eslint-plugin-promise": "5.2.0" + "luxon": "3.7.1" } } diff --git a/dev/del-old-packages/README.md b/dev/del-old-packages/README.md new file mode 100644 index 0000000000..5f8f88e09e --- /dev/null +++ b/dev/del-old-packages/README.md @@ -0,0 +1,15 @@ +## Tool to delete old versions in GitHub Packages container registry + +This tool will fetch all versions for packages `datatracker-db` and `datatracker-db-pg` and delete all versions that are not latest and older than 7 days. + +### Requirements + +- Node 18.x or later +- Must provide a valid token in ENV variable `GITHUB_TOKEN` with read and delete packages permissions. + +### Usage + +```sh +npm install +node index +``` \ No newline at end of file diff --git a/dev/del-old-packages/index.js b/dev/del-old-packages/index.js new file mode 100644 index 0000000000..ff5ab649a2 --- /dev/null +++ b/dev/del-old-packages/index.js @@ -0,0 +1,54 @@ +import { Octokit } from '@octokit/core' +import { setTimeout } from 'node:timers/promises' +import { DateTime } from 'luxon' + +const octokit = new Octokit({ + auth: process.env.GITHUB_TOKEN +}) + +const oldestDate = DateTime.utc().minus({ days: 7 }) + +for (const pkgName of ['datatracker-db']) { + let hasMore = true + let currentPage = 1 + + while (hasMore) { + try { + console.info(`Fetching page ${currentPage}...`) + const versions = await octokit.request('GET /orgs/{org}/packages/{package_type}/{package_name}/versions{?page,per_page,state}', { + package_type: 'container', + package_name: pkgName, + org: 'ietf-tools', + page: currentPage, + per_page: 100 + }) + if (versions?.data?.length > 0) { + for (const ver of versions?.data) { + const verDate = DateTime.fromISO(ver.created_at) + if (ver?.metadata?.container?.tags?.includes('latest') || ver?.metadata?.container?.tags?.includes('latest-arm64') || ver?.metadata?.container?.tags?.includes('latest-x64')) { + console.info(`Latest package (${ver.id})... Skipping...`) + } else if (verDate > oldestDate) { + console.info(`Recent package (${ver.id}, ${verDate.toRelative()})... Skipping...`) + } else { + console.info(`Deleting package version ${ver.id}...`) + await octokit.request('DELETE /orgs/{org}/packages/{package_type}/{package_name}/versions/{package_version_id}', { + package_type: 'container', + package_name: pkgName, + org: 'ietf-tools', + package_version_id: ver.id + }) + await setTimeout(250) + } + } + currentPage++ + hasMore = true + } else { + hasMore = false + console.info('No more versions for this package.') + } + } catch (err) { + console.error(err) + hasMore = false + } + } +} diff --git a/dev/del-old-packages/package-lock.json b/dev/del-old-packages/package-lock.json new file mode 100644 index 0000000000..9899b290fb --- /dev/null +++ b/dev/del-old-packages/package-lock.json @@ -0,0 +1,368 @@ +{ + "name": "del-packages", + "version": "1.0.0", + "lockfileVersion": 2, + "requires": true, + "packages": { + "": { + "name": "del-packages", + "version": "1.0.0", + "license": "ISC", + "dependencies": { + "@octokit/core": "^4.2.4", + "luxon": "^3.4.4" + } + }, + "node_modules/@octokit/auth-token": { + "version": "3.0.2", + "resolved": "https://registry.npmjs.org/@octokit/auth-token/-/auth-token-3.0.2.tgz", + "integrity": "sha512-pq7CwIMV1kmzkFTimdwjAINCXKTajZErLB4wMLYapR2nuB/Jpr66+05wOTZMSCBXP6n4DdDWT2W19Bm17vU69Q==", + "dependencies": { + "@octokit/types": "^8.0.0" + }, + "engines": { + "node": ">= 14" + } + }, + "node_modules/@octokit/core": { + "version": "4.2.4", + "resolved": "https://registry.npmjs.org/@octokit/core/-/core-4.2.4.tgz", + "integrity": "sha512-rYKilwgzQ7/imScn3M9/pFfUf4I1AZEH3KhyJmtPdE2zfaXAn2mFfUy4FbKewzc2We5y/LlKLj36fWJLKC2SIQ==", + "dependencies": { + "@octokit/auth-token": "^3.0.0", + "@octokit/graphql": "^5.0.0", + "@octokit/request": "^6.0.0", + "@octokit/request-error": "^3.0.0", + "@octokit/types": "^9.0.0", + "before-after-hook": "^2.2.0", + "universal-user-agent": "^6.0.0" + }, + "engines": { + "node": ">= 14" + } + }, + "node_modules/@octokit/core/node_modules/@octokit/openapi-types": { + "version": "16.0.0", + "resolved": "https://registry.npmjs.org/@octokit/openapi-types/-/openapi-types-16.0.0.tgz", + "integrity": "sha512-JbFWOqTJVLHZSUUoF4FzAZKYtqdxWu9Z5m2QQnOyEa04fOFljvyh7D3GYKbfuaSWisqehImiVIMG4eyJeP5VEA==" + }, + "node_modules/@octokit/core/node_modules/@octokit/types": { + "version": "9.0.0", + "resolved": "https://registry.npmjs.org/@octokit/types/-/types-9.0.0.tgz", + "integrity": "sha512-LUewfj94xCMH2rbD5YJ+6AQ4AVjFYTgpp6rboWM5T7N3IsIF65SBEOVcYMGAEzO/kKNiNaW4LoWtoThOhH06gw==", + "dependencies": { + "@octokit/openapi-types": "^16.0.0" + } + }, + "node_modules/@octokit/endpoint": { + "version": "7.0.3", + "resolved": "https://registry.npmjs.org/@octokit/endpoint/-/endpoint-7.0.3.tgz", + "integrity": "sha512-57gRlb28bwTsdNXq+O3JTQ7ERmBTuik9+LelgcLIVfYwf235VHbN9QNo4kXExtp/h8T423cR5iJThKtFYxC7Lw==", + "dependencies": { + "@octokit/types": "^8.0.0", + "is-plain-object": "^5.0.0", + "universal-user-agent": "^6.0.0" + }, + "engines": { + "node": ">= 14" + } + }, + "node_modules/@octokit/graphql": { + "version": "5.0.4", + "resolved": "https://registry.npmjs.org/@octokit/graphql/-/graphql-5.0.4.tgz", + "integrity": "sha512-amO1M5QUQgYQo09aStR/XO7KAl13xpigcy/kI8/N1PnZYSS69fgte+xA4+c2DISKqUZfsh0wwjc2FaCt99L41A==", + "dependencies": { + "@octokit/request": "^6.0.0", + "@octokit/types": "^8.0.0", + "universal-user-agent": "^6.0.0" + }, + "engines": { + "node": ">= 14" + } + }, + "node_modules/@octokit/openapi-types": { + "version": "14.0.0", + "resolved": "https://registry.npmjs.org/@octokit/openapi-types/-/openapi-types-14.0.0.tgz", + "integrity": "sha512-HNWisMYlR8VCnNurDU6os2ikx0s0VyEjDYHNS/h4cgb8DeOxQ0n72HyinUtdDVxJhFy3FWLGl0DJhfEWk3P5Iw==" + }, + "node_modules/@octokit/request": { + "version": "6.2.2", + "resolved": "https://registry.npmjs.org/@octokit/request/-/request-6.2.2.tgz", + "integrity": "sha512-6VDqgj0HMc2FUX2awIs+sM6OwLgwHvAi4KCK3mT2H2IKRt6oH9d0fej5LluF5mck1lRR/rFWN0YIDSYXYSylbw==", + "dependencies": { + "@octokit/endpoint": "^7.0.0", + "@octokit/request-error": "^3.0.0", + "@octokit/types": "^8.0.0", + "is-plain-object": "^5.0.0", + "node-fetch": "^2.6.7", + "universal-user-agent": "^6.0.0" + }, + "engines": { + "node": ">= 14" + } + }, + "node_modules/@octokit/request-error": { + "version": "3.0.2", + "resolved": "https://registry.npmjs.org/@octokit/request-error/-/request-error-3.0.2.tgz", + "integrity": "sha512-WMNOFYrSaX8zXWoJg9u/pKgWPo94JXilMLb2VManNOby9EZxrQaBe/QSC4a1TzpAlpxofg2X/jMnCyZgL6y7eg==", + "dependencies": { + "@octokit/types": "^8.0.0", + "deprecation": "^2.0.0", + "once": "^1.4.0" + }, + "engines": { + "node": ">= 14" + } + }, + "node_modules/@octokit/types": { + "version": "8.1.0", + "resolved": "https://registry.npmjs.org/@octokit/types/-/types-8.1.0.tgz", + "integrity": "sha512-N4nLjzkiWBqVQqljTTsCrbvHGoWdWfcCeZjbHdggw7a9HbJMnxbK8A+UWdqwR4out30JarlSa3eqKyVK0n5aBg==", + "dependencies": { + "@octokit/openapi-types": "^14.0.0" + } + }, + "node_modules/before-after-hook": { + "version": "2.2.3", + "resolved": "https://registry.npmjs.org/before-after-hook/-/before-after-hook-2.2.3.tgz", + "integrity": "sha512-NzUnlZexiaH/46WDhANlyR2bXRopNg4F/zuSA3OpZnllCUgRaOF2znDioDWrmbNVsuZk6l9pMquQB38cfBZwkQ==" + }, + "node_modules/deprecation": { + "version": "2.3.1", + "resolved": "https://registry.npmjs.org/deprecation/-/deprecation-2.3.1.tgz", + "integrity": "sha512-xmHIy4F3scKVwMsQ4WnVaS8bHOx0DmVwRywosKhaILI0ywMDWPtBSku2HNxRvF7jtwDRsoEwYQSfbxj8b7RlJQ==" + }, + "node_modules/is-plain-object": { + "version": "5.0.0", + "resolved": "https://registry.npmjs.org/is-plain-object/-/is-plain-object-5.0.0.tgz", + "integrity": "sha512-VRSzKkbMm5jMDoKLbltAkFQ5Qr7VDiTFGXxYFXXowVj387GeGNOCsOH6Msy00SGZ3Fp84b1Naa1psqgcCIEP5Q==", + "engines": { + "node": ">=0.10.0" + } + }, + "node_modules/luxon": { + "version": "3.4.4", + "resolved": "https://registry.npmjs.org/luxon/-/luxon-3.4.4.tgz", + "integrity": "sha512-zobTr7akeGHnv7eBOXcRgMeCP6+uyYsczwmeRCauvpvaAltgNyTbLH/+VaEAPUeWBT+1GuNmz4wC/6jtQzbbVA==", + "engines": { + "node": ">=12" + } + }, + "node_modules/node-fetch": { + "version": "2.6.7", + "resolved": "https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.7.tgz", + "integrity": "sha512-ZjMPFEfVx5j+y2yF35Kzx5sF7kDzxuDj6ziH4FFbOp87zKDZNx8yExJIb05OGF4Nlt9IHFIMBkRl41VdvcNdbQ==", + "dependencies": { + "whatwg-url": "^5.0.0" + }, + "engines": { + "node": "4.x || >=6.0.0" + }, + "peerDependencies": { + "encoding": "^0.1.0" + }, + "peerDependenciesMeta": { + "encoding": { + "optional": true + } + } + }, + "node_modules/once": { + "version": "1.4.0", + "resolved": "https://registry.npmjs.org/once/-/once-1.4.0.tgz", + "integrity": "sha512-lNaJgI+2Q5URQBkccEKHTQOPaXdUxnZZElQTZY0MFUAuaEqe1E+Nyvgdz/aIyNi6Z9MzO5dv1H8n58/GELp3+w==", + "dependencies": { + "wrappy": "1" + } + }, + "node_modules/tr46": { + "version": "0.0.3", + "resolved": "https://registry.npmjs.org/tr46/-/tr46-0.0.3.tgz", + "integrity": "sha512-N3WMsuqV66lT30CrXNbEjx4GEwlow3v6rr4mCcv6prnfwhS01rkgyFdjPNBYd9br7LpXV1+Emh01fHnq2Gdgrw==" + }, + "node_modules/universal-user-agent": { + "version": "6.0.0", + "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-6.0.0.tgz", + "integrity": "sha512-isyNax3wXoKaulPDZWHQqbmIx1k2tb9fb3GGDBRxCscfYV2Ch7WxPArBsFEG8s/safwXTT7H4QGhaIkTp9447w==" + }, + "node_modules/webidl-conversions": { + "version": "3.0.1", + "resolved": "https://registry.npmjs.org/webidl-conversions/-/webidl-conversions-3.0.1.tgz", + "integrity": "sha512-2JAn3z8AR6rjK8Sm8orRC0h/bcl/DqL7tRPdGZ4I1CjdF+EaMLmYxBHyXuKL849eucPFhvBoxMsflfOb8kxaeQ==" + }, + "node_modules/whatwg-url": { + "version": "5.0.0", + "resolved": "https://registry.npmjs.org/whatwg-url/-/whatwg-url-5.0.0.tgz", + "integrity": "sha512-saE57nupxk6v3HY35+jzBwYa0rKSy0XR8JSxZPwgLr7ys0IBzhGviA1/TUGJLmSVqs8pb9AnvICXEuOHLprYTw==", + "dependencies": { + "tr46": "~0.0.3", + "webidl-conversions": "^3.0.0" + } + }, + "node_modules/wrappy": { + "version": "1.0.2", + "resolved": "https://registry.npmjs.org/wrappy/-/wrappy-1.0.2.tgz", + "integrity": "sha512-l4Sp/DRseor9wL6EvV2+TuQn63dMkPjZ/sp9XkghTEbV9KlPS1xUsZ3u7/IQO4wxtcFB4bgpQPRcR3QCvezPcQ==" + } + }, + "dependencies": { + "@octokit/auth-token": { + "version": "3.0.2", + "resolved": "https://registry.npmjs.org/@octokit/auth-token/-/auth-token-3.0.2.tgz", + "integrity": "sha512-pq7CwIMV1kmzkFTimdwjAINCXKTajZErLB4wMLYapR2nuB/Jpr66+05wOTZMSCBXP6n4DdDWT2W19Bm17vU69Q==", + "requires": { + "@octokit/types": "^8.0.0" + } + }, + "@octokit/core": { + "version": "4.2.4", + "resolved": "https://registry.npmjs.org/@octokit/core/-/core-4.2.4.tgz", + "integrity": "sha512-rYKilwgzQ7/imScn3M9/pFfUf4I1AZEH3KhyJmtPdE2zfaXAn2mFfUy4FbKewzc2We5y/LlKLj36fWJLKC2SIQ==", + "requires": { + "@octokit/auth-token": "^3.0.0", + "@octokit/graphql": "^5.0.0", + "@octokit/request": "^6.0.0", + "@octokit/request-error": "^3.0.0", + "@octokit/types": "^9.0.0", + "before-after-hook": "^2.2.0", + "universal-user-agent": "^6.0.0" + }, + "dependencies": { + "@octokit/openapi-types": { + "version": "16.0.0", + "resolved": "https://registry.npmjs.org/@octokit/openapi-types/-/openapi-types-16.0.0.tgz", + "integrity": "sha512-JbFWOqTJVLHZSUUoF4FzAZKYtqdxWu9Z5m2QQnOyEa04fOFljvyh7D3GYKbfuaSWisqehImiVIMG4eyJeP5VEA==" + }, + "@octokit/types": { + "version": "9.0.0", + "resolved": "https://registry.npmjs.org/@octokit/types/-/types-9.0.0.tgz", + "integrity": "sha512-LUewfj94xCMH2rbD5YJ+6AQ4AVjFYTgpp6rboWM5T7N3IsIF65SBEOVcYMGAEzO/kKNiNaW4LoWtoThOhH06gw==", + "requires": { + "@octokit/openapi-types": "^16.0.0" + } + } + } + }, + "@octokit/endpoint": { + "version": "7.0.3", + "resolved": "https://registry.npmjs.org/@octokit/endpoint/-/endpoint-7.0.3.tgz", + "integrity": "sha512-57gRlb28bwTsdNXq+O3JTQ7ERmBTuik9+LelgcLIVfYwf235VHbN9QNo4kXExtp/h8T423cR5iJThKtFYxC7Lw==", + "requires": { + "@octokit/types": "^8.0.0", + "is-plain-object": "^5.0.0", + "universal-user-agent": "^6.0.0" + } + }, + "@octokit/graphql": { + "version": "5.0.4", + "resolved": "https://registry.npmjs.org/@octokit/graphql/-/graphql-5.0.4.tgz", + "integrity": "sha512-amO1M5QUQgYQo09aStR/XO7KAl13xpigcy/kI8/N1PnZYSS69fgte+xA4+c2DISKqUZfsh0wwjc2FaCt99L41A==", + "requires": { + "@octokit/request": "^6.0.0", + "@octokit/types": "^8.0.0", + "universal-user-agent": "^6.0.0" + } + }, + "@octokit/openapi-types": { + "version": "14.0.0", + "resolved": "https://registry.npmjs.org/@octokit/openapi-types/-/openapi-types-14.0.0.tgz", + "integrity": "sha512-HNWisMYlR8VCnNurDU6os2ikx0s0VyEjDYHNS/h4cgb8DeOxQ0n72HyinUtdDVxJhFy3FWLGl0DJhfEWk3P5Iw==" + }, + "@octokit/request": { + "version": "6.2.2", + "resolved": "https://registry.npmjs.org/@octokit/request/-/request-6.2.2.tgz", + "integrity": "sha512-6VDqgj0HMc2FUX2awIs+sM6OwLgwHvAi4KCK3mT2H2IKRt6oH9d0fej5LluF5mck1lRR/rFWN0YIDSYXYSylbw==", + "requires": { + "@octokit/endpoint": "^7.0.0", + "@octokit/request-error": "^3.0.0", + "@octokit/types": "^8.0.0", + "is-plain-object": "^5.0.0", + "node-fetch": "^2.6.7", + "universal-user-agent": "^6.0.0" + } + }, + "@octokit/request-error": { + "version": "3.0.2", + "resolved": "https://registry.npmjs.org/@octokit/request-error/-/request-error-3.0.2.tgz", + "integrity": "sha512-WMNOFYrSaX8zXWoJg9u/pKgWPo94JXilMLb2VManNOby9EZxrQaBe/QSC4a1TzpAlpxofg2X/jMnCyZgL6y7eg==", + "requires": { + "@octokit/types": "^8.0.0", + "deprecation": "^2.0.0", + "once": "^1.4.0" + } + }, + "@octokit/types": { + "version": "8.1.0", + "resolved": "https://registry.npmjs.org/@octokit/types/-/types-8.1.0.tgz", + "integrity": "sha512-N4nLjzkiWBqVQqljTTsCrbvHGoWdWfcCeZjbHdggw7a9HbJMnxbK8A+UWdqwR4out30JarlSa3eqKyVK0n5aBg==", + "requires": { + "@octokit/openapi-types": "^14.0.0" + } + }, + "before-after-hook": { + "version": "2.2.3", + "resolved": "https://registry.npmjs.org/before-after-hook/-/before-after-hook-2.2.3.tgz", + "integrity": "sha512-NzUnlZexiaH/46WDhANlyR2bXRopNg4F/zuSA3OpZnllCUgRaOF2znDioDWrmbNVsuZk6l9pMquQB38cfBZwkQ==" + }, + "deprecation": { + "version": "2.3.1", + "resolved": "https://registry.npmjs.org/deprecation/-/deprecation-2.3.1.tgz", + "integrity": "sha512-xmHIy4F3scKVwMsQ4WnVaS8bHOx0DmVwRywosKhaILI0ywMDWPtBSku2HNxRvF7jtwDRsoEwYQSfbxj8b7RlJQ==" + }, + "is-plain-object": { + "version": "5.0.0", + "resolved": "https://registry.npmjs.org/is-plain-object/-/is-plain-object-5.0.0.tgz", + "integrity": "sha512-VRSzKkbMm5jMDoKLbltAkFQ5Qr7VDiTFGXxYFXXowVj387GeGNOCsOH6Msy00SGZ3Fp84b1Naa1psqgcCIEP5Q==" + }, + "luxon": { + "version": "3.4.4", + "resolved": "https://registry.npmjs.org/luxon/-/luxon-3.4.4.tgz", + "integrity": "sha512-zobTr7akeGHnv7eBOXcRgMeCP6+uyYsczwmeRCauvpvaAltgNyTbLH/+VaEAPUeWBT+1GuNmz4wC/6jtQzbbVA==" + }, + "node-fetch": { + "version": "2.6.7", + "resolved": "https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.7.tgz", + "integrity": "sha512-ZjMPFEfVx5j+y2yF35Kzx5sF7kDzxuDj6ziH4FFbOp87zKDZNx8yExJIb05OGF4Nlt9IHFIMBkRl41VdvcNdbQ==", + "requires": { + "whatwg-url": "^5.0.0" + } + }, + "once": { + "version": "1.4.0", + "resolved": "https://registry.npmjs.org/once/-/once-1.4.0.tgz", + "integrity": "sha512-lNaJgI+2Q5URQBkccEKHTQOPaXdUxnZZElQTZY0MFUAuaEqe1E+Nyvgdz/aIyNi6Z9MzO5dv1H8n58/GELp3+w==", + "requires": { + "wrappy": "1" + } + }, + "tr46": { + "version": "0.0.3", + "resolved": "https://registry.npmjs.org/tr46/-/tr46-0.0.3.tgz", + "integrity": "sha512-N3WMsuqV66lT30CrXNbEjx4GEwlow3v6rr4mCcv6prnfwhS01rkgyFdjPNBYd9br7LpXV1+Emh01fHnq2Gdgrw==" + }, + "universal-user-agent": { + "version": "6.0.0", + "resolved": "https://registry.npmjs.org/universal-user-agent/-/universal-user-agent-6.0.0.tgz", + "integrity": "sha512-isyNax3wXoKaulPDZWHQqbmIx1k2tb9fb3GGDBRxCscfYV2Ch7WxPArBsFEG8s/safwXTT7H4QGhaIkTp9447w==" + }, + "webidl-conversions": { + "version": "3.0.1", + "resolved": "https://registry.npmjs.org/webidl-conversions/-/webidl-conversions-3.0.1.tgz", + "integrity": "sha512-2JAn3z8AR6rjK8Sm8orRC0h/bcl/DqL7tRPdGZ4I1CjdF+EaMLmYxBHyXuKL849eucPFhvBoxMsflfOb8kxaeQ==" + }, + "whatwg-url": { + "version": "5.0.0", + "resolved": "https://registry.npmjs.org/whatwg-url/-/whatwg-url-5.0.0.tgz", + "integrity": "sha512-saE57nupxk6v3HY35+jzBwYa0rKSy0XR8JSxZPwgLr7ys0IBzhGviA1/TUGJLmSVqs8pb9AnvICXEuOHLprYTw==", + "requires": { + "tr46": "~0.0.3", + "webidl-conversions": "^3.0.0" + } + }, + "wrappy": { + "version": "1.0.2", + "resolved": "https://registry.npmjs.org/wrappy/-/wrappy-1.0.2.tgz", + "integrity": "sha512-l4Sp/DRseor9wL6EvV2+TuQn63dMkPjZ/sp9XkghTEbV9KlPS1xUsZ3u7/IQO4wxtcFB4bgpQPRcR3QCvezPcQ==" + } + } +} diff --git a/dev/del-old-packages/package.json b/dev/del-old-packages/package.json new file mode 100644 index 0000000000..c0b57b7f7b --- /dev/null +++ b/dev/del-old-packages/package.json @@ -0,0 +1,16 @@ +{ + "name": "del-packages", + "version": "1.0.0", + "description": "", + "type": "module", + "main": "index.js", + "scripts": { + "test": "echo \"Error: no test specified\" && exit 1" + }, + "author": "", + "license": "ISC", + "dependencies": { + "@octokit/core": "^4.2.4", + "luxon": "^3.4.4" + } +} diff --git a/dev/deploy-to-container/cli.js b/dev/deploy-to-container/cli.js index f804215a20..1a2d993ac4 100644 --- a/dev/deploy-to-container/cli.js +++ b/dev/deploy-to-container/cli.js @@ -3,7 +3,7 @@ import Docker from 'dockerode' import path from 'path' import fs from 'fs-extra' -import tar from 'tar' +import * as tar from 'tar' import yargs from 'yargs/yargs' import { hideBin } from 'yargs/helpers' import slugify from 'slugify' @@ -23,7 +23,7 @@ async function main () { throw new Error('Missing --branch argument!') } if (branch.indexOf('/') >= 0) { - branch = branch.split('/')[1] + branch = branch.split('/').slice(1).join('-') } branch = slugify(branch, { lower: true, strict: true }) if (branch.length < 1) { @@ -67,8 +67,10 @@ async function main () { .replace('__DBHOST__', `dt-db-${branch}`) .replace('__SECRETKEY__', nanoid(36)) .replace('__MQCONNSTR__', `amqp://datatracker:${mqKey}@dt-mq-${branch}/dt`) + .replace('__HOSTNAME__', hostname) ) await fs.copy(path.join(basePath, 'docker/scripts/app-create-dirs.sh'), path.join(releasePath, 'app-create-dirs.sh')) + await fs.copy(path.join(basePath, 'docker/scripts/app-init-celery.sh'), path.join(releasePath, 'app-init-celery.sh')) await fs.copy(path.join(basePath, 'dev/deploy-to-container/start.sh'), path.join(releasePath, 'start.sh')) await fs.copy(path.join(basePath, 'test/data'), path.join(releasePath, 'test/data')) console.info('Updated configuration files.') @@ -97,14 +99,6 @@ async function main () { }) console.info('Pulled latest MQ docker image.') - // Pull latest Celery image - console.info('Pulling latest Celery docker image...') - const celeryImagePullStream = await dock.pull('ghcr.io/ietf-tools/datatracker-celery:latest') - await new Promise((resolve, reject) => { - dock.modem.followProgress(celeryImagePullStream, (err, res) => err ? reject(err) : resolve(res)) - }) - console.info('Pulled latest Celery docker image.') - // Terminate existing containers console.info('Ensuring existing containers with same name are terminated...') const containers = await dock.listContainers({ all: true }) @@ -175,6 +169,9 @@ async function main () { Image: 'ghcr.io/ietf-tools/datatracker-db:latest', name: `dt-db-${branch}`, Hostname: `dt-db-${branch}`, + Labels: { + ...argv.nodbrefresh === 'true' && { nodbrefresh: '1' } + }, HostConfig: { NetworkMode: 'shared', RestartPolicy: { @@ -194,6 +191,9 @@ async function main () { Env: [ `CELERY_PASSWORD=${mqKey}` ], + Labels: { + ...argv.nodbrefresh === 'true' && { nodbrefresh: '1' } + }, HostConfig: { Memory: 4 * (1024 ** 3), // in bytes NetworkMode: 'shared', @@ -214,7 +214,7 @@ async function main () { const celeryContainers = {} for (const conConf of conConfs) { celeryContainers[conConf.name] = await dock.createContainer({ - Image: 'ghcr.io/ietf-tools/datatracker-celery:latest', + Image: 'ghcr.io/ietf-tools/datatracker-app-base:latest', name: `dt-${conConf.name}-${branch}`, Hostname: `dt-${conConf.name}-${branch}`, Env: [ @@ -222,6 +222,9 @@ async function main () { `CELERY_ROLE=${conConf.role}`, 'UPDATE_REQUIREMENTS_FROM=requirements.txt' ], + Labels: { + ...argv.nodbrefresh === 'true' && { nodbrefresh: '1' } + }, HostConfig: { Binds: [ 'dt-assets:/assets', @@ -233,7 +236,7 @@ async function main () { Name: 'unless-stopped' } }, - Cmd: ['--loglevel=INFO'] + Entrypoint: ['bash', '-c', 'chmod +x ./app-init-celery.sh && ./app-init-celery.sh'] }) } console.info('Created Celery docker containers successfully.') @@ -245,15 +248,17 @@ async function main () { name: `dt-app-${branch}`, Hostname: `dt-app-${branch}`, Env: [ - `LETSENCRYPT_HOST=${hostname}`, + // `LETSENCRYPT_HOST=${hostname}`, `VIRTUAL_HOST=${hostname}`, - `VIRTUAL_PORT=8000` + `VIRTUAL_PORT=8000`, + `PGHOST=dt-db-${branch}` ], Labels: { appversion: `${argv.appversion}` ?? '0.0.0', commit: `${argv.commit}` ?? 'unknown', ghrunid: `${argv.ghrunid}` ?? '0', - hostname + hostname, + ...argv.nodbrefresh === 'true' && { nodbrefresh: '1' } }, HostConfig: { Binds: [ diff --git a/dev/deploy-to-container/package-lock.json b/dev/deploy-to-container/package-lock.json index 4a1bc40460..5d5bef5604 100644 --- a/dev/deploy-to-container/package-lock.json +++ b/dev/deploy-to-container/package-lock.json @@ -6,18 +6,156 @@ "": { "name": "deploy-to-container", "dependencies": { - "dockerode": "^3.3.3", - "fs-extra": "^10.1.0", - "nanoid": "4.0.0", - "nanoid-dictionary": "5.0.0-beta.1", - "slugify": "1.6.5", - "tar": "^6.1.11", - "yargs": "^17.5.1" + "dockerode": "^4.0.10", + "fs-extra": "^11.3.4", + "nanoid": "5.1.7", + "nanoid-dictionary": "5.0.0", + "slugify": "1.6.9", + "tar": "^7.5.13", + "yargs": "^17.7.2" }, "engines": { "node": ">=16" } }, + "node_modules/@balena/dockerignore": { + "version": "1.0.2", + "resolved": "https://registry.npmjs.org/@balena/dockerignore/-/dockerignore-1.0.2.tgz", + "integrity": "sha512-wMue2Sy4GAVTk6Ic4tJVcnfdau+gx2EnG7S+uAEe+TWJFqE4YoWN4/H8MSLj4eYJKxGg26lZwboEniNiNwZQ6Q==" + }, + "node_modules/@grpc/grpc-js": { + "version": "1.12.5", + "resolved": "https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-1.12.5.tgz", + "integrity": "sha512-d3iiHxdpg5+ZcJ6jnDSOT8Z0O0VMVGy34jAnYLUX8yd36b1qn8f1TwOA/Lc7TsOh03IkPJ38eGI5qD2EjNkoEA==", + "dependencies": { + "@grpc/proto-loader": "^0.7.13", + "@js-sdsl/ordered-map": "^4.4.2" + }, + "engines": { + "node": ">=12.10.0" + } + }, + "node_modules/@grpc/proto-loader": { + "version": "0.7.13", + "resolved": "https://registry.npmjs.org/@grpc/proto-loader/-/proto-loader-0.7.13.tgz", + "integrity": "sha512-AiXO/bfe9bmxBjxxtYxFAXGZvMaN5s8kO+jBHAJCON8rJoB5YS/D6X7ZNc6XQkuHNmyl4CYaMI1fJ/Gn27RGGw==", + "dependencies": { + "lodash.camelcase": "^4.3.0", + "long": "^5.0.0", + "protobufjs": "^7.2.5", + "yargs": "^17.7.2" + }, + "bin": { + "proto-loader-gen-types": "build/bin/proto-loader-gen-types.js" + }, + "engines": { + "node": ">=6" + } + }, + "node_modules/@isaacs/fs-minipass": { + "version": "4.0.1", + "resolved": "https://registry.npmjs.org/@isaacs/fs-minipass/-/fs-minipass-4.0.1.tgz", + "integrity": "sha512-wgm9Ehl2jpeqP3zw/7mo3kRHFp5MEDhqAdwy1fTGkHAwnkGOVsgpvQhL8B5n1qlb01jV3n/bI0ZfZp5lWA1k4w==", + "dependencies": { + "minipass": "^7.0.4" + }, + "engines": { + "node": ">=18.0.0" + } + }, + "node_modules/@js-sdsl/ordered-map": { + "version": "4.4.2", + "resolved": "https://registry.npmjs.org/@js-sdsl/ordered-map/-/ordered-map-4.4.2.tgz", + "integrity": "sha512-iUKgm52T8HOE/makSxjqoWhe95ZJA1/G1sYsGev2JDKUSS14KAgg1LHb+Ba+IPow0xflbnSkOsZcO08C7w1gYw==", + "funding": { + "type": "opencollective", + "url": "https://opencollective.com/js-sdsl" + } + }, + "node_modules/@protobufjs/aspromise": { + "version": "1.1.2", + "resolved": "https://registry.npmjs.org/@protobufjs/aspromise/-/aspromise-1.1.2.tgz", + "integrity": "sha512-j+gKExEuLmKwvz3OgROXtrJ2UG2x8Ch2YZUxahh+s1F2HZ+wAceUNLkvy6zKCPVRkU++ZWQrdxsUeQXmcg4uoQ==" + }, + "node_modules/@protobufjs/base64": { + "version": "1.1.2", + "resolved": "https://registry.npmjs.org/@protobufjs/base64/-/base64-1.1.2.tgz", + "integrity": "sha512-AZkcAA5vnN/v4PDqKyMR5lx7hZttPDgClv83E//FMNhR2TMcLUhfRUBHCmSl0oi9zMgDDqRUJkSxO3wm85+XLg==" + }, + "node_modules/@protobufjs/codegen": { + "version": "2.0.4", + "resolved": "https://registry.npmjs.org/@protobufjs/codegen/-/codegen-2.0.4.tgz", + "integrity": "sha512-YyFaikqM5sH0ziFZCN3xDC7zeGaB/d0IUb9CATugHWbd1FRFwWwt4ld4OYMPWu5a3Xe01mGAULCdqhMlPl29Jg==" + }, + "node_modules/@protobufjs/eventemitter": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/eventemitter/-/eventemitter-1.1.0.tgz", + "integrity": "sha512-j9ednRT81vYJ9OfVuXG6ERSTdEL1xVsNgqpkxMsbIabzSo3goCjDIveeGv5d03om39ML71RdmrGNjG5SReBP/Q==" + }, + "node_modules/@protobufjs/fetch": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/fetch/-/fetch-1.1.0.tgz", + "integrity": "sha512-lljVXpqXebpsijW71PZaCYeIcE5on1w5DlQy5WH6GLbFryLUrBD4932W/E2BSpfRJWseIL4v/KPgBFxDOIdKpQ==", + "dependencies": { + "@protobufjs/aspromise": "^1.1.1", + "@protobufjs/inquire": "^1.1.0" + } + }, + "node_modules/@protobufjs/float": { + "version": "1.0.2", + "resolved": "https://registry.npmjs.org/@protobufjs/float/-/float-1.0.2.tgz", + "integrity": "sha512-Ddb+kVXlXst9d+R9PfTIxh1EdNkgoRe5tOX6t01f1lYWOvJnSPDBlG241QLzcyPdoNTsblLUdujGSE4RzrTZGQ==" + }, + "node_modules/@protobufjs/inquire": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/inquire/-/inquire-1.1.0.tgz", + "integrity": "sha512-kdSefcPdruJiFMVSbn801t4vFK7KB/5gd2fYvrxhuJYg8ILrmn9SKSX2tZdV6V+ksulWqS7aXjBcRXl3wHoD9Q==" + }, + "node_modules/@protobufjs/path": { + "version": "1.1.2", + "resolved": "https://registry.npmjs.org/@protobufjs/path/-/path-1.1.2.tgz", + "integrity": "sha512-6JOcJ5Tm08dOHAbdR3GrvP+yUUfkjG5ePsHYczMFLq3ZmMkAD98cDgcT2iA1lJ9NVwFd4tH/iSSoe44YWkltEA==" + }, + "node_modules/@protobufjs/pool": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/pool/-/pool-1.1.0.tgz", + "integrity": "sha512-0kELaGSIDBKvcgS4zkjz1PeddatrjYcmMWOlAuAPwAeccUrPHdUqo/J6LiymHHEiJT5NrF1UVwxY14f+fy4WQw==" + }, + "node_modules/@protobufjs/utf8": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/utf8/-/utf8-1.1.0.tgz", + "integrity": "sha512-Vvn3zZrhQZkkBE8LSuW3em98c0FwgO4nxzv6OdSxPKJIEKY2bGbHn+mhGIPerzI4twdxaP8/0+06HBpwf345Lw==" + }, + "node_modules/@types/node": { + "version": "22.10.5", + "resolved": "https://registry.npmjs.org/@types/node/-/node-22.10.5.tgz", + "integrity": "sha512-F8Q+SeGimwOo86fiovQh8qiXfFEh2/ocYv7tU5pJ3EXMSSxk1Joj5wefpFK2fHTf/N6HKGSxIDBT9f3gCxXPkQ==", + "dependencies": { + "undici-types": "~6.20.0" + } + }, + "node_modules/ansi-regex": { + "version": "5.0.1", + "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", + "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==", + "engines": { + "node": ">=8" + } + }, + "node_modules/ansi-styles": { + "version": "4.3.0", + "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz", + "integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==", + "dependencies": { + "color-convert": "^2.0.1" + }, + "engines": { + "node": ">=8" + }, + "funding": { + "url": "https://github.com/chalk/ansi-styles?sponsor=1" + } + }, "node_modules/asn1": { "version": "0.2.6", "resolved": "https://registry.npmjs.org/asn1/-/asn1-0.2.6.tgz", @@ -53,10 +191,43 @@ "tweetnacl": "^0.14.3" } }, + "node_modules/bl": { + "version": "4.1.0", + "resolved": "https://registry.npmjs.org/bl/-/bl-4.1.0.tgz", + "integrity": "sha512-1W07cM9gS6DcLperZfFSj+bWLtaPGSOHWhPiGzXmvVJbRLdG82sH/Kn8EtW1VqWVA54AKf2h5k5BbnIbwF3h6w==", + "dependencies": { + "buffer": "^5.5.0", + "inherits": "^2.0.4", + "readable-stream": "^3.4.0" + } + }, + "node_modules/buffer": { + "version": "5.7.1", + "resolved": "https://registry.npmjs.org/buffer/-/buffer-5.7.1.tgz", + "integrity": "sha512-EHcyIPBQ4BSGlvjB16k5KgAJ27CIsHY/2JBmCRReo48y9rQ3MaUzWX3KVlBa4U7MyX02HdVj0K7C3WaB3ju7FQ==", + "funding": [ + { + "type": "github", + "url": "https://github.com/sponsors/feross" + }, + { + "type": "patreon", + "url": "https://www.patreon.com/feross" + }, + { + "type": "consulting", + "url": "https://feross.org/support" + } + ], + "dependencies": { + "base64-js": "^1.3.1", + "ieee754": "^1.1.13" + } + }, "node_modules/buildcheck": { - "version": "0.0.3", - "resolved": "https://registry.npmjs.org/buildcheck/-/buildcheck-0.0.3.tgz", - "integrity": "sha512-pziaA+p/wdVImfcbsZLNF32EiWyujlQLwolMqUQE8xpKNOH7KmZQaY8sXN7DGOEzPAElo9QTaeNRfGnf3iOJbA==", + "version": "0.0.7", + "resolved": "https://registry.npmjs.org/buildcheck/-/buildcheck-0.0.7.tgz", + "integrity": "sha512-lHblz4ahamxpTmnsk+MNTRWsjYKv965MwOrSJyeD588rR3Jcu7swE+0wN5F+PbL5cjgu/9ObkhfzEPuofEMwLA==", "optional": true, "engines": { "node": ">=10.0.0" @@ -68,80 +239,16 @@ "integrity": "sha512-jJ0bqzaylmJtVnNgzTeSOs8DPavpbYgEr/b0YL8/2GO3xJEhInFmhKMUnEJQjZumK7KXGFhUy89PrsJWlakBVg==" }, "node_modules/cliui": { - "version": "7.0.4", - "resolved": "https://registry.npmjs.org/cliui/-/cliui-7.0.4.tgz", - "integrity": "sha512-OcRE68cOsVMXp1Yvonl/fzkQOyjLSu/8bhPDfQt0e0/Eb283TKP20Fs2MqoPsr9SwA595rRCA+QMzYc9nBP+JQ==", + "version": "8.0.1", + "resolved": "https://registry.npmjs.org/cliui/-/cliui-8.0.1.tgz", + "integrity": "sha512-BSeNnyus75C4//NQ9gQt1/csTXyo/8Sb+afLAkzAptFuMsod9HFokGNudZpi/oQV73hnVK+sR+5PVRMd+Dr7YQ==", "dependencies": { "string-width": "^4.2.0", - "strip-ansi": "^6.0.0", + "strip-ansi": "^6.0.1", "wrap-ansi": "^7.0.0" - } - }, - "node_modules/cliui/node_modules/ansi-regex": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", - "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==", - "engines": { - "node": ">=8" - } - }, - "node_modules/cliui/node_modules/ansi-styles": { - "version": "4.3.0", - "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz", - "integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==", - "dependencies": { - "color-convert": "^2.0.1" - }, - "engines": { - "node": ">=8" - }, - "funding": { - "url": "https://github.com/chalk/ansi-styles?sponsor=1" - } - }, - "node_modules/cliui/node_modules/emoji-regex": { - "version": "8.0.0", - "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", - "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" - }, - "node_modules/cliui/node_modules/string-width": { - "version": "4.2.3", - "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", - "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", - "dependencies": { - "emoji-regex": "^8.0.0", - "is-fullwidth-code-point": "^3.0.0", - "strip-ansi": "^6.0.1" - }, - "engines": { - "node": ">=8" - } - }, - "node_modules/cliui/node_modules/strip-ansi": { - "version": "6.0.1", - "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", - "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", - "dependencies": { - "ansi-regex": "^5.0.1" }, "engines": { - "node": ">=8" - } - }, - "node_modules/cliui/node_modules/wrap-ansi": { - "version": "7.0.0", - "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-7.0.0.tgz", - "integrity": "sha512-YVGIj2kamLSTxw6NsZjoBxfSwsn0ycdesmc4p+Q21c5zPuZ1pl+NfxVdxPtdHvmNVOQ6XSYG4AUtyt/Fi7D16Q==", - "dependencies": { - "ansi-styles": "^4.0.0", - "string-width": "^4.1.0", - "strip-ansi": "^6.0.0" - }, - "engines": { - "node": ">=10" - }, - "funding": { - "url": "https://github.com/chalk/wrap-ansi?sponsor=1" + "node": ">=12" } }, "node_modules/color-convert": { @@ -161,25 +268,25 @@ "integrity": "sha512-dOy+3AuW3a2wNbZHIuMZpTcgjGuLU/uBL/ubcZF9OXbDo8ff4O8yVp5Bf0efS8uEoYo5q4Fx7dY9OgQGXgAsQA==" }, "node_modules/cpu-features": { - "version": "0.0.4", - "resolved": "https://registry.npmjs.org/cpu-features/-/cpu-features-0.0.4.tgz", - "integrity": "sha512-fKiZ/zp1mUwQbnzb9IghXtHtDoTMtNeb8oYGx6kX2SYfhnG0HNdBEBIzB9b5KlXu5DQPhfy3mInbBxFcgwAr3A==", + "version": "0.0.10", + "resolved": "https://registry.npmjs.org/cpu-features/-/cpu-features-0.0.10.tgz", + "integrity": "sha512-9IkYqtX3YHPCzoVg1Py+o9057a3i0fp7S530UWokCSaFVTc7CwXPRiOjRjBQQ18ZCNafx78YfnG+HALxtVmOGA==", "hasInstallScript": true, "optional": true, "dependencies": { - "buildcheck": "0.0.3", - "nan": "^2.15.0" + "buildcheck": "~0.0.6", + "nan": "^2.19.0" }, "engines": { "node": ">=10.0.0" } }, "node_modules/debug": { - "version": "4.3.4", - "resolved": "https://registry.npmjs.org/debug/-/debug-4.3.4.tgz", - "integrity": "sha512-PRWFHuSU3eDtQJPvnNY7Jcket1j0t5OuOsFzPPzsekD52Zl8qUfFIPEiswXqIvHWGVHOgX+7G/vCNNhehwxfkQ==", + "version": "4.4.3", + "resolved": "https://registry.npmjs.org/debug/-/debug-4.4.3.tgz", + "integrity": "sha512-RGwwWnwQvkVfavKVt22FGLw+xYSdzARwm0ru6DhTVA3umU5hZc28V3kO4stgYryrTlLpuvgI9GiijltAjNbcqA==", "dependencies": { - "ms": "2.1.2" + "ms": "^2.1.3" }, "engines": { "node": ">=6.0" @@ -191,35 +298,45 @@ } }, "node_modules/docker-modem": { - "version": "3.0.5", - "resolved": "https://registry.npmjs.org/docker-modem/-/docker-modem-3.0.5.tgz", - "integrity": "sha512-x1E6jxWdtoK3+ifAUWj4w5egPdTDGBpesSCErm+aKET5BnnEOvDtTP6GxcnMB1zZiv2iQ0qJZvJie+1wfIRg6Q==", + "version": "5.0.7", + "resolved": "https://registry.npmjs.org/docker-modem/-/docker-modem-5.0.7.tgz", + "integrity": "sha512-XJgGhoR/CLpqshm4d3L7rzH6t8NgDFUIIpztYlLHIApeJjMZKYJMz2zxPsYxnejq5h3ELYSw/RBsi3t5h7gNTA==", "dependencies": { "debug": "^4.1.1", "readable-stream": "^3.5.0", "split-ca": "^1.0.1", - "ssh2": "^1.4.0" + "ssh2": "^1.15.0" }, "engines": { "node": ">= 8.0" } }, "node_modules/dockerode": { - "version": "3.3.3", - "resolved": "https://registry.npmjs.org/dockerode/-/dockerode-3.3.3.tgz", - "integrity": "sha512-lvKV6/NGf2/CYLt5V4c0fd6Fl9XZSCo1Z2HBT9ioKrKLMB2o+gA62Uza8RROpzGvYv57KJx2dKu+ZwSpB//OIA==", + "version": "4.0.10", + "resolved": "https://registry.npmjs.org/dockerode/-/dockerode-4.0.10.tgz", + "integrity": "sha512-8L/P9JynLBiG7/coiA4FlQXegHltRqS0a+KqI44P1zgQh8QLHTg7FKOwhkBgSJwZTeHsq30WRoVFLuwkfK0YFg==", "dependencies": { - "docker-modem": "^3.0.0", - "tar-fs": "~2.0.1" + "@balena/dockerignore": "^1.0.2", + "@grpc/grpc-js": "^1.11.1", + "@grpc/proto-loader": "^0.7.13", + "docker-modem": "^5.0.7", + "protobufjs": "^7.3.2", + "tar-fs": "^2.1.4", + "uuid": "^10.0.0" }, "engines": { "node": ">= 8.0" } }, + "node_modules/emoji-regex": { + "version": "8.0.0", + "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", + "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" + }, "node_modules/end-of-stream": { - "version": "1.4.4", - "resolved": "https://registry.npmjs.org/end-of-stream/-/end-of-stream-1.4.4.tgz", - "integrity": "sha512-+uw1inIHVPQoaVuHzRyXd21icM+cnt4CzD5rW+NC1wjOUSTOs+Te7FOv7AhN7vS9x/oIyhLP5PR1H+phQAHu5Q==", + "version": "1.4.5", + "resolved": "https://registry.npmjs.org/end-of-stream/-/end-of-stream-1.4.5.tgz", + "integrity": "sha512-ooEGc6HP26xXq/N+GCGOT0JKCLDGrq2bQUZrQ7gyrJiZANJ/8YDTxTpQBXGMn+WbIQXNVpyWymm7KYVICQnyOg==", "dependencies": { "once": "^1.4.0" } @@ -238,27 +355,16 @@ "integrity": "sha512-y6OAwoSIf7FyjMIv94u+b5rdheZEjzR63GTyZJm5qh4Bi+2YgwLCcI/fPFZkL5PSixOt6ZNKm+w+Hfp/Bciwow==" }, "node_modules/fs-extra": { - "version": "10.1.0", - "resolved": "https://registry.npmjs.org/fs-extra/-/fs-extra-10.1.0.tgz", - "integrity": "sha512-oRXApq54ETRj4eMiFzGnHWGy+zo5raudjuxN0b8H7s/RU2oW0Wvsx9O0ACRN/kRq9E8Vu/ReskGB5o3ji+FzHQ==", + "version": "11.3.4", + "resolved": "https://registry.npmjs.org/fs-extra/-/fs-extra-11.3.4.tgz", + "integrity": "sha512-CTXd6rk/M3/ULNQj8FBqBWHYBVYybQ3VPBw0xGKFe3tuH7ytT6ACnvzpIQ3UZtB8yvUKC2cXn1a+x+5EVQLovA==", "dependencies": { "graceful-fs": "^4.2.0", "jsonfile": "^6.0.1", "universalify": "^2.0.0" }, "engines": { - "node": ">=12" - } - }, - "node_modules/fs-minipass": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/fs-minipass/-/fs-minipass-2.1.0.tgz", - "integrity": "sha512-V/JgOLFCS+R6Vcq0slCuaeWEdNC3ouDlJMNIsacH2VtALiu9mV4LPrHc5cDl8k5aw6J8jwgWWpiTo5RYhmIzvg==", - "dependencies": { - "minipass": "^3.0.0" - }, - "engines": { - "node": ">= 8" + "node": ">=14.14" } }, "node_modules/get-caller-file": { @@ -317,38 +423,33 @@ "graceful-fs": "^4.1.6" } }, + "node_modules/lodash.camelcase": { + "version": "4.3.0", + "resolved": "https://registry.npmjs.org/lodash.camelcase/-/lodash.camelcase-4.3.0.tgz", + "integrity": "sha512-TwuEnCnxbc3rAvhf/LbG7tJUDzhqXyFnv3dtzLOPgCG/hODL7WFnsbwktkD7yUV0RrreP/l1PALq/YSg6VvjlA==" + }, + "node_modules/long": { + "version": "5.2.4", + "resolved": "https://registry.npmjs.org/long/-/long-5.2.4.tgz", + "integrity": "sha512-qtzLbJE8hq7VabR3mISmVGtoXP8KGc2Z/AT8OuqlYD7JTR3oqrgwdjnk07wpj1twXxYmgDXgoKVWUG/fReSzHg==" + }, "node_modules/minipass": { - "version": "3.3.4", - "resolved": "https://registry.npmjs.org/minipass/-/minipass-3.3.4.tgz", - "integrity": "sha512-I9WPbWHCGu8W+6k1ZiGpPu0GkoKBeorkfKNuAFBNS1HNFJvke82sxvI5bzcCNpWPorkOO5QQ+zomzzwRxejXiw==", - "dependencies": { - "yallist": "^4.0.0" - }, + "version": "7.1.2", + "resolved": "https://registry.npmjs.org/minipass/-/minipass-7.1.2.tgz", + "integrity": "sha512-qOOzS1cBTWYF4BH8fVePDBOO9iptMnGUEZwNc/cMWnTV2nVLZ7VoNWEPHkYczZA0pdoA7dl6e7FL659nX9S2aw==", "engines": { - "node": ">=8" + "node": ">=16 || 14 >=14.17" } }, "node_modules/minizlib": { - "version": "2.1.2", - "resolved": "https://registry.npmjs.org/minizlib/-/minizlib-2.1.2.tgz", - "integrity": "sha512-bAxsR8BVfj60DWXHE3u30oHzfl4G7khkSuPW+qvpd7jFRHm7dLxOjUk1EHACJ/hxLY8phGJ0YhYHZo7jil7Qdg==", + "version": "3.1.0", + "resolved": "https://registry.npmjs.org/minizlib/-/minizlib-3.1.0.tgz", + "integrity": "sha512-KZxYo1BUkWD2TVFLr0MQoM8vUUigWD3LlD83a/75BqC+4qE0Hb1Vo5v1FgcfaNXvfXzr+5EhQ6ing/CaBijTlw==", "dependencies": { - "minipass": "^3.0.0", - "yallist": "^4.0.0" + "minipass": "^7.1.2" }, "engines": { - "node": ">= 8" - } - }, - "node_modules/mkdirp": { - "version": "1.0.4", - "resolved": "https://registry.npmjs.org/mkdirp/-/mkdirp-1.0.4.tgz", - "integrity": "sha512-vVqVZQyf3WLx2Shd0qJ9xuvqgAyKPLAiqITEtqW0oIUjzo3PePDd6fW9iFz30ef7Ysp/oiWqbhszeGWW2T6Gzw==", - "bin": { - "mkdirp": "bin/cmd.js" - }, - "engines": { - "node": ">=10" + "node": ">= 18" } }, "node_modules/mkdirp-classic": { @@ -357,31 +458,38 @@ "integrity": "sha512-gKLcREMhtuZRwRAfqP3RFW+TK4JqApVBtOIftVgjuABpAtpxhPGaDcfvbhNvD0B8iD1oUr/txX35NjcaY6Ns/A==" }, "node_modules/ms": { - "version": "2.1.2", - "resolved": "https://registry.npmjs.org/ms/-/ms-2.1.2.tgz", - "integrity": "sha512-sGkPx+VjMtmA6MX27oA4FBFELFCZZ4S4XqeGOXCv68tT+jb3vk/RyaKWP0PTKyWtmLSM0b+adUTEvbs1PEaH2w==" + "version": "2.1.3", + "resolved": "https://registry.npmjs.org/ms/-/ms-2.1.3.tgz", + "integrity": "sha512-6FlzubTLZG3J2a/NVCAleEhjzq5oxgHyaCU9yYXvcLsvoVaHJq/s5xXI6/XXP6tz7R9xAOtHnSO/tXtF3WRTlA==" }, "node_modules/nan": { - "version": "2.16.0", - "resolved": "https://registry.npmjs.org/nan/-/nan-2.16.0.tgz", - "integrity": "sha512-UdAqHyFngu7TfQKsCBgAA6pWDkT8MAO7d0jyOecVhN5354xbLqdn8mV9Tat9gepAupm0bt2DbeaSC8vS52MuFA==", + "version": "2.26.2", + "resolved": "https://registry.npmjs.org/nan/-/nan-2.26.2.tgz", + "integrity": "sha512-0tTvBTYkt3tdGw22nrAy50x7gpbGCCFH3AFcyS5WiUu7Eu4vWlri1woE6qHBSfy11vksDqkiwjOnlR7WV8G1Hw==", "optional": true }, "node_modules/nanoid": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/nanoid/-/nanoid-4.0.0.tgz", - "integrity": "sha512-IgBP8piMxe/gf73RTQx7hmnhwz0aaEXYakvqZyE302IXW3HyVNhdNGC+O2MwMAVhLEnvXlvKtGbtJf6wvHihCg==", + "version": "5.1.7", + "resolved": "https://registry.npmjs.org/nanoid/-/nanoid-5.1.7.tgz", + "integrity": "sha512-ua3NDgISf6jdwezAheMOk4mbE1LXjm1DfMUDMuJf4AqxLFK3ccGpgWizwa5YV7Yz9EpXwEaWoRXSb/BnV0t5dQ==", + "funding": [ + { + "type": "github", + "url": "https://github.com/sponsors/ai" + } + ], "bin": { "nanoid": "bin/nanoid.js" }, "engines": { - "node": "^14 || ^16 || >=18" + "node": "^18 || >=20" } }, "node_modules/nanoid-dictionary": { - "version": "5.0.0-beta.1", - "resolved": "https://registry.npmjs.org/nanoid-dictionary/-/nanoid-dictionary-5.0.0-beta.1.tgz", - "integrity": "sha512-xBkL9zzkNjzJ/UnmWyiOUDVX/COoi05eS0oU28RYKFFQhdnzO5dTOPbVZ/fCFgIOGr1zNinDHJ68mm/KQfcgcw==" + "version": "5.0.0", + "resolved": "https://registry.npmjs.org/nanoid-dictionary/-/nanoid-dictionary-5.0.0.tgz", + "integrity": "sha512-/iCyQHwt36XkaIvSE9fcC8p6DiMPCZMTSMj9UT56Cv6T7f5CuxvOMhpNncaNieQ4z4d32p7ruEtAfRsb7Ya8Gw==", + "license": "MIT" }, "node_modules/once": { "version": "1.4.0", @@ -391,10 +499,33 @@ "wrappy": "1" } }, + "node_modules/protobufjs": { + "version": "7.4.0", + "resolved": "https://registry.npmjs.org/protobufjs/-/protobufjs-7.4.0.tgz", + "integrity": "sha512-mRUWCc3KUU4w1jU8sGxICXH/gNS94DvI1gxqDvBzhj1JpcsimQkYiOJfwsPUykUI5ZaspFbSgmBLER8IrQ3tqw==", + "hasInstallScript": true, + "dependencies": { + "@protobufjs/aspromise": "^1.1.2", + "@protobufjs/base64": "^1.1.2", + "@protobufjs/codegen": "^2.0.4", + "@protobufjs/eventemitter": "^1.1.0", + "@protobufjs/fetch": "^1.1.0", + "@protobufjs/float": "^1.0.2", + "@protobufjs/inquire": "^1.1.0", + "@protobufjs/path": "^1.1.2", + "@protobufjs/pool": "^1.1.0", + "@protobufjs/utf8": "^1.1.0", + "@types/node": ">=13.7.0", + "long": "^5.0.0" + }, + "engines": { + "node": ">=12.0.0" + } + }, "node_modules/pump": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/pump/-/pump-3.0.0.tgz", - "integrity": "sha512-LwZy+p3SFs1Pytd/jYct4wpv49HiYCqd9Rlc5ZVdk0V+8Yzv6jR5Blk3TRmPL1ft69TxP0IMZGJ+WPFU2BFhww==", + "version": "3.0.4", + "resolved": "https://registry.npmjs.org/pump/-/pump-3.0.4.tgz", + "integrity": "sha512-VS7sjc6KR7e1ukRFhQSY5LM2uBWAUPiOPa/A3mkKmiMwSmRFUITt0xuj+/lesgnCv+dPIEYlkzrcyXgquIHMcA==", "dependencies": { "end-of-stream": "^1.1.0", "once": "^1.3.1" @@ -446,9 +577,9 @@ "integrity": "sha512-YZo3K82SD7Riyi0E1EQPojLz7kpepnSQI9IyPbHHg1XXXevb5dJI7tpyN2ADxGcQbHG7vcyRHk0cbwqcQriUtg==" }, "node_modules/slugify": { - "version": "1.6.5", - "resolved": "https://registry.npmjs.org/slugify/-/slugify-1.6.5.tgz", - "integrity": "sha512-8mo9bslnBO3tr5PEVFzMPIWwWnipGS0xVbYf65zxDqfNwmzYn1LpiKNrR6DlClusuvo+hDHd1zKpmfAe83NQSQ==", + "version": "1.6.9", + "resolved": "https://registry.npmjs.org/slugify/-/slugify-1.6.9.tgz", + "integrity": "sha512-vZ7rfeehZui7wQs438JXBckYLkIIdfHOXsaVEUMyS5fHo1483l1bMdo0EDSWYclY0yZKFOipDy4KHuKs6ssvdg==", "engines": { "node": ">=8.0.0" } @@ -459,20 +590,20 @@ "integrity": "sha512-Q5thBSxp5t8WPTTJQS59LrGqOZqOsrhDGDVm8azCqIBjSBd7nd9o2PM+mDulQQkh8h//4U6hFZnc/mul8t5pWQ==" }, "node_modules/ssh2": { - "version": "1.11.0", - "resolved": "https://registry.npmjs.org/ssh2/-/ssh2-1.11.0.tgz", - "integrity": "sha512-nfg0wZWGSsfUe/IBJkXVll3PEZ//YH2guww+mP88gTpuSU4FtZN7zu9JoeTGOyCNx2dTDtT9fOpWwlzyj4uOOw==", + "version": "1.17.0", + "resolved": "https://registry.npmjs.org/ssh2/-/ssh2-1.17.0.tgz", + "integrity": "sha512-wPldCk3asibAjQ/kziWQQt1Wh3PgDFpC0XpwclzKcdT1vql6KeYxf5LIt4nlFkUeR8WuphYMKqUA56X4rjbfgQ==", "hasInstallScript": true, "dependencies": { - "asn1": "^0.2.4", + "asn1": "^0.2.6", "bcrypt-pbkdf": "^1.0.2" }, "engines": { "node": ">=10.16.0" }, "optionalDependencies": { - "cpu-features": "~0.0.4", - "nan": "^2.16.0" + "cpu-features": "~0.0.10", + "nan": "^2.23.0" } }, "node_modules/string_decoder": { @@ -483,31 +614,54 @@ "safe-buffer": "~5.2.0" } }, + "node_modules/string-width": { + "version": "4.2.3", + "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", + "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", + "dependencies": { + "emoji-regex": "^8.0.0", + "is-fullwidth-code-point": "^3.0.0", + "strip-ansi": "^6.0.1" + }, + "engines": { + "node": ">=8" + } + }, + "node_modules/strip-ansi": { + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", + "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", + "dependencies": { + "ansi-regex": "^5.0.1" + }, + "engines": { + "node": ">=8" + } + }, "node_modules/tar": { - "version": "6.1.11", - "resolved": "https://registry.npmjs.org/tar/-/tar-6.1.11.tgz", - "integrity": "sha512-an/KZQzQUkZCkuoAA64hM92X0Urb6VpRhAFllDzz44U2mcD5scmT3zBc4VgVpkugF580+DQn8eAFSyoQt0tznA==", + "version": "7.5.13", + "resolved": "https://registry.npmjs.org/tar/-/tar-7.5.13.tgz", + "integrity": "sha512-tOG/7GyXpFevhXVh8jOPJrmtRpOTsYqUIkVdVooZYJS/z8WhfQUX8RJILmeuJNinGAMSu1veBr4asSHFt5/hng==", "dependencies": { - "chownr": "^2.0.0", - "fs-minipass": "^2.0.0", - "minipass": "^3.0.0", - "minizlib": "^2.1.1", - "mkdirp": "^1.0.3", - "yallist": "^4.0.0" + "@isaacs/fs-minipass": "^4.0.0", + "chownr": "^3.0.0", + "minipass": "^7.1.2", + "minizlib": "^3.1.0", + "yallist": "^5.0.0" }, "engines": { - "node": ">= 10" + "node": ">=18" } }, "node_modules/tar-fs": { - "version": "2.0.1", - "resolved": "https://registry.npmjs.org/tar-fs/-/tar-fs-2.0.1.tgz", - "integrity": "sha512-6tzWDMeroL87uF/+lin46k+Q+46rAJ0SyPGz7OW7wTgblI273hsBqk2C1j0/xNadNLKDTUL9BukSjB7cwgmlPA==", + "version": "2.1.4", + "resolved": "https://registry.npmjs.org/tar-fs/-/tar-fs-2.1.4.tgz", + "integrity": "sha512-mDAjwmZdh7LTT6pNleZ05Yt65HC3E+NiQzl672vQG38jIrehtJk/J3mNwIg+vShQPcLF/LV7CMnDW6vjj6sfYQ==", "dependencies": { "chownr": "^1.1.1", "mkdirp-classic": "^0.5.2", "pump": "^3.0.0", - "tar-stream": "^2.0.0" + "tar-stream": "^2.1.4" } }, "node_modules/tar-stream": { @@ -525,45 +679,12 @@ "node": ">=6" } }, - "node_modules/tar-stream/node_modules/bl": { - "version": "4.1.0", - "resolved": "https://registry.npmjs.org/bl/-/bl-4.1.0.tgz", - "integrity": "sha512-1W07cM9gS6DcLperZfFSj+bWLtaPGSOHWhPiGzXmvVJbRLdG82sH/Kn8EtW1VqWVA54AKf2h5k5BbnIbwF3h6w==", - "dependencies": { - "buffer": "^5.5.0", - "inherits": "^2.0.4", - "readable-stream": "^3.4.0" - } - }, - "node_modules/tar-stream/node_modules/buffer": { - "version": "5.7.1", - "resolved": "https://registry.npmjs.org/buffer/-/buffer-5.7.1.tgz", - "integrity": "sha512-EHcyIPBQ4BSGlvjB16k5KgAJ27CIsHY/2JBmCRReo48y9rQ3MaUzWX3KVlBa4U7MyX02HdVj0K7C3WaB3ju7FQ==", - "funding": [ - { - "type": "github", - "url": "https://github.com/sponsors/feross" - }, - { - "type": "patreon", - "url": "https://www.patreon.com/feross" - }, - { - "type": "consulting", - "url": "https://feross.org/support" - } - ], - "dependencies": { - "base64-js": "^1.3.1", - "ieee754": "^1.1.13" - } - }, "node_modules/tar/node_modules/chownr": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/chownr/-/chownr-2.0.0.tgz", - "integrity": "sha512-bIomtDF5KGpdogkLd9VspvFzk9KfpyyGlS8YFVZl7TGPBHL5snIOnxeshwVgPteQ9b4Eydl+pVbIyE1DcvCWgQ==", + "version": "3.0.0", + "resolved": "https://registry.npmjs.org/chownr/-/chownr-3.0.0.tgz", + "integrity": "sha512-+IxzY9BZOQd/XuYPRmrvEVjF/nqj5kgT4kEq7VofrDoM1MxoRjEWkrCC3EtLi59TVawxTAn+orJwFQcrqEN1+g==", "engines": { - "node": ">=10" + "node": ">=18" } }, "node_modules/tweetnacl": { @@ -571,6 +692,11 @@ "resolved": "https://registry.npmjs.org/tweetnacl/-/tweetnacl-0.14.5.tgz", "integrity": "sha512-KXXFFdAbFXY4geFIwoyNK+f5Z1b7swfXABfL7HXCmoIWMKU3dmS26672A4EeQtDzLKy7SXmfBu51JolvEKwtGA==" }, + "node_modules/undici-types": { + "version": "6.20.0", + "resolved": "https://registry.npmjs.org/undici-types/-/undici-types-6.20.0.tgz", + "integrity": "sha512-Ny6QZ2Nju20vw1SRHe3d9jVu6gJ+4e3+MMpqu7pqE5HT6WsTSlce++GQmK5UXS8mzV8DSYHrQH+Xrf2jVcuKNg==" + }, "node_modules/universalify": { "version": "2.0.0", "resolved": "https://registry.npmjs.org/universalify/-/universalify-2.0.0.tgz", @@ -584,6 +710,34 @@ "resolved": "https://registry.npmjs.org/util-deprecate/-/util-deprecate-1.0.2.tgz", "integrity": "sha512-EPD5q1uXyFxJpCrLnCc1nHnq3gOa6DZBocAIiI2TaSCA7VCJ1UJDMagCzIkXNsUYfD1daK//LTEQ8xiIbrHtcw==" }, + "node_modules/uuid": { + "version": "10.0.0", + "resolved": "https://registry.npmjs.org/uuid/-/uuid-10.0.0.tgz", + "integrity": "sha512-8XkAphELsDnEGrDxUOHB3RGvXz6TeuYSGEZBOjtTtPm2lwhGBjLgOzLHB63IUWfBpNucQjND6d3AOudO+H3RWQ==", + "funding": [ + "https://github.com/sponsors/broofa", + "https://github.com/sponsors/ctavan" + ], + "bin": { + "uuid": "dist/bin/uuid" + } + }, + "node_modules/wrap-ansi": { + "version": "7.0.0", + "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-7.0.0.tgz", + "integrity": "sha512-YVGIj2kamLSTxw6NsZjoBxfSwsn0ycdesmc4p+Q21c5zPuZ1pl+NfxVdxPtdHvmNVOQ6XSYG4AUtyt/Fi7D16Q==", + "dependencies": { + "ansi-styles": "^4.0.0", + "string-width": "^4.1.0", + "strip-ansi": "^6.0.0" + }, + "engines": { + "node": ">=10" + }, + "funding": { + "url": "https://github.com/chalk/wrap-ansi?sponsor=1" + } + }, "node_modules/wrappy": { "version": "1.0.2", "resolved": "https://registry.npmjs.org/wrappy/-/wrappy-1.0.2.tgz", @@ -598,22 +752,25 @@ } }, "node_modules/yallist": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/yallist/-/yallist-4.0.0.tgz", - "integrity": "sha512-3wdGidZyq5PB084XLES5TpOSRA3wjXAlIWMhum2kRcv/41Sn2emQ0dycQW4uZXLejwKvg6EsvbdlVL+FYEct7A==" + "version": "5.0.0", + "resolved": "https://registry.npmjs.org/yallist/-/yallist-5.0.0.tgz", + "integrity": "sha512-YgvUTfwqyc7UXVMrB+SImsVYSmTS8X/tSrtdNZMImM+n7+QTriRXyXim0mBrTXNeqzVF0KWGgHPeiyViFFrNDw==", + "engines": { + "node": ">=18" + } }, "node_modules/yargs": { - "version": "17.5.1", - "resolved": "https://registry.npmjs.org/yargs/-/yargs-17.5.1.tgz", - "integrity": "sha512-t6YAJcxDkNX7NFYiVtKvWUz8l+PaKTLiL63mJYWR2GnHq2gjEWISzsLp9wg3aY36dY1j+gfIEL3pIF+XlJJfbA==", + "version": "17.7.2", + "resolved": "https://registry.npmjs.org/yargs/-/yargs-17.7.2.tgz", + "integrity": "sha512-7dSzzRQ++CKnNI/krKnYRV7JKKPUXMEh61soaHKg9mrWEhzFWhFnxPxGl+69cD1Ou63C13NUPCnmIcrvqCuM6w==", "dependencies": { - "cliui": "^7.0.2", + "cliui": "^8.0.1", "escalade": "^3.1.1", "get-caller-file": "^2.0.5", "require-directory": "^2.1.1", "string-width": "^4.2.3", "y18n": "^5.0.5", - "yargs-parser": "^21.0.0" + "yargs-parser": "^21.1.1" }, "engines": { "node": ">=12" @@ -626,46 +783,122 @@ "engines": { "node": ">=12" } + } + }, + "dependencies": { + "@balena/dockerignore": { + "version": "1.0.2", + "resolved": "https://registry.npmjs.org/@balena/dockerignore/-/dockerignore-1.0.2.tgz", + "integrity": "sha512-wMue2Sy4GAVTk6Ic4tJVcnfdau+gx2EnG7S+uAEe+TWJFqE4YoWN4/H8MSLj4eYJKxGg26lZwboEniNiNwZQ6Q==" }, - "node_modules/yargs/node_modules/ansi-regex": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", - "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==", - "engines": { - "node": ">=8" + "@grpc/grpc-js": { + "version": "1.12.5", + "resolved": "https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-1.12.5.tgz", + "integrity": "sha512-d3iiHxdpg5+ZcJ6jnDSOT8Z0O0VMVGy34jAnYLUX8yd36b1qn8f1TwOA/Lc7TsOh03IkPJ38eGI5qD2EjNkoEA==", + "requires": { + "@grpc/proto-loader": "^0.7.13", + "@js-sdsl/ordered-map": "^4.4.2" } }, - "node_modules/yargs/node_modules/emoji-regex": { - "version": "8.0.0", - "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", - "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" + "@grpc/proto-loader": { + "version": "0.7.13", + "resolved": "https://registry.npmjs.org/@grpc/proto-loader/-/proto-loader-0.7.13.tgz", + "integrity": "sha512-AiXO/bfe9bmxBjxxtYxFAXGZvMaN5s8kO+jBHAJCON8rJoB5YS/D6X7ZNc6XQkuHNmyl4CYaMI1fJ/Gn27RGGw==", + "requires": { + "lodash.camelcase": "^4.3.0", + "long": "^5.0.0", + "protobufjs": "^7.2.5", + "yargs": "^17.7.2" + } }, - "node_modules/yargs/node_modules/string-width": { - "version": "4.2.3", - "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", - "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", - "dependencies": { - "emoji-regex": "^8.0.0", - "is-fullwidth-code-point": "^3.0.0", - "strip-ansi": "^6.0.1" - }, - "engines": { - "node": ">=8" + "@isaacs/fs-minipass": { + "version": "4.0.1", + "resolved": "https://registry.npmjs.org/@isaacs/fs-minipass/-/fs-minipass-4.0.1.tgz", + "integrity": "sha512-wgm9Ehl2jpeqP3zw/7mo3kRHFp5MEDhqAdwy1fTGkHAwnkGOVsgpvQhL8B5n1qlb01jV3n/bI0ZfZp5lWA1k4w==", + "requires": { + "minipass": "^7.0.4" } }, - "node_modules/yargs/node_modules/strip-ansi": { - "version": "6.0.1", - "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", - "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", - "dependencies": { - "ansi-regex": "^5.0.1" - }, - "engines": { - "node": ">=8" + "@js-sdsl/ordered-map": { + "version": "4.4.2", + "resolved": "https://registry.npmjs.org/@js-sdsl/ordered-map/-/ordered-map-4.4.2.tgz", + "integrity": "sha512-iUKgm52T8HOE/makSxjqoWhe95ZJA1/G1sYsGev2JDKUSS14KAgg1LHb+Ba+IPow0xflbnSkOsZcO08C7w1gYw==" + }, + "@protobufjs/aspromise": { + "version": "1.1.2", + "resolved": "https://registry.npmjs.org/@protobufjs/aspromise/-/aspromise-1.1.2.tgz", + "integrity": "sha512-j+gKExEuLmKwvz3OgROXtrJ2UG2x8Ch2YZUxahh+s1F2HZ+wAceUNLkvy6zKCPVRkU++ZWQrdxsUeQXmcg4uoQ==" + }, + "@protobufjs/base64": { + "version": "1.1.2", + "resolved": "https://registry.npmjs.org/@protobufjs/base64/-/base64-1.1.2.tgz", + "integrity": "sha512-AZkcAA5vnN/v4PDqKyMR5lx7hZttPDgClv83E//FMNhR2TMcLUhfRUBHCmSl0oi9zMgDDqRUJkSxO3wm85+XLg==" + }, + "@protobufjs/codegen": { + "version": "2.0.4", + "resolved": "https://registry.npmjs.org/@protobufjs/codegen/-/codegen-2.0.4.tgz", + "integrity": "sha512-YyFaikqM5sH0ziFZCN3xDC7zeGaB/d0IUb9CATugHWbd1FRFwWwt4ld4OYMPWu5a3Xe01mGAULCdqhMlPl29Jg==" + }, + "@protobufjs/eventemitter": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/eventemitter/-/eventemitter-1.1.0.tgz", + "integrity": "sha512-j9ednRT81vYJ9OfVuXG6ERSTdEL1xVsNgqpkxMsbIabzSo3goCjDIveeGv5d03om39ML71RdmrGNjG5SReBP/Q==" + }, + "@protobufjs/fetch": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/fetch/-/fetch-1.1.0.tgz", + "integrity": "sha512-lljVXpqXebpsijW71PZaCYeIcE5on1w5DlQy5WH6GLbFryLUrBD4932W/E2BSpfRJWseIL4v/KPgBFxDOIdKpQ==", + "requires": { + "@protobufjs/aspromise": "^1.1.1", + "@protobufjs/inquire": "^1.1.0" } - } - }, - "dependencies": { + }, + "@protobufjs/float": { + "version": "1.0.2", + "resolved": "https://registry.npmjs.org/@protobufjs/float/-/float-1.0.2.tgz", + "integrity": "sha512-Ddb+kVXlXst9d+R9PfTIxh1EdNkgoRe5tOX6t01f1lYWOvJnSPDBlG241QLzcyPdoNTsblLUdujGSE4RzrTZGQ==" + }, + "@protobufjs/inquire": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/inquire/-/inquire-1.1.0.tgz", + "integrity": "sha512-kdSefcPdruJiFMVSbn801t4vFK7KB/5gd2fYvrxhuJYg8ILrmn9SKSX2tZdV6V+ksulWqS7aXjBcRXl3wHoD9Q==" + }, + "@protobufjs/path": { + "version": "1.1.2", + "resolved": "https://registry.npmjs.org/@protobufjs/path/-/path-1.1.2.tgz", + "integrity": "sha512-6JOcJ5Tm08dOHAbdR3GrvP+yUUfkjG5ePsHYczMFLq3ZmMkAD98cDgcT2iA1lJ9NVwFd4tH/iSSoe44YWkltEA==" + }, + "@protobufjs/pool": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/pool/-/pool-1.1.0.tgz", + "integrity": "sha512-0kELaGSIDBKvcgS4zkjz1PeddatrjYcmMWOlAuAPwAeccUrPHdUqo/J6LiymHHEiJT5NrF1UVwxY14f+fy4WQw==" + }, + "@protobufjs/utf8": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/utf8/-/utf8-1.1.0.tgz", + "integrity": "sha512-Vvn3zZrhQZkkBE8LSuW3em98c0FwgO4nxzv6OdSxPKJIEKY2bGbHn+mhGIPerzI4twdxaP8/0+06HBpwf345Lw==" + }, + "@types/node": { + "version": "22.10.5", + "resolved": "https://registry.npmjs.org/@types/node/-/node-22.10.5.tgz", + "integrity": "sha512-F8Q+SeGimwOo86fiovQh8qiXfFEh2/ocYv7tU5pJ3EXMSSxk1Joj5wefpFK2fHTf/N6HKGSxIDBT9f3gCxXPkQ==", + "requires": { + "undici-types": "~6.20.0" + } + }, + "ansi-regex": { + "version": "5.0.1", + "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", + "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==" + }, + "ansi-styles": { + "version": "4.3.0", + "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz", + "integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==", + "requires": { + "color-convert": "^2.0.1" + } + }, "asn1": { "version": "0.2.6", "resolved": "https://registry.npmjs.org/asn1/-/asn1-0.2.6.tgz", @@ -687,10 +920,29 @@ "tweetnacl": "^0.14.3" } }, + "bl": { + "version": "4.1.0", + "resolved": "https://registry.npmjs.org/bl/-/bl-4.1.0.tgz", + "integrity": "sha512-1W07cM9gS6DcLperZfFSj+bWLtaPGSOHWhPiGzXmvVJbRLdG82sH/Kn8EtW1VqWVA54AKf2h5k5BbnIbwF3h6w==", + "requires": { + "buffer": "^5.5.0", + "inherits": "^2.0.4", + "readable-stream": "^3.4.0" + } + }, + "buffer": { + "version": "5.7.1", + "resolved": "https://registry.npmjs.org/buffer/-/buffer-5.7.1.tgz", + "integrity": "sha512-EHcyIPBQ4BSGlvjB16k5KgAJ27CIsHY/2JBmCRReo48y9rQ3MaUzWX3KVlBa4U7MyX02HdVj0K7C3WaB3ju7FQ==", + "requires": { + "base64-js": "^1.3.1", + "ieee754": "^1.1.13" + } + }, "buildcheck": { - "version": "0.0.3", - "resolved": "https://registry.npmjs.org/buildcheck/-/buildcheck-0.0.3.tgz", - "integrity": "sha512-pziaA+p/wdVImfcbsZLNF32EiWyujlQLwolMqUQE8xpKNOH7KmZQaY8sXN7DGOEzPAElo9QTaeNRfGnf3iOJbA==", + "version": "0.0.7", + "resolved": "https://registry.npmjs.org/buildcheck/-/buildcheck-0.0.7.tgz", + "integrity": "sha512-lHblz4ahamxpTmnsk+MNTRWsjYKv965MwOrSJyeD588rR3Jcu7swE+0wN5F+PbL5cjgu/9ObkhfzEPuofEMwLA==", "optional": true }, "chownr": { @@ -699,61 +951,13 @@ "integrity": "sha512-jJ0bqzaylmJtVnNgzTeSOs8DPavpbYgEr/b0YL8/2GO3xJEhInFmhKMUnEJQjZumK7KXGFhUy89PrsJWlakBVg==" }, "cliui": { - "version": "7.0.4", - "resolved": "https://registry.npmjs.org/cliui/-/cliui-7.0.4.tgz", - "integrity": "sha512-OcRE68cOsVMXp1Yvonl/fzkQOyjLSu/8bhPDfQt0e0/Eb283TKP20Fs2MqoPsr9SwA595rRCA+QMzYc9nBP+JQ==", + "version": "8.0.1", + "resolved": "https://registry.npmjs.org/cliui/-/cliui-8.0.1.tgz", + "integrity": "sha512-BSeNnyus75C4//NQ9gQt1/csTXyo/8Sb+afLAkzAptFuMsod9HFokGNudZpi/oQV73hnVK+sR+5PVRMd+Dr7YQ==", "requires": { "string-width": "^4.2.0", - "strip-ansi": "^6.0.0", + "strip-ansi": "^6.0.1", "wrap-ansi": "^7.0.0" - }, - "dependencies": { - "ansi-regex": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", - "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==" - }, - "ansi-styles": { - "version": "4.3.0", - "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz", - "integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==", - "requires": { - "color-convert": "^2.0.1" - } - }, - "emoji-regex": { - "version": "8.0.0", - "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", - "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" - }, - "string-width": { - "version": "4.2.3", - "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", - "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", - "requires": { - "emoji-regex": "^8.0.0", - "is-fullwidth-code-point": "^3.0.0", - "strip-ansi": "^6.0.1" - } - }, - "strip-ansi": { - "version": "6.0.1", - "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", - "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", - "requires": { - "ansi-regex": "^5.0.1" - } - }, - "wrap-ansi": { - "version": "7.0.0", - "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-7.0.0.tgz", - "integrity": "sha512-YVGIj2kamLSTxw6NsZjoBxfSwsn0ycdesmc4p+Q21c5zPuZ1pl+NfxVdxPtdHvmNVOQ6XSYG4AUtyt/Fi7D16Q==", - "requires": { - "ansi-styles": "^4.0.0", - "string-width": "^4.1.0", - "strip-ansi": "^6.0.0" - } - } } }, "color-convert": { @@ -770,47 +974,57 @@ "integrity": "sha512-dOy+3AuW3a2wNbZHIuMZpTcgjGuLU/uBL/ubcZF9OXbDo8ff4O8yVp5Bf0efS8uEoYo5q4Fx7dY9OgQGXgAsQA==" }, "cpu-features": { - "version": "0.0.4", - "resolved": "https://registry.npmjs.org/cpu-features/-/cpu-features-0.0.4.tgz", - "integrity": "sha512-fKiZ/zp1mUwQbnzb9IghXtHtDoTMtNeb8oYGx6kX2SYfhnG0HNdBEBIzB9b5KlXu5DQPhfy3mInbBxFcgwAr3A==", + "version": "0.0.10", + "resolved": "https://registry.npmjs.org/cpu-features/-/cpu-features-0.0.10.tgz", + "integrity": "sha512-9IkYqtX3YHPCzoVg1Py+o9057a3i0fp7S530UWokCSaFVTc7CwXPRiOjRjBQQ18ZCNafx78YfnG+HALxtVmOGA==", "optional": true, "requires": { - "buildcheck": "0.0.3", - "nan": "^2.15.0" + "buildcheck": "~0.0.6", + "nan": "^2.19.0" } }, "debug": { - "version": "4.3.4", - "resolved": "https://registry.npmjs.org/debug/-/debug-4.3.4.tgz", - "integrity": "sha512-PRWFHuSU3eDtQJPvnNY7Jcket1j0t5OuOsFzPPzsekD52Zl8qUfFIPEiswXqIvHWGVHOgX+7G/vCNNhehwxfkQ==", + "version": "4.4.3", + "resolved": "https://registry.npmjs.org/debug/-/debug-4.4.3.tgz", + "integrity": "sha512-RGwwWnwQvkVfavKVt22FGLw+xYSdzARwm0ru6DhTVA3umU5hZc28V3kO4stgYryrTlLpuvgI9GiijltAjNbcqA==", "requires": { - "ms": "2.1.2" + "ms": "^2.1.3" } }, "docker-modem": { - "version": "3.0.5", - "resolved": "https://registry.npmjs.org/docker-modem/-/docker-modem-3.0.5.tgz", - "integrity": "sha512-x1E6jxWdtoK3+ifAUWj4w5egPdTDGBpesSCErm+aKET5BnnEOvDtTP6GxcnMB1zZiv2iQ0qJZvJie+1wfIRg6Q==", + "version": "5.0.7", + "resolved": "https://registry.npmjs.org/docker-modem/-/docker-modem-5.0.7.tgz", + "integrity": "sha512-XJgGhoR/CLpqshm4d3L7rzH6t8NgDFUIIpztYlLHIApeJjMZKYJMz2zxPsYxnejq5h3ELYSw/RBsi3t5h7gNTA==", "requires": { "debug": "^4.1.1", "readable-stream": "^3.5.0", "split-ca": "^1.0.1", - "ssh2": "^1.4.0" + "ssh2": "^1.15.0" } }, "dockerode": { - "version": "3.3.3", - "resolved": "https://registry.npmjs.org/dockerode/-/dockerode-3.3.3.tgz", - "integrity": "sha512-lvKV6/NGf2/CYLt5V4c0fd6Fl9XZSCo1Z2HBT9ioKrKLMB2o+gA62Uza8RROpzGvYv57KJx2dKu+ZwSpB//OIA==", + "version": "4.0.10", + "resolved": "https://registry.npmjs.org/dockerode/-/dockerode-4.0.10.tgz", + "integrity": "sha512-8L/P9JynLBiG7/coiA4FlQXegHltRqS0a+KqI44P1zgQh8QLHTg7FKOwhkBgSJwZTeHsq30WRoVFLuwkfK0YFg==", "requires": { - "docker-modem": "^3.0.0", - "tar-fs": "~2.0.1" + "@balena/dockerignore": "^1.0.2", + "@grpc/grpc-js": "^1.11.1", + "@grpc/proto-loader": "^0.7.13", + "docker-modem": "^5.0.7", + "protobufjs": "^7.3.2", + "tar-fs": "^2.1.4", + "uuid": "^10.0.0" } }, + "emoji-regex": { + "version": "8.0.0", + "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", + "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" + }, "end-of-stream": { - "version": "1.4.4", - "resolved": "https://registry.npmjs.org/end-of-stream/-/end-of-stream-1.4.4.tgz", - "integrity": "sha512-+uw1inIHVPQoaVuHzRyXd21icM+cnt4CzD5rW+NC1wjOUSTOs+Te7FOv7AhN7vS9x/oIyhLP5PR1H+phQAHu5Q==", + "version": "1.4.5", + "resolved": "https://registry.npmjs.org/end-of-stream/-/end-of-stream-1.4.5.tgz", + "integrity": "sha512-ooEGc6HP26xXq/N+GCGOT0JKCLDGrq2bQUZrQ7gyrJiZANJ/8YDTxTpQBXGMn+WbIQXNVpyWymm7KYVICQnyOg==", "requires": { "once": "^1.4.0" } @@ -826,23 +1040,15 @@ "integrity": "sha512-y6OAwoSIf7FyjMIv94u+b5rdheZEjzR63GTyZJm5qh4Bi+2YgwLCcI/fPFZkL5PSixOt6ZNKm+w+Hfp/Bciwow==" }, "fs-extra": { - "version": "10.1.0", - "resolved": "https://registry.npmjs.org/fs-extra/-/fs-extra-10.1.0.tgz", - "integrity": "sha512-oRXApq54ETRj4eMiFzGnHWGy+zo5raudjuxN0b8H7s/RU2oW0Wvsx9O0ACRN/kRq9E8Vu/ReskGB5o3ji+FzHQ==", + "version": "11.3.4", + "resolved": "https://registry.npmjs.org/fs-extra/-/fs-extra-11.3.4.tgz", + "integrity": "sha512-CTXd6rk/M3/ULNQj8FBqBWHYBVYybQ3VPBw0xGKFe3tuH7ytT6ACnvzpIQ3UZtB8yvUKC2cXn1a+x+5EVQLovA==", "requires": { "graceful-fs": "^4.2.0", "jsonfile": "^6.0.1", "universalify": "^2.0.0" } }, - "fs-minipass": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/fs-minipass/-/fs-minipass-2.1.0.tgz", - "integrity": "sha512-V/JgOLFCS+R6Vcq0slCuaeWEdNC3ouDlJMNIsacH2VtALiu9mV4LPrHc5cDl8k5aw6J8jwgWWpiTo5RYhmIzvg==", - "requires": { - "minipass": "^3.0.0" - } - }, "get-caller-file": { "version": "2.0.5", "resolved": "https://registry.npmjs.org/get-caller-file/-/get-caller-file-2.0.5.tgz", @@ -877,53 +1083,54 @@ "universalify": "^2.0.0" } }, + "lodash.camelcase": { + "version": "4.3.0", + "resolved": "https://registry.npmjs.org/lodash.camelcase/-/lodash.camelcase-4.3.0.tgz", + "integrity": "sha512-TwuEnCnxbc3rAvhf/LbG7tJUDzhqXyFnv3dtzLOPgCG/hODL7WFnsbwktkD7yUV0RrreP/l1PALq/YSg6VvjlA==" + }, + "long": { + "version": "5.2.4", + "resolved": "https://registry.npmjs.org/long/-/long-5.2.4.tgz", + "integrity": "sha512-qtzLbJE8hq7VabR3mISmVGtoXP8KGc2Z/AT8OuqlYD7JTR3oqrgwdjnk07wpj1twXxYmgDXgoKVWUG/fReSzHg==" + }, "minipass": { - "version": "3.3.4", - "resolved": "https://registry.npmjs.org/minipass/-/minipass-3.3.4.tgz", - "integrity": "sha512-I9WPbWHCGu8W+6k1ZiGpPu0GkoKBeorkfKNuAFBNS1HNFJvke82sxvI5bzcCNpWPorkOO5QQ+zomzzwRxejXiw==", - "requires": { - "yallist": "^4.0.0" - } + "version": "7.1.2", + "resolved": "https://registry.npmjs.org/minipass/-/minipass-7.1.2.tgz", + "integrity": "sha512-qOOzS1cBTWYF4BH8fVePDBOO9iptMnGUEZwNc/cMWnTV2nVLZ7VoNWEPHkYczZA0pdoA7dl6e7FL659nX9S2aw==" }, "minizlib": { - "version": "2.1.2", - "resolved": "https://registry.npmjs.org/minizlib/-/minizlib-2.1.2.tgz", - "integrity": "sha512-bAxsR8BVfj60DWXHE3u30oHzfl4G7khkSuPW+qvpd7jFRHm7dLxOjUk1EHACJ/hxLY8phGJ0YhYHZo7jil7Qdg==", + "version": "3.1.0", + "resolved": "https://registry.npmjs.org/minizlib/-/minizlib-3.1.0.tgz", + "integrity": "sha512-KZxYo1BUkWD2TVFLr0MQoM8vUUigWD3LlD83a/75BqC+4qE0Hb1Vo5v1FgcfaNXvfXzr+5EhQ6ing/CaBijTlw==", "requires": { - "minipass": "^3.0.0", - "yallist": "^4.0.0" + "minipass": "^7.1.2" } }, - "mkdirp": { - "version": "1.0.4", - "resolved": "https://registry.npmjs.org/mkdirp/-/mkdirp-1.0.4.tgz", - "integrity": "sha512-vVqVZQyf3WLx2Shd0qJ9xuvqgAyKPLAiqITEtqW0oIUjzo3PePDd6fW9iFz30ef7Ysp/oiWqbhszeGWW2T6Gzw==" - }, "mkdirp-classic": { "version": "0.5.3", "resolved": "https://registry.npmjs.org/mkdirp-classic/-/mkdirp-classic-0.5.3.tgz", "integrity": "sha512-gKLcREMhtuZRwRAfqP3RFW+TK4JqApVBtOIftVgjuABpAtpxhPGaDcfvbhNvD0B8iD1oUr/txX35NjcaY6Ns/A==" }, "ms": { - "version": "2.1.2", - "resolved": "https://registry.npmjs.org/ms/-/ms-2.1.2.tgz", - "integrity": "sha512-sGkPx+VjMtmA6MX27oA4FBFELFCZZ4S4XqeGOXCv68tT+jb3vk/RyaKWP0PTKyWtmLSM0b+adUTEvbs1PEaH2w==" + "version": "2.1.3", + "resolved": "https://registry.npmjs.org/ms/-/ms-2.1.3.tgz", + "integrity": "sha512-6FlzubTLZG3J2a/NVCAleEhjzq5oxgHyaCU9yYXvcLsvoVaHJq/s5xXI6/XXP6tz7R9xAOtHnSO/tXtF3WRTlA==" }, "nan": { - "version": "2.16.0", - "resolved": "https://registry.npmjs.org/nan/-/nan-2.16.0.tgz", - "integrity": "sha512-UdAqHyFngu7TfQKsCBgAA6pWDkT8MAO7d0jyOecVhN5354xbLqdn8mV9Tat9gepAupm0bt2DbeaSC8vS52MuFA==", + "version": "2.26.2", + "resolved": "https://registry.npmjs.org/nan/-/nan-2.26.2.tgz", + "integrity": "sha512-0tTvBTYkt3tdGw22nrAy50x7gpbGCCFH3AFcyS5WiUu7Eu4vWlri1woE6qHBSfy11vksDqkiwjOnlR7WV8G1Hw==", "optional": true }, "nanoid": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/nanoid/-/nanoid-4.0.0.tgz", - "integrity": "sha512-IgBP8piMxe/gf73RTQx7hmnhwz0aaEXYakvqZyE302IXW3HyVNhdNGC+O2MwMAVhLEnvXlvKtGbtJf6wvHihCg==" + "version": "5.1.7", + "resolved": "https://registry.npmjs.org/nanoid/-/nanoid-5.1.7.tgz", + "integrity": "sha512-ua3NDgISf6jdwezAheMOk4mbE1LXjm1DfMUDMuJf4AqxLFK3ccGpgWizwa5YV7Yz9EpXwEaWoRXSb/BnV0t5dQ==" }, "nanoid-dictionary": { - "version": "5.0.0-beta.1", - "resolved": "https://registry.npmjs.org/nanoid-dictionary/-/nanoid-dictionary-5.0.0-beta.1.tgz", - "integrity": "sha512-xBkL9zzkNjzJ/UnmWyiOUDVX/COoi05eS0oU28RYKFFQhdnzO5dTOPbVZ/fCFgIOGr1zNinDHJ68mm/KQfcgcw==" + "version": "5.0.0", + "resolved": "https://registry.npmjs.org/nanoid-dictionary/-/nanoid-dictionary-5.0.0.tgz", + "integrity": "sha512-/iCyQHwt36XkaIvSE9fcC8p6DiMPCZMTSMj9UT56Cv6T7f5CuxvOMhpNncaNieQ4z4d32p7ruEtAfRsb7Ya8Gw==" }, "once": { "version": "1.4.0", @@ -933,10 +1140,29 @@ "wrappy": "1" } }, + "protobufjs": { + "version": "7.4.0", + "resolved": "https://registry.npmjs.org/protobufjs/-/protobufjs-7.4.0.tgz", + "integrity": "sha512-mRUWCc3KUU4w1jU8sGxICXH/gNS94DvI1gxqDvBzhj1JpcsimQkYiOJfwsPUykUI5ZaspFbSgmBLER8IrQ3tqw==", + "requires": { + "@protobufjs/aspromise": "^1.1.2", + "@protobufjs/base64": "^1.1.2", + "@protobufjs/codegen": "^2.0.4", + "@protobufjs/eventemitter": "^1.1.0", + "@protobufjs/fetch": "^1.1.0", + "@protobufjs/float": "^1.0.2", + "@protobufjs/inquire": "^1.1.0", + "@protobufjs/path": "^1.1.2", + "@protobufjs/pool": "^1.1.0", + "@protobufjs/utf8": "^1.1.0", + "@types/node": ">=13.7.0", + "long": "^5.0.0" + } + }, "pump": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/pump/-/pump-3.0.0.tgz", - "integrity": "sha512-LwZy+p3SFs1Pytd/jYct4wpv49HiYCqd9Rlc5ZVdk0V+8Yzv6jR5Blk3TRmPL1ft69TxP0IMZGJ+WPFU2BFhww==", + "version": "3.0.4", + "resolved": "https://registry.npmjs.org/pump/-/pump-3.0.4.tgz", + "integrity": "sha512-VS7sjc6KR7e1ukRFhQSY5LM2uBWAUPiOPa/A3mkKmiMwSmRFUITt0xuj+/lesgnCv+dPIEYlkzrcyXgquIHMcA==", "requires": { "end-of-stream": "^1.1.0", "once": "^1.3.1" @@ -968,9 +1194,9 @@ "integrity": "sha512-YZo3K82SD7Riyi0E1EQPojLz7kpepnSQI9IyPbHHg1XXXevb5dJI7tpyN2ADxGcQbHG7vcyRHk0cbwqcQriUtg==" }, "slugify": { - "version": "1.6.5", - "resolved": "https://registry.npmjs.org/slugify/-/slugify-1.6.5.tgz", - "integrity": "sha512-8mo9bslnBO3tr5PEVFzMPIWwWnipGS0xVbYf65zxDqfNwmzYn1LpiKNrR6DlClusuvo+hDHd1zKpmfAe83NQSQ==" + "version": "1.6.9", + "resolved": "https://registry.npmjs.org/slugify/-/slugify-1.6.9.tgz", + "integrity": "sha512-vZ7rfeehZui7wQs438JXBckYLkIIdfHOXsaVEUMyS5fHo1483l1bMdo0EDSWYclY0yZKFOipDy4KHuKs6ssvdg==" }, "split-ca": { "version": "1.0.1", @@ -978,14 +1204,14 @@ "integrity": "sha512-Q5thBSxp5t8WPTTJQS59LrGqOZqOsrhDGDVm8azCqIBjSBd7nd9o2PM+mDulQQkh8h//4U6hFZnc/mul8t5pWQ==" }, "ssh2": { - "version": "1.11.0", - "resolved": "https://registry.npmjs.org/ssh2/-/ssh2-1.11.0.tgz", - "integrity": "sha512-nfg0wZWGSsfUe/IBJkXVll3PEZ//YH2guww+mP88gTpuSU4FtZN7zu9JoeTGOyCNx2dTDtT9fOpWwlzyj4uOOw==", + "version": "1.17.0", + "resolved": "https://registry.npmjs.org/ssh2/-/ssh2-1.17.0.tgz", + "integrity": "sha512-wPldCk3asibAjQ/kziWQQt1Wh3PgDFpC0XpwclzKcdT1vql6KeYxf5LIt4nlFkUeR8WuphYMKqUA56X4rjbfgQ==", "requires": { - "asn1": "^0.2.4", + "asn1": "^0.2.6", "bcrypt-pbkdf": "^1.0.2", - "cpu-features": "~0.0.4", - "nan": "^2.16.0" + "cpu-features": "~0.0.10", + "nan": "^2.23.0" } }, "string_decoder": { @@ -996,35 +1222,52 @@ "safe-buffer": "~5.2.0" } }, + "string-width": { + "version": "4.2.3", + "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", + "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", + "requires": { + "emoji-regex": "^8.0.0", + "is-fullwidth-code-point": "^3.0.0", + "strip-ansi": "^6.0.1" + } + }, + "strip-ansi": { + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", + "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", + "requires": { + "ansi-regex": "^5.0.1" + } + }, "tar": { - "version": "6.1.11", - "resolved": "https://registry.npmjs.org/tar/-/tar-6.1.11.tgz", - "integrity": "sha512-an/KZQzQUkZCkuoAA64hM92X0Urb6VpRhAFllDzz44U2mcD5scmT3zBc4VgVpkugF580+DQn8eAFSyoQt0tznA==", + "version": "7.5.13", + "resolved": "https://registry.npmjs.org/tar/-/tar-7.5.13.tgz", + "integrity": "sha512-tOG/7GyXpFevhXVh8jOPJrmtRpOTsYqUIkVdVooZYJS/z8WhfQUX8RJILmeuJNinGAMSu1veBr4asSHFt5/hng==", "requires": { - "chownr": "^2.0.0", - "fs-minipass": "^2.0.0", - "minipass": "^3.0.0", - "minizlib": "^2.1.1", - "mkdirp": "^1.0.3", - "yallist": "^4.0.0" + "@isaacs/fs-minipass": "^4.0.0", + "chownr": "^3.0.0", + "minipass": "^7.1.2", + "minizlib": "^3.1.0", + "yallist": "^5.0.0" }, "dependencies": { "chownr": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/chownr/-/chownr-2.0.0.tgz", - "integrity": "sha512-bIomtDF5KGpdogkLd9VspvFzk9KfpyyGlS8YFVZl7TGPBHL5snIOnxeshwVgPteQ9b4Eydl+pVbIyE1DcvCWgQ==" + "version": "3.0.0", + "resolved": "https://registry.npmjs.org/chownr/-/chownr-3.0.0.tgz", + "integrity": "sha512-+IxzY9BZOQd/XuYPRmrvEVjF/nqj5kgT4kEq7VofrDoM1MxoRjEWkrCC3EtLi59TVawxTAn+orJwFQcrqEN1+g==" } } }, "tar-fs": { - "version": "2.0.1", - "resolved": "https://registry.npmjs.org/tar-fs/-/tar-fs-2.0.1.tgz", - "integrity": "sha512-6tzWDMeroL87uF/+lin46k+Q+46rAJ0SyPGz7OW7wTgblI273hsBqk2C1j0/xNadNLKDTUL9BukSjB7cwgmlPA==", + "version": "2.1.4", + "resolved": "https://registry.npmjs.org/tar-fs/-/tar-fs-2.1.4.tgz", + "integrity": "sha512-mDAjwmZdh7LTT6pNleZ05Yt65HC3E+NiQzl672vQG38jIrehtJk/J3mNwIg+vShQPcLF/LV7CMnDW6vjj6sfYQ==", "requires": { "chownr": "^1.1.1", "mkdirp-classic": "^0.5.2", "pump": "^3.0.0", - "tar-stream": "^2.0.0" + "tar-stream": "^2.1.4" } }, "tar-stream": { @@ -1037,27 +1280,6 @@ "fs-constants": "^1.0.0", "inherits": "^2.0.3", "readable-stream": "^3.1.1" - }, - "dependencies": { - "bl": { - "version": "4.1.0", - "resolved": "https://registry.npmjs.org/bl/-/bl-4.1.0.tgz", - "integrity": "sha512-1W07cM9gS6DcLperZfFSj+bWLtaPGSOHWhPiGzXmvVJbRLdG82sH/Kn8EtW1VqWVA54AKf2h5k5BbnIbwF3h6w==", - "requires": { - "buffer": "^5.5.0", - "inherits": "^2.0.4", - "readable-stream": "^3.4.0" - } - }, - "buffer": { - "version": "5.7.1", - "resolved": "https://registry.npmjs.org/buffer/-/buffer-5.7.1.tgz", - "integrity": "sha512-EHcyIPBQ4BSGlvjB16k5KgAJ27CIsHY/2JBmCRReo48y9rQ3MaUzWX3KVlBa4U7MyX02HdVj0K7C3WaB3ju7FQ==", - "requires": { - "base64-js": "^1.3.1", - "ieee754": "^1.1.13" - } - } } }, "tweetnacl": { @@ -1065,6 +1287,11 @@ "resolved": "https://registry.npmjs.org/tweetnacl/-/tweetnacl-0.14.5.tgz", "integrity": "sha512-KXXFFdAbFXY4geFIwoyNK+f5Z1b7swfXABfL7HXCmoIWMKU3dmS26672A4EeQtDzLKy7SXmfBu51JolvEKwtGA==" }, + "undici-types": { + "version": "6.20.0", + "resolved": "https://registry.npmjs.org/undici-types/-/undici-types-6.20.0.tgz", + "integrity": "sha512-Ny6QZ2Nju20vw1SRHe3d9jVu6gJ+4e3+MMpqu7pqE5HT6WsTSlce++GQmK5UXS8mzV8DSYHrQH+Xrf2jVcuKNg==" + }, "universalify": { "version": "2.0.0", "resolved": "https://registry.npmjs.org/universalify/-/universalify-2.0.0.tgz", @@ -1075,6 +1302,21 @@ "resolved": "https://registry.npmjs.org/util-deprecate/-/util-deprecate-1.0.2.tgz", "integrity": "sha512-EPD5q1uXyFxJpCrLnCc1nHnq3gOa6DZBocAIiI2TaSCA7VCJ1UJDMagCzIkXNsUYfD1daK//LTEQ8xiIbrHtcw==" }, + "uuid": { + "version": "10.0.0", + "resolved": "https://registry.npmjs.org/uuid/-/uuid-10.0.0.tgz", + "integrity": "sha512-8XkAphELsDnEGrDxUOHB3RGvXz6TeuYSGEZBOjtTtPm2lwhGBjLgOzLHB63IUWfBpNucQjND6d3AOudO+H3RWQ==" + }, + "wrap-ansi": { + "version": "7.0.0", + "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-7.0.0.tgz", + "integrity": "sha512-YVGIj2kamLSTxw6NsZjoBxfSwsn0ycdesmc4p+Q21c5zPuZ1pl+NfxVdxPtdHvmNVOQ6XSYG4AUtyt/Fi7D16Q==", + "requires": { + "ansi-styles": "^4.0.0", + "string-width": "^4.1.0", + "strip-ansi": "^6.0.0" + } + }, "wrappy": { "version": "1.0.2", "resolved": "https://registry.npmjs.org/wrappy/-/wrappy-1.0.2.tgz", @@ -1086,52 +1328,22 @@ "integrity": "sha512-0pfFzegeDWJHJIAmTLRP2DwHjdF5s7jo9tuztdQxAhINCdvS+3nGINqPd00AphqJR/0LhANUS6/+7SCb98YOfA==" }, "yallist": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/yallist/-/yallist-4.0.0.tgz", - "integrity": "sha512-3wdGidZyq5PB084XLES5TpOSRA3wjXAlIWMhum2kRcv/41Sn2emQ0dycQW4uZXLejwKvg6EsvbdlVL+FYEct7A==" + "version": "5.0.0", + "resolved": "https://registry.npmjs.org/yallist/-/yallist-5.0.0.tgz", + "integrity": "sha512-YgvUTfwqyc7UXVMrB+SImsVYSmTS8X/tSrtdNZMImM+n7+QTriRXyXim0mBrTXNeqzVF0KWGgHPeiyViFFrNDw==" }, "yargs": { - "version": "17.5.1", - "resolved": "https://registry.npmjs.org/yargs/-/yargs-17.5.1.tgz", - "integrity": "sha512-t6YAJcxDkNX7NFYiVtKvWUz8l+PaKTLiL63mJYWR2GnHq2gjEWISzsLp9wg3aY36dY1j+gfIEL3pIF+XlJJfbA==", + "version": "17.7.2", + "resolved": "https://registry.npmjs.org/yargs/-/yargs-17.7.2.tgz", + "integrity": "sha512-7dSzzRQ++CKnNI/krKnYRV7JKKPUXMEh61soaHKg9mrWEhzFWhFnxPxGl+69cD1Ou63C13NUPCnmIcrvqCuM6w==", "requires": { - "cliui": "^7.0.2", + "cliui": "^8.0.1", "escalade": "^3.1.1", "get-caller-file": "^2.0.5", "require-directory": "^2.1.1", "string-width": "^4.2.3", "y18n": "^5.0.5", - "yargs-parser": "^21.0.0" - }, - "dependencies": { - "ansi-regex": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", - "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==" - }, - "emoji-regex": { - "version": "8.0.0", - "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", - "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" - }, - "string-width": { - "version": "4.2.3", - "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", - "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", - "requires": { - "emoji-regex": "^8.0.0", - "is-fullwidth-code-point": "^3.0.0", - "strip-ansi": "^6.0.1" - } - }, - "strip-ansi": { - "version": "6.0.1", - "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", - "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", - "requires": { - "ansi-regex": "^5.0.1" - } - } + "yargs-parser": "^21.1.1" } }, "yargs-parser": { diff --git a/dev/deploy-to-container/package.json b/dev/deploy-to-container/package.json index d3c14207e8..ccc78fc63b 100644 --- a/dev/deploy-to-container/package.json +++ b/dev/deploy-to-container/package.json @@ -2,13 +2,13 @@ "name": "deploy-to-container", "type": "module", "dependencies": { - "dockerode": "^3.3.3", - "fs-extra": "^10.1.0", - "nanoid": "4.0.0", - "nanoid-dictionary": "5.0.0-beta.1", - "slugify": "1.6.5", - "tar": "^6.1.11", - "yargs": "^17.5.1" + "dockerode": "^4.0.10", + "fs-extra": "^11.3.4", + "nanoid": "5.1.7", + "nanoid-dictionary": "5.0.0", + "slugify": "1.6.9", + "tar": "^7.5.13", + "yargs": "^17.7.2" }, "engines": { "node": ">=16" diff --git a/dev/deploy-to-container/refresh.js b/dev/deploy-to-container/refresh.js new file mode 100644 index 0000000000..7ea13c885a --- /dev/null +++ b/dev/deploy-to-container/refresh.js @@ -0,0 +1,84 @@ +#!/usr/bin/env node + +import Docker from 'dockerode' + +async function main () { + // Connect to Docker Engine API + console.info('Connecting to Docker Engine API...') + const dock = new Docker() + await dock.ping() + console.info('Connected to Docker Engine API.') + + // Pull latest DB image + console.info('Pulling latest DB docker image...') + const dbImagePullStream = await dock.pull('ghcr.io/ietf-tools/datatracker-db:latest') + await new Promise((resolve, reject) => { + dock.modem.followProgress(dbImagePullStream, (err, res) => err ? reject(err) : resolve(res)) + }) + console.info('Pulled latest DB docker image successfully.') + + // Terminate existing containers + console.info('Terminating DB containers and stopping app containers...') + const containers = await dock.listContainers({ all: true }) + const dbContainersToCreate = [] + const containersToRestart = [] + for (const container of containers) { + if ( + container.Names.some(n => n.startsWith('/dt-db-')) && + container.Labels?.nodbrefresh !== '1' + ) { + console.info(`Terminating DB container ${container.Id}...`) + dbContainersToCreate.push(container.Names.find(n => n.startsWith('/dt-db-')).substring(1)) + const oldContainer = dock.getContainer(container.Id) + if (container.State === 'running') { + await oldContainer.stop({ t: 5 }) + } + await oldContainer.remove({ + force: true, + v: true + }) + } else if ( + ( + container.Names.some(n => n.startsWith('/dt-app-')) || + container.Names.some(n => n.startsWith('/dt-celery-')) || + container.Names.some(n => n.startsWith('/dt-beat-')) + ) && container.Labels?.nodbrefresh !== '1' + ) { + if (container.State === 'running') { + const appContainer = dock.getContainer(container.Id) + containersToRestart.push(appContainer) + console.info(`Stopping app / celery container ${container.Id}...`) + await appContainer.stop({ t: 5 }) + } + } + } + console.info('DB containers have been terminated.') + + // Create DB containers + for (const dbContainerName of dbContainersToCreate) { + console.info(`Recreating DB docker container... [${dbContainerName}]`) + const dbContainer = await dock.createContainer({ + Image: 'ghcr.io/ietf-tools/datatracker-db:latest', + name: dbContainerName, + Hostname: dbContainerName, + HostConfig: { + NetworkMode: 'shared', + RestartPolicy: { + Name: 'unless-stopped' + } + } + }) + await dbContainer.start() + } + console.info('Recreated and started DB docker containers successfully.') + + console.info('Restarting app / celery containers...') + for (const appContainer of containersToRestart) { + await appContainer.start() + } + console.info('Done.') + + process.exit(0) +} + +main() diff --git a/dev/deploy-to-container/settings_local.py b/dev/deploy-to-container/settings_local.py index 5b67ce55f4..055b48d0f5 100644 --- a/dev/deploy-to-container/settings_local.py +++ b/dev/deploy-to-container/settings_local.py @@ -1,29 +1,21 @@ # Copyright The IETF Trust 2007-2019, All Rights Reserved # -*- coding: utf-8 -*- -from ietf.settings import * # pyflakes:ignore +from ietf.settings import * # pyflakes:ignore ALLOWED_HOSTS = ['*'] DATABASES = { 'default': { 'HOST': '__DBHOST__', - 'PORT': 3306, - 'NAME': 'ietf_utf8', - 'ENGINE': 'django.db.backends.mysql', + 'PORT': 5432, + 'NAME': 'datatracker', + 'ENGINE': 'django.db.backends.postgresql', 'USER': 'django', 'PASSWORD': 'RkTkDPFnKpko', - 'OPTIONS': { - 'sql_mode': 'STRICT_TRANS_TABLES', - 'init_command': 'SET storage_engine=InnoDB; SET names "utf8"', - }, }, } -DATABASE_TEST_OPTIONS = { - 'init_command': 'SET storage_engine=InnoDB', -} - SECRET_KEY = "__SECRETKEY__" CELERY_BROKER_URL = '__MQCONNSTR__' @@ -31,9 +23,6 @@ IDSUBMIT_IDNITS_BINARY = "/usr/local/bin/idnits" IDSUBMIT_REPOSITORY_PATH = "/test/id/" IDSUBMIT_STAGING_PATH = "/test/staging/" -INTERNET_DRAFT_ARCHIVE_DIR = "/test/archive/" -INTERNET_ALL_DRAFTS_ARCHIVE_DIR = "/test/archive/" -RFC_PATH = "/test/rfc/" AGENDA_PATH = '/assets/www6s/proceedings/' MEETINGHOST_LOGO_PATH = AGENDA_PATH @@ -51,7 +40,6 @@ SUBMIT_YANG_CATALOG_MODEL_DIR = '/assets/ietf-ftp/yang/catalogmod/' SUBMIT_YANG_DRAFT_MODEL_DIR = '/assets/ietf-ftp/yang/draftmod/' -SUBMIT_YANG_INVAL_MODEL_DIR = '/assets/ietf-ftp/yang/invalmod/' SUBMIT_YANG_IANA_MODEL_DIR = '/assets/ietf-ftp/yang/ianamod/' SUBMIT_YANG_RFC_MODEL_DIR = '/assets/ietf-ftp/yang/rfcmod/' @@ -64,17 +52,30 @@ # 'ietf.context_processors.sql_debug', # ] -DOCUMENT_PATH_PATTERN = '/assets/ietf-ftp/{doc.type_id}/' +DOCUMENT_PATH_PATTERN = '/assets/ietfdata/doc/{doc.type_id}/' INTERNET_DRAFT_PATH = '/assets/ietf-ftp/internet-drafts/' RFC_PATH = '/assets/ietf-ftp/rfc/' CHARTER_PATH = '/assets/ietf-ftp/charter/' BOFREQ_PATH = '/assets/ietf-ftp/bofreq/' CONFLICT_REVIEW_PATH = '/assets/ietf-ftp/conflict-reviews/' STATUS_CHANGE_PATH = '/assets/ietf-ftp/status-changes/' -INTERNET_DRAFT_ARCHIVE_DIR = '/assets/ietf-ftp/internet-drafts/' -INTERNET_ALL_DRAFTS_ARCHIVE_DIR = '/assets/ietf-ftp/internet-drafts/' +INTERNET_DRAFT_ARCHIVE_DIR = '/assets/collection/draft-archive' +INTERNET_ALL_DRAFTS_ARCHIVE_DIR = '/assets/archive/id' +BIBXML_BASE_PATH = '/assets/ietfdata/derived/bibxml' +IDSUBMIT_REPOSITORY_PATH = INTERNET_DRAFT_PATH +FTP_DIR = '/assets/ftp' +NFS_METRICS_TMP_DIR = '/assets/tmp' NOMCOM_PUBLIC_KEYS_DIR = 'data/nomcom_keys/public_keys/' SLIDE_STAGING_PATH = '/test/staging/' DE_GFM_BINARY = '/usr/local/bin/de-gfm' + +APP_API_TOKENS = { + "ietf.api.red_api" : ["devtoken", "redtoken"], # Not a real secret + "ietf.api.views.ingest_email_test": ["ingestion-test-token"], # Not a real secret + "ietf.api.views_rpc" : ["devtoken"], # Not a real secret +} + +# OIDC configuration +SITE_URL = 'https://__HOSTNAME__' diff --git a/dev/deploy-to-container/start.sh b/dev/deploy-to-container/start.sh index 3091de0d2d..5d976f80ea 100644 --- a/dev/deploy-to-container/start.sh +++ b/dev/deploy-to-container/start.sh @@ -21,11 +21,32 @@ pip --disable-pip-version-check --no-cache-dir install -r requirements.txt echo "Creating data directories..." chmod +x ./app-create-dirs.sh ./app-create-dirs.sh + +if [ -n "$PGHOST" ]; then + echo "Altering PG search path..." + psql -U django -h $PGHOST -d datatracker -v ON_ERROR_STOP=1 -c '\x' -c 'ALTER USER django set search_path=datatracker,public;' +fi + +echo "Starting memcached..." +/usr/bin/memcached -d -u root + echo "Running Datatracker checks..." ./ietf/manage.py check # Migrate, adjusting to what the current state of the underlying database might be: +# On production, the blobdb tables are in a separate database. Manipulate migration +# history to ensure that they're created for the sandbox environment that runs it +# all from a single database. +echo "Ensuring blobdb relations exist..." +/usr/local/bin/python ./ietf/manage.py migrate --settings=settings_local --fake blobdb zero +if ! /usr/local/bin/python ./ietf/manage.py migrate --settings=settings_local blobdb; then + # If we are restarting a sandbox, the migration may already have run and re-running + # it will fail. Assume that happened and fake it. + /usr/local/bin/python ./ietf/manage.py migrate --settings=settings_local --fake blobdb +fi + +# Now run the migrations for real echo "Running Datatracker migrations..." /usr/local/bin/python ./ietf/manage.py migrate --settings=settings_local diff --git a/dev/deploy/build.sh b/dev/deploy/build.sh deleted file mode 100644 index a802acb46b..0000000000 --- a/dev/deploy/build.sh +++ /dev/null @@ -1,11 +0,0 @@ -#!/bin/bash - -echo "Compiling native node packages..." -yarn rebuild -echo "Packaging static assets..." -if [ "${SHOULD_DEPLOY}" = "true" ]; then - yarn build --base=https://www.ietf.org/lib/dt/$PKG_VERSION/ -else - yarn build -fi -yarn legacy:build diff --git a/dev/diff/package-lock.json b/dev/diff/package-lock.json index 96ce2b5481..d1c2fbd763 100644 --- a/dev/diff/package-lock.json +++ b/dev/diff/package-lock.json @@ -6,24 +6,243 @@ "": { "name": "diff", "dependencies": { - "chalk": "^5.0.1", - "dockerode": "^3.3.3", - "enquirer": "^2.3.6", + "chalk": "^5.4.1", + "dockerode": "^4.0.6", + "enquirer": "^2.4.1", "extract-zip": "^2.0.1", - "fs-extra": "^10.1.0", - "got": "^12.3.1", + "fs-extra": "^11.3.0", + "got": "^13.0.0", "keypress": "^0.2.1", - "listr2": "^5.0.2", + "listr2": "^6.6.1", "lodash-es": "^4.17.21", - "luxon": "^3.0.1", - "pretty-bytes": "^6.0.0", - "tar": "^6.1.11", - "yargs": "^17.5.1" + "luxon": "^3.6.1", + "pretty-bytes": "^6.1.1", + "tar": "^7.4.3", + "yargs": "^17.7.2" }, "engines": { "node": ">=16" } }, + "node_modules/@balena/dockerignore": { + "version": "1.0.2", + "resolved": "https://registry.npmjs.org/@balena/dockerignore/-/dockerignore-1.0.2.tgz", + "integrity": "sha512-wMue2Sy4GAVTk6Ic4tJVcnfdau+gx2EnG7S+uAEe+TWJFqE4YoWN4/H8MSLj4eYJKxGg26lZwboEniNiNwZQ6Q==" + }, + "node_modules/@grpc/grpc-js": { + "version": "1.12.5", + "resolved": "https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-1.12.5.tgz", + "integrity": "sha512-d3iiHxdpg5+ZcJ6jnDSOT8Z0O0VMVGy34jAnYLUX8yd36b1qn8f1TwOA/Lc7TsOh03IkPJ38eGI5qD2EjNkoEA==", + "license": "Apache-2.0", + "dependencies": { + "@grpc/proto-loader": "^0.7.13", + "@js-sdsl/ordered-map": "^4.4.2" + }, + "engines": { + "node": ">=12.10.0" + } + }, + "node_modules/@grpc/proto-loader": { + "version": "0.7.13", + "resolved": "https://registry.npmjs.org/@grpc/proto-loader/-/proto-loader-0.7.13.tgz", + "integrity": "sha512-AiXO/bfe9bmxBjxxtYxFAXGZvMaN5s8kO+jBHAJCON8rJoB5YS/D6X7ZNc6XQkuHNmyl4CYaMI1fJ/Gn27RGGw==", + "license": "Apache-2.0", + "dependencies": { + "lodash.camelcase": "^4.3.0", + "long": "^5.0.0", + "protobufjs": "^7.2.5", + "yargs": "^17.7.2" + }, + "bin": { + "proto-loader-gen-types": "build/bin/proto-loader-gen-types.js" + }, + "engines": { + "node": ">=6" + } + }, + "node_modules/@isaacs/cliui": { + "version": "8.0.2", + "resolved": "https://registry.npmjs.org/@isaacs/cliui/-/cliui-8.0.2.tgz", + "integrity": "sha512-O8jcjabXaleOG9DQ0+ARXWZBTfnP4WNAqzuiJK7ll44AmxGKv/J2M4TPjxjY3znBCfvBXFzucm1twdyFybFqEA==", + "dependencies": { + "string-width": "^5.1.2", + "string-width-cjs": "npm:string-width@^4.2.0", + "strip-ansi": "^7.0.1", + "strip-ansi-cjs": "npm:strip-ansi@^6.0.1", + "wrap-ansi": "^8.1.0", + "wrap-ansi-cjs": "npm:wrap-ansi@^7.0.0" + }, + "engines": { + "node": ">=12" + } + }, + "node_modules/@isaacs/cliui/node_modules/ansi-regex": { + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-6.0.1.tgz", + "integrity": "sha512-n5M855fKb2SsfMIiFFoVrABHJC8QtHwVx+mHWP3QcEqBHYienj5dHSgjbxtC0WEZXYt4wcD6zrQElDPhFuZgfA==", + "engines": { + "node": ">=12" + }, + "funding": { + "url": "https://github.com/chalk/ansi-regex?sponsor=1" + } + }, + "node_modules/@isaacs/cliui/node_modules/ansi-styles": { + "version": "6.2.1", + "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-6.2.1.tgz", + "integrity": "sha512-bN798gFfQX+viw3R7yrGWRqnrN2oRkEkUjjl4JNn4E8GxxbjtG3FbrEIIY3l8/hrwUwIeCZvi4QuOTP4MErVug==", + "engines": { + "node": ">=12" + }, + "funding": { + "url": "https://github.com/chalk/ansi-styles?sponsor=1" + } + }, + "node_modules/@isaacs/cliui/node_modules/emoji-regex": { + "version": "9.2.2", + "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-9.2.2.tgz", + "integrity": "sha512-L18DaJsXSUk2+42pv8mLs5jJT2hqFkFE4j21wOmgbUqsZ2hL72NsUU785g9RXgo3s0ZNgVl42TiHp3ZtOv/Vyg==" + }, + "node_modules/@isaacs/cliui/node_modules/string-width": { + "version": "5.1.2", + "resolved": "https://registry.npmjs.org/string-width/-/string-width-5.1.2.tgz", + "integrity": "sha512-HnLOCR3vjcY8beoNLtcjZ5/nxn2afmME6lhrDrebokqMap+XbeW8n9TXpPDOqdGK5qcI3oT0GKTW6wC7EMiVqA==", + "dependencies": { + "eastasianwidth": "^0.2.0", + "emoji-regex": "^9.2.2", + "strip-ansi": "^7.0.1" + }, + "engines": { + "node": ">=12" + }, + "funding": { + "url": "https://github.com/sponsors/sindresorhus" + } + }, + "node_modules/@isaacs/cliui/node_modules/strip-ansi": { + "version": "7.1.0", + "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-7.1.0.tgz", + "integrity": "sha512-iq6eVVI64nQQTRYq2KtEg2d2uU7LElhTJwsH4YzIHZshxlgZms/wIc4VoDQTlG/IvVIrBKG06CrZnp0qv7hkcQ==", + "dependencies": { + "ansi-regex": "^6.0.1" + }, + "engines": { + "node": ">=12" + }, + "funding": { + "url": "https://github.com/chalk/strip-ansi?sponsor=1" + } + }, + "node_modules/@isaacs/cliui/node_modules/wrap-ansi": { + "version": "8.1.0", + "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-8.1.0.tgz", + "integrity": "sha512-si7QWI6zUMq56bESFvagtmzMdGOtoxfR+Sez11Mobfc7tm+VkUckk9bW2UeffTGVUbOksxmSw0AA2gs8g71NCQ==", + "dependencies": { + "ansi-styles": "^6.1.0", + "string-width": "^5.0.1", + "strip-ansi": "^7.0.1" + }, + "engines": { + "node": ">=12" + }, + "funding": { + "url": "https://github.com/chalk/wrap-ansi?sponsor=1" + } + }, + "node_modules/@isaacs/fs-minipass": { + "version": "4.0.1", + "resolved": "https://registry.npmjs.org/@isaacs/fs-minipass/-/fs-minipass-4.0.1.tgz", + "integrity": "sha512-wgm9Ehl2jpeqP3zw/7mo3kRHFp5MEDhqAdwy1fTGkHAwnkGOVsgpvQhL8B5n1qlb01jV3n/bI0ZfZp5lWA1k4w==", + "dependencies": { + "minipass": "^7.0.4" + }, + "engines": { + "node": ">=18.0.0" + } + }, + "node_modules/@js-sdsl/ordered-map": { + "version": "4.4.2", + "resolved": "https://registry.npmjs.org/@js-sdsl/ordered-map/-/ordered-map-4.4.2.tgz", + "integrity": "sha512-iUKgm52T8HOE/makSxjqoWhe95ZJA1/G1sYsGev2JDKUSS14KAgg1LHb+Ba+IPow0xflbnSkOsZcO08C7w1gYw==", + "license": "MIT", + "funding": { + "type": "opencollective", + "url": "https://opencollective.com/js-sdsl" + } + }, + "node_modules/@pkgjs/parseargs": { + "version": "0.11.0", + "resolved": "https://registry.npmjs.org/@pkgjs/parseargs/-/parseargs-0.11.0.tgz", + "integrity": "sha512-+1VkjdD0QBLPodGrJUeqarH8VAIvQODIbwh9XpP5Syisf7YoQgsJKPNFoqqLQlu+VQ/tVSshMR6loPMn8U+dPg==", + "optional": true, + "engines": { + "node": ">=14" + } + }, + "node_modules/@protobufjs/aspromise": { + "version": "1.1.2", + "resolved": "https://registry.npmjs.org/@protobufjs/aspromise/-/aspromise-1.1.2.tgz", + "integrity": "sha512-j+gKExEuLmKwvz3OgROXtrJ2UG2x8Ch2YZUxahh+s1F2HZ+wAceUNLkvy6zKCPVRkU++ZWQrdxsUeQXmcg4uoQ==", + "license": "BSD-3-Clause" + }, + "node_modules/@protobufjs/base64": { + "version": "1.1.2", + "resolved": "https://registry.npmjs.org/@protobufjs/base64/-/base64-1.1.2.tgz", + "integrity": "sha512-AZkcAA5vnN/v4PDqKyMR5lx7hZttPDgClv83E//FMNhR2TMcLUhfRUBHCmSl0oi9zMgDDqRUJkSxO3wm85+XLg==", + "license": "BSD-3-Clause" + }, + "node_modules/@protobufjs/codegen": { + "version": "2.0.4", + "resolved": "https://registry.npmjs.org/@protobufjs/codegen/-/codegen-2.0.4.tgz", + "integrity": "sha512-YyFaikqM5sH0ziFZCN3xDC7zeGaB/d0IUb9CATugHWbd1FRFwWwt4ld4OYMPWu5a3Xe01mGAULCdqhMlPl29Jg==", + "license": "BSD-3-Clause" + }, + "node_modules/@protobufjs/eventemitter": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/eventemitter/-/eventemitter-1.1.0.tgz", + "integrity": "sha512-j9ednRT81vYJ9OfVuXG6ERSTdEL1xVsNgqpkxMsbIabzSo3goCjDIveeGv5d03om39ML71RdmrGNjG5SReBP/Q==", + "license": "BSD-3-Clause" + }, + "node_modules/@protobufjs/fetch": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/fetch/-/fetch-1.1.0.tgz", + "integrity": "sha512-lljVXpqXebpsijW71PZaCYeIcE5on1w5DlQy5WH6GLbFryLUrBD4932W/E2BSpfRJWseIL4v/KPgBFxDOIdKpQ==", + "license": "BSD-3-Clause", + "dependencies": { + "@protobufjs/aspromise": "^1.1.1", + "@protobufjs/inquire": "^1.1.0" + } + }, + "node_modules/@protobufjs/float": { + "version": "1.0.2", + "resolved": "https://registry.npmjs.org/@protobufjs/float/-/float-1.0.2.tgz", + "integrity": "sha512-Ddb+kVXlXst9d+R9PfTIxh1EdNkgoRe5tOX6t01f1lYWOvJnSPDBlG241QLzcyPdoNTsblLUdujGSE4RzrTZGQ==", + "license": "BSD-3-Clause" + }, + "node_modules/@protobufjs/inquire": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/inquire/-/inquire-1.1.0.tgz", + "integrity": "sha512-kdSefcPdruJiFMVSbn801t4vFK7KB/5gd2fYvrxhuJYg8ILrmn9SKSX2tZdV6V+ksulWqS7aXjBcRXl3wHoD9Q==", + "license": "BSD-3-Clause" + }, + "node_modules/@protobufjs/path": { + "version": "1.1.2", + "resolved": "https://registry.npmjs.org/@protobufjs/path/-/path-1.1.2.tgz", + "integrity": "sha512-6JOcJ5Tm08dOHAbdR3GrvP+yUUfkjG5ePsHYczMFLq3ZmMkAD98cDgcT2iA1lJ9NVwFd4tH/iSSoe44YWkltEA==", + "license": "BSD-3-Clause" + }, + "node_modules/@protobufjs/pool": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/pool/-/pool-1.1.0.tgz", + "integrity": "sha512-0kELaGSIDBKvcgS4zkjz1PeddatrjYcmMWOlAuAPwAeccUrPHdUqo/J6LiymHHEiJT5NrF1UVwxY14f+fy4WQw==", + "license": "BSD-3-Clause" + }, + "node_modules/@protobufjs/utf8": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/utf8/-/utf8-1.1.0.tgz", + "integrity": "sha512-Vvn3zZrhQZkkBE8LSuW3em98c0FwgO4nxzv6OdSxPKJIEKY2bGbHn+mhGIPerzI4twdxaP8/0+06HBpwf345Lw==", + "license": "BSD-3-Clause" + }, "node_modules/@sindresorhus/is": { "version": "5.3.0", "resolved": "https://registry.npmjs.org/@sindresorhus/is/-/is-5.3.0.tgz", @@ -46,48 +265,16 @@ "node": ">=14.16" } }, - "node_modules/@types/cacheable-request": { - "version": "6.0.2", - "resolved": "https://registry.npmjs.org/@types/cacheable-request/-/cacheable-request-6.0.2.tgz", - "integrity": "sha512-B3xVo+dlKM6nnKTcmm5ZtY/OL8bOAOd2Olee9M1zft65ox50OzjEHW91sDiU9j6cvW8Ejg1/Qkf4xd2kugApUA==", - "dependencies": { - "@types/http-cache-semantics": "*", - "@types/keyv": "*", - "@types/node": "*", - "@types/responselike": "*" - } - }, "node_modules/@types/http-cache-semantics": { "version": "4.0.1", "resolved": "https://registry.npmjs.org/@types/http-cache-semantics/-/http-cache-semantics-4.0.1.tgz", "integrity": "sha512-SZs7ekbP8CN0txVG2xVRH6EgKmEm31BOxA07vkFaETzZz1xh+cbt8BcI0slpymvwhx5dlFnQG2rTlPVQn+iRPQ==" }, - "node_modules/@types/json-buffer": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/@types/json-buffer/-/json-buffer-3.0.0.tgz", - "integrity": "sha512-3YP80IxxFJB4b5tYC2SUPwkg0XQLiu0nWvhRgEatgjf+29IcWO9X1k8xRv5DGssJ/lCrjYTjQPcobJr2yWIVuQ==" - }, - "node_modules/@types/keyv": { - "version": "3.1.4", - "resolved": "https://registry.npmjs.org/@types/keyv/-/keyv-3.1.4.tgz", - "integrity": "sha512-BQ5aZNSCpj7D6K2ksrRCTmKRLEpnPvWDiLPfoGyhZ++8YtiK9d/3DBKPJgry359X/P1PfruyYwvnvwFjuEiEIg==", - "dependencies": { - "@types/node": "*" - } - }, "node_modules/@types/node": { "version": "18.6.5", "resolved": "https://registry.npmjs.org/@types/node/-/node-18.6.5.tgz", "integrity": "sha512-Xjt5ZGUa5WusGZJ4WJPbOT8QOqp6nDynVFRKcUt32bOgvXEoc6o085WNkYTMO7ifAj2isEfQQ2cseE+wT6jsRw==" }, - "node_modules/@types/responselike": { - "version": "1.0.0", - "resolved": "https://registry.npmjs.org/@types/responselike/-/responselike-1.0.0.tgz", - "integrity": "sha512-85Y2BjiufFzaMIlvJDvTTB8Fxl2xfLo4HgmHzVBz08w4wDePCTjYw66PdrolO0kzli3yam/YCgRufyo1DdQVTA==", - "dependencies": { - "@types/node": "*" - } - }, "node_modules/@types/yauzl": { "version": "2.10.0", "resolved": "https://registry.npmjs.org/@types/yauzl/-/yauzl-2.10.0.tgz", @@ -97,41 +284,63 @@ "@types/node": "*" } }, - "node_modules/aggregate-error": { - "version": "3.1.0", - "resolved": "https://registry.npmjs.org/aggregate-error/-/aggregate-error-3.1.0.tgz", - "integrity": "sha512-4I7Td01quW/RpocfNayFdFVk1qSuoh0E7JrbRJ16nH01HhKFQ88INq9Sd+nd72zqRySlr9BmDA8xlEJ6vJMrYA==", + "node_modules/ansi-colors": { + "version": "4.1.3", + "resolved": "https://registry.npmjs.org/ansi-colors/-/ansi-colors-4.1.3.tgz", + "integrity": "sha512-/6w/C21Pm1A7aZitlI5Ni/2J6FFQN8i1Cvz3kHABAAbw93v/NlvKdVOqz7CCWz/3iv/JplRSEEZ83XION15ovw==", + "engines": { + "node": ">=6" + } + }, + "node_modules/ansi-escapes": { + "version": "5.0.0", + "resolved": "https://registry.npmjs.org/ansi-escapes/-/ansi-escapes-5.0.0.tgz", + "integrity": "sha512-5GFMVX8HqE/TB+FuBJGuO5XG0WrsA6ptUqoODaT/n9mmUaZFkqnBueB4leqGBCmrUHnCnC4PCZTCd0E7QQ83bA==", "dependencies": { - "clean-stack": "^2.0.0", - "indent-string": "^4.0.0" + "type-fest": "^1.0.2" + }, + "engines": { + "node": ">=12" }, + "funding": { + "url": "https://github.com/sponsors/sindresorhus" + } + }, + "node_modules/ansi-regex": { + "version": "5.0.1", + "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", + "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==", "engines": { "node": ">=8" } }, - "node_modules/ansi-colors": { - "version": "4.1.3", - "resolved": "https://registry.npmjs.org/ansi-colors/-/ansi-colors-4.1.3.tgz", - "integrity": "sha512-/6w/C21Pm1A7aZitlI5Ni/2J6FFQN8i1Cvz3kHABAAbw93v/NlvKdVOqz7CCWz/3iv/JplRSEEZ83XION15ovw==", + "node_modules/ansi-styles": { + "version": "4.3.0", + "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz", + "integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==", + "dependencies": { + "color-convert": "^2.0.1" + }, "engines": { - "node": ">=6" + "node": ">=8" + }, + "funding": { + "url": "https://github.com/chalk/ansi-styles?sponsor=1" } }, "node_modules/asn1": { "version": "0.2.6", "resolved": "https://registry.npmjs.org/asn1/-/asn1-0.2.6.tgz", "integrity": "sha512-ix/FxPn0MDjeyJ7i/yoHGFt/EX6LyNbxSEhPPXODPL+KB0VPk86UYfL0lMdy+KCnv+fmvIzySwaK5COwqVbWTQ==", + "license": "MIT", "dependencies": { "safer-buffer": "~2.1.0" } }, - "node_modules/astral-regex": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/astral-regex/-/astral-regex-2.0.0.tgz", - "integrity": "sha512-Z7tMw1ytTXt5jqMcOP+OQteU1VuNK9Y02uuJtKQ1Sv69jXQKKg5cibLwGJow8yzZP+eAc18EmLGPal0bp36rvQ==", - "engines": { - "node": ">=8" - } + "node_modules/balanced-match": { + "version": "1.0.2", + "resolved": "https://registry.npmjs.org/balanced-match/-/balanced-match-1.0.2.tgz", + "integrity": "sha512-3oSeUO0TMV67hN1AmbXsK4yaqU7tjiHlbxRDZOpH0KW9+CeX4bRAaX0Anxt0tx2MrpRpWwQaPwIlISEJhYU5Pw==" }, "node_modules/base64-js": { "version": "1.5.1", @@ -150,16 +359,61 @@ "type": "consulting", "url": "https://feross.org/support" } - ] + ], + "license": "MIT" }, "node_modules/bcrypt-pbkdf": { "version": "1.0.2", "resolved": "https://registry.npmjs.org/bcrypt-pbkdf/-/bcrypt-pbkdf-1.0.2.tgz", "integrity": "sha512-qeFIXtP4MSoi6NLqO12WfqARWWuCKi2Rn/9hJLEmtB5yTNr9DqFWkJRCf2qShWzPeAMRnOgCrq0sg/KLv5ES9w==", + "license": "BSD-3-Clause", "dependencies": { "tweetnacl": "^0.14.3" } }, + "node_modules/bl": { + "version": "4.1.0", + "resolved": "https://registry.npmjs.org/bl/-/bl-4.1.0.tgz", + "integrity": "sha512-1W07cM9gS6DcLperZfFSj+bWLtaPGSOHWhPiGzXmvVJbRLdG82sH/Kn8EtW1VqWVA54AKf2h5k5BbnIbwF3h6w==", + "license": "MIT", + "dependencies": { + "buffer": "^5.5.0", + "inherits": "^2.0.4", + "readable-stream": "^3.4.0" + } + }, + "node_modules/brace-expansion": { + "version": "2.0.1", + "resolved": "https://registry.npmjs.org/brace-expansion/-/brace-expansion-2.0.1.tgz", + "integrity": "sha512-XnAIvQ8eM+kC6aULx6wuQiwVsnzsi9d3WxzV3FpWTGA19F621kwdbsAcFKXgKUHZWsy+mY6iL1sHTxWEFCytDA==", + "dependencies": { + "balanced-match": "^1.0.0" + } + }, + "node_modules/buffer": { + "version": "5.7.1", + "resolved": "https://registry.npmjs.org/buffer/-/buffer-5.7.1.tgz", + "integrity": "sha512-EHcyIPBQ4BSGlvjB16k5KgAJ27CIsHY/2JBmCRReo48y9rQ3MaUzWX3KVlBa4U7MyX02HdVj0K7C3WaB3ju7FQ==", + "funding": [ + { + "type": "github", + "url": "https://github.com/sponsors/feross" + }, + { + "type": "patreon", + "url": "https://www.patreon.com/feross" + }, + { + "type": "consulting", + "url": "https://feross.org/support" + } + ], + "license": "MIT", + "dependencies": { + "base64-js": "^1.3.1", + "ieee754": "^1.1.13" + } + }, "node_modules/buffer-crc32": { "version": "0.2.13", "resolved": "https://registry.npmjs.org/buffer-crc32/-/buffer-crc32-0.2.13.tgz", @@ -169,65 +423,44 @@ } }, "node_modules/buildcheck": { - "version": "0.0.3", - "resolved": "https://registry.npmjs.org/buildcheck/-/buildcheck-0.0.3.tgz", - "integrity": "sha512-pziaA+p/wdVImfcbsZLNF32EiWyujlQLwolMqUQE8xpKNOH7KmZQaY8sXN7DGOEzPAElo9QTaeNRfGnf3iOJbA==", + "version": "0.0.6", + "resolved": "https://registry.npmjs.org/buildcheck/-/buildcheck-0.0.6.tgz", + "integrity": "sha512-8f9ZJCUXyT1M35Jx7MkBgmBMo3oHTTBIPLiY9xyL0pl3T5RwcPEY8cUHr5LBNfu/fk6c2T4DJZuVM/8ZZT2D2A==", "optional": true, "engines": { "node": ">=10.0.0" } }, "node_modules/cacheable-lookup": { - "version": "6.1.0", - "resolved": "https://registry.npmjs.org/cacheable-lookup/-/cacheable-lookup-6.1.0.tgz", - "integrity": "sha512-KJ/Dmo1lDDhmW2XDPMo+9oiy/CeqosPguPCrgcVzKyZrL6pM1gU2GmPY/xo6OQPTUaA/c0kwHuywB4E6nmT9ww==", + "version": "7.0.0", + "resolved": "https://registry.npmjs.org/cacheable-lookup/-/cacheable-lookup-7.0.0.tgz", + "integrity": "sha512-+qJyx4xiKra8mZrcwhjMRMUhD5NR1R8esPkzIYxX96JiecFoxAXFuz/GpR3+ev4PE1WamHip78wV0vcmPQtp8w==", "engines": { - "node": ">=10.6.0" + "node": ">=14.16" } }, "node_modules/cacheable-request": { - "version": "7.0.2", - "resolved": "https://registry.npmjs.org/cacheable-request/-/cacheable-request-7.0.2.tgz", - "integrity": "sha512-pouW8/FmiPQbuGpkXQ9BAPv/Mo5xDGANgSNXzTzJ8DrKGuXOssM4wIQRjfanNRh3Yu5cfYPvcorqbhg2KIJtew==", - "dependencies": { - "clone-response": "^1.0.2", - "get-stream": "^5.1.0", - "http-cache-semantics": "^4.0.0", - "keyv": "^4.0.0", - "lowercase-keys": "^2.0.0", - "normalize-url": "^6.0.1", - "responselike": "^2.0.0" - }, - "engines": { - "node": ">=8" - } - }, - "node_modules/cacheable-request/node_modules/get-stream": { - "version": "5.2.0", - "resolved": "https://registry.npmjs.org/get-stream/-/get-stream-5.2.0.tgz", - "integrity": "sha512-nBF+F1rAZVCu/p7rjzgA+Yb4lfYXrpl7a6VmJrU8wF9I1CKvP/QwPNZHnOlwbTkY6dvtFIzFMSyQXbLoTQPRpA==", + "version": "10.2.8", + "resolved": "https://registry.npmjs.org/cacheable-request/-/cacheable-request-10.2.8.tgz", + "integrity": "sha512-IDVO5MJ4LItE6HKFQTqT2ocAQsisOoCTUDu1ddCmnhyiwFQjXNPp4081Xj23N4tO+AFEFNzGuNEf/c8Gwwt15A==", "dependencies": { - "pump": "^3.0.0" - }, - "engines": { - "node": ">=8" + "@types/http-cache-semantics": "^4.0.1", + "get-stream": "^6.0.1", + "http-cache-semantics": "^4.1.1", + "keyv": "^4.5.2", + "mimic-response": "^4.0.0", + "normalize-url": "^8.0.0", + "responselike": "^3.0.0" }, - "funding": { - "url": "https://github.com/sponsors/sindresorhus" - } - }, - "node_modules/cacheable-request/node_modules/lowercase-keys": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/lowercase-keys/-/lowercase-keys-2.0.0.tgz", - "integrity": "sha512-tqNXrS78oMOE73NMxK4EMLQsQowWf8jKooH9g7xPavRT706R6bkQJ6DY2Te7QukaZsulxa30wQ7bk0pm4XiHmA==", "engines": { - "node": ">=8" + "node": ">=14.16" } }, "node_modules/chalk": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/chalk/-/chalk-5.0.1.tgz", - "integrity": "sha512-Fo07WOYGqMfCWHOzSXOt2CxDbC6skS/jO9ynEcmpANMoPrD+W1r1K6Vx7iNm+AQmETU1Xr2t+n8nzkV9t6xh3w==", + "version": "5.4.1", + "resolved": "https://registry.npmjs.org/chalk/-/chalk-5.4.1.tgz", + "integrity": "sha512-zgVZuo2WcZgfUEmsn6eO3kINexW8RAE4maiQ8QNs8CtpPCSyMiYsULR3HQYkm3w8FIA3SberyMJMSldGsW+U3w==", + "license": "MIT", "engines": { "node": "^12.17.0 || ^14.13 || >=16.0.0" }, @@ -238,154 +471,95 @@ "node_modules/chownr": { "version": "1.1.4", "resolved": "https://registry.npmjs.org/chownr/-/chownr-1.1.4.tgz", - "integrity": "sha512-jJ0bqzaylmJtVnNgzTeSOs8DPavpbYgEr/b0YL8/2GO3xJEhInFmhKMUnEJQjZumK7KXGFhUy89PrsJWlakBVg==" + "integrity": "sha512-jJ0bqzaylmJtVnNgzTeSOs8DPavpbYgEr/b0YL8/2GO3xJEhInFmhKMUnEJQjZumK7KXGFhUy89PrsJWlakBVg==", + "license": "ISC" }, - "node_modules/clean-stack": { - "version": "2.2.0", - "resolved": "https://registry.npmjs.org/clean-stack/-/clean-stack-2.2.0.tgz", - "integrity": "sha512-4diC9HaTE+KRAMWhDhrGOECgWZxoevMc5TlkObMqNSsVU62PYzXZ/SMTjzyGAFF1YusgxGcSWTEXBhp0CPwQ1A==", + "node_modules/cli-cursor": { + "version": "4.0.0", + "resolved": "https://registry.npmjs.org/cli-cursor/-/cli-cursor-4.0.0.tgz", + "integrity": "sha512-VGtlMu3x/4DOtIUwEkRezxUZ2lBacNJCHash0N0WeZDBS+7Ux1dm3XWAgWYxLJFMMdOeXMHXorshEFhbMSGelg==", + "dependencies": { + "restore-cursor": "^4.0.0" + }, "engines": { - "node": ">=6" + "node": "^12.20.0 || ^14.13.1 || >=16.0.0" + }, + "funding": { + "url": "https://github.com/sponsors/sindresorhus" } }, "node_modules/cli-truncate": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/cli-truncate/-/cli-truncate-2.1.0.tgz", - "integrity": "sha512-n8fOixwDD6b/ObinzTrp1ZKFzbgvKZvuz/TvejnLn1aQfC6r52XEx85FmuC+3HI+JM7coBRXUvNqEU2PHVrHpg==", + "version": "3.1.0", + "resolved": "https://registry.npmjs.org/cli-truncate/-/cli-truncate-3.1.0.tgz", + "integrity": "sha512-wfOBkjXteqSnI59oPcJkcPl/ZmwvMMOj340qUIY1SKZCv0B9Cf4D4fAucRkIKQmsIuYK3x1rrgU7MeGRruiuiA==", "dependencies": { - "slice-ansi": "^3.0.0", - "string-width": "^4.2.0" + "slice-ansi": "^5.0.0", + "string-width": "^5.0.0" }, "engines": { - "node": ">=8" + "node": "^12.20.0 || ^14.13.1 || >=16.0.0" }, "funding": { "url": "https://github.com/sponsors/sindresorhus" } }, "node_modules/cli-truncate/node_modules/ansi-regex": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", - "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==", + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-6.0.1.tgz", + "integrity": "sha512-n5M855fKb2SsfMIiFFoVrABHJC8QtHwVx+mHWP3QcEqBHYienj5dHSgjbxtC0WEZXYt4wcD6zrQElDPhFuZgfA==", "engines": { - "node": ">=8" + "node": ">=12" + }, + "funding": { + "url": "https://github.com/chalk/ansi-regex?sponsor=1" } }, "node_modules/cli-truncate/node_modules/emoji-regex": { - "version": "8.0.0", - "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", - "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" + "version": "9.2.2", + "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-9.2.2.tgz", + "integrity": "sha512-L18DaJsXSUk2+42pv8mLs5jJT2hqFkFE4j21wOmgbUqsZ2hL72NsUU785g9RXgo3s0ZNgVl42TiHp3ZtOv/Vyg==" }, "node_modules/cli-truncate/node_modules/string-width": { - "version": "4.2.3", - "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", - "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", + "version": "5.1.2", + "resolved": "https://registry.npmjs.org/string-width/-/string-width-5.1.2.tgz", + "integrity": "sha512-HnLOCR3vjcY8beoNLtcjZ5/nxn2afmME6lhrDrebokqMap+XbeW8n9TXpPDOqdGK5qcI3oT0GKTW6wC7EMiVqA==", "dependencies": { - "emoji-regex": "^8.0.0", - "is-fullwidth-code-point": "^3.0.0", - "strip-ansi": "^6.0.1" + "eastasianwidth": "^0.2.0", + "emoji-regex": "^9.2.2", + "strip-ansi": "^7.0.1" }, "engines": { - "node": ">=8" + "node": ">=12" + }, + "funding": { + "url": "https://github.com/sponsors/sindresorhus" } }, "node_modules/cli-truncate/node_modules/strip-ansi": { - "version": "6.0.1", - "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", - "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", + "version": "7.0.1", + "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-7.0.1.tgz", + "integrity": "sha512-cXNxvT8dFNRVfhVME3JAe98mkXDYN2O1l7jmcwMnOslDeESg1rF/OZMtK0nRAhiari1unG5cD4jG3rapUAkLbw==", "dependencies": { - "ansi-regex": "^5.0.1" + "ansi-regex": "^6.0.1" }, "engines": { - "node": ">=8" + "node": ">=12" + }, + "funding": { + "url": "https://github.com/chalk/strip-ansi?sponsor=1" } }, "node_modules/cliui": { - "version": "7.0.4", - "resolved": "https://registry.npmjs.org/cliui/-/cliui-7.0.4.tgz", - "integrity": "sha512-OcRE68cOsVMXp1Yvonl/fzkQOyjLSu/8bhPDfQt0e0/Eb283TKP20Fs2MqoPsr9SwA595rRCA+QMzYc9nBP+JQ==", + "version": "8.0.1", + "resolved": "https://registry.npmjs.org/cliui/-/cliui-8.0.1.tgz", + "integrity": "sha512-BSeNnyus75C4//NQ9gQt1/csTXyo/8Sb+afLAkzAptFuMsod9HFokGNudZpi/oQV73hnVK+sR+5PVRMd+Dr7YQ==", "dependencies": { "string-width": "^4.2.0", - "strip-ansi": "^6.0.0", + "strip-ansi": "^6.0.1", "wrap-ansi": "^7.0.0" - } - }, - "node_modules/cliui/node_modules/ansi-regex": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", - "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==", - "engines": { - "node": ">=8" - } - }, - "node_modules/cliui/node_modules/ansi-styles": { - "version": "4.3.0", - "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz", - "integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==", - "dependencies": { - "color-convert": "^2.0.1" - }, - "engines": { - "node": ">=8" - }, - "funding": { - "url": "https://github.com/chalk/ansi-styles?sponsor=1" - } - }, - "node_modules/cliui/node_modules/emoji-regex": { - "version": "8.0.0", - "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", - "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" - }, - "node_modules/cliui/node_modules/string-width": { - "version": "4.2.3", - "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", - "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", - "dependencies": { - "emoji-regex": "^8.0.0", - "is-fullwidth-code-point": "^3.0.0", - "strip-ansi": "^6.0.1" - }, - "engines": { - "node": ">=8" - } - }, - "node_modules/cliui/node_modules/strip-ansi": { - "version": "6.0.1", - "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", - "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", - "dependencies": { - "ansi-regex": "^5.0.1" - }, - "engines": { - "node": ">=8" - } - }, - "node_modules/cliui/node_modules/wrap-ansi": { - "version": "7.0.0", - "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-7.0.0.tgz", - "integrity": "sha512-YVGIj2kamLSTxw6NsZjoBxfSwsn0ycdesmc4p+Q21c5zPuZ1pl+NfxVdxPtdHvmNVOQ6XSYG4AUtyt/Fi7D16Q==", - "dependencies": { - "ansi-styles": "^4.0.0", - "string-width": "^4.1.0", - "strip-ansi": "^6.0.0" }, "engines": { - "node": ">=10" - }, - "funding": { - "url": "https://github.com/chalk/wrap-ansi?sponsor=1" - } - }, - "node_modules/clone-response": { - "version": "1.0.3", - "resolved": "https://registry.npmjs.org/clone-response/-/clone-response-1.0.3.tgz", - "integrity": "sha512-ROoL94jJH2dUVML2Y/5PEDNaSHgeOdSDicUyS7izcF63G6sTc/FTjLub4b8Il9S8S0beOfYt0TaA5qvFK+w0wA==", - "dependencies": { - "mimic-response": "^1.0.0" - }, - "funding": { - "url": "https://github.com/sponsors/sindresorhus" + "node": ">=12" } }, "node_modules/color-convert": { @@ -405,34 +579,35 @@ "integrity": "sha512-dOy+3AuW3a2wNbZHIuMZpTcgjGuLU/uBL/ubcZF9OXbDo8ff4O8yVp5Bf0efS8uEoYo5q4Fx7dY9OgQGXgAsQA==" }, "node_modules/colorette": { - "version": "2.0.19", - "resolved": "https://registry.npmjs.org/colorette/-/colorette-2.0.19.tgz", - "integrity": "sha512-3tlv/dIP7FWvj3BsbHrGLJ6l/oKh1O3TcgBqMn+yyCagOxc23fyzDS6HypQbgxWbkpDnf52p1LuR4eWDQ/K9WQ==" + "version": "2.0.20", + "resolved": "https://registry.npmjs.org/colorette/-/colorette-2.0.20.tgz", + "integrity": "sha512-IfEDxwoWIjkeXL1eXcDiow4UbKjhLdq6/EuSVR9GMN7KVH3r9gQ83e73hsz1Nd1T3ijd5xv1wcWRYO+D6kCI2w==" }, - "node_modules/compress-brotli": { - "version": "1.3.8", - "resolved": "https://registry.npmjs.org/compress-brotli/-/compress-brotli-1.3.8.tgz", - "integrity": "sha512-lVcQsjhxhIXsuupfy9fmZUFtAIdBmXA7EGY6GBdgZ++qkM9zG4YFT8iU7FoBxzryNDMOpD1HIFHUSX4D87oqhQ==", + "node_modules/cpu-features": { + "version": "0.0.10", + "resolved": "https://registry.npmjs.org/cpu-features/-/cpu-features-0.0.10.tgz", + "integrity": "sha512-9IkYqtX3YHPCzoVg1Py+o9057a3i0fp7S530UWokCSaFVTc7CwXPRiOjRjBQQ18ZCNafx78YfnG+HALxtVmOGA==", + "hasInstallScript": true, + "optional": true, "dependencies": { - "@types/json-buffer": "~3.0.0", - "json-buffer": "~3.0.1" + "buildcheck": "~0.0.6", + "nan": "^2.19.0" }, "engines": { - "node": ">= 12" + "node": ">=10.0.0" } }, - "node_modules/cpu-features": { - "version": "0.0.4", - "resolved": "https://registry.npmjs.org/cpu-features/-/cpu-features-0.0.4.tgz", - "integrity": "sha512-fKiZ/zp1mUwQbnzb9IghXtHtDoTMtNeb8oYGx6kX2SYfhnG0HNdBEBIzB9b5KlXu5DQPhfy3mInbBxFcgwAr3A==", - "hasInstallScript": true, - "optional": true, + "node_modules/cross-spawn": { + "version": "7.0.3", + "resolved": "https://registry.npmjs.org/cross-spawn/-/cross-spawn-7.0.3.tgz", + "integrity": "sha512-iRDPJKUPVEND7dHPO8rkbOnPpyDygcDFtWjpeWNCgy8WP2rXcxXL8TskReQl6OrB2G7+UJrags1q15Fudc7G6w==", "dependencies": { - "buildcheck": "0.0.3", - "nan": "^2.15.0" + "path-key": "^3.1.0", + "shebang-command": "^2.0.0", + "which": "^2.0.1" }, "engines": { - "node": ">=10.0.0" + "node": ">= 8" } }, "node_modules/debug": { @@ -485,31 +660,48 @@ } }, "node_modules/docker-modem": { - "version": "3.0.5", - "resolved": "https://registry.npmjs.org/docker-modem/-/docker-modem-3.0.5.tgz", - "integrity": "sha512-x1E6jxWdtoK3+ifAUWj4w5egPdTDGBpesSCErm+aKET5BnnEOvDtTP6GxcnMB1zZiv2iQ0qJZvJie+1wfIRg6Q==", + "version": "5.0.6", + "resolved": "https://registry.npmjs.org/docker-modem/-/docker-modem-5.0.6.tgz", + "integrity": "sha512-ens7BiayssQz/uAxGzH8zGXCtiV24rRWXdjNha5V4zSOcxmAZsfGVm/PPFbwQdqEkDnhG+SyR9E3zSHUbOKXBQ==", + "license": "Apache-2.0", "dependencies": { "debug": "^4.1.1", "readable-stream": "^3.5.0", "split-ca": "^1.0.1", - "ssh2": "^1.4.0" + "ssh2": "^1.15.0" }, "engines": { "node": ">= 8.0" } }, "node_modules/dockerode": { - "version": "3.3.3", - "resolved": "https://registry.npmjs.org/dockerode/-/dockerode-3.3.3.tgz", - "integrity": "sha512-lvKV6/NGf2/CYLt5V4c0fd6Fl9XZSCo1Z2HBT9ioKrKLMB2o+gA62Uza8RROpzGvYv57KJx2dKu+ZwSpB//OIA==", + "version": "4.0.6", + "resolved": "https://registry.npmjs.org/dockerode/-/dockerode-4.0.6.tgz", + "integrity": "sha512-FbVf3Z8fY/kALB9s+P9epCpWhfi/r0N2DgYYcYpsAUlaTxPjdsitsFobnltb+lyCgAIvf9C+4PSWlTnHlJMf1w==", + "license": "Apache-2.0", "dependencies": { - "docker-modem": "^3.0.0", - "tar-fs": "~2.0.1" + "@balena/dockerignore": "^1.0.2", + "@grpc/grpc-js": "^1.11.1", + "@grpc/proto-loader": "^0.7.13", + "docker-modem": "^5.0.6", + "protobufjs": "^7.3.2", + "tar-fs": "~2.1.2", + "uuid": "^10.0.0" }, "engines": { "node": ">= 8.0" } }, + "node_modules/eastasianwidth": { + "version": "0.2.0", + "resolved": "https://registry.npmjs.org/eastasianwidth/-/eastasianwidth-0.2.0.tgz", + "integrity": "sha512-I88TYZWc9XiYHRQ4/3c5rjjfgkjhLyW2luGIheGERbNQ6OY7yTybanSpDXZa8y7VUP9YmDcYa+eyq4ca7iLqWA==" + }, + "node_modules/emoji-regex": { + "version": "8.0.0", + "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", + "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" + }, "node_modules/end-of-stream": { "version": "1.4.4", "resolved": "https://registry.npmjs.org/end-of-stream/-/end-of-stream-1.4.4.tgz", @@ -519,11 +711,12 @@ } }, "node_modules/enquirer": { - "version": "2.3.6", - "resolved": "https://registry.npmjs.org/enquirer/-/enquirer-2.3.6.tgz", - "integrity": "sha512-yjNnPr315/FjS4zIsUxYguYUPP2e1NK4d7E7ZOLiyYCcbFBiTMyID+2wvm2w6+pZ/odMA7cRkjhsPbltwBOrLg==", + "version": "2.4.1", + "resolved": "https://registry.npmjs.org/enquirer/-/enquirer-2.4.1.tgz", + "integrity": "sha512-rRqJg/6gd538VHvR3PSrdRBb/1Vy2YfzHqzvbhGIQpDRKIa4FgV/54b5Q1xYSxOOwKvjXweS26E0Q+nAMwp2pQ==", "dependencies": { - "ansi-colors": "^4.1.1" + "ansi-colors": "^4.1.1", + "strip-ansi": "^6.0.1" }, "engines": { "node": ">=8.6" @@ -537,6 +730,11 @@ "node": ">=6" } }, + "node_modules/eventemitter3": { + "version": "5.0.1", + "resolved": "https://registry.npmjs.org/eventemitter3/-/eventemitter3-5.0.1.tgz", + "integrity": "sha512-GWkBvjiSZK87ELrYOSESUYeVIc9mvLLf/nXalMOS5dYrgZq9o5OVkbZAVM06CVxYsCwH9BDZFPlQTlPA1j4ahA==" + }, "node_modules/extract-zip": { "version": "2.0.1", "resolved": "https://registry.npmjs.org/extract-zip/-/extract-zip-2.0.1.tgz", @@ -578,10 +776,36 @@ "pend": "~1.2.0" } }, + "node_modules/foreground-child": { + "version": "3.1.1", + "resolved": "https://registry.npmjs.org/foreground-child/-/foreground-child-3.1.1.tgz", + "integrity": "sha512-TMKDUnIte6bfb5nWv7V/caI169OHgvwjb7V4WkeUvbQQdjr5rWKqHFiKWb/fcOwB+CzBT+qbWjvj+DVwRskpIg==", + "dependencies": { + "cross-spawn": "^7.0.0", + "signal-exit": "^4.0.1" + }, + "engines": { + "node": ">=14" + }, + "funding": { + "url": "https://github.com/sponsors/isaacs" + } + }, + "node_modules/foreground-child/node_modules/signal-exit": { + "version": "4.1.0", + "resolved": "https://registry.npmjs.org/signal-exit/-/signal-exit-4.1.0.tgz", + "integrity": "sha512-bzyZ1e88w9O1iNJbKnOlvYTrWPDl46O1bG0D3XInv+9tkPrxrN8jUUTiFlDkkmKWgn1M6CfIA13SuGqOa9Korw==", + "engines": { + "node": ">=14" + }, + "funding": { + "url": "https://github.com/sponsors/isaacs" + } + }, "node_modules/form-data-encoder": { - "version": "2.0.1", - "resolved": "https://registry.npmjs.org/form-data-encoder/-/form-data-encoder-2.0.1.tgz", - "integrity": "sha512-Oy+P9w5mnO4TWXVgUiQvggNKPI9/ummcSt5usuIV6HkaLKigwzPpoenhEqmGmx3zHqm6ZLJ+CR/99N8JLinaEw==", + "version": "2.1.4", + "resolved": "https://registry.npmjs.org/form-data-encoder/-/form-data-encoder-2.1.4.tgz", + "integrity": "sha512-yDYSgNMraqvnxiEXO4hi88+YZxaHC6QKzb5N84iRCTDeRO7ZALpir/lVmf/uXUhnwUr2O4HU8s/n6x+yNjQkHw==", "engines": { "node": ">= 14.17" } @@ -589,30 +813,21 @@ "node_modules/fs-constants": { "version": "1.0.0", "resolved": "https://registry.npmjs.org/fs-constants/-/fs-constants-1.0.0.tgz", - "integrity": "sha512-y6OAwoSIf7FyjMIv94u+b5rdheZEjzR63GTyZJm5qh4Bi+2YgwLCcI/fPFZkL5PSixOt6ZNKm+w+Hfp/Bciwow==" + "integrity": "sha512-y6OAwoSIf7FyjMIv94u+b5rdheZEjzR63GTyZJm5qh4Bi+2YgwLCcI/fPFZkL5PSixOt6ZNKm+w+Hfp/Bciwow==", + "license": "MIT" }, "node_modules/fs-extra": { - "version": "10.1.0", - "resolved": "https://registry.npmjs.org/fs-extra/-/fs-extra-10.1.0.tgz", - "integrity": "sha512-oRXApq54ETRj4eMiFzGnHWGy+zo5raudjuxN0b8H7s/RU2oW0Wvsx9O0ACRN/kRq9E8Vu/ReskGB5o3ji+FzHQ==", + "version": "11.3.0", + "resolved": "https://registry.npmjs.org/fs-extra/-/fs-extra-11.3.0.tgz", + "integrity": "sha512-Z4XaCL6dUDHfP/jT25jJKMmtxvuwbkrD1vNSMFlo9lNLY2c5FHYSQgHPRZUjAB26TpDEoW9HCOgplrdbaPV/ew==", + "license": "MIT", "dependencies": { "graceful-fs": "^4.2.0", "jsonfile": "^6.0.1", "universalify": "^2.0.0" }, "engines": { - "node": ">=12" - } - }, - "node_modules/fs-minipass": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/fs-minipass/-/fs-minipass-2.1.0.tgz", - "integrity": "sha512-V/JgOLFCS+R6Vcq0slCuaeWEdNC3ouDlJMNIsacH2VtALiu9mV4LPrHc5cDl8k5aw6J8jwgWWpiTo5RYhmIzvg==", - "dependencies": { - "minipass": "^3.0.0" - }, - "engines": { - "node": ">= 8" + "node": ">=14.14" } }, "node_modules/get-caller-file": { @@ -634,27 +849,46 @@ "url": "https://github.com/sponsors/sindresorhus" } }, + "node_modules/glob": { + "version": "10.3.12", + "resolved": "https://registry.npmjs.org/glob/-/glob-10.3.12.tgz", + "integrity": "sha512-TCNv8vJ+xz4QiqTpfOJA7HvYv+tNIRHKfUWw/q+v2jdgN4ebz+KY9tGx5J4rHP0o84mNP+ApH66HRX8us3Khqg==", + "dependencies": { + "foreground-child": "^3.1.0", + "jackspeak": "^2.3.6", + "minimatch": "^9.0.1", + "minipass": "^7.0.4", + "path-scurry": "^1.10.2" + }, + "bin": { + "glob": "dist/esm/bin.mjs" + }, + "engines": { + "node": ">=16 || 14 >=14.17" + }, + "funding": { + "url": "https://github.com/sponsors/isaacs" + } + }, "node_modules/got": { - "version": "12.3.1", - "resolved": "https://registry.npmjs.org/got/-/got-12.3.1.tgz", - "integrity": "sha512-tS6+JMhBh4iXMSXF6KkIsRxmloPln31QHDlcb6Ec3bzxjjFJFr/8aXdpyuLmVc9I4i2HyBHYw1QU5K1ruUdpkw==", + "version": "13.0.0", + "resolved": "https://registry.npmjs.org/got/-/got-13.0.0.tgz", + "integrity": "sha512-XfBk1CxOOScDcMr9O1yKkNaQyy865NbYs+F7dr4H0LZMVgCj2Le59k6PqbNHoL5ToeaEQUYh6c6yMfVcc6SJxA==", "dependencies": { "@sindresorhus/is": "^5.2.0", "@szmarczak/http-timer": "^5.0.1", - "@types/cacheable-request": "^6.0.2", - "@types/responselike": "^1.0.0", - "cacheable-lookup": "^6.0.4", - "cacheable-request": "^7.0.2", + "cacheable-lookup": "^7.0.0", + "cacheable-request": "^10.2.8", "decompress-response": "^6.0.0", - "form-data-encoder": "^2.0.1", + "form-data-encoder": "^2.1.2", "get-stream": "^6.0.1", "http2-wrapper": "^2.1.10", "lowercase-keys": "^3.0.0", "p-cancelable": "^3.0.0", - "responselike": "^2.0.0" + "responselike": "^3.0.0" }, "engines": { - "node": ">=14.16" + "node": ">=16" }, "funding": { "url": "https://github.com/sindresorhus/got?sponsor=1" @@ -666,9 +900,9 @@ "integrity": "sha512-9ByhssR2fPVsNZj478qUUbKfmL0+t5BDVyjShtyZZLiK7ZDAArFFfopyOTj0M05wE2tJPisA4iTnnXl2YoPvOA==" }, "node_modules/http-cache-semantics": { - "version": "4.1.0", - "resolved": "https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-4.1.0.tgz", - "integrity": "sha512-carPklcUh7ROWRK7Cv27RPtdhYhUsela/ue5/jKzjegVvXDqM2ILE9Q2BGn9JZJh1g87cp56su/FgQSzcWS8cQ==" + "version": "4.1.1", + "resolved": "https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-4.1.1.tgz", + "integrity": "sha512-er295DKPVsV82j5kw1Gjt+ADA/XYHsajl82cGNQG2eyoPkvgUhX+nDIyelzhIWbbsXP39EHcI6l5tYs2FYqYXQ==" }, "node_modules/http2-wrapper": { "version": "2.1.11", @@ -699,15 +933,8 @@ "type": "consulting", "url": "https://feross.org/support" } - ] - }, - "node_modules/indent-string": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/indent-string/-/indent-string-4.0.0.tgz", - "integrity": "sha512-EdDDZu4A2OyIK7Lr/2zG+w5jmbuk1DVBnEwREQvBzspBJkCEbRa8GxU1lghYcaGJCnRWibjDXlq779X1/y5xwg==", - "engines": { - "node": ">=8" - } + ], + "license": "BSD-3-Clause" }, "node_modules/inherits": { "version": "2.0.4", @@ -722,6 +949,28 @@ "node": ">=8" } }, + "node_modules/isexe": { + "version": "2.0.0", + "resolved": "https://registry.npmjs.org/isexe/-/isexe-2.0.0.tgz", + "integrity": "sha512-RHxMLp9lnKHGHRng9QFhRCMbYAcVpn69smSGcq3f36xjgVVWThj4qqLbTLlq7Ssj8B+fIQ1EuCEGI2lKsyQeIw==" + }, + "node_modules/jackspeak": { + "version": "2.3.6", + "resolved": "https://registry.npmjs.org/jackspeak/-/jackspeak-2.3.6.tgz", + "integrity": "sha512-N3yCS/NegsOBokc8GAdM8UcmfsKiSS8cipheD/nivzr700H+nsMOxJjQnvwOcRYVuFkdH0wGUvW2WbXGmrZGbQ==", + "dependencies": { + "@isaacs/cliui": "^8.0.2" + }, + "engines": { + "node": ">=14" + }, + "funding": { + "url": "https://github.com/sponsors/isaacs" + }, + "optionalDependencies": { + "@pkgjs/parseargs": "^0.11.0" + } + }, "node_modules/json-buffer": { "version": "3.0.1", "resolved": "https://registry.npmjs.org/json-buffer/-/json-buffer-3.0.1.tgz", @@ -744,30 +993,27 @@ "integrity": "sha512-HjorDJFNhnM4SicvaUXac0X77NiskggxJdesG72+O5zBKpSqKFCrqmndKVqpu3pFqkla0St6uGk8Ju0sCurrmg==" }, "node_modules/keyv": { - "version": "4.3.3", - "resolved": "https://registry.npmjs.org/keyv/-/keyv-4.3.3.tgz", - "integrity": "sha512-AcysI17RvakTh8ir03+a3zJr5r0ovnAH/XTXei/4HIv3bL2K/jzvgivLK9UuI/JbU1aJjM3NSAnVvVVd3n+4DQ==", + "version": "4.5.2", + "resolved": "https://registry.npmjs.org/keyv/-/keyv-4.5.2.tgz", + "integrity": "sha512-5MHbFaKn8cNSmVW7BYnijeAVlE4cYA/SVkifVgrh7yotnfhKmjuXpDKjrABLnT0SfHWV21P8ow07OGfRrNDg8g==", "dependencies": { - "compress-brotli": "^1.3.8", "json-buffer": "3.0.1" } }, "node_modules/listr2": { - "version": "5.0.2", - "resolved": "https://registry.npmjs.org/listr2/-/listr2-5.0.2.tgz", - "integrity": "sha512-/yIYitiJgIkzVIKqQc+Dpcpwaj4PjeW/hSi7/xoGJ5mil7vQYedr5kKuEyLhANlnifCcumafnIR6azpb8loODw==", - "dependencies": { - "cli-truncate": "^2.1.0", - "colorette": "^2.0.19", - "log-update": "^4.0.0", - "p-map": "^4.0.0", + "version": "6.6.1", + "resolved": "https://registry.npmjs.org/listr2/-/listr2-6.6.1.tgz", + "integrity": "sha512-+rAXGHh0fkEWdXBmX+L6mmfmXmXvDGEKzkjxO+8mP3+nI/r/CWznVBvsibXdxda9Zz0OW2e2ikphN3OwCT/jSg==", + "dependencies": { + "cli-truncate": "^3.1.0", + "colorette": "^2.0.20", + "eventemitter3": "^5.0.1", + "log-update": "^5.0.1", "rfdc": "^1.3.0", - "rxjs": "^7.5.6", - "through": "^2.3.8", - "wrap-ansi": "^7.0.0" + "wrap-ansi": "^8.1.0" }, "engines": { - "node": "^14.13.1 || >=16.0.0" + "node": ">=16.0.0" }, "peerDependencies": { "enquirer": ">= 2.3.0 < 3" @@ -779,67 +1025,73 @@ } }, "node_modules/listr2/node_modules/ansi-regex": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", - "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==", + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-6.0.1.tgz", + "integrity": "sha512-n5M855fKb2SsfMIiFFoVrABHJC8QtHwVx+mHWP3QcEqBHYienj5dHSgjbxtC0WEZXYt4wcD6zrQElDPhFuZgfA==", "engines": { - "node": ">=8" + "node": ">=12" + }, + "funding": { + "url": "https://github.com/chalk/ansi-regex?sponsor=1" } }, "node_modules/listr2/node_modules/ansi-styles": { - "version": "4.3.0", - "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz", - "integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==", - "dependencies": { - "color-convert": "^2.0.1" - }, + "version": "6.2.1", + "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-6.2.1.tgz", + "integrity": "sha512-bN798gFfQX+viw3R7yrGWRqnrN2oRkEkUjjl4JNn4E8GxxbjtG3FbrEIIY3l8/hrwUwIeCZvi4QuOTP4MErVug==", "engines": { - "node": ">=8" + "node": ">=12" }, "funding": { "url": "https://github.com/chalk/ansi-styles?sponsor=1" } }, "node_modules/listr2/node_modules/emoji-regex": { - "version": "8.0.0", - "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", - "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" + "version": "9.2.2", + "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-9.2.2.tgz", + "integrity": "sha512-L18DaJsXSUk2+42pv8mLs5jJT2hqFkFE4j21wOmgbUqsZ2hL72NsUU785g9RXgo3s0ZNgVl42TiHp3ZtOv/Vyg==" }, "node_modules/listr2/node_modules/string-width": { - "version": "4.2.3", - "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", - "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", + "version": "5.1.2", + "resolved": "https://registry.npmjs.org/string-width/-/string-width-5.1.2.tgz", + "integrity": "sha512-HnLOCR3vjcY8beoNLtcjZ5/nxn2afmME6lhrDrebokqMap+XbeW8n9TXpPDOqdGK5qcI3oT0GKTW6wC7EMiVqA==", "dependencies": { - "emoji-regex": "^8.0.0", - "is-fullwidth-code-point": "^3.0.0", - "strip-ansi": "^6.0.1" + "eastasianwidth": "^0.2.0", + "emoji-regex": "^9.2.2", + "strip-ansi": "^7.0.1" }, "engines": { - "node": ">=8" + "node": ">=12" + }, + "funding": { + "url": "https://github.com/sponsors/sindresorhus" } }, "node_modules/listr2/node_modules/strip-ansi": { - "version": "6.0.1", - "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", - "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", + "version": "7.0.1", + "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-7.0.1.tgz", + "integrity": "sha512-cXNxvT8dFNRVfhVME3JAe98mkXDYN2O1l7jmcwMnOslDeESg1rF/OZMtK0nRAhiari1unG5cD4jG3rapUAkLbw==", "dependencies": { - "ansi-regex": "^5.0.1" + "ansi-regex": "^6.0.1" }, "engines": { - "node": ">=8" + "node": ">=12" + }, + "funding": { + "url": "https://github.com/chalk/strip-ansi?sponsor=1" } }, "node_modules/listr2/node_modules/wrap-ansi": { - "version": "7.0.0", - "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-7.0.0.tgz", - "integrity": "sha512-YVGIj2kamLSTxw6NsZjoBxfSwsn0ycdesmc4p+Q21c5zPuZ1pl+NfxVdxPtdHvmNVOQ6XSYG4AUtyt/Fi7D16Q==", + "version": "8.1.0", + "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-8.1.0.tgz", + "integrity": "sha512-si7QWI6zUMq56bESFvagtmzMdGOtoxfR+Sez11Mobfc7tm+VkUckk9bW2UeffTGVUbOksxmSw0AA2gs8g71NCQ==", "dependencies": { - "ansi-styles": "^4.0.0", - "string-width": "^4.1.0", - "strip-ansi": "^6.0.0" + "ansi-styles": "^6.1.0", + "string-width": "^5.0.1", + "strip-ansi": "^7.0.1" }, "engines": { - "node": ">=10" + "node": ">=12" }, "funding": { "url": "https://github.com/chalk/wrap-ansi?sponsor=1" @@ -850,150 +1102,108 @@ "resolved": "https://registry.npmjs.org/lodash-es/-/lodash-es-4.17.21.tgz", "integrity": "sha512-mKnC+QJ9pWVzv+C4/U3rRsHapFfHvQFoFB92e52xeyGMcX6/OlIl78je1u8vePzYZSkkogMPJ2yjxxsb89cxyw==" }, + "node_modules/lodash.camelcase": { + "version": "4.3.0", + "resolved": "https://registry.npmjs.org/lodash.camelcase/-/lodash.camelcase-4.3.0.tgz", + "integrity": "sha512-TwuEnCnxbc3rAvhf/LbG7tJUDzhqXyFnv3dtzLOPgCG/hODL7WFnsbwktkD7yUV0RrreP/l1PALq/YSg6VvjlA==", + "license": "MIT" + }, "node_modules/log-update": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/log-update/-/log-update-4.0.0.tgz", - "integrity": "sha512-9fkkDevMefjg0mmzWFBW8YkFP91OrizzkW3diF7CpG+S2EYdy4+TVfGwz1zeF8x7hCx1ovSPTOE9Ngib74qqUg==", + "version": "5.0.1", + "resolved": "https://registry.npmjs.org/log-update/-/log-update-5.0.1.tgz", + "integrity": "sha512-5UtUDQ/6edw4ofyljDNcOVJQ4c7OjDro4h3y8e1GQL5iYElYclVHJ3zeWchylvMaKnDbDilC8irOVyexnA/Slw==", "dependencies": { - "ansi-escapes": "^4.3.0", - "cli-cursor": "^3.1.0", - "slice-ansi": "^4.0.0", - "wrap-ansi": "^6.2.0" + "ansi-escapes": "^5.0.0", + "cli-cursor": "^4.0.0", + "slice-ansi": "^5.0.0", + "strip-ansi": "^7.0.1", + "wrap-ansi": "^8.0.1" }, "engines": { - "node": ">=10" + "node": "^12.20.0 || ^14.13.1 || >=16.0.0" }, "funding": { "url": "https://github.com/sponsors/sindresorhus" } }, - "node_modules/log-update/node_modules/ansi-escapes": { - "version": "4.3.2", - "resolved": "https://registry.npmjs.org/ansi-escapes/-/ansi-escapes-4.3.2.tgz", - "integrity": "sha512-gKXj5ALrKWQLsYG9jlTRmR/xKluxHV+Z9QEwNIgCfM1/uwPMCuzVVnh5mwTd+OuBZcwSIMbqssNWRm1lE51QaQ==", - "dependencies": { - "type-fest": "^0.21.3" - }, + "node_modules/log-update/node_modules/ansi-regex": { + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-6.0.1.tgz", + "integrity": "sha512-n5M855fKb2SsfMIiFFoVrABHJC8QtHwVx+mHWP3QcEqBHYienj5dHSgjbxtC0WEZXYt4wcD6zrQElDPhFuZgfA==", "engines": { - "node": ">=8" + "node": ">=12" }, "funding": { - "url": "https://github.com/sponsors/sindresorhus" - } - }, - "node_modules/log-update/node_modules/ansi-regex": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", - "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==", - "engines": { - "node": ">=8" + "url": "https://github.com/chalk/ansi-regex?sponsor=1" } }, "node_modules/log-update/node_modules/ansi-styles": { - "version": "4.3.0", - "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz", - "integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==", - "dependencies": { - "color-convert": "^2.0.1" - }, + "version": "6.2.1", + "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-6.2.1.tgz", + "integrity": "sha512-bN798gFfQX+viw3R7yrGWRqnrN2oRkEkUjjl4JNn4E8GxxbjtG3FbrEIIY3l8/hrwUwIeCZvi4QuOTP4MErVug==", "engines": { - "node": ">=8" + "node": ">=12" }, "funding": { "url": "https://github.com/chalk/ansi-styles?sponsor=1" } }, - "node_modules/log-update/node_modules/cli-cursor": { - "version": "3.1.0", - "resolved": "https://registry.npmjs.org/cli-cursor/-/cli-cursor-3.1.0.tgz", - "integrity": "sha512-I/zHAwsKf9FqGoXM4WWRACob9+SNukZTd94DWF57E4toouRulbCxcUh6RKUEOQlYTHJnzkPMySvPNaaSLNfLZw==", - "dependencies": { - "restore-cursor": "^3.1.0" - }, - "engines": { - "node": ">=8" - } - }, "node_modules/log-update/node_modules/emoji-regex": { - "version": "8.0.0", - "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", - "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" + "version": "9.2.2", + "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-9.2.2.tgz", + "integrity": "sha512-L18DaJsXSUk2+42pv8mLs5jJT2hqFkFE4j21wOmgbUqsZ2hL72NsUU785g9RXgo3s0ZNgVl42TiHp3ZtOv/Vyg==" }, - "node_modules/log-update/node_modules/restore-cursor": { - "version": "3.1.0", - "resolved": "https://registry.npmjs.org/restore-cursor/-/restore-cursor-3.1.0.tgz", - "integrity": "sha512-l+sSefzHpj5qimhFSE5a8nufZYAM3sBSVMAPtYkmC+4EH2anSGaEMXSD0izRQbu9nfyQ9y5JrVmp7E8oZrUjvA==", + "node_modules/log-update/node_modules/string-width": { + "version": "5.1.2", + "resolved": "https://registry.npmjs.org/string-width/-/string-width-5.1.2.tgz", + "integrity": "sha512-HnLOCR3vjcY8beoNLtcjZ5/nxn2afmME6lhrDrebokqMap+XbeW8n9TXpPDOqdGK5qcI3oT0GKTW6wC7EMiVqA==", "dependencies": { - "onetime": "^5.1.0", - "signal-exit": "^3.0.2" + "eastasianwidth": "^0.2.0", + "emoji-regex": "^9.2.2", + "strip-ansi": "^7.0.1" }, "engines": { - "node": ">=8" + "node": ">=12" + }, + "funding": { + "url": "https://github.com/sponsors/sindresorhus" } }, - "node_modules/log-update/node_modules/slice-ansi": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/slice-ansi/-/slice-ansi-4.0.0.tgz", - "integrity": "sha512-qMCMfhY040cVHT43K9BFygqYbUPFZKHOg7K73mtTWJRb8pyP3fzf4Ixd5SzdEJQ6MRUg/WBnOLxghZtKKurENQ==", + "node_modules/log-update/node_modules/strip-ansi": { + "version": "7.0.1", + "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-7.0.1.tgz", + "integrity": "sha512-cXNxvT8dFNRVfhVME3JAe98mkXDYN2O1l7jmcwMnOslDeESg1rF/OZMtK0nRAhiari1unG5cD4jG3rapUAkLbw==", "dependencies": { - "ansi-styles": "^4.0.0", - "astral-regex": "^2.0.0", - "is-fullwidth-code-point": "^3.0.0" + "ansi-regex": "^6.0.1" }, "engines": { - "node": ">=10" + "node": ">=12" }, "funding": { - "url": "https://github.com/chalk/slice-ansi?sponsor=1" + "url": "https://github.com/chalk/strip-ansi?sponsor=1" } }, - "node_modules/log-update/node_modules/string-width": { - "version": "4.2.3", - "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", - "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", + "node_modules/log-update/node_modules/wrap-ansi": { + "version": "8.1.0", + "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-8.1.0.tgz", + "integrity": "sha512-si7QWI6zUMq56bESFvagtmzMdGOtoxfR+Sez11Mobfc7tm+VkUckk9bW2UeffTGVUbOksxmSw0AA2gs8g71NCQ==", "dependencies": { - "emoji-regex": "^8.0.0", - "is-fullwidth-code-point": "^3.0.0", - "strip-ansi": "^6.0.1" + "ansi-styles": "^6.1.0", + "string-width": "^5.0.1", + "strip-ansi": "^7.0.1" }, "engines": { - "node": ">=8" + "node": ">=12" + }, + "funding": { + "url": "https://github.com/chalk/wrap-ansi?sponsor=1" } }, - "node_modules/log-update/node_modules/strip-ansi": { - "version": "6.0.1", - "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", - "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", - "dependencies": { - "ansi-regex": "^5.0.1" - }, - "engines": { - "node": ">=8" - } - }, - "node_modules/log-update/node_modules/type-fest": { - "version": "0.21.3", - "resolved": "https://registry.npmjs.org/type-fest/-/type-fest-0.21.3.tgz", - "integrity": "sha512-t0rzBq87m3fVcduHDUFhKmyyX+9eo6WQjZvf51Ea/M0Q7+T374Jp1aUiyUl0GKxp8M/OETVHSDvmkyPgvX+X2w==", - "engines": { - "node": ">=10" - }, - "funding": { - "url": "https://github.com/sponsors/sindresorhus" - } - }, - "node_modules/log-update/node_modules/wrap-ansi": { - "version": "6.2.0", - "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-6.2.0.tgz", - "integrity": "sha512-r6lPcBGxZXlIcymEu7InxDMhdW0KDxpLgoFLcguasxCaJ/SOIZwINatK9KY/tf+ZrlywOKU0UDj3ATXUBfxJXA==", - "dependencies": { - "ansi-styles": "^4.0.0", - "string-width": "^4.1.0", - "strip-ansi": "^6.0.0" - }, - "engines": { - "node": ">=8" - } + "node_modules/long": { + "version": "5.2.4", + "resolved": "https://registry.npmjs.org/long/-/long-5.2.4.tgz", + "integrity": "sha512-qtzLbJE8hq7VabR3mISmVGtoXP8KGc2Z/AT8OuqlYD7JTR3oqrgwdjnk07wpj1twXxYmgDXgoKVWUG/fReSzHg==", + "license": "Apache-2.0" }, "node_modules/lowercase-keys": { "version": "3.0.0", @@ -1006,10 +1216,19 @@ "url": "https://github.com/sponsors/sindresorhus" } }, + "node_modules/lru-cache": { + "version": "10.2.2", + "resolved": "https://registry.npmjs.org/lru-cache/-/lru-cache-10.2.2.tgz", + "integrity": "sha512-9hp3Vp2/hFQUiIwKo8XCeFVnrg8Pk3TYNPIR7tJADKi5YfcF7vEaK7avFHTlSy3kOKYaJQaalfEo6YuXdceBOQ==", + "engines": { + "node": "14 || >=16.14" + } + }, "node_modules/luxon": { - "version": "3.0.1", - "resolved": "https://registry.npmjs.org/luxon/-/luxon-3.0.1.tgz", - "integrity": "sha512-hF3kv0e5gwHQZKz4wtm4c+inDtyc7elkanAsBq+fundaCdUBNJB1dHEGUZIM6SfSBUlbVFduPwEtNjFK8wLtcw==", + "version": "3.6.1", + "resolved": "https://registry.npmjs.org/luxon/-/luxon-3.6.1.tgz", + "integrity": "sha512-tJLxrKJhO2ukZ5z0gyjY1zPh3Rh88Ej9P7jNrZiHMUXHae1yvI2imgOZtL1TO8TW6biMMKfTtAOoEJANgtWBMQ==", + "license": "MIT", "engines": { "node": ">=12" } @@ -1023,51 +1242,69 @@ } }, "node_modules/mimic-response": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/mimic-response/-/mimic-response-1.0.1.tgz", - "integrity": "sha512-j5EctnkH7amfV/q5Hgmoal1g2QHFJRraOtmx0JpIqkxhBhI/lJSl1nMpQ45hVarwNETOoWEimndZ4QK0RHxuxQ==", + "version": "4.0.0", + "resolved": "https://registry.npmjs.org/mimic-response/-/mimic-response-4.0.0.tgz", + "integrity": "sha512-e5ISH9xMYU0DzrT+jl8q2ze9D6eWBto+I8CNpe+VI+K2J/F/k3PdkdTdz4wvGVH4NTpo+NRYTVIuMQEMMcsLqg==", "engines": { - "node": ">=4" + "node": "^12.20.0 || ^14.13.1 || >=16.0.0" + }, + "funding": { + "url": "https://github.com/sponsors/sindresorhus" } }, - "node_modules/minipass": { - "version": "3.3.4", - "resolved": "https://registry.npmjs.org/minipass/-/minipass-3.3.4.tgz", - "integrity": "sha512-I9WPbWHCGu8W+6k1ZiGpPu0GkoKBeorkfKNuAFBNS1HNFJvke82sxvI5bzcCNpWPorkOO5QQ+zomzzwRxejXiw==", + "node_modules/minimatch": { + "version": "9.0.4", + "resolved": "https://registry.npmjs.org/minimatch/-/minimatch-9.0.4.tgz", + "integrity": "sha512-KqWh+VchfxcMNRAJjj2tnsSJdNbHsVgnkBhTNrW7AjVo6OvLtxw8zfT9oLw1JSohlFzJ8jCoTgaoXvJ+kHt6fw==", "dependencies": { - "yallist": "^4.0.0" + "brace-expansion": "^2.0.1" }, "engines": { - "node": ">=8" + "node": ">=16 || 14 >=14.17" + }, + "funding": { + "url": "https://github.com/sponsors/isaacs" + } + }, + "node_modules/minipass": { + "version": "7.1.2", + "resolved": "https://registry.npmjs.org/minipass/-/minipass-7.1.2.tgz", + "integrity": "sha512-qOOzS1cBTWYF4BH8fVePDBOO9iptMnGUEZwNc/cMWnTV2nVLZ7VoNWEPHkYczZA0pdoA7dl6e7FL659nX9S2aw==", + "engines": { + "node": ">=16 || 14 >=14.17" } }, "node_modules/minizlib": { - "version": "2.1.2", - "resolved": "https://registry.npmjs.org/minizlib/-/minizlib-2.1.2.tgz", - "integrity": "sha512-bAxsR8BVfj60DWXHE3u30oHzfl4G7khkSuPW+qvpd7jFRHm7dLxOjUk1EHACJ/hxLY8phGJ0YhYHZo7jil7Qdg==", + "version": "3.0.1", + "resolved": "https://registry.npmjs.org/minizlib/-/minizlib-3.0.1.tgz", + "integrity": "sha512-umcy022ILvb5/3Djuu8LWeqUa8D68JaBzlttKeMWen48SjabqS3iY5w/vzeMzMUNhLDifyhbOwKDSznB1vvrwg==", "dependencies": { - "minipass": "^3.0.0", - "yallist": "^4.0.0" + "minipass": "^7.0.4", + "rimraf": "^5.0.5" }, "engines": { - "node": ">= 8" + "node": ">= 18" } }, "node_modules/mkdirp": { - "version": "1.0.4", - "resolved": "https://registry.npmjs.org/mkdirp/-/mkdirp-1.0.4.tgz", - "integrity": "sha512-vVqVZQyf3WLx2Shd0qJ9xuvqgAyKPLAiqITEtqW0oIUjzo3PePDd6fW9iFz30ef7Ysp/oiWqbhszeGWW2T6Gzw==", + "version": "3.0.1", + "resolved": "https://registry.npmjs.org/mkdirp/-/mkdirp-3.0.1.tgz", + "integrity": "sha512-+NsyUUAZDmo6YVHzL/stxSu3t9YS1iljliy3BSDrXJ/dkn1KYdmtZODGGjLcc9XLgVVpH4KshHB8XmZgMhaBXg==", "bin": { - "mkdirp": "bin/cmd.js" + "mkdirp": "dist/cjs/src/bin.js" }, "engines": { "node": ">=10" + }, + "funding": { + "url": "https://github.com/sponsors/isaacs" } }, "node_modules/mkdirp-classic": { "version": "0.5.3", "resolved": "https://registry.npmjs.org/mkdirp-classic/-/mkdirp-classic-0.5.3.tgz", - "integrity": "sha512-gKLcREMhtuZRwRAfqP3RFW+TK4JqApVBtOIftVgjuABpAtpxhPGaDcfvbhNvD0B8iD1oUr/txX35NjcaY6Ns/A==" + "integrity": "sha512-gKLcREMhtuZRwRAfqP3RFW+TK4JqApVBtOIftVgjuABpAtpxhPGaDcfvbhNvD0B8iD1oUr/txX35NjcaY6Ns/A==", + "license": "MIT" }, "node_modules/ms": { "version": "2.1.2", @@ -1075,17 +1312,18 @@ "integrity": "sha512-sGkPx+VjMtmA6MX27oA4FBFELFCZZ4S4XqeGOXCv68tT+jb3vk/RyaKWP0PTKyWtmLSM0b+adUTEvbs1PEaH2w==" }, "node_modules/nan": { - "version": "2.16.0", - "resolved": "https://registry.npmjs.org/nan/-/nan-2.16.0.tgz", - "integrity": "sha512-UdAqHyFngu7TfQKsCBgAA6pWDkT8MAO7d0jyOecVhN5354xbLqdn8mV9Tat9gepAupm0bt2DbeaSC8vS52MuFA==", + "version": "2.22.0", + "resolved": "https://registry.npmjs.org/nan/-/nan-2.22.0.tgz", + "integrity": "sha512-nbajikzWTMwsW+eSsNm3QwlOs7het9gGJU5dDZzRTQGk03vyBOauxgI4VakDzE0PtsGTmXPsXTbbjVhRwR5mpw==", + "license": "MIT", "optional": true }, "node_modules/normalize-url": { - "version": "6.1.0", - "resolved": "https://registry.npmjs.org/normalize-url/-/normalize-url-6.1.0.tgz", - "integrity": "sha512-DlL+XwOy3NxAQ8xuC0okPgK46iuVNAK01YN7RueYBqqFeGsBjV9XmCAzAdgt+667bCl5kPh9EqKKDwnaPG1I7A==", + "version": "8.0.0", + "resolved": "https://registry.npmjs.org/normalize-url/-/normalize-url-8.0.0.tgz", + "integrity": "sha512-uVFpKhj5MheNBJRTiMZ9pE/7hD1QTeEvugSJW/OmLzAp78PB5O6adfMNTvmfKhXBkvCzC+rqifWcVYpGFwTjnw==", "engines": { - "node": ">=10" + "node": ">=14.16" }, "funding": { "url": "https://github.com/sponsors/sindresorhus" @@ -1121,18 +1359,27 @@ "node": ">=12.20" } }, - "node_modules/p-map": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/p-map/-/p-map-4.0.0.tgz", - "integrity": "sha512-/bjOqmgETBYB5BoEeGVea8dmvHb2m9GLy1E9W43yeyfP6QQCZGFNa+XRceJEuDB6zqr+gKpIAmlLebMpykw/MQ==", + "node_modules/path-key": { + "version": "3.1.1", + "resolved": "https://registry.npmjs.org/path-key/-/path-key-3.1.1.tgz", + "integrity": "sha512-ojmeN0qd+y0jszEtoY48r0Peq5dwMEkIlCOu6Q5f41lfkswXuKtYrhgoTpLnyIcHm24Uhqx+5Tqm2InSwLhE6Q==", + "engines": { + "node": ">=8" + } + }, + "node_modules/path-scurry": { + "version": "1.10.2", + "resolved": "https://registry.npmjs.org/path-scurry/-/path-scurry-1.10.2.tgz", + "integrity": "sha512-7xTavNy5RQXnsjANvVvMkEjvloOinkAjv/Z6Ildz9v2RinZ4SBKTWFOVRbaF8p0vpHnyjV/UwNDdKuUv6M5qcA==", "dependencies": { - "aggregate-error": "^3.0.0" + "lru-cache": "^10.2.0", + "minipass": "^5.0.0 || ^6.0.2 || ^7.0.0" }, "engines": { - "node": ">=10" + "node": ">=16 || 14 >=14.17" }, "funding": { - "url": "https://github.com/sponsors/sindresorhus" + "url": "https://github.com/sponsors/isaacs" } }, "node_modules/pend": { @@ -1141,9 +1388,9 @@ "integrity": "sha512-F3asv42UuXchdzt+xXqfW1OGlVBe+mxa2mqI0pg5yAHZPvFmY3Y6drSf/GQ1A86WgWEN9Kzh/WrgKa6iGcHXLg==" }, "node_modules/pretty-bytes": { - "version": "6.0.0", - "resolved": "https://registry.npmjs.org/pretty-bytes/-/pretty-bytes-6.0.0.tgz", - "integrity": "sha512-6UqkYefdogmzqAZWzJ7laYeJnaXDy2/J+ZqiiMtS7t7OfpXWTlaeGMwX8U6EFvPV/YWWEKRkS8hKS4k60WHTOg==", + "version": "6.1.1", + "resolved": "https://registry.npmjs.org/pretty-bytes/-/pretty-bytes-6.1.1.tgz", + "integrity": "sha512-mQUvGU6aUFQ+rNvTIAcZuWGRT9a6f6Yrg9bHs4ImKF+HZCEK+plBvnAZYSIQztknZF2qnzNtr6F8s0+IuptdlQ==", "engines": { "node": "^14.13.1 || >=16.0.0" }, @@ -1151,6 +1398,30 @@ "url": "https://github.com/sponsors/sindresorhus" } }, + "node_modules/protobufjs": { + "version": "7.4.0", + "resolved": "https://registry.npmjs.org/protobufjs/-/protobufjs-7.4.0.tgz", + "integrity": "sha512-mRUWCc3KUU4w1jU8sGxICXH/gNS94DvI1gxqDvBzhj1JpcsimQkYiOJfwsPUykUI5ZaspFbSgmBLER8IrQ3tqw==", + "hasInstallScript": true, + "license": "BSD-3-Clause", + "dependencies": { + "@protobufjs/aspromise": "^1.1.2", + "@protobufjs/base64": "^1.1.2", + "@protobufjs/codegen": "^2.0.4", + "@protobufjs/eventemitter": "^1.1.0", + "@protobufjs/fetch": "^1.1.0", + "@protobufjs/float": "^1.0.2", + "@protobufjs/inquire": "^1.1.0", + "@protobufjs/path": "^1.1.2", + "@protobufjs/pool": "^1.1.0", + "@protobufjs/utf8": "^1.1.0", + "@types/node": ">=13.7.0", + "long": "^5.0.0" + }, + "engines": { + "node": ">=12.0.0" + } + }, "node_modules/pump": { "version": "3.0.0", "resolved": "https://registry.npmjs.org/pump/-/pump-3.0.0.tgz", @@ -1198,22 +1469,32 @@ "integrity": "sha512-0a1F4l73/ZFZOakJnQ3FvkJ2+gSTQWz/r2KE5OdDY0TxPm5h4GkqkWWfM47T7HsbnOtcJVEF4epCVy6u7Q3K+g==" }, "node_modules/responselike": { - "version": "2.0.1", - "resolved": "https://registry.npmjs.org/responselike/-/responselike-2.0.1.tgz", - "integrity": "sha512-4gl03wn3hj1HP3yzgdI7d3lCkF95F21Pz4BPGvKHinyQzALR5CapwC8yIi0Rh58DEMQ/SguC03wFj2k0M/mHhw==", + "version": "3.0.0", + "resolved": "https://registry.npmjs.org/responselike/-/responselike-3.0.0.tgz", + "integrity": "sha512-40yHxbNcl2+rzXvZuVkrYohathsSJlMTXKryG5y8uciHv1+xDLHQpgjG64JUO9nrEq2jGLH6IZ8BcZyw3wrweg==", "dependencies": { - "lowercase-keys": "^2.0.0" + "lowercase-keys": "^3.0.0" + }, + "engines": { + "node": ">=14.16" }, "funding": { "url": "https://github.com/sponsors/sindresorhus" } }, - "node_modules/responselike/node_modules/lowercase-keys": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/lowercase-keys/-/lowercase-keys-2.0.0.tgz", - "integrity": "sha512-tqNXrS78oMOE73NMxK4EMLQsQowWf8jKooH9g7xPavRT706R6bkQJ6DY2Te7QukaZsulxa30wQ7bk0pm4XiHmA==", + "node_modules/restore-cursor": { + "version": "4.0.0", + "resolved": "https://registry.npmjs.org/restore-cursor/-/restore-cursor-4.0.0.tgz", + "integrity": "sha512-I9fPXU9geO9bHOt9pHHOhOkYerIMsmVaWB0rA2AI9ERh/+x/i7MV5HKBNrg+ljO5eoPVgCcnFuRjJ9uH6I/3eg==", + "dependencies": { + "onetime": "^5.1.0", + "signal-exit": "^3.0.2" + }, "engines": { - "node": ">=8" + "node": "^12.20.0 || ^14.13.1 || >=16.0.0" + }, + "funding": { + "url": "https://github.com/sponsors/sindresorhus" } }, "node_modules/rfdc": { @@ -1221,12 +1502,21 @@ "resolved": "https://registry.npmjs.org/rfdc/-/rfdc-1.3.0.tgz", "integrity": "sha512-V2hovdzFbOi77/WajaSMXk2OLm+xNIeQdMMuB7icj7bk6zi2F8GGAxigcnDFpJHbNyNcgyJDiP+8nOrY5cZGrA==" }, - "node_modules/rxjs": { - "version": "7.5.6", - "resolved": "https://registry.npmjs.org/rxjs/-/rxjs-7.5.6.tgz", - "integrity": "sha512-dnyv2/YsXhnm461G+R/Pe5bWP41Nm6LBXEYWI6eiFP4fiwx6WRI/CD0zbdVAudd9xwLEF2IDcKXLHit0FYjUzw==", + "node_modules/rimraf": { + "version": "5.0.5", + "resolved": "https://registry.npmjs.org/rimraf/-/rimraf-5.0.5.tgz", + "integrity": "sha512-CqDakW+hMe/Bz202FPEymy68P+G50RfMQK+Qo5YUqc9SPipvbGjCGKd0RSKEelbsfQuw3g5NZDSrlZZAJurH1A==", "dependencies": { - "tslib": "^2.1.0" + "glob": "^10.3.7" + }, + "bin": { + "rimraf": "dist/esm/bin.mjs" + }, + "engines": { + "node": ">=14" + }, + "funding": { + "url": "https://github.com/sponsors/isaacs" } }, "node_modules/safe-buffer": { @@ -1251,7 +1541,27 @@ "node_modules/safer-buffer": { "version": "2.1.2", "resolved": "https://registry.npmjs.org/safer-buffer/-/safer-buffer-2.1.2.tgz", - "integrity": "sha512-YZo3K82SD7Riyi0E1EQPojLz7kpepnSQI9IyPbHHg1XXXevb5dJI7tpyN2ADxGcQbHG7vcyRHk0cbwqcQriUtg==" + "integrity": "sha512-YZo3K82SD7Riyi0E1EQPojLz7kpepnSQI9IyPbHHg1XXXevb5dJI7tpyN2ADxGcQbHG7vcyRHk0cbwqcQriUtg==", + "license": "MIT" + }, + "node_modules/shebang-command": { + "version": "2.0.0", + "resolved": "https://registry.npmjs.org/shebang-command/-/shebang-command-2.0.0.tgz", + "integrity": "sha512-kHxr2zZpYtdmrN1qDjrrX/Z1rR1kG8Dx+gkpK1G4eXmvXswmcE1hTWBWYUzlraYw1/yZp6YuDY77YtvbN0dmDA==", + "dependencies": { + "shebang-regex": "^3.0.0" + }, + "engines": { + "node": ">=8" + } + }, + "node_modules/shebang-regex": { + "version": "3.0.0", + "resolved": "https://registry.npmjs.org/shebang-regex/-/shebang-regex-3.0.0.tgz", + "integrity": "sha512-7++dFhtcx3353uBaq8DDR4NuxBetBzC7ZQOhmTQInHEd6bSrXdiEyzCvG07Z44UYdLShWUyXt5M/yhz8ekcb1A==", + "engines": { + "node": ">=8" + } }, "node_modules/signal-exit": { "version": "3.0.7", @@ -1259,52 +1569,63 @@ "integrity": "sha512-wnD2ZE+l+SPC/uoS0vXeE9L1+0wuaMqKlfz9AMUo38JsyLSBWSFcHR1Rri62LZc12vLr1gb3jl7iwQhgwpAbGQ==" }, "node_modules/slice-ansi": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/slice-ansi/-/slice-ansi-3.0.0.tgz", - "integrity": "sha512-pSyv7bSTC7ig9Dcgbw9AuRNUb5k5V6oDudjZoMBSr13qpLBG7tB+zgCkARjq7xIUgdz5P1Qe8u+rSGdouOOIyQ==", + "version": "5.0.0", + "resolved": "https://registry.npmjs.org/slice-ansi/-/slice-ansi-5.0.0.tgz", + "integrity": "sha512-FC+lgizVPfie0kkhqUScwRu1O/lF6NOgJmlCgK+/LYxDCTk8sGelYaHDhFcDN+Sn3Cv+3VSa4Byeo+IMCzpMgQ==", "dependencies": { - "ansi-styles": "^4.0.0", - "astral-regex": "^2.0.0", - "is-fullwidth-code-point": "^3.0.0" + "ansi-styles": "^6.0.0", + "is-fullwidth-code-point": "^4.0.0" }, "engines": { - "node": ">=8" + "node": ">=12" + }, + "funding": { + "url": "https://github.com/chalk/slice-ansi?sponsor=1" } }, "node_modules/slice-ansi/node_modules/ansi-styles": { - "version": "4.3.0", - "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz", - "integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==", - "dependencies": { - "color-convert": "^2.0.1" - }, + "version": "6.2.1", + "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-6.2.1.tgz", + "integrity": "sha512-bN798gFfQX+viw3R7yrGWRqnrN2oRkEkUjjl4JNn4E8GxxbjtG3FbrEIIY3l8/hrwUwIeCZvi4QuOTP4MErVug==", "engines": { - "node": ">=8" + "node": ">=12" }, "funding": { "url": "https://github.com/chalk/ansi-styles?sponsor=1" } }, + "node_modules/slice-ansi/node_modules/is-fullwidth-code-point": { + "version": "4.0.0", + "resolved": "https://registry.npmjs.org/is-fullwidth-code-point/-/is-fullwidth-code-point-4.0.0.tgz", + "integrity": "sha512-O4L094N2/dZ7xqVdrXhh9r1KODPJpFms8B5sGdJLPy664AgvXsreZUyCQQNItZRDlYug4xStLjNp/sz3HvBowQ==", + "engines": { + "node": ">=12" + }, + "funding": { + "url": "https://github.com/sponsors/sindresorhus" + } + }, "node_modules/split-ca": { "version": "1.0.1", "resolved": "https://registry.npmjs.org/split-ca/-/split-ca-1.0.1.tgz", - "integrity": "sha512-Q5thBSxp5t8WPTTJQS59LrGqOZqOsrhDGDVm8azCqIBjSBd7nd9o2PM+mDulQQkh8h//4U6hFZnc/mul8t5pWQ==" + "integrity": "sha512-Q5thBSxp5t8WPTTJQS59LrGqOZqOsrhDGDVm8azCqIBjSBd7nd9o2PM+mDulQQkh8h//4U6hFZnc/mul8t5pWQ==", + "license": "ISC" }, "node_modules/ssh2": { - "version": "1.11.0", - "resolved": "https://registry.npmjs.org/ssh2/-/ssh2-1.11.0.tgz", - "integrity": "sha512-nfg0wZWGSsfUe/IBJkXVll3PEZ//YH2guww+mP88gTpuSU4FtZN7zu9JoeTGOyCNx2dTDtT9fOpWwlzyj4uOOw==", + "version": "1.16.0", + "resolved": "https://registry.npmjs.org/ssh2/-/ssh2-1.16.0.tgz", + "integrity": "sha512-r1X4KsBGedJqo7h8F5c4Ybpcr5RjyP+aWIG007uBPRjmdQWfEiVLzSK71Zji1B9sKxwaCvD8y8cwSkYrlLiRRg==", "hasInstallScript": true, "dependencies": { - "asn1": "^0.2.4", + "asn1": "^0.2.6", "bcrypt-pbkdf": "^1.0.2" }, "engines": { "node": ">=10.16.0" }, "optionalDependencies": { - "cpu-features": "~0.0.4", - "nan": "^2.16.0" + "cpu-features": "~0.0.10", + "nan": "^2.20.0" } }, "node_modules/string_decoder": { @@ -1315,37 +1636,89 @@ "safe-buffer": "~5.2.0" } }, + "node_modules/string-width": { + "version": "4.2.3", + "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", + "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", + "dependencies": { + "emoji-regex": "^8.0.0", + "is-fullwidth-code-point": "^3.0.0", + "strip-ansi": "^6.0.1" + }, + "engines": { + "node": ">=8" + } + }, + "node_modules/string-width-cjs": { + "name": "string-width", + "version": "4.2.3", + "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", + "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", + "dependencies": { + "emoji-regex": "^8.0.0", + "is-fullwidth-code-point": "^3.0.0", + "strip-ansi": "^6.0.1" + }, + "engines": { + "node": ">=8" + } + }, + "node_modules/strip-ansi": { + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", + "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", + "dependencies": { + "ansi-regex": "^5.0.1" + }, + "engines": { + "node": ">=8" + } + }, + "node_modules/strip-ansi-cjs": { + "name": "strip-ansi", + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", + "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", + "dependencies": { + "ansi-regex": "^5.0.1" + }, + "engines": { + "node": ">=8" + } + }, "node_modules/tar": { - "version": "6.1.11", - "resolved": "https://registry.npmjs.org/tar/-/tar-6.1.11.tgz", - "integrity": "sha512-an/KZQzQUkZCkuoAA64hM92X0Urb6VpRhAFllDzz44U2mcD5scmT3zBc4VgVpkugF580+DQn8eAFSyoQt0tznA==", + "version": "7.4.3", + "resolved": "https://registry.npmjs.org/tar/-/tar-7.4.3.tgz", + "integrity": "sha512-5S7Va8hKfV7W5U6g3aYxXmlPoZVAwUMy9AOKyF2fVuZa2UD3qZjg578OrLRt8PcNN1PleVaL/5/yYATNL0ICUw==", "dependencies": { - "chownr": "^2.0.0", - "fs-minipass": "^2.0.0", - "minipass": "^3.0.0", - "minizlib": "^2.1.1", - "mkdirp": "^1.0.3", - "yallist": "^4.0.0" + "@isaacs/fs-minipass": "^4.0.0", + "chownr": "^3.0.0", + "minipass": "^7.1.2", + "minizlib": "^3.0.1", + "mkdirp": "^3.0.1", + "yallist": "^5.0.0" }, "engines": { - "node": ">= 10" + "node": ">=18" } }, "node_modules/tar-fs": { - "version": "2.0.1", - "resolved": "https://registry.npmjs.org/tar-fs/-/tar-fs-2.0.1.tgz", - "integrity": "sha512-6tzWDMeroL87uF/+lin46k+Q+46rAJ0SyPGz7OW7wTgblI273hsBqk2C1j0/xNadNLKDTUL9BukSjB7cwgmlPA==", + "version": "2.1.2", + "resolved": "https://registry.npmjs.org/tar-fs/-/tar-fs-2.1.2.tgz", + "integrity": "sha512-EsaAXwxmx8UB7FRKqeozqEPop69DXcmYwTQwXvyAPF352HJsPdkVhvTaDPYqfNgruveJIJy3TA2l+2zj8LJIJA==", + "license": "MIT", "dependencies": { "chownr": "^1.1.1", "mkdirp-classic": "^0.5.2", "pump": "^3.0.0", - "tar-stream": "^2.0.0" + "tar-stream": "^2.1.4" } }, "node_modules/tar-stream": { "version": "2.2.0", "resolved": "https://registry.npmjs.org/tar-stream/-/tar-stream-2.2.0.tgz", "integrity": "sha512-ujeqbceABgwMZxEJnk2HDY2DlnUZ+9oEcb1KzTVfYHio0UE6dG71n60d8D2I4qNvleWrrXpmjpt7vZeF1LnMZQ==", + "license": "MIT", "dependencies": { "bl": "^4.0.3", "end-of-stream": "^1.4.1", @@ -1357,61 +1730,30 @@ "node": ">=6" } }, - "node_modules/tar-stream/node_modules/bl": { - "version": "4.1.0", - "resolved": "https://registry.npmjs.org/bl/-/bl-4.1.0.tgz", - "integrity": "sha512-1W07cM9gS6DcLperZfFSj+bWLtaPGSOHWhPiGzXmvVJbRLdG82sH/Kn8EtW1VqWVA54AKf2h5k5BbnIbwF3h6w==", - "dependencies": { - "buffer": "^5.5.0", - "inherits": "^2.0.4", - "readable-stream": "^3.4.0" - } - }, - "node_modules/tar-stream/node_modules/buffer": { - "version": "5.7.1", - "resolved": "https://registry.npmjs.org/buffer/-/buffer-5.7.1.tgz", - "integrity": "sha512-EHcyIPBQ4BSGlvjB16k5KgAJ27CIsHY/2JBmCRReo48y9rQ3MaUzWX3KVlBa4U7MyX02HdVj0K7C3WaB3ju7FQ==", - "funding": [ - { - "type": "github", - "url": "https://github.com/sponsors/feross" - }, - { - "type": "patreon", - "url": "https://www.patreon.com/feross" - }, - { - "type": "consulting", - "url": "https://feross.org/support" - } - ], - "dependencies": { - "base64-js": "^1.3.1", - "ieee754": "^1.1.13" - } - }, "node_modules/tar/node_modules/chownr": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/chownr/-/chownr-2.0.0.tgz", - "integrity": "sha512-bIomtDF5KGpdogkLd9VspvFzk9KfpyyGlS8YFVZl7TGPBHL5snIOnxeshwVgPteQ9b4Eydl+pVbIyE1DcvCWgQ==", + "version": "3.0.0", + "resolved": "https://registry.npmjs.org/chownr/-/chownr-3.0.0.tgz", + "integrity": "sha512-+IxzY9BZOQd/XuYPRmrvEVjF/nqj5kgT4kEq7VofrDoM1MxoRjEWkrCC3EtLi59TVawxTAn+orJwFQcrqEN1+g==", "engines": { - "node": ">=10" + "node": ">=18" } }, - "node_modules/through": { - "version": "2.3.8", - "resolved": "https://registry.npmjs.org/through/-/through-2.3.8.tgz", - "integrity": "sha512-w89qg7PI8wAdvX60bMDP+bFoD5Dvhm9oLheFp5O4a2QF0cSBGsBX4qZmadPMvVqlLJBBci+WqGGOAPvcDeNSVg==" - }, - "node_modules/tslib": { - "version": "2.4.0", - "resolved": "https://registry.npmjs.org/tslib/-/tslib-2.4.0.tgz", - "integrity": "sha512-d6xOpEDfsi2CZVlPQzGeux8XMwLT9hssAsaPYExaQMuYskwb+x1x7J371tWlbBdWHroy99KnVB6qIkUbs5X3UQ==" - }, "node_modules/tweetnacl": { "version": "0.14.5", "resolved": "https://registry.npmjs.org/tweetnacl/-/tweetnacl-0.14.5.tgz", - "integrity": "sha512-KXXFFdAbFXY4geFIwoyNK+f5Z1b7swfXABfL7HXCmoIWMKU3dmS26672A4EeQtDzLKy7SXmfBu51JolvEKwtGA==" + "integrity": "sha512-KXXFFdAbFXY4geFIwoyNK+f5Z1b7swfXABfL7HXCmoIWMKU3dmS26672A4EeQtDzLKy7SXmfBu51JolvEKwtGA==", + "license": "Unlicense" + }, + "node_modules/type-fest": { + "version": "1.4.0", + "resolved": "https://registry.npmjs.org/type-fest/-/type-fest-1.4.0.tgz", + "integrity": "sha512-yGSza74xk0UG8k+pLh5oeoYirvIiWo5t0/o3zHHAO2tRDiZcxWP7fywNlXhqb6/r6sWvwi+RsyQMWhVLe4BVuA==", + "engines": { + "node": ">=10" + }, + "funding": { + "url": "https://github.com/sponsors/sindresorhus" + } }, "node_modules/universalify": { "version": "2.0.0", @@ -1426,6 +1768,66 @@ "resolved": "https://registry.npmjs.org/util-deprecate/-/util-deprecate-1.0.2.tgz", "integrity": "sha512-EPD5q1uXyFxJpCrLnCc1nHnq3gOa6DZBocAIiI2TaSCA7VCJ1UJDMagCzIkXNsUYfD1daK//LTEQ8xiIbrHtcw==" }, + "node_modules/uuid": { + "version": "10.0.0", + "resolved": "https://registry.npmjs.org/uuid/-/uuid-10.0.0.tgz", + "integrity": "sha512-8XkAphELsDnEGrDxUOHB3RGvXz6TeuYSGEZBOjtTtPm2lwhGBjLgOzLHB63IUWfBpNucQjND6d3AOudO+H3RWQ==", + "funding": [ + "https://github.com/sponsors/broofa", + "https://github.com/sponsors/ctavan" + ], + "license": "MIT", + "bin": { + "uuid": "dist/bin/uuid" + } + }, + "node_modules/which": { + "version": "2.0.2", + "resolved": "https://registry.npmjs.org/which/-/which-2.0.2.tgz", + "integrity": "sha512-BLI3Tl1TW3Pvl70l3yq3Y64i+awpwXqsGBYWkkqMtnbXgrMD+yj7rhW0kuEDxzJaYXGjEW5ogapKNMEKNMjibA==", + "dependencies": { + "isexe": "^2.0.0" + }, + "bin": { + "node-which": "bin/node-which" + }, + "engines": { + "node": ">= 8" + } + }, + "node_modules/wrap-ansi": { + "version": "7.0.0", + "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-7.0.0.tgz", + "integrity": "sha512-YVGIj2kamLSTxw6NsZjoBxfSwsn0ycdesmc4p+Q21c5zPuZ1pl+NfxVdxPtdHvmNVOQ6XSYG4AUtyt/Fi7D16Q==", + "dependencies": { + "ansi-styles": "^4.0.0", + "string-width": "^4.1.0", + "strip-ansi": "^6.0.0" + }, + "engines": { + "node": ">=10" + }, + "funding": { + "url": "https://github.com/chalk/wrap-ansi?sponsor=1" + } + }, + "node_modules/wrap-ansi-cjs": { + "name": "wrap-ansi", + "version": "7.0.0", + "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-7.0.0.tgz", + "integrity": "sha512-YVGIj2kamLSTxw6NsZjoBxfSwsn0ycdesmc4p+Q21c5zPuZ1pl+NfxVdxPtdHvmNVOQ6XSYG4AUtyt/Fi7D16Q==", + "dependencies": { + "ansi-styles": "^4.0.0", + "string-width": "^4.1.0", + "strip-ansi": "^6.0.0" + }, + "engines": { + "node": ">=10" + }, + "funding": { + "url": "https://github.com/chalk/wrap-ansi?sponsor=1" + } + }, "node_modules/wrappy": { "version": "1.0.2", "resolved": "https://registry.npmjs.org/wrappy/-/wrappy-1.0.2.tgz", @@ -1440,22 +1842,25 @@ } }, "node_modules/yallist": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/yallist/-/yallist-4.0.0.tgz", - "integrity": "sha512-3wdGidZyq5PB084XLES5TpOSRA3wjXAlIWMhum2kRcv/41Sn2emQ0dycQW4uZXLejwKvg6EsvbdlVL+FYEct7A==" + "version": "5.0.0", + "resolved": "https://registry.npmjs.org/yallist/-/yallist-5.0.0.tgz", + "integrity": "sha512-YgvUTfwqyc7UXVMrB+SImsVYSmTS8X/tSrtdNZMImM+n7+QTriRXyXim0mBrTXNeqzVF0KWGgHPeiyViFFrNDw==", + "engines": { + "node": ">=18" + } }, "node_modules/yargs": { - "version": "17.5.1", - "resolved": "https://registry.npmjs.org/yargs/-/yargs-17.5.1.tgz", - "integrity": "sha512-t6YAJcxDkNX7NFYiVtKvWUz8l+PaKTLiL63mJYWR2GnHq2gjEWISzsLp9wg3aY36dY1j+gfIEL3pIF+XlJJfbA==", + "version": "17.7.2", + "resolved": "https://registry.npmjs.org/yargs/-/yargs-17.7.2.tgz", + "integrity": "sha512-7dSzzRQ++CKnNI/krKnYRV7JKKPUXMEh61soaHKg9mrWEhzFWhFnxPxGl+69cD1Ou63C13NUPCnmIcrvqCuM6w==", "dependencies": { - "cliui": "^7.0.2", + "cliui": "^8.0.1", "escalade": "^3.1.1", "get-caller-file": "^2.0.5", "require-directory": "^2.1.1", "string-width": "^4.2.3", "y18n": "^5.0.5", - "yargs-parser": "^21.0.0" + "yargs-parser": "^21.1.1" }, "engines": { "node": ">=12" @@ -1469,43 +1874,6 @@ "node": ">=12" } }, - "node_modules/yargs/node_modules/ansi-regex": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", - "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==", - "engines": { - "node": ">=8" - } - }, - "node_modules/yargs/node_modules/emoji-regex": { - "version": "8.0.0", - "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", - "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" - }, - "node_modules/yargs/node_modules/string-width": { - "version": "4.2.3", - "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", - "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", - "dependencies": { - "emoji-regex": "^8.0.0", - "is-fullwidth-code-point": "^3.0.0", - "strip-ansi": "^6.0.1" - }, - "engines": { - "node": ">=8" - } - }, - "node_modules/yargs/node_modules/strip-ansi": { - "version": "6.0.1", - "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", - "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", - "dependencies": { - "ansi-regex": "^5.0.1" - }, - "engines": { - "node": ">=8" - } - }, "node_modules/yauzl": { "version": "2.10.0", "resolved": "https://registry.npmjs.org/yauzl/-/yauzl-2.10.0.tgz", @@ -1517,6 +1885,162 @@ } }, "dependencies": { + "@balena/dockerignore": { + "version": "1.0.2", + "resolved": "https://registry.npmjs.org/@balena/dockerignore/-/dockerignore-1.0.2.tgz", + "integrity": "sha512-wMue2Sy4GAVTk6Ic4tJVcnfdau+gx2EnG7S+uAEe+TWJFqE4YoWN4/H8MSLj4eYJKxGg26lZwboEniNiNwZQ6Q==" + }, + "@grpc/grpc-js": { + "version": "1.12.5", + "resolved": "https://registry.npmjs.org/@grpc/grpc-js/-/grpc-js-1.12.5.tgz", + "integrity": "sha512-d3iiHxdpg5+ZcJ6jnDSOT8Z0O0VMVGy34jAnYLUX8yd36b1qn8f1TwOA/Lc7TsOh03IkPJ38eGI5qD2EjNkoEA==", + "requires": { + "@grpc/proto-loader": "^0.7.13", + "@js-sdsl/ordered-map": "^4.4.2" + } + }, + "@grpc/proto-loader": { + "version": "0.7.13", + "resolved": "https://registry.npmjs.org/@grpc/proto-loader/-/proto-loader-0.7.13.tgz", + "integrity": "sha512-AiXO/bfe9bmxBjxxtYxFAXGZvMaN5s8kO+jBHAJCON8rJoB5YS/D6X7ZNc6XQkuHNmyl4CYaMI1fJ/Gn27RGGw==", + "requires": { + "lodash.camelcase": "^4.3.0", + "long": "^5.0.0", + "protobufjs": "^7.2.5", + "yargs": "^17.7.2" + } + }, + "@isaacs/cliui": { + "version": "8.0.2", + "resolved": "https://registry.npmjs.org/@isaacs/cliui/-/cliui-8.0.2.tgz", + "integrity": "sha512-O8jcjabXaleOG9DQ0+ARXWZBTfnP4WNAqzuiJK7ll44AmxGKv/J2M4TPjxjY3znBCfvBXFzucm1twdyFybFqEA==", + "requires": { + "string-width": "^5.1.2", + "string-width-cjs": "npm:string-width@^4.2.0", + "strip-ansi": "^7.0.1", + "strip-ansi-cjs": "npm:strip-ansi@^6.0.1", + "wrap-ansi": "^8.1.0", + "wrap-ansi-cjs": "npm:wrap-ansi@^7.0.0" + }, + "dependencies": { + "ansi-regex": { + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-6.0.1.tgz", + "integrity": "sha512-n5M855fKb2SsfMIiFFoVrABHJC8QtHwVx+mHWP3QcEqBHYienj5dHSgjbxtC0WEZXYt4wcD6zrQElDPhFuZgfA==" + }, + "ansi-styles": { + "version": "6.2.1", + "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-6.2.1.tgz", + "integrity": "sha512-bN798gFfQX+viw3R7yrGWRqnrN2oRkEkUjjl4JNn4E8GxxbjtG3FbrEIIY3l8/hrwUwIeCZvi4QuOTP4MErVug==" + }, + "emoji-regex": { + "version": "9.2.2", + "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-9.2.2.tgz", + "integrity": "sha512-L18DaJsXSUk2+42pv8mLs5jJT2hqFkFE4j21wOmgbUqsZ2hL72NsUU785g9RXgo3s0ZNgVl42TiHp3ZtOv/Vyg==" + }, + "string-width": { + "version": "5.1.2", + "resolved": "https://registry.npmjs.org/string-width/-/string-width-5.1.2.tgz", + "integrity": "sha512-HnLOCR3vjcY8beoNLtcjZ5/nxn2afmME6lhrDrebokqMap+XbeW8n9TXpPDOqdGK5qcI3oT0GKTW6wC7EMiVqA==", + "requires": { + "eastasianwidth": "^0.2.0", + "emoji-regex": "^9.2.2", + "strip-ansi": "^7.0.1" + } + }, + "strip-ansi": { + "version": "7.1.0", + "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-7.1.0.tgz", + "integrity": "sha512-iq6eVVI64nQQTRYq2KtEg2d2uU7LElhTJwsH4YzIHZshxlgZms/wIc4VoDQTlG/IvVIrBKG06CrZnp0qv7hkcQ==", + "requires": { + "ansi-regex": "^6.0.1" + } + }, + "wrap-ansi": { + "version": "8.1.0", + "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-8.1.0.tgz", + "integrity": "sha512-si7QWI6zUMq56bESFvagtmzMdGOtoxfR+Sez11Mobfc7tm+VkUckk9bW2UeffTGVUbOksxmSw0AA2gs8g71NCQ==", + "requires": { + "ansi-styles": "^6.1.0", + "string-width": "^5.0.1", + "strip-ansi": "^7.0.1" + } + } + } + }, + "@isaacs/fs-minipass": { + "version": "4.0.1", + "resolved": "https://registry.npmjs.org/@isaacs/fs-minipass/-/fs-minipass-4.0.1.tgz", + "integrity": "sha512-wgm9Ehl2jpeqP3zw/7mo3kRHFp5MEDhqAdwy1fTGkHAwnkGOVsgpvQhL8B5n1qlb01jV3n/bI0ZfZp5lWA1k4w==", + "requires": { + "minipass": "^7.0.4" + } + }, + "@js-sdsl/ordered-map": { + "version": "4.4.2", + "resolved": "https://registry.npmjs.org/@js-sdsl/ordered-map/-/ordered-map-4.4.2.tgz", + "integrity": "sha512-iUKgm52T8HOE/makSxjqoWhe95ZJA1/G1sYsGev2JDKUSS14KAgg1LHb+Ba+IPow0xflbnSkOsZcO08C7w1gYw==" + }, + "@pkgjs/parseargs": { + "version": "0.11.0", + "resolved": "https://registry.npmjs.org/@pkgjs/parseargs/-/parseargs-0.11.0.tgz", + "integrity": "sha512-+1VkjdD0QBLPodGrJUeqarH8VAIvQODIbwh9XpP5Syisf7YoQgsJKPNFoqqLQlu+VQ/tVSshMR6loPMn8U+dPg==", + "optional": true + }, + "@protobufjs/aspromise": { + "version": "1.1.2", + "resolved": "https://registry.npmjs.org/@protobufjs/aspromise/-/aspromise-1.1.2.tgz", + "integrity": "sha512-j+gKExEuLmKwvz3OgROXtrJ2UG2x8Ch2YZUxahh+s1F2HZ+wAceUNLkvy6zKCPVRkU++ZWQrdxsUeQXmcg4uoQ==" + }, + "@protobufjs/base64": { + "version": "1.1.2", + "resolved": "https://registry.npmjs.org/@protobufjs/base64/-/base64-1.1.2.tgz", + "integrity": "sha512-AZkcAA5vnN/v4PDqKyMR5lx7hZttPDgClv83E//FMNhR2TMcLUhfRUBHCmSl0oi9zMgDDqRUJkSxO3wm85+XLg==" + }, + "@protobufjs/codegen": { + "version": "2.0.4", + "resolved": "https://registry.npmjs.org/@protobufjs/codegen/-/codegen-2.0.4.tgz", + "integrity": "sha512-YyFaikqM5sH0ziFZCN3xDC7zeGaB/d0IUb9CATugHWbd1FRFwWwt4ld4OYMPWu5a3Xe01mGAULCdqhMlPl29Jg==" + }, + "@protobufjs/eventemitter": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/eventemitter/-/eventemitter-1.1.0.tgz", + "integrity": "sha512-j9ednRT81vYJ9OfVuXG6ERSTdEL1xVsNgqpkxMsbIabzSo3goCjDIveeGv5d03om39ML71RdmrGNjG5SReBP/Q==" + }, + "@protobufjs/fetch": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/fetch/-/fetch-1.1.0.tgz", + "integrity": "sha512-lljVXpqXebpsijW71PZaCYeIcE5on1w5DlQy5WH6GLbFryLUrBD4932W/E2BSpfRJWseIL4v/KPgBFxDOIdKpQ==", + "requires": { + "@protobufjs/aspromise": "^1.1.1", + "@protobufjs/inquire": "^1.1.0" + } + }, + "@protobufjs/float": { + "version": "1.0.2", + "resolved": "https://registry.npmjs.org/@protobufjs/float/-/float-1.0.2.tgz", + "integrity": "sha512-Ddb+kVXlXst9d+R9PfTIxh1EdNkgoRe5tOX6t01f1lYWOvJnSPDBlG241QLzcyPdoNTsblLUdujGSE4RzrTZGQ==" + }, + "@protobufjs/inquire": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/inquire/-/inquire-1.1.0.tgz", + "integrity": "sha512-kdSefcPdruJiFMVSbn801t4vFK7KB/5gd2fYvrxhuJYg8ILrmn9SKSX2tZdV6V+ksulWqS7aXjBcRXl3wHoD9Q==" + }, + "@protobufjs/path": { + "version": "1.1.2", + "resolved": "https://registry.npmjs.org/@protobufjs/path/-/path-1.1.2.tgz", + "integrity": "sha512-6JOcJ5Tm08dOHAbdR3GrvP+yUUfkjG5ePsHYczMFLq3ZmMkAD98cDgcT2iA1lJ9NVwFd4tH/iSSoe44YWkltEA==" + }, + "@protobufjs/pool": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/pool/-/pool-1.1.0.tgz", + "integrity": "sha512-0kELaGSIDBKvcgS4zkjz1PeddatrjYcmMWOlAuAPwAeccUrPHdUqo/J6LiymHHEiJT5NrF1UVwxY14f+fy4WQw==" + }, + "@protobufjs/utf8": { + "version": "1.1.0", + "resolved": "https://registry.npmjs.org/@protobufjs/utf8/-/utf8-1.1.0.tgz", + "integrity": "sha512-Vvn3zZrhQZkkBE8LSuW3em98c0FwgO4nxzv6OdSxPKJIEKY2bGbHn+mhGIPerzI4twdxaP8/0+06HBpwf345Lw==" + }, "@sindresorhus/is": { "version": "5.3.0", "resolved": "https://registry.npmjs.org/@sindresorhus/is/-/is-5.3.0.tgz", @@ -1530,48 +2054,16 @@ "defer-to-connect": "^2.0.1" } }, - "@types/cacheable-request": { - "version": "6.0.2", - "resolved": "https://registry.npmjs.org/@types/cacheable-request/-/cacheable-request-6.0.2.tgz", - "integrity": "sha512-B3xVo+dlKM6nnKTcmm5ZtY/OL8bOAOd2Olee9M1zft65ox50OzjEHW91sDiU9j6cvW8Ejg1/Qkf4xd2kugApUA==", - "requires": { - "@types/http-cache-semantics": "*", - "@types/keyv": "*", - "@types/node": "*", - "@types/responselike": "*" - } - }, "@types/http-cache-semantics": { "version": "4.0.1", "resolved": "https://registry.npmjs.org/@types/http-cache-semantics/-/http-cache-semantics-4.0.1.tgz", "integrity": "sha512-SZs7ekbP8CN0txVG2xVRH6EgKmEm31BOxA07vkFaETzZz1xh+cbt8BcI0slpymvwhx5dlFnQG2rTlPVQn+iRPQ==" }, - "@types/json-buffer": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/@types/json-buffer/-/json-buffer-3.0.0.tgz", - "integrity": "sha512-3YP80IxxFJB4b5tYC2SUPwkg0XQLiu0nWvhRgEatgjf+29IcWO9X1k8xRv5DGssJ/lCrjYTjQPcobJr2yWIVuQ==" - }, - "@types/keyv": { - "version": "3.1.4", - "resolved": "https://registry.npmjs.org/@types/keyv/-/keyv-3.1.4.tgz", - "integrity": "sha512-BQ5aZNSCpj7D6K2ksrRCTmKRLEpnPvWDiLPfoGyhZ++8YtiK9d/3DBKPJgry359X/P1PfruyYwvnvwFjuEiEIg==", - "requires": { - "@types/node": "*" - } - }, "@types/node": { "version": "18.6.5", "resolved": "https://registry.npmjs.org/@types/node/-/node-18.6.5.tgz", "integrity": "sha512-Xjt5ZGUa5WusGZJ4WJPbOT8QOqp6nDynVFRKcUt32bOgvXEoc6o085WNkYTMO7ifAj2isEfQQ2cseE+wT6jsRw==" }, - "@types/responselike": { - "version": "1.0.0", - "resolved": "https://registry.npmjs.org/@types/responselike/-/responselike-1.0.0.tgz", - "integrity": "sha512-85Y2BjiufFzaMIlvJDvTTB8Fxl2xfLo4HgmHzVBz08w4wDePCTjYw66PdrolO0kzli3yam/YCgRufyo1DdQVTA==", - "requires": { - "@types/node": "*" - } - }, "@types/yauzl": { "version": "2.10.0", "resolved": "https://registry.npmjs.org/@types/yauzl/-/yauzl-2.10.0.tgz", @@ -1581,20 +2073,32 @@ "@types/node": "*" } }, - "aggregate-error": { - "version": "3.1.0", - "resolved": "https://registry.npmjs.org/aggregate-error/-/aggregate-error-3.1.0.tgz", - "integrity": "sha512-4I7Td01quW/RpocfNayFdFVk1qSuoh0E7JrbRJ16nH01HhKFQ88INq9Sd+nd72zqRySlr9BmDA8xlEJ6vJMrYA==", - "requires": { - "clean-stack": "^2.0.0", - "indent-string": "^4.0.0" - } - }, "ansi-colors": { "version": "4.1.3", "resolved": "https://registry.npmjs.org/ansi-colors/-/ansi-colors-4.1.3.tgz", "integrity": "sha512-/6w/C21Pm1A7aZitlI5Ni/2J6FFQN8i1Cvz3kHABAAbw93v/NlvKdVOqz7CCWz/3iv/JplRSEEZ83XION15ovw==" }, + "ansi-escapes": { + "version": "5.0.0", + "resolved": "https://registry.npmjs.org/ansi-escapes/-/ansi-escapes-5.0.0.tgz", + "integrity": "sha512-5GFMVX8HqE/TB+FuBJGuO5XG0WrsA6ptUqoODaT/n9mmUaZFkqnBueB4leqGBCmrUHnCnC4PCZTCd0E7QQ83bA==", + "requires": { + "type-fest": "^1.0.2" + } + }, + "ansi-regex": { + "version": "5.0.1", + "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", + "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==" + }, + "ansi-styles": { + "version": "4.3.0", + "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz", + "integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==", + "requires": { + "color-convert": "^2.0.1" + } + }, "asn1": { "version": "0.2.6", "resolved": "https://registry.npmjs.org/asn1/-/asn1-0.2.6.tgz", @@ -1603,22 +2107,49 @@ "safer-buffer": "~2.1.0" } }, - "astral-regex": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/astral-regex/-/astral-regex-2.0.0.tgz", - "integrity": "sha512-Z7tMw1ytTXt5jqMcOP+OQteU1VuNK9Y02uuJtKQ1Sv69jXQKKg5cibLwGJow8yzZP+eAc18EmLGPal0bp36rvQ==" + "balanced-match": { + "version": "1.0.2", + "resolved": "https://registry.npmjs.org/balanced-match/-/balanced-match-1.0.2.tgz", + "integrity": "sha512-3oSeUO0TMV67hN1AmbXsK4yaqU7tjiHlbxRDZOpH0KW9+CeX4bRAaX0Anxt0tx2MrpRpWwQaPwIlISEJhYU5Pw==" }, "base64-js": { "version": "1.5.1", "resolved": "https://registry.npmjs.org/base64-js/-/base64-js-1.5.1.tgz", "integrity": "sha512-AKpaYlHn8t4SVbOHCy+b5+KKgvR4vrsD8vbvrbiQJps7fKDTkjkDry6ji0rUJjC0kzbNePLwzxq8iypo41qeWA==" }, - "bcrypt-pbkdf": { - "version": "1.0.2", - "resolved": "https://registry.npmjs.org/bcrypt-pbkdf/-/bcrypt-pbkdf-1.0.2.tgz", - "integrity": "sha512-qeFIXtP4MSoi6NLqO12WfqARWWuCKi2Rn/9hJLEmtB5yTNr9DqFWkJRCf2qShWzPeAMRnOgCrq0sg/KLv5ES9w==", + "bcrypt-pbkdf": { + "version": "1.0.2", + "resolved": "https://registry.npmjs.org/bcrypt-pbkdf/-/bcrypt-pbkdf-1.0.2.tgz", + "integrity": "sha512-qeFIXtP4MSoi6NLqO12WfqARWWuCKi2Rn/9hJLEmtB5yTNr9DqFWkJRCf2qShWzPeAMRnOgCrq0sg/KLv5ES9w==", + "requires": { + "tweetnacl": "^0.14.3" + } + }, + "bl": { + "version": "4.1.0", + "resolved": "https://registry.npmjs.org/bl/-/bl-4.1.0.tgz", + "integrity": "sha512-1W07cM9gS6DcLperZfFSj+bWLtaPGSOHWhPiGzXmvVJbRLdG82sH/Kn8EtW1VqWVA54AKf2h5k5BbnIbwF3h6w==", + "requires": { + "buffer": "^5.5.0", + "inherits": "^2.0.4", + "readable-stream": "^3.4.0" + } + }, + "brace-expansion": { + "version": "2.0.1", + "resolved": "https://registry.npmjs.org/brace-expansion/-/brace-expansion-2.0.1.tgz", + "integrity": "sha512-XnAIvQ8eM+kC6aULx6wuQiwVsnzsi9d3WxzV3FpWTGA19F621kwdbsAcFKXgKUHZWsy+mY6iL1sHTxWEFCytDA==", "requires": { - "tweetnacl": "^0.14.3" + "balanced-match": "^1.0.0" + } + }, + "buffer": { + "version": "5.7.1", + "resolved": "https://registry.npmjs.org/buffer/-/buffer-5.7.1.tgz", + "integrity": "sha512-EHcyIPBQ4BSGlvjB16k5KgAJ27CIsHY/2JBmCRReo48y9rQ3MaUzWX3KVlBa4U7MyX02HdVj0K7C3WaB3ju7FQ==", + "requires": { + "base64-js": "^1.3.1", + "ieee754": "^1.1.13" } }, "buffer-crc32": { @@ -1627,163 +2158,95 @@ "integrity": "sha512-VO9Ht/+p3SN7SKWqcrgEzjGbRSJYTx+Q1pTQC0wrWqHx0vpJraQ6GtHx8tvcg1rlK1byhU5gccxgOgj7B0TDkQ==" }, "buildcheck": { - "version": "0.0.3", - "resolved": "https://registry.npmjs.org/buildcheck/-/buildcheck-0.0.3.tgz", - "integrity": "sha512-pziaA+p/wdVImfcbsZLNF32EiWyujlQLwolMqUQE8xpKNOH7KmZQaY8sXN7DGOEzPAElo9QTaeNRfGnf3iOJbA==", + "version": "0.0.6", + "resolved": "https://registry.npmjs.org/buildcheck/-/buildcheck-0.0.6.tgz", + "integrity": "sha512-8f9ZJCUXyT1M35Jx7MkBgmBMo3oHTTBIPLiY9xyL0pl3T5RwcPEY8cUHr5LBNfu/fk6c2T4DJZuVM/8ZZT2D2A==", "optional": true }, "cacheable-lookup": { - "version": "6.1.0", - "resolved": "https://registry.npmjs.org/cacheable-lookup/-/cacheable-lookup-6.1.0.tgz", - "integrity": "sha512-KJ/Dmo1lDDhmW2XDPMo+9oiy/CeqosPguPCrgcVzKyZrL6pM1gU2GmPY/xo6OQPTUaA/c0kwHuywB4E6nmT9ww==" + "version": "7.0.0", + "resolved": "https://registry.npmjs.org/cacheable-lookup/-/cacheable-lookup-7.0.0.tgz", + "integrity": "sha512-+qJyx4xiKra8mZrcwhjMRMUhD5NR1R8esPkzIYxX96JiecFoxAXFuz/GpR3+ev4PE1WamHip78wV0vcmPQtp8w==" }, "cacheable-request": { - "version": "7.0.2", - "resolved": "https://registry.npmjs.org/cacheable-request/-/cacheable-request-7.0.2.tgz", - "integrity": "sha512-pouW8/FmiPQbuGpkXQ9BAPv/Mo5xDGANgSNXzTzJ8DrKGuXOssM4wIQRjfanNRh3Yu5cfYPvcorqbhg2KIJtew==", + "version": "10.2.8", + "resolved": "https://registry.npmjs.org/cacheable-request/-/cacheable-request-10.2.8.tgz", + "integrity": "sha512-IDVO5MJ4LItE6HKFQTqT2ocAQsisOoCTUDu1ddCmnhyiwFQjXNPp4081Xj23N4tO+AFEFNzGuNEf/c8Gwwt15A==", "requires": { - "clone-response": "^1.0.2", - "get-stream": "^5.1.0", - "http-cache-semantics": "^4.0.0", - "keyv": "^4.0.0", - "lowercase-keys": "^2.0.0", - "normalize-url": "^6.0.1", - "responselike": "^2.0.0" - }, - "dependencies": { - "get-stream": { - "version": "5.2.0", - "resolved": "https://registry.npmjs.org/get-stream/-/get-stream-5.2.0.tgz", - "integrity": "sha512-nBF+F1rAZVCu/p7rjzgA+Yb4lfYXrpl7a6VmJrU8wF9I1CKvP/QwPNZHnOlwbTkY6dvtFIzFMSyQXbLoTQPRpA==", - "requires": { - "pump": "^3.0.0" - } - }, - "lowercase-keys": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/lowercase-keys/-/lowercase-keys-2.0.0.tgz", - "integrity": "sha512-tqNXrS78oMOE73NMxK4EMLQsQowWf8jKooH9g7xPavRT706R6bkQJ6DY2Te7QukaZsulxa30wQ7bk0pm4XiHmA==" - } + "@types/http-cache-semantics": "^4.0.1", + "get-stream": "^6.0.1", + "http-cache-semantics": "^4.1.1", + "keyv": "^4.5.2", + "mimic-response": "^4.0.0", + "normalize-url": "^8.0.0", + "responselike": "^3.0.0" } }, "chalk": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/chalk/-/chalk-5.0.1.tgz", - "integrity": "sha512-Fo07WOYGqMfCWHOzSXOt2CxDbC6skS/jO9ynEcmpANMoPrD+W1r1K6Vx7iNm+AQmETU1Xr2t+n8nzkV9t6xh3w==" + "version": "5.4.1", + "resolved": "https://registry.npmjs.org/chalk/-/chalk-5.4.1.tgz", + "integrity": "sha512-zgVZuo2WcZgfUEmsn6eO3kINexW8RAE4maiQ8QNs8CtpPCSyMiYsULR3HQYkm3w8FIA3SberyMJMSldGsW+U3w==" }, "chownr": { "version": "1.1.4", "resolved": "https://registry.npmjs.org/chownr/-/chownr-1.1.4.tgz", "integrity": "sha512-jJ0bqzaylmJtVnNgzTeSOs8DPavpbYgEr/b0YL8/2GO3xJEhInFmhKMUnEJQjZumK7KXGFhUy89PrsJWlakBVg==" }, - "clean-stack": { - "version": "2.2.0", - "resolved": "https://registry.npmjs.org/clean-stack/-/clean-stack-2.2.0.tgz", - "integrity": "sha512-4diC9HaTE+KRAMWhDhrGOECgWZxoevMc5TlkObMqNSsVU62PYzXZ/SMTjzyGAFF1YusgxGcSWTEXBhp0CPwQ1A==" + "cli-cursor": { + "version": "4.0.0", + "resolved": "https://registry.npmjs.org/cli-cursor/-/cli-cursor-4.0.0.tgz", + "integrity": "sha512-VGtlMu3x/4DOtIUwEkRezxUZ2lBacNJCHash0N0WeZDBS+7Ux1dm3XWAgWYxLJFMMdOeXMHXorshEFhbMSGelg==", + "requires": { + "restore-cursor": "^4.0.0" + } }, "cli-truncate": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/cli-truncate/-/cli-truncate-2.1.0.tgz", - "integrity": "sha512-n8fOixwDD6b/ObinzTrp1ZKFzbgvKZvuz/TvejnLn1aQfC6r52XEx85FmuC+3HI+JM7coBRXUvNqEU2PHVrHpg==", + "version": "3.1.0", + "resolved": "https://registry.npmjs.org/cli-truncate/-/cli-truncate-3.1.0.tgz", + "integrity": "sha512-wfOBkjXteqSnI59oPcJkcPl/ZmwvMMOj340qUIY1SKZCv0B9Cf4D4fAucRkIKQmsIuYK3x1rrgU7MeGRruiuiA==", "requires": { - "slice-ansi": "^3.0.0", - "string-width": "^4.2.0" + "slice-ansi": "^5.0.0", + "string-width": "^5.0.0" }, "dependencies": { "ansi-regex": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", - "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==" + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-6.0.1.tgz", + "integrity": "sha512-n5M855fKb2SsfMIiFFoVrABHJC8QtHwVx+mHWP3QcEqBHYienj5dHSgjbxtC0WEZXYt4wcD6zrQElDPhFuZgfA==" }, "emoji-regex": { - "version": "8.0.0", - "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", - "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" + "version": "9.2.2", + "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-9.2.2.tgz", + "integrity": "sha512-L18DaJsXSUk2+42pv8mLs5jJT2hqFkFE4j21wOmgbUqsZ2hL72NsUU785g9RXgo3s0ZNgVl42TiHp3ZtOv/Vyg==" }, "string-width": { - "version": "4.2.3", - "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", - "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", + "version": "5.1.2", + "resolved": "https://registry.npmjs.org/string-width/-/string-width-5.1.2.tgz", + "integrity": "sha512-HnLOCR3vjcY8beoNLtcjZ5/nxn2afmME6lhrDrebokqMap+XbeW8n9TXpPDOqdGK5qcI3oT0GKTW6wC7EMiVqA==", "requires": { - "emoji-regex": "^8.0.0", - "is-fullwidth-code-point": "^3.0.0", - "strip-ansi": "^6.0.1" + "eastasianwidth": "^0.2.0", + "emoji-regex": "^9.2.2", + "strip-ansi": "^7.0.1" } }, "strip-ansi": { - "version": "6.0.1", - "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", - "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", + "version": "7.0.1", + "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-7.0.1.tgz", + "integrity": "sha512-cXNxvT8dFNRVfhVME3JAe98mkXDYN2O1l7jmcwMnOslDeESg1rF/OZMtK0nRAhiari1unG5cD4jG3rapUAkLbw==", "requires": { - "ansi-regex": "^5.0.1" + "ansi-regex": "^6.0.1" } } } }, "cliui": { - "version": "7.0.4", - "resolved": "https://registry.npmjs.org/cliui/-/cliui-7.0.4.tgz", - "integrity": "sha512-OcRE68cOsVMXp1Yvonl/fzkQOyjLSu/8bhPDfQt0e0/Eb283TKP20Fs2MqoPsr9SwA595rRCA+QMzYc9nBP+JQ==", + "version": "8.0.1", + "resolved": "https://registry.npmjs.org/cliui/-/cliui-8.0.1.tgz", + "integrity": "sha512-BSeNnyus75C4//NQ9gQt1/csTXyo/8Sb+afLAkzAptFuMsod9HFokGNudZpi/oQV73hnVK+sR+5PVRMd+Dr7YQ==", "requires": { "string-width": "^4.2.0", - "strip-ansi": "^6.0.0", + "strip-ansi": "^6.0.1", "wrap-ansi": "^7.0.0" - }, - "dependencies": { - "ansi-regex": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", - "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==" - }, - "ansi-styles": { - "version": "4.3.0", - "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz", - "integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==", - "requires": { - "color-convert": "^2.0.1" - } - }, - "emoji-regex": { - "version": "8.0.0", - "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", - "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" - }, - "string-width": { - "version": "4.2.3", - "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", - "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", - "requires": { - "emoji-regex": "^8.0.0", - "is-fullwidth-code-point": "^3.0.0", - "strip-ansi": "^6.0.1" - } - }, - "strip-ansi": { - "version": "6.0.1", - "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", - "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", - "requires": { - "ansi-regex": "^5.0.1" - } - }, - "wrap-ansi": { - "version": "7.0.0", - "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-7.0.0.tgz", - "integrity": "sha512-YVGIj2kamLSTxw6NsZjoBxfSwsn0ycdesmc4p+Q21c5zPuZ1pl+NfxVdxPtdHvmNVOQ6XSYG4AUtyt/Fi7D16Q==", - "requires": { - "ansi-styles": "^4.0.0", - "string-width": "^4.1.0", - "strip-ansi": "^6.0.0" - } - } - } - }, - "clone-response": { - "version": "1.0.3", - "resolved": "https://registry.npmjs.org/clone-response/-/clone-response-1.0.3.tgz", - "integrity": "sha512-ROoL94jJH2dUVML2Y/5PEDNaSHgeOdSDicUyS7izcF63G6sTc/FTjLub4b8Il9S8S0beOfYt0TaA5qvFK+w0wA==", - "requires": { - "mimic-response": "^1.0.0" } }, "color-convert": { @@ -1800,27 +2263,28 @@ "integrity": "sha512-dOy+3AuW3a2wNbZHIuMZpTcgjGuLU/uBL/ubcZF9OXbDo8ff4O8yVp5Bf0efS8uEoYo5q4Fx7dY9OgQGXgAsQA==" }, "colorette": { - "version": "2.0.19", - "resolved": "https://registry.npmjs.org/colorette/-/colorette-2.0.19.tgz", - "integrity": "sha512-3tlv/dIP7FWvj3BsbHrGLJ6l/oKh1O3TcgBqMn+yyCagOxc23fyzDS6HypQbgxWbkpDnf52p1LuR4eWDQ/K9WQ==" + "version": "2.0.20", + "resolved": "https://registry.npmjs.org/colorette/-/colorette-2.0.20.tgz", + "integrity": "sha512-IfEDxwoWIjkeXL1eXcDiow4UbKjhLdq6/EuSVR9GMN7KVH3r9gQ83e73hsz1Nd1T3ijd5xv1wcWRYO+D6kCI2w==" }, - "compress-brotli": { - "version": "1.3.8", - "resolved": "https://registry.npmjs.org/compress-brotli/-/compress-brotli-1.3.8.tgz", - "integrity": "sha512-lVcQsjhxhIXsuupfy9fmZUFtAIdBmXA7EGY6GBdgZ++qkM9zG4YFT8iU7FoBxzryNDMOpD1HIFHUSX4D87oqhQ==", + "cpu-features": { + "version": "0.0.10", + "resolved": "https://registry.npmjs.org/cpu-features/-/cpu-features-0.0.10.tgz", + "integrity": "sha512-9IkYqtX3YHPCzoVg1Py+o9057a3i0fp7S530UWokCSaFVTc7CwXPRiOjRjBQQ18ZCNafx78YfnG+HALxtVmOGA==", + "optional": true, "requires": { - "@types/json-buffer": "~3.0.0", - "json-buffer": "~3.0.1" + "buildcheck": "~0.0.6", + "nan": "^2.19.0" } }, - "cpu-features": { - "version": "0.0.4", - "resolved": "https://registry.npmjs.org/cpu-features/-/cpu-features-0.0.4.tgz", - "integrity": "sha512-fKiZ/zp1mUwQbnzb9IghXtHtDoTMtNeb8oYGx6kX2SYfhnG0HNdBEBIzB9b5KlXu5DQPhfy3mInbBxFcgwAr3A==", - "optional": true, + "cross-spawn": { + "version": "7.0.3", + "resolved": "https://registry.npmjs.org/cross-spawn/-/cross-spawn-7.0.3.tgz", + "integrity": "sha512-iRDPJKUPVEND7dHPO8rkbOnPpyDygcDFtWjpeWNCgy8WP2rXcxXL8TskReQl6OrB2G7+UJrags1q15Fudc7G6w==", "requires": { - "buildcheck": "0.0.3", - "nan": "^2.15.0" + "path-key": "^3.1.0", + "shebang-command": "^2.0.0", + "which": "^2.0.1" } }, "debug": { @@ -1852,25 +2316,40 @@ "integrity": "sha512-4tvttepXG1VaYGrRibk5EwJd1t4udunSOVMdLSAL6mId1ix438oPwPZMALY41FCijukO1L0twNcGsdzS7dHgDg==" }, "docker-modem": { - "version": "3.0.5", - "resolved": "https://registry.npmjs.org/docker-modem/-/docker-modem-3.0.5.tgz", - "integrity": "sha512-x1E6jxWdtoK3+ifAUWj4w5egPdTDGBpesSCErm+aKET5BnnEOvDtTP6GxcnMB1zZiv2iQ0qJZvJie+1wfIRg6Q==", + "version": "5.0.6", + "resolved": "https://registry.npmjs.org/docker-modem/-/docker-modem-5.0.6.tgz", + "integrity": "sha512-ens7BiayssQz/uAxGzH8zGXCtiV24rRWXdjNha5V4zSOcxmAZsfGVm/PPFbwQdqEkDnhG+SyR9E3zSHUbOKXBQ==", "requires": { "debug": "^4.1.1", "readable-stream": "^3.5.0", "split-ca": "^1.0.1", - "ssh2": "^1.4.0" + "ssh2": "^1.15.0" } }, "dockerode": { - "version": "3.3.3", - "resolved": "https://registry.npmjs.org/dockerode/-/dockerode-3.3.3.tgz", - "integrity": "sha512-lvKV6/NGf2/CYLt5V4c0fd6Fl9XZSCo1Z2HBT9ioKrKLMB2o+gA62Uza8RROpzGvYv57KJx2dKu+ZwSpB//OIA==", + "version": "4.0.6", + "resolved": "https://registry.npmjs.org/dockerode/-/dockerode-4.0.6.tgz", + "integrity": "sha512-FbVf3Z8fY/kALB9s+P9epCpWhfi/r0N2DgYYcYpsAUlaTxPjdsitsFobnltb+lyCgAIvf9C+4PSWlTnHlJMf1w==", "requires": { - "docker-modem": "^3.0.0", - "tar-fs": "~2.0.1" + "@balena/dockerignore": "^1.0.2", + "@grpc/grpc-js": "^1.11.1", + "@grpc/proto-loader": "^0.7.13", + "docker-modem": "^5.0.6", + "protobufjs": "^7.3.2", + "tar-fs": "~2.1.2", + "uuid": "^10.0.0" } }, + "eastasianwidth": { + "version": "0.2.0", + "resolved": "https://registry.npmjs.org/eastasianwidth/-/eastasianwidth-0.2.0.tgz", + "integrity": "sha512-I88TYZWc9XiYHRQ4/3c5rjjfgkjhLyW2luGIheGERbNQ6OY7yTybanSpDXZa8y7VUP9YmDcYa+eyq4ca7iLqWA==" + }, + "emoji-regex": { + "version": "8.0.0", + "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", + "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" + }, "end-of-stream": { "version": "1.4.4", "resolved": "https://registry.npmjs.org/end-of-stream/-/end-of-stream-1.4.4.tgz", @@ -1880,11 +2359,12 @@ } }, "enquirer": { - "version": "2.3.6", - "resolved": "https://registry.npmjs.org/enquirer/-/enquirer-2.3.6.tgz", - "integrity": "sha512-yjNnPr315/FjS4zIsUxYguYUPP2e1NK4d7E7ZOLiyYCcbFBiTMyID+2wvm2w6+pZ/odMA7cRkjhsPbltwBOrLg==", + "version": "2.4.1", + "resolved": "https://registry.npmjs.org/enquirer/-/enquirer-2.4.1.tgz", + "integrity": "sha512-rRqJg/6gd538VHvR3PSrdRBb/1Vy2YfzHqzvbhGIQpDRKIa4FgV/54b5Q1xYSxOOwKvjXweS26E0Q+nAMwp2pQ==", "requires": { - "ansi-colors": "^4.1.1" + "ansi-colors": "^4.1.1", + "strip-ansi": "^6.0.1" } }, "escalade": { @@ -1892,6 +2372,11 @@ "resolved": "https://registry.npmjs.org/escalade/-/escalade-3.1.1.tgz", "integrity": "sha512-k0er2gUkLf8O0zKJiAhmkTnJlTvINGv7ygDNPbeIsX/TJjGJZHuh9B2UxbsaEkmlEo9MfhrSzmhIlhRlI2GXnw==" }, + "eventemitter3": { + "version": "5.0.1", + "resolved": "https://registry.npmjs.org/eventemitter3/-/eventemitter3-5.0.1.tgz", + "integrity": "sha512-GWkBvjiSZK87ELrYOSESUYeVIc9mvLLf/nXalMOS5dYrgZq9o5OVkbZAVM06CVxYsCwH9BDZFPlQTlPA1j4ahA==" + }, "extract-zip": { "version": "2.0.1", "resolved": "https://registry.npmjs.org/extract-zip/-/extract-zip-2.0.1.tgz", @@ -1921,10 +2406,26 @@ "pend": "~1.2.0" } }, + "foreground-child": { + "version": "3.1.1", + "resolved": "https://registry.npmjs.org/foreground-child/-/foreground-child-3.1.1.tgz", + "integrity": "sha512-TMKDUnIte6bfb5nWv7V/caI169OHgvwjb7V4WkeUvbQQdjr5rWKqHFiKWb/fcOwB+CzBT+qbWjvj+DVwRskpIg==", + "requires": { + "cross-spawn": "^7.0.0", + "signal-exit": "^4.0.1" + }, + "dependencies": { + "signal-exit": { + "version": "4.1.0", + "resolved": "https://registry.npmjs.org/signal-exit/-/signal-exit-4.1.0.tgz", + "integrity": "sha512-bzyZ1e88w9O1iNJbKnOlvYTrWPDl46O1bG0D3XInv+9tkPrxrN8jUUTiFlDkkmKWgn1M6CfIA13SuGqOa9Korw==" + } + } + }, "form-data-encoder": { - "version": "2.0.1", - "resolved": "https://registry.npmjs.org/form-data-encoder/-/form-data-encoder-2.0.1.tgz", - "integrity": "sha512-Oy+P9w5mnO4TWXVgUiQvggNKPI9/ummcSt5usuIV6HkaLKigwzPpoenhEqmGmx3zHqm6ZLJ+CR/99N8JLinaEw==" + "version": "2.1.4", + "resolved": "https://registry.npmjs.org/form-data-encoder/-/form-data-encoder-2.1.4.tgz", + "integrity": "sha512-yDYSgNMraqvnxiEXO4hi88+YZxaHC6QKzb5N84iRCTDeRO7ZALpir/lVmf/uXUhnwUr2O4HU8s/n6x+yNjQkHw==" }, "fs-constants": { "version": "1.0.0", @@ -1932,23 +2433,15 @@ "integrity": "sha512-y6OAwoSIf7FyjMIv94u+b5rdheZEjzR63GTyZJm5qh4Bi+2YgwLCcI/fPFZkL5PSixOt6ZNKm+w+Hfp/Bciwow==" }, "fs-extra": { - "version": "10.1.0", - "resolved": "https://registry.npmjs.org/fs-extra/-/fs-extra-10.1.0.tgz", - "integrity": "sha512-oRXApq54ETRj4eMiFzGnHWGy+zo5raudjuxN0b8H7s/RU2oW0Wvsx9O0ACRN/kRq9E8Vu/ReskGB5o3ji+FzHQ==", + "version": "11.3.0", + "resolved": "https://registry.npmjs.org/fs-extra/-/fs-extra-11.3.0.tgz", + "integrity": "sha512-Z4XaCL6dUDHfP/jT25jJKMmtxvuwbkrD1vNSMFlo9lNLY2c5FHYSQgHPRZUjAB26TpDEoW9HCOgplrdbaPV/ew==", "requires": { "graceful-fs": "^4.2.0", "jsonfile": "^6.0.1", "universalify": "^2.0.0" } }, - "fs-minipass": { - "version": "2.1.0", - "resolved": "https://registry.npmjs.org/fs-minipass/-/fs-minipass-2.1.0.tgz", - "integrity": "sha512-V/JgOLFCS+R6Vcq0slCuaeWEdNC3ouDlJMNIsacH2VtALiu9mV4LPrHc5cDl8k5aw6J8jwgWWpiTo5RYhmIzvg==", - "requires": { - "minipass": "^3.0.0" - } - }, "get-caller-file": { "version": "2.0.5", "resolved": "https://registry.npmjs.org/get-caller-file/-/get-caller-file-2.0.5.tgz", @@ -1959,24 +2452,34 @@ "resolved": "https://registry.npmjs.org/get-stream/-/get-stream-6.0.1.tgz", "integrity": "sha512-ts6Wi+2j3jQjqi70w5AlN8DFnkSwC+MqmxEzdEALB2qXZYV3X/b1CTfgPLGJNMeAWxdPfU8FO1ms3NUfaHCPYg==" }, + "glob": { + "version": "10.3.12", + "resolved": "https://registry.npmjs.org/glob/-/glob-10.3.12.tgz", + "integrity": "sha512-TCNv8vJ+xz4QiqTpfOJA7HvYv+tNIRHKfUWw/q+v2jdgN4ebz+KY9tGx5J4rHP0o84mNP+ApH66HRX8us3Khqg==", + "requires": { + "foreground-child": "^3.1.0", + "jackspeak": "^2.3.6", + "minimatch": "^9.0.1", + "minipass": "^7.0.4", + "path-scurry": "^1.10.2" + } + }, "got": { - "version": "12.3.1", - "resolved": "https://registry.npmjs.org/got/-/got-12.3.1.tgz", - "integrity": "sha512-tS6+JMhBh4iXMSXF6KkIsRxmloPln31QHDlcb6Ec3bzxjjFJFr/8aXdpyuLmVc9I4i2HyBHYw1QU5K1ruUdpkw==", + "version": "13.0.0", + "resolved": "https://registry.npmjs.org/got/-/got-13.0.0.tgz", + "integrity": "sha512-XfBk1CxOOScDcMr9O1yKkNaQyy865NbYs+F7dr4H0LZMVgCj2Le59k6PqbNHoL5ToeaEQUYh6c6yMfVcc6SJxA==", "requires": { "@sindresorhus/is": "^5.2.0", "@szmarczak/http-timer": "^5.0.1", - "@types/cacheable-request": "^6.0.2", - "@types/responselike": "^1.0.0", - "cacheable-lookup": "^6.0.4", - "cacheable-request": "^7.0.2", + "cacheable-lookup": "^7.0.0", + "cacheable-request": "^10.2.8", "decompress-response": "^6.0.0", - "form-data-encoder": "^2.0.1", + "form-data-encoder": "^2.1.2", "get-stream": "^6.0.1", "http2-wrapper": "^2.1.10", "lowercase-keys": "^3.0.0", "p-cancelable": "^3.0.0", - "responselike": "^2.0.0" + "responselike": "^3.0.0" } }, "graceful-fs": { @@ -1985,9 +2488,9 @@ "integrity": "sha512-9ByhssR2fPVsNZj478qUUbKfmL0+t5BDVyjShtyZZLiK7ZDAArFFfopyOTj0M05wE2tJPisA4iTnnXl2YoPvOA==" }, "http-cache-semantics": { - "version": "4.1.0", - "resolved": "https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-4.1.0.tgz", - "integrity": "sha512-carPklcUh7ROWRK7Cv27RPtdhYhUsela/ue5/jKzjegVvXDqM2ILE9Q2BGn9JZJh1g87cp56su/FgQSzcWS8cQ==" + "version": "4.1.1", + "resolved": "https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-4.1.1.tgz", + "integrity": "sha512-er295DKPVsV82j5kw1Gjt+ADA/XYHsajl82cGNQG2eyoPkvgUhX+nDIyelzhIWbbsXP39EHcI6l5tYs2FYqYXQ==" }, "http2-wrapper": { "version": "2.1.11", @@ -2003,11 +2506,6 @@ "resolved": "https://registry.npmjs.org/ieee754/-/ieee754-1.2.1.tgz", "integrity": "sha512-dcyqhDvX1C46lXZcVqCpK+FtMRQVdIMN6/Df5js2zouUsqG7I6sFxitIC+7KYK29KdXOLHdu9zL4sFnoVQnqaA==" }, - "indent-string": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/indent-string/-/indent-string-4.0.0.tgz", - "integrity": "sha512-EdDDZu4A2OyIK7Lr/2zG+w5jmbuk1DVBnEwREQvBzspBJkCEbRa8GxU1lghYcaGJCnRWibjDXlq779X1/y5xwg==" - }, "inherits": { "version": "2.0.4", "resolved": "https://registry.npmjs.org/inherits/-/inherits-2.0.4.tgz", @@ -2018,6 +2516,20 @@ "resolved": "https://registry.npmjs.org/is-fullwidth-code-point/-/is-fullwidth-code-point-3.0.0.tgz", "integrity": "sha512-zymm5+u+sCsSWyD9qNaejV3DFvhCKclKdizYaJUuHA83RLjb7nSuGnddCHGv0hk+KY7BMAlsWeK4Ueg6EV6XQg==" }, + "isexe": { + "version": "2.0.0", + "resolved": "https://registry.npmjs.org/isexe/-/isexe-2.0.0.tgz", + "integrity": "sha512-RHxMLp9lnKHGHRng9QFhRCMbYAcVpn69smSGcq3f36xjgVVWThj4qqLbTLlq7Ssj8B+fIQ1EuCEGI2lKsyQeIw==" + }, + "jackspeak": { + "version": "2.3.6", + "resolved": "https://registry.npmjs.org/jackspeak/-/jackspeak-2.3.6.tgz", + "integrity": "sha512-N3yCS/NegsOBokc8GAdM8UcmfsKiSS8cipheD/nivzr700H+nsMOxJjQnvwOcRYVuFkdH0wGUvW2WbXGmrZGbQ==", + "requires": { + "@isaacs/cliui": "^8.0.2", + "@pkgjs/parseargs": "^0.11.0" + } + }, "json-buffer": { "version": "3.0.1", "resolved": "https://registry.npmjs.org/json-buffer/-/json-buffer-3.0.1.tgz", @@ -2038,73 +2550,67 @@ "integrity": "sha512-HjorDJFNhnM4SicvaUXac0X77NiskggxJdesG72+O5zBKpSqKFCrqmndKVqpu3pFqkla0St6uGk8Ju0sCurrmg==" }, "keyv": { - "version": "4.3.3", - "resolved": "https://registry.npmjs.org/keyv/-/keyv-4.3.3.tgz", - "integrity": "sha512-AcysI17RvakTh8ir03+a3zJr5r0ovnAH/XTXei/4HIv3bL2K/jzvgivLK9UuI/JbU1aJjM3NSAnVvVVd3n+4DQ==", + "version": "4.5.2", + "resolved": "https://registry.npmjs.org/keyv/-/keyv-4.5.2.tgz", + "integrity": "sha512-5MHbFaKn8cNSmVW7BYnijeAVlE4cYA/SVkifVgrh7yotnfhKmjuXpDKjrABLnT0SfHWV21P8ow07OGfRrNDg8g==", "requires": { - "compress-brotli": "^1.3.8", "json-buffer": "3.0.1" } }, "listr2": { - "version": "5.0.2", - "resolved": "https://registry.npmjs.org/listr2/-/listr2-5.0.2.tgz", - "integrity": "sha512-/yIYitiJgIkzVIKqQc+Dpcpwaj4PjeW/hSi7/xoGJ5mil7vQYedr5kKuEyLhANlnifCcumafnIR6azpb8loODw==", - "requires": { - "cli-truncate": "^2.1.0", - "colorette": "^2.0.19", - "log-update": "^4.0.0", - "p-map": "^4.0.0", + "version": "6.6.1", + "resolved": "https://registry.npmjs.org/listr2/-/listr2-6.6.1.tgz", + "integrity": "sha512-+rAXGHh0fkEWdXBmX+L6mmfmXmXvDGEKzkjxO+8mP3+nI/r/CWznVBvsibXdxda9Zz0OW2e2ikphN3OwCT/jSg==", + "requires": { + "cli-truncate": "^3.1.0", + "colorette": "^2.0.20", + "eventemitter3": "^5.0.1", + "log-update": "^5.0.1", "rfdc": "^1.3.0", - "rxjs": "^7.5.6", - "through": "^2.3.8", - "wrap-ansi": "^7.0.0" + "wrap-ansi": "^8.1.0" }, "dependencies": { "ansi-regex": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", - "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==" + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-6.0.1.tgz", + "integrity": "sha512-n5M855fKb2SsfMIiFFoVrABHJC8QtHwVx+mHWP3QcEqBHYienj5dHSgjbxtC0WEZXYt4wcD6zrQElDPhFuZgfA==" }, "ansi-styles": { - "version": "4.3.0", - "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz", - "integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==", - "requires": { - "color-convert": "^2.0.1" - } + "version": "6.2.1", + "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-6.2.1.tgz", + "integrity": "sha512-bN798gFfQX+viw3R7yrGWRqnrN2oRkEkUjjl4JNn4E8GxxbjtG3FbrEIIY3l8/hrwUwIeCZvi4QuOTP4MErVug==" }, "emoji-regex": { - "version": "8.0.0", - "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", - "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" + "version": "9.2.2", + "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-9.2.2.tgz", + "integrity": "sha512-L18DaJsXSUk2+42pv8mLs5jJT2hqFkFE4j21wOmgbUqsZ2hL72NsUU785g9RXgo3s0ZNgVl42TiHp3ZtOv/Vyg==" }, "string-width": { - "version": "4.2.3", - "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", - "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", + "version": "5.1.2", + "resolved": "https://registry.npmjs.org/string-width/-/string-width-5.1.2.tgz", + "integrity": "sha512-HnLOCR3vjcY8beoNLtcjZ5/nxn2afmME6lhrDrebokqMap+XbeW8n9TXpPDOqdGK5qcI3oT0GKTW6wC7EMiVqA==", "requires": { - "emoji-regex": "^8.0.0", - "is-fullwidth-code-point": "^3.0.0", - "strip-ansi": "^6.0.1" + "eastasianwidth": "^0.2.0", + "emoji-regex": "^9.2.2", + "strip-ansi": "^7.0.1" } }, "strip-ansi": { - "version": "6.0.1", - "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", - "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", + "version": "7.0.1", + "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-7.0.1.tgz", + "integrity": "sha512-cXNxvT8dFNRVfhVME3JAe98mkXDYN2O1l7jmcwMnOslDeESg1rF/OZMtK0nRAhiari1unG5cD4jG3rapUAkLbw==", "requires": { - "ansi-regex": "^5.0.1" + "ansi-regex": "^6.0.1" } }, "wrap-ansi": { - "version": "7.0.0", - "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-7.0.0.tgz", - "integrity": "sha512-YVGIj2kamLSTxw6NsZjoBxfSwsn0ycdesmc4p+Q21c5zPuZ1pl+NfxVdxPtdHvmNVOQ6XSYG4AUtyt/Fi7D16Q==", + "version": "8.1.0", + "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-8.1.0.tgz", + "integrity": "sha512-si7QWI6zUMq56bESFvagtmzMdGOtoxfR+Sez11Mobfc7tm+VkUckk9bW2UeffTGVUbOksxmSw0AA2gs8g71NCQ==", "requires": { - "ansi-styles": "^4.0.0", - "string-width": "^4.1.0", - "strip-ansi": "^6.0.0" + "ansi-styles": "^6.1.0", + "string-width": "^5.0.1", + "strip-ansi": "^7.0.1" } } } @@ -2114,114 +2620,87 @@ "resolved": "https://registry.npmjs.org/lodash-es/-/lodash-es-4.17.21.tgz", "integrity": "sha512-mKnC+QJ9pWVzv+C4/U3rRsHapFfHvQFoFB92e52xeyGMcX6/OlIl78je1u8vePzYZSkkogMPJ2yjxxsb89cxyw==" }, + "lodash.camelcase": { + "version": "4.3.0", + "resolved": "https://registry.npmjs.org/lodash.camelcase/-/lodash.camelcase-4.3.0.tgz", + "integrity": "sha512-TwuEnCnxbc3rAvhf/LbG7tJUDzhqXyFnv3dtzLOPgCG/hODL7WFnsbwktkD7yUV0RrreP/l1PALq/YSg6VvjlA==" + }, "log-update": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/log-update/-/log-update-4.0.0.tgz", - "integrity": "sha512-9fkkDevMefjg0mmzWFBW8YkFP91OrizzkW3diF7CpG+S2EYdy4+TVfGwz1zeF8x7hCx1ovSPTOE9Ngib74qqUg==", + "version": "5.0.1", + "resolved": "https://registry.npmjs.org/log-update/-/log-update-5.0.1.tgz", + "integrity": "sha512-5UtUDQ/6edw4ofyljDNcOVJQ4c7OjDro4h3y8e1GQL5iYElYclVHJ3zeWchylvMaKnDbDilC8irOVyexnA/Slw==", "requires": { - "ansi-escapes": "^4.3.0", - "cli-cursor": "^3.1.0", - "slice-ansi": "^4.0.0", - "wrap-ansi": "^6.2.0" + "ansi-escapes": "^5.0.0", + "cli-cursor": "^4.0.0", + "slice-ansi": "^5.0.0", + "strip-ansi": "^7.0.1", + "wrap-ansi": "^8.0.1" }, "dependencies": { - "ansi-escapes": { - "version": "4.3.2", - "resolved": "https://registry.npmjs.org/ansi-escapes/-/ansi-escapes-4.3.2.tgz", - "integrity": "sha512-gKXj5ALrKWQLsYG9jlTRmR/xKluxHV+Z9QEwNIgCfM1/uwPMCuzVVnh5mwTd+OuBZcwSIMbqssNWRm1lE51QaQ==", - "requires": { - "type-fest": "^0.21.3" - } - }, "ansi-regex": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", - "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==" + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-6.0.1.tgz", + "integrity": "sha512-n5M855fKb2SsfMIiFFoVrABHJC8QtHwVx+mHWP3QcEqBHYienj5dHSgjbxtC0WEZXYt4wcD6zrQElDPhFuZgfA==" }, "ansi-styles": { - "version": "4.3.0", - "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz", - "integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==", - "requires": { - "color-convert": "^2.0.1" - } - }, - "cli-cursor": { - "version": "3.1.0", - "resolved": "https://registry.npmjs.org/cli-cursor/-/cli-cursor-3.1.0.tgz", - "integrity": "sha512-I/zHAwsKf9FqGoXM4WWRACob9+SNukZTd94DWF57E4toouRulbCxcUh6RKUEOQlYTHJnzkPMySvPNaaSLNfLZw==", - "requires": { - "restore-cursor": "^3.1.0" - } + "version": "6.2.1", + "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-6.2.1.tgz", + "integrity": "sha512-bN798gFfQX+viw3R7yrGWRqnrN2oRkEkUjjl4JNn4E8GxxbjtG3FbrEIIY3l8/hrwUwIeCZvi4QuOTP4MErVug==" }, "emoji-regex": { - "version": "8.0.0", - "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", - "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" - }, - "restore-cursor": { - "version": "3.1.0", - "resolved": "https://registry.npmjs.org/restore-cursor/-/restore-cursor-3.1.0.tgz", - "integrity": "sha512-l+sSefzHpj5qimhFSE5a8nufZYAM3sBSVMAPtYkmC+4EH2anSGaEMXSD0izRQbu9nfyQ9y5JrVmp7E8oZrUjvA==", - "requires": { - "onetime": "^5.1.0", - "signal-exit": "^3.0.2" - } - }, - "slice-ansi": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/slice-ansi/-/slice-ansi-4.0.0.tgz", - "integrity": "sha512-qMCMfhY040cVHT43K9BFygqYbUPFZKHOg7K73mtTWJRb8pyP3fzf4Ixd5SzdEJQ6MRUg/WBnOLxghZtKKurENQ==", - "requires": { - "ansi-styles": "^4.0.0", - "astral-regex": "^2.0.0", - "is-fullwidth-code-point": "^3.0.0" - } + "version": "9.2.2", + "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-9.2.2.tgz", + "integrity": "sha512-L18DaJsXSUk2+42pv8mLs5jJT2hqFkFE4j21wOmgbUqsZ2hL72NsUU785g9RXgo3s0ZNgVl42TiHp3ZtOv/Vyg==" }, "string-width": { - "version": "4.2.3", - "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", - "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", + "version": "5.1.2", + "resolved": "https://registry.npmjs.org/string-width/-/string-width-5.1.2.tgz", + "integrity": "sha512-HnLOCR3vjcY8beoNLtcjZ5/nxn2afmME6lhrDrebokqMap+XbeW8n9TXpPDOqdGK5qcI3oT0GKTW6wC7EMiVqA==", "requires": { - "emoji-regex": "^8.0.0", - "is-fullwidth-code-point": "^3.0.0", - "strip-ansi": "^6.0.1" + "eastasianwidth": "^0.2.0", + "emoji-regex": "^9.2.2", + "strip-ansi": "^7.0.1" } }, "strip-ansi": { - "version": "6.0.1", - "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", - "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", + "version": "7.0.1", + "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-7.0.1.tgz", + "integrity": "sha512-cXNxvT8dFNRVfhVME3JAe98mkXDYN2O1l7jmcwMnOslDeESg1rF/OZMtK0nRAhiari1unG5cD4jG3rapUAkLbw==", "requires": { - "ansi-regex": "^5.0.1" + "ansi-regex": "^6.0.1" } }, - "type-fest": { - "version": "0.21.3", - "resolved": "https://registry.npmjs.org/type-fest/-/type-fest-0.21.3.tgz", - "integrity": "sha512-t0rzBq87m3fVcduHDUFhKmyyX+9eo6WQjZvf51Ea/M0Q7+T374Jp1aUiyUl0GKxp8M/OETVHSDvmkyPgvX+X2w==" - }, "wrap-ansi": { - "version": "6.2.0", - "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-6.2.0.tgz", - "integrity": "sha512-r6lPcBGxZXlIcymEu7InxDMhdW0KDxpLgoFLcguasxCaJ/SOIZwINatK9KY/tf+ZrlywOKU0UDj3ATXUBfxJXA==", + "version": "8.1.0", + "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-8.1.0.tgz", + "integrity": "sha512-si7QWI6zUMq56bESFvagtmzMdGOtoxfR+Sez11Mobfc7tm+VkUckk9bW2UeffTGVUbOksxmSw0AA2gs8g71NCQ==", "requires": { - "ansi-styles": "^4.0.0", - "string-width": "^4.1.0", - "strip-ansi": "^6.0.0" + "ansi-styles": "^6.1.0", + "string-width": "^5.0.1", + "strip-ansi": "^7.0.1" } } } }, + "long": { + "version": "5.2.4", + "resolved": "https://registry.npmjs.org/long/-/long-5.2.4.tgz", + "integrity": "sha512-qtzLbJE8hq7VabR3mISmVGtoXP8KGc2Z/AT8OuqlYD7JTR3oqrgwdjnk07wpj1twXxYmgDXgoKVWUG/fReSzHg==" + }, "lowercase-keys": { "version": "3.0.0", "resolved": "https://registry.npmjs.org/lowercase-keys/-/lowercase-keys-3.0.0.tgz", "integrity": "sha512-ozCC6gdQ+glXOQsveKD0YsDy8DSQFjDTz4zyzEHNV5+JP5D62LmfDZ6o1cycFx9ouG940M5dE8C8CTewdj2YWQ==" }, + "lru-cache": { + "version": "10.2.2", + "resolved": "https://registry.npmjs.org/lru-cache/-/lru-cache-10.2.2.tgz", + "integrity": "sha512-9hp3Vp2/hFQUiIwKo8XCeFVnrg8Pk3TYNPIR7tJADKi5YfcF7vEaK7avFHTlSy3kOKYaJQaalfEo6YuXdceBOQ==" + }, "luxon": { - "version": "3.0.1", - "resolved": "https://registry.npmjs.org/luxon/-/luxon-3.0.1.tgz", - "integrity": "sha512-hF3kv0e5gwHQZKz4wtm4c+inDtyc7elkanAsBq+fundaCdUBNJB1dHEGUZIM6SfSBUlbVFduPwEtNjFK8wLtcw==" + "version": "3.6.1", + "resolved": "https://registry.npmjs.org/luxon/-/luxon-3.6.1.tgz", + "integrity": "sha512-tJLxrKJhO2ukZ5z0gyjY1zPh3Rh88Ej9P7jNrZiHMUXHae1yvI2imgOZtL1TO8TW6biMMKfTtAOoEJANgtWBMQ==" }, "mimic-fn": { "version": "2.1.0", @@ -2229,31 +2708,36 @@ "integrity": "sha512-OqbOk5oEQeAZ8WXWydlu9HJjz9WVdEIvamMCcXmuqUYjTknH/sqsWvhQ3vgwKFRR1HpjvNBKQ37nbJgYzGqGcg==" }, "mimic-response": { - "version": "1.0.1", - "resolved": "https://registry.npmjs.org/mimic-response/-/mimic-response-1.0.1.tgz", - "integrity": "sha512-j5EctnkH7amfV/q5Hgmoal1g2QHFJRraOtmx0JpIqkxhBhI/lJSl1nMpQ45hVarwNETOoWEimndZ4QK0RHxuxQ==" + "version": "4.0.0", + "resolved": "https://registry.npmjs.org/mimic-response/-/mimic-response-4.0.0.tgz", + "integrity": "sha512-e5ISH9xMYU0DzrT+jl8q2ze9D6eWBto+I8CNpe+VI+K2J/F/k3PdkdTdz4wvGVH4NTpo+NRYTVIuMQEMMcsLqg==" }, - "minipass": { - "version": "3.3.4", - "resolved": "https://registry.npmjs.org/minipass/-/minipass-3.3.4.tgz", - "integrity": "sha512-I9WPbWHCGu8W+6k1ZiGpPu0GkoKBeorkfKNuAFBNS1HNFJvke82sxvI5bzcCNpWPorkOO5QQ+zomzzwRxejXiw==", + "minimatch": { + "version": "9.0.4", + "resolved": "https://registry.npmjs.org/minimatch/-/minimatch-9.0.4.tgz", + "integrity": "sha512-KqWh+VchfxcMNRAJjj2tnsSJdNbHsVgnkBhTNrW7AjVo6OvLtxw8zfT9oLw1JSohlFzJ8jCoTgaoXvJ+kHt6fw==", "requires": { - "yallist": "^4.0.0" + "brace-expansion": "^2.0.1" } }, + "minipass": { + "version": "7.1.2", + "resolved": "https://registry.npmjs.org/minipass/-/minipass-7.1.2.tgz", + "integrity": "sha512-qOOzS1cBTWYF4BH8fVePDBOO9iptMnGUEZwNc/cMWnTV2nVLZ7VoNWEPHkYczZA0pdoA7dl6e7FL659nX9S2aw==" + }, "minizlib": { - "version": "2.1.2", - "resolved": "https://registry.npmjs.org/minizlib/-/minizlib-2.1.2.tgz", - "integrity": "sha512-bAxsR8BVfj60DWXHE3u30oHzfl4G7khkSuPW+qvpd7jFRHm7dLxOjUk1EHACJ/hxLY8phGJ0YhYHZo7jil7Qdg==", + "version": "3.0.1", + "resolved": "https://registry.npmjs.org/minizlib/-/minizlib-3.0.1.tgz", + "integrity": "sha512-umcy022ILvb5/3Djuu8LWeqUa8D68JaBzlttKeMWen48SjabqS3iY5w/vzeMzMUNhLDifyhbOwKDSznB1vvrwg==", "requires": { - "minipass": "^3.0.0", - "yallist": "^4.0.0" + "minipass": "^7.0.4", + "rimraf": "^5.0.5" } }, "mkdirp": { - "version": "1.0.4", - "resolved": "https://registry.npmjs.org/mkdirp/-/mkdirp-1.0.4.tgz", - "integrity": "sha512-vVqVZQyf3WLx2Shd0qJ9xuvqgAyKPLAiqITEtqW0oIUjzo3PePDd6fW9iFz30ef7Ysp/oiWqbhszeGWW2T6Gzw==" + "version": "3.0.1", + "resolved": "https://registry.npmjs.org/mkdirp/-/mkdirp-3.0.1.tgz", + "integrity": "sha512-+NsyUUAZDmo6YVHzL/stxSu3t9YS1iljliy3BSDrXJ/dkn1KYdmtZODGGjLcc9XLgVVpH4KshHB8XmZgMhaBXg==" }, "mkdirp-classic": { "version": "0.5.3", @@ -2266,15 +2750,15 @@ "integrity": "sha512-sGkPx+VjMtmA6MX27oA4FBFELFCZZ4S4XqeGOXCv68tT+jb3vk/RyaKWP0PTKyWtmLSM0b+adUTEvbs1PEaH2w==" }, "nan": { - "version": "2.16.0", - "resolved": "https://registry.npmjs.org/nan/-/nan-2.16.0.tgz", - "integrity": "sha512-UdAqHyFngu7TfQKsCBgAA6pWDkT8MAO7d0jyOecVhN5354xbLqdn8mV9Tat9gepAupm0bt2DbeaSC8vS52MuFA==", + "version": "2.22.0", + "resolved": "https://registry.npmjs.org/nan/-/nan-2.22.0.tgz", + "integrity": "sha512-nbajikzWTMwsW+eSsNm3QwlOs7het9gGJU5dDZzRTQGk03vyBOauxgI4VakDzE0PtsGTmXPsXTbbjVhRwR5mpw==", "optional": true }, "normalize-url": { - "version": "6.1.0", - "resolved": "https://registry.npmjs.org/normalize-url/-/normalize-url-6.1.0.tgz", - "integrity": "sha512-DlL+XwOy3NxAQ8xuC0okPgK46iuVNAK01YN7RueYBqqFeGsBjV9XmCAzAdgt+667bCl5kPh9EqKKDwnaPG1I7A==" + "version": "8.0.0", + "resolved": "https://registry.npmjs.org/normalize-url/-/normalize-url-8.0.0.tgz", + "integrity": "sha512-uVFpKhj5MheNBJRTiMZ9pE/7hD1QTeEvugSJW/OmLzAp78PB5O6adfMNTvmfKhXBkvCzC+rqifWcVYpGFwTjnw==" }, "once": { "version": "1.4.0", @@ -2297,12 +2781,18 @@ "resolved": "https://registry.npmjs.org/p-cancelable/-/p-cancelable-3.0.0.tgz", "integrity": "sha512-mlVgR3PGuzlo0MmTdk4cXqXWlwQDLnONTAg6sm62XkMJEiRxN3GL3SffkYvqwonbkJBcrI7Uvv5Zh9yjvn2iUw==" }, - "p-map": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/p-map/-/p-map-4.0.0.tgz", - "integrity": "sha512-/bjOqmgETBYB5BoEeGVea8dmvHb2m9GLy1E9W43yeyfP6QQCZGFNa+XRceJEuDB6zqr+gKpIAmlLebMpykw/MQ==", + "path-key": { + "version": "3.1.1", + "resolved": "https://registry.npmjs.org/path-key/-/path-key-3.1.1.tgz", + "integrity": "sha512-ojmeN0qd+y0jszEtoY48r0Peq5dwMEkIlCOu6Q5f41lfkswXuKtYrhgoTpLnyIcHm24Uhqx+5Tqm2InSwLhE6Q==" + }, + "path-scurry": { + "version": "1.10.2", + "resolved": "https://registry.npmjs.org/path-scurry/-/path-scurry-1.10.2.tgz", + "integrity": "sha512-7xTavNy5RQXnsjANvVvMkEjvloOinkAjv/Z6Ildz9v2RinZ4SBKTWFOVRbaF8p0vpHnyjV/UwNDdKuUv6M5qcA==", "requires": { - "aggregate-error": "^3.0.0" + "lru-cache": "^10.2.0", + "minipass": "^5.0.0 || ^6.0.2 || ^7.0.0" } }, "pend": { @@ -2311,9 +2801,28 @@ "integrity": "sha512-F3asv42UuXchdzt+xXqfW1OGlVBe+mxa2mqI0pg5yAHZPvFmY3Y6drSf/GQ1A86WgWEN9Kzh/WrgKa6iGcHXLg==" }, "pretty-bytes": { - "version": "6.0.0", - "resolved": "https://registry.npmjs.org/pretty-bytes/-/pretty-bytes-6.0.0.tgz", - "integrity": "sha512-6UqkYefdogmzqAZWzJ7laYeJnaXDy2/J+ZqiiMtS7t7OfpXWTlaeGMwX8U6EFvPV/YWWEKRkS8hKS4k60WHTOg==" + "version": "6.1.1", + "resolved": "https://registry.npmjs.org/pretty-bytes/-/pretty-bytes-6.1.1.tgz", + "integrity": "sha512-mQUvGU6aUFQ+rNvTIAcZuWGRT9a6f6Yrg9bHs4ImKF+HZCEK+plBvnAZYSIQztknZF2qnzNtr6F8s0+IuptdlQ==" + }, + "protobufjs": { + "version": "7.4.0", + "resolved": "https://registry.npmjs.org/protobufjs/-/protobufjs-7.4.0.tgz", + "integrity": "sha512-mRUWCc3KUU4w1jU8sGxICXH/gNS94DvI1gxqDvBzhj1JpcsimQkYiOJfwsPUykUI5ZaspFbSgmBLER8IrQ3tqw==", + "requires": { + "@protobufjs/aspromise": "^1.1.2", + "@protobufjs/base64": "^1.1.2", + "@protobufjs/codegen": "^2.0.4", + "@protobufjs/eventemitter": "^1.1.0", + "@protobufjs/fetch": "^1.1.0", + "@protobufjs/float": "^1.0.2", + "@protobufjs/inquire": "^1.1.0", + "@protobufjs/path": "^1.1.2", + "@protobufjs/pool": "^1.1.0", + "@protobufjs/utf8": "^1.1.0", + "@types/node": ">=13.7.0", + "long": "^5.0.0" + } }, "pump": { "version": "3.0.0", @@ -2350,18 +2859,20 @@ "integrity": "sha512-0a1F4l73/ZFZOakJnQ3FvkJ2+gSTQWz/r2KE5OdDY0TxPm5h4GkqkWWfM47T7HsbnOtcJVEF4epCVy6u7Q3K+g==" }, "responselike": { - "version": "2.0.1", - "resolved": "https://registry.npmjs.org/responselike/-/responselike-2.0.1.tgz", - "integrity": "sha512-4gl03wn3hj1HP3yzgdI7d3lCkF95F21Pz4BPGvKHinyQzALR5CapwC8yIi0Rh58DEMQ/SguC03wFj2k0M/mHhw==", + "version": "3.0.0", + "resolved": "https://registry.npmjs.org/responselike/-/responselike-3.0.0.tgz", + "integrity": "sha512-40yHxbNcl2+rzXvZuVkrYohathsSJlMTXKryG5y8uciHv1+xDLHQpgjG64JUO9nrEq2jGLH6IZ8BcZyw3wrweg==", "requires": { - "lowercase-keys": "^2.0.0" - }, - "dependencies": { - "lowercase-keys": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/lowercase-keys/-/lowercase-keys-2.0.0.tgz", - "integrity": "sha512-tqNXrS78oMOE73NMxK4EMLQsQowWf8jKooH9g7xPavRT706R6bkQJ6DY2Te7QukaZsulxa30wQ7bk0pm4XiHmA==" - } + "lowercase-keys": "^3.0.0" + } + }, + "restore-cursor": { + "version": "4.0.0", + "resolved": "https://registry.npmjs.org/restore-cursor/-/restore-cursor-4.0.0.tgz", + "integrity": "sha512-I9fPXU9geO9bHOt9pHHOhOkYerIMsmVaWB0rA2AI9ERh/+x/i7MV5HKBNrg+ljO5eoPVgCcnFuRjJ9uH6I/3eg==", + "requires": { + "onetime": "^5.1.0", + "signal-exit": "^3.0.2" } }, "rfdc": { @@ -2369,12 +2880,12 @@ "resolved": "https://registry.npmjs.org/rfdc/-/rfdc-1.3.0.tgz", "integrity": "sha512-V2hovdzFbOi77/WajaSMXk2OLm+xNIeQdMMuB7icj7bk6zi2F8GGAxigcnDFpJHbNyNcgyJDiP+8nOrY5cZGrA==" }, - "rxjs": { - "version": "7.5.6", - "resolved": "https://registry.npmjs.org/rxjs/-/rxjs-7.5.6.tgz", - "integrity": "sha512-dnyv2/YsXhnm461G+R/Pe5bWP41Nm6LBXEYWI6eiFP4fiwx6WRI/CD0zbdVAudd9xwLEF2IDcKXLHit0FYjUzw==", + "rimraf": { + "version": "5.0.5", + "resolved": "https://registry.npmjs.org/rimraf/-/rimraf-5.0.5.tgz", + "integrity": "sha512-CqDakW+hMe/Bz202FPEymy68P+G50RfMQK+Qo5YUqc9SPipvbGjCGKd0RSKEelbsfQuw3g5NZDSrlZZAJurH1A==", "requires": { - "tslib": "^2.1.0" + "glob": "^10.3.7" } }, "safe-buffer": { @@ -2387,28 +2898,42 @@ "resolved": "https://registry.npmjs.org/safer-buffer/-/safer-buffer-2.1.2.tgz", "integrity": "sha512-YZo3K82SD7Riyi0E1EQPojLz7kpepnSQI9IyPbHHg1XXXevb5dJI7tpyN2ADxGcQbHG7vcyRHk0cbwqcQriUtg==" }, + "shebang-command": { + "version": "2.0.0", + "resolved": "https://registry.npmjs.org/shebang-command/-/shebang-command-2.0.0.tgz", + "integrity": "sha512-kHxr2zZpYtdmrN1qDjrrX/Z1rR1kG8Dx+gkpK1G4eXmvXswmcE1hTWBWYUzlraYw1/yZp6YuDY77YtvbN0dmDA==", + "requires": { + "shebang-regex": "^3.0.0" + } + }, + "shebang-regex": { + "version": "3.0.0", + "resolved": "https://registry.npmjs.org/shebang-regex/-/shebang-regex-3.0.0.tgz", + "integrity": "sha512-7++dFhtcx3353uBaq8DDR4NuxBetBzC7ZQOhmTQInHEd6bSrXdiEyzCvG07Z44UYdLShWUyXt5M/yhz8ekcb1A==" + }, "signal-exit": { "version": "3.0.7", "resolved": "https://registry.npmjs.org/signal-exit/-/signal-exit-3.0.7.tgz", "integrity": "sha512-wnD2ZE+l+SPC/uoS0vXeE9L1+0wuaMqKlfz9AMUo38JsyLSBWSFcHR1Rri62LZc12vLr1gb3jl7iwQhgwpAbGQ==" }, "slice-ansi": { - "version": "3.0.0", - "resolved": "https://registry.npmjs.org/slice-ansi/-/slice-ansi-3.0.0.tgz", - "integrity": "sha512-pSyv7bSTC7ig9Dcgbw9AuRNUb5k5V6oDudjZoMBSr13qpLBG7tB+zgCkARjq7xIUgdz5P1Qe8u+rSGdouOOIyQ==", + "version": "5.0.0", + "resolved": "https://registry.npmjs.org/slice-ansi/-/slice-ansi-5.0.0.tgz", + "integrity": "sha512-FC+lgizVPfie0kkhqUScwRu1O/lF6NOgJmlCgK+/LYxDCTk8sGelYaHDhFcDN+Sn3Cv+3VSa4Byeo+IMCzpMgQ==", "requires": { - "ansi-styles": "^4.0.0", - "astral-regex": "^2.0.0", - "is-fullwidth-code-point": "^3.0.0" + "ansi-styles": "^6.0.0", + "is-fullwidth-code-point": "^4.0.0" }, "dependencies": { "ansi-styles": { - "version": "4.3.0", - "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz", - "integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==", - "requires": { - "color-convert": "^2.0.1" - } + "version": "6.2.1", + "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-6.2.1.tgz", + "integrity": "sha512-bN798gFfQX+viw3R7yrGWRqnrN2oRkEkUjjl4JNn4E8GxxbjtG3FbrEIIY3l8/hrwUwIeCZvi4QuOTP4MErVug==" + }, + "is-fullwidth-code-point": { + "version": "4.0.0", + "resolved": "https://registry.npmjs.org/is-fullwidth-code-point/-/is-fullwidth-code-point-4.0.0.tgz", + "integrity": "sha512-O4L094N2/dZ7xqVdrXhh9r1KODPJpFms8B5sGdJLPy664AgvXsreZUyCQQNItZRDlYug4xStLjNp/sz3HvBowQ==" } } }, @@ -2418,14 +2943,14 @@ "integrity": "sha512-Q5thBSxp5t8WPTTJQS59LrGqOZqOsrhDGDVm8azCqIBjSBd7nd9o2PM+mDulQQkh8h//4U6hFZnc/mul8t5pWQ==" }, "ssh2": { - "version": "1.11.0", - "resolved": "https://registry.npmjs.org/ssh2/-/ssh2-1.11.0.tgz", - "integrity": "sha512-nfg0wZWGSsfUe/IBJkXVll3PEZ//YH2guww+mP88gTpuSU4FtZN7zu9JoeTGOyCNx2dTDtT9fOpWwlzyj4uOOw==", + "version": "1.16.0", + "resolved": "https://registry.npmjs.org/ssh2/-/ssh2-1.16.0.tgz", + "integrity": "sha512-r1X4KsBGedJqo7h8F5c4Ybpcr5RjyP+aWIG007uBPRjmdQWfEiVLzSK71Zji1B9sKxwaCvD8y8cwSkYrlLiRRg==", "requires": { - "asn1": "^0.2.4", + "asn1": "^0.2.6", "bcrypt-pbkdf": "^1.0.2", - "cpu-features": "~0.0.4", - "nan": "^2.16.0" + "cpu-features": "~0.0.10", + "nan": "^2.20.0" } }, "string_decoder": { @@ -2436,35 +2961,71 @@ "safe-buffer": "~5.2.0" } }, + "string-width": { + "version": "4.2.3", + "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", + "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", + "requires": { + "emoji-regex": "^8.0.0", + "is-fullwidth-code-point": "^3.0.0", + "strip-ansi": "^6.0.1" + } + }, + "string-width-cjs": { + "version": "npm:string-width@4.2.3", + "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", + "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", + "requires": { + "emoji-regex": "^8.0.0", + "is-fullwidth-code-point": "^3.0.0", + "strip-ansi": "^6.0.1" + } + }, + "strip-ansi": { + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", + "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", + "requires": { + "ansi-regex": "^5.0.1" + } + }, + "strip-ansi-cjs": { + "version": "npm:strip-ansi@6.0.1", + "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", + "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", + "requires": { + "ansi-regex": "^5.0.1" + } + }, "tar": { - "version": "6.1.11", - "resolved": "https://registry.npmjs.org/tar/-/tar-6.1.11.tgz", - "integrity": "sha512-an/KZQzQUkZCkuoAA64hM92X0Urb6VpRhAFllDzz44U2mcD5scmT3zBc4VgVpkugF580+DQn8eAFSyoQt0tznA==", + "version": "7.4.3", + "resolved": "https://registry.npmjs.org/tar/-/tar-7.4.3.tgz", + "integrity": "sha512-5S7Va8hKfV7W5U6g3aYxXmlPoZVAwUMy9AOKyF2fVuZa2UD3qZjg578OrLRt8PcNN1PleVaL/5/yYATNL0ICUw==", "requires": { - "chownr": "^2.0.0", - "fs-minipass": "^2.0.0", - "minipass": "^3.0.0", - "minizlib": "^2.1.1", - "mkdirp": "^1.0.3", - "yallist": "^4.0.0" + "@isaacs/fs-minipass": "^4.0.0", + "chownr": "^3.0.0", + "minipass": "^7.1.2", + "minizlib": "^3.0.1", + "mkdirp": "^3.0.1", + "yallist": "^5.0.0" }, "dependencies": { "chownr": { - "version": "2.0.0", - "resolved": "https://registry.npmjs.org/chownr/-/chownr-2.0.0.tgz", - "integrity": "sha512-bIomtDF5KGpdogkLd9VspvFzk9KfpyyGlS8YFVZl7TGPBHL5snIOnxeshwVgPteQ9b4Eydl+pVbIyE1DcvCWgQ==" + "version": "3.0.0", + "resolved": "https://registry.npmjs.org/chownr/-/chownr-3.0.0.tgz", + "integrity": "sha512-+IxzY9BZOQd/XuYPRmrvEVjF/nqj5kgT4kEq7VofrDoM1MxoRjEWkrCC3EtLi59TVawxTAn+orJwFQcrqEN1+g==" } } }, "tar-fs": { - "version": "2.0.1", - "resolved": "https://registry.npmjs.org/tar-fs/-/tar-fs-2.0.1.tgz", - "integrity": "sha512-6tzWDMeroL87uF/+lin46k+Q+46rAJ0SyPGz7OW7wTgblI273hsBqk2C1j0/xNadNLKDTUL9BukSjB7cwgmlPA==", + "version": "2.1.2", + "resolved": "https://registry.npmjs.org/tar-fs/-/tar-fs-2.1.2.tgz", + "integrity": "sha512-EsaAXwxmx8UB7FRKqeozqEPop69DXcmYwTQwXvyAPF352HJsPdkVhvTaDPYqfNgruveJIJy3TA2l+2zj8LJIJA==", "requires": { "chownr": "^1.1.1", "mkdirp-classic": "^0.5.2", "pump": "^3.0.0", - "tar-stream": "^2.0.0" + "tar-stream": "^2.1.4" } }, "tar-stream": { @@ -2477,44 +3038,18 @@ "fs-constants": "^1.0.0", "inherits": "^2.0.3", "readable-stream": "^3.1.1" - }, - "dependencies": { - "bl": { - "version": "4.1.0", - "resolved": "https://registry.npmjs.org/bl/-/bl-4.1.0.tgz", - "integrity": "sha512-1W07cM9gS6DcLperZfFSj+bWLtaPGSOHWhPiGzXmvVJbRLdG82sH/Kn8EtW1VqWVA54AKf2h5k5BbnIbwF3h6w==", - "requires": { - "buffer": "^5.5.0", - "inherits": "^2.0.4", - "readable-stream": "^3.4.0" - } - }, - "buffer": { - "version": "5.7.1", - "resolved": "https://registry.npmjs.org/buffer/-/buffer-5.7.1.tgz", - "integrity": "sha512-EHcyIPBQ4BSGlvjB16k5KgAJ27CIsHY/2JBmCRReo48y9rQ3MaUzWX3KVlBa4U7MyX02HdVj0K7C3WaB3ju7FQ==", - "requires": { - "base64-js": "^1.3.1", - "ieee754": "^1.1.13" - } - } } }, - "through": { - "version": "2.3.8", - "resolved": "https://registry.npmjs.org/through/-/through-2.3.8.tgz", - "integrity": "sha512-w89qg7PI8wAdvX60bMDP+bFoD5Dvhm9oLheFp5O4a2QF0cSBGsBX4qZmadPMvVqlLJBBci+WqGGOAPvcDeNSVg==" - }, - "tslib": { - "version": "2.4.0", - "resolved": "https://registry.npmjs.org/tslib/-/tslib-2.4.0.tgz", - "integrity": "sha512-d6xOpEDfsi2CZVlPQzGeux8XMwLT9hssAsaPYExaQMuYskwb+x1x7J371tWlbBdWHroy99KnVB6qIkUbs5X3UQ==" - }, "tweetnacl": { "version": "0.14.5", "resolved": "https://registry.npmjs.org/tweetnacl/-/tweetnacl-0.14.5.tgz", "integrity": "sha512-KXXFFdAbFXY4geFIwoyNK+f5Z1b7swfXABfL7HXCmoIWMKU3dmS26672A4EeQtDzLKy7SXmfBu51JolvEKwtGA==" }, + "type-fest": { + "version": "1.4.0", + "resolved": "https://registry.npmjs.org/type-fest/-/type-fest-1.4.0.tgz", + "integrity": "sha512-yGSza74xk0UG8k+pLh5oeoYirvIiWo5t0/o3zHHAO2tRDiZcxWP7fywNlXhqb6/r6sWvwi+RsyQMWhVLe4BVuA==" + }, "universalify": { "version": "2.0.0", "resolved": "https://registry.npmjs.org/universalify/-/universalify-2.0.0.tgz", @@ -2525,6 +3060,39 @@ "resolved": "https://registry.npmjs.org/util-deprecate/-/util-deprecate-1.0.2.tgz", "integrity": "sha512-EPD5q1uXyFxJpCrLnCc1nHnq3gOa6DZBocAIiI2TaSCA7VCJ1UJDMagCzIkXNsUYfD1daK//LTEQ8xiIbrHtcw==" }, + "uuid": { + "version": "10.0.0", + "resolved": "https://registry.npmjs.org/uuid/-/uuid-10.0.0.tgz", + "integrity": "sha512-8XkAphELsDnEGrDxUOHB3RGvXz6TeuYSGEZBOjtTtPm2lwhGBjLgOzLHB63IUWfBpNucQjND6d3AOudO+H3RWQ==" + }, + "which": { + "version": "2.0.2", + "resolved": "https://registry.npmjs.org/which/-/which-2.0.2.tgz", + "integrity": "sha512-BLI3Tl1TW3Pvl70l3yq3Y64i+awpwXqsGBYWkkqMtnbXgrMD+yj7rhW0kuEDxzJaYXGjEW5ogapKNMEKNMjibA==", + "requires": { + "isexe": "^2.0.0" + } + }, + "wrap-ansi": { + "version": "7.0.0", + "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-7.0.0.tgz", + "integrity": "sha512-YVGIj2kamLSTxw6NsZjoBxfSwsn0ycdesmc4p+Q21c5zPuZ1pl+NfxVdxPtdHvmNVOQ6XSYG4AUtyt/Fi7D16Q==", + "requires": { + "ansi-styles": "^4.0.0", + "string-width": "^4.1.0", + "strip-ansi": "^6.0.0" + } + }, + "wrap-ansi-cjs": { + "version": "npm:wrap-ansi@7.0.0", + "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-7.0.0.tgz", + "integrity": "sha512-YVGIj2kamLSTxw6NsZjoBxfSwsn0ycdesmc4p+Q21c5zPuZ1pl+NfxVdxPtdHvmNVOQ6XSYG4AUtyt/Fi7D16Q==", + "requires": { + "ansi-styles": "^4.0.0", + "string-width": "^4.1.0", + "strip-ansi": "^6.0.0" + } + }, "wrappy": { "version": "1.0.2", "resolved": "https://registry.npmjs.org/wrappy/-/wrappy-1.0.2.tgz", @@ -2536,52 +3104,22 @@ "integrity": "sha512-0pfFzegeDWJHJIAmTLRP2DwHjdF5s7jo9tuztdQxAhINCdvS+3nGINqPd00AphqJR/0LhANUS6/+7SCb98YOfA==" }, "yallist": { - "version": "4.0.0", - "resolved": "https://registry.npmjs.org/yallist/-/yallist-4.0.0.tgz", - "integrity": "sha512-3wdGidZyq5PB084XLES5TpOSRA3wjXAlIWMhum2kRcv/41Sn2emQ0dycQW4uZXLejwKvg6EsvbdlVL+FYEct7A==" + "version": "5.0.0", + "resolved": "https://registry.npmjs.org/yallist/-/yallist-5.0.0.tgz", + "integrity": "sha512-YgvUTfwqyc7UXVMrB+SImsVYSmTS8X/tSrtdNZMImM+n7+QTriRXyXim0mBrTXNeqzVF0KWGgHPeiyViFFrNDw==" }, "yargs": { - "version": "17.5.1", - "resolved": "https://registry.npmjs.org/yargs/-/yargs-17.5.1.tgz", - "integrity": "sha512-t6YAJcxDkNX7NFYiVtKvWUz8l+PaKTLiL63mJYWR2GnHq2gjEWISzsLp9wg3aY36dY1j+gfIEL3pIF+XlJJfbA==", + "version": "17.7.2", + "resolved": "https://registry.npmjs.org/yargs/-/yargs-17.7.2.tgz", + "integrity": "sha512-7dSzzRQ++CKnNI/krKnYRV7JKKPUXMEh61soaHKg9mrWEhzFWhFnxPxGl+69cD1Ou63C13NUPCnmIcrvqCuM6w==", "requires": { - "cliui": "^7.0.2", + "cliui": "^8.0.1", "escalade": "^3.1.1", "get-caller-file": "^2.0.5", "require-directory": "^2.1.1", "string-width": "^4.2.3", "y18n": "^5.0.5", - "yargs-parser": "^21.0.0" - }, - "dependencies": { - "ansi-regex": { - "version": "5.0.1", - "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", - "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==" - }, - "emoji-regex": { - "version": "8.0.0", - "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", - "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" - }, - "string-width": { - "version": "4.2.3", - "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", - "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", - "requires": { - "emoji-regex": "^8.0.0", - "is-fullwidth-code-point": "^3.0.0", - "strip-ansi": "^6.0.1" - } - }, - "strip-ansi": { - "version": "6.0.1", - "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", - "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", - "requires": { - "ansi-regex": "^5.0.1" - } - } + "yargs-parser": "^21.1.1" } }, "yargs-parser": { diff --git a/dev/diff/package.json b/dev/diff/package.json index 43d37545f9..a5a7beb0df 100644 --- a/dev/diff/package.json +++ b/dev/diff/package.json @@ -2,19 +2,19 @@ "name": "diff", "type": "module", "dependencies": { - "chalk": "^5.0.1", - "dockerode": "^3.3.3", - "enquirer": "^2.3.6", + "chalk": "^5.4.1", + "dockerode": "^4.0.6", + "enquirer": "^2.4.1", "extract-zip": "^2.0.1", - "fs-extra": "^10.1.0", - "got": "^12.3.1", + "fs-extra": "^11.3.0", + "got": "^13.0.0", "keypress": "^0.2.1", - "listr2": "^5.0.2", + "listr2": "^6.6.1", "lodash-es": "^4.17.21", - "luxon": "^3.0.1", - "pretty-bytes": "^6.0.0", - "tar": "^6.1.11", - "yargs": "^17.5.1" + "luxon": "^3.6.1", + "pretty-bytes": "^6.1.1", + "tar": "^7.4.3", + "yargs": "^17.7.2" }, "engines": { "node": ">=16" diff --git a/dev/diff/prepare.sh b/dev/diff/prepare.sh index ad281c1831..5fce2b0564 100644 --- a/dev/diff/prepare.sh +++ b/dev/diff/prepare.sh @@ -14,5 +14,5 @@ chmod +x ./docker/scripts/app-create-dirs.sh ./docker/scripts/app-create-dirs.sh ./ietf/manage.py check -./ietf/manage.py migrate +./ietf/manage.py migrate --fake-initial diff --git a/dev/diff/settings_local.py b/dev/diff/settings_local.py index 902fc54d14..c255cac23d 100644 --- a/dev/diff/settings_local.py +++ b/dev/diff/settings_local.py @@ -1,35 +1,25 @@ # Copyright The IETF Trust 2007-2019, All Rights Reserved # -*- coding: utf-8 -*- -from ietf.settings import * # pyflakes:ignore +from ietf.settings import * # pyflakes:ignore ALLOWED_HOSTS = ['*'] DATABASES = { 'default': { 'HOST': '__DBHOST__', - 'PORT': 3306, - 'NAME': 'ietf_utf8', - 'ENGINE': 'django.db.backends.mysql', + 'PORT': 5432, + 'NAME': 'datatracker', + 'ENGINE': 'django.db.backends.postgresql', 'USER': 'django', 'PASSWORD': 'RkTkDPFnKpko', - 'OPTIONS': { - 'sql_mode': 'STRICT_TRANS_TABLES', - 'init_command': 'SET storage_engine=InnoDB; SET names "utf8"', - }, }, } -DATABASE_TEST_OPTIONS = { - 'init_command': 'SET storage_engine=InnoDB', -} IDSUBMIT_IDNITS_BINARY = "/usr/local/bin/idnits" IDSUBMIT_REPOSITORY_PATH = "test/id/" IDSUBMIT_STAGING_PATH = "test/staging/" -INTERNET_DRAFT_ARCHIVE_DIR = "test/archive/" -INTERNET_ALL_DRAFTS_ARCHIVE_DIR = "test/archive/" -RFC_PATH = "test/rfc/" AGENDA_PATH = '/assets/www6s/proceedings/' MEETINGHOST_LOGO_PATH = AGENDA_PATH @@ -47,7 +37,6 @@ SUBMIT_YANG_CATALOG_MODEL_DIR = '/assets/ietf-ftp/yang/catalogmod/' SUBMIT_YANG_DRAFT_MODEL_DIR = '/assets/ietf-ftp/yang/draftmod/' -SUBMIT_YANG_INVAL_MODEL_DIR = '/assets/ietf-ftp/yang/invalmod/' SUBMIT_YANG_IANA_MODEL_DIR = '/assets/ietf-ftp/yang/ianamod/' SUBMIT_YANG_RFC_MODEL_DIR = '/assets/ietf-ftp/yang/rfcmod/' @@ -67,8 +56,11 @@ BOFREQ_PATH = '/assets/ietf-ftp/bofreq/' CONFLICT_REVIEW_PATH = '/assets/ietf-ftp/conflict-reviews/' STATUS_CHANGE_PATH = '/assets/ietf-ftp/status-changes/' -INTERNET_DRAFT_ARCHIVE_DIR = '/assets/ietf-ftp/internet-drafts/' +INTERNET_DRAFT_ARCHIVE_DIR = '/assets/collection/draft-archive' INTERNET_ALL_DRAFTS_ARCHIVE_DIR = '/assets/ietf-ftp/internet-drafts/' +BIBXML_BASE_PATH = '/assets/ietfdata/derived/bibxml' +FTP_DIR = '/assets/ftp' +NFS_METRICS_TMP_DIR = '/assets/tmp' NOMCOM_PUBLIC_KEYS_DIR = 'data/nomcom_keys/public_keys/' SLIDE_STAGING_PATH = 'test/staging/' diff --git a/dev/k8s-get-deploy-name/.editorconfig b/dev/k8s-get-deploy-name/.editorconfig new file mode 100644 index 0000000000..fec5c66519 --- /dev/null +++ b/dev/k8s-get-deploy-name/.editorconfig @@ -0,0 +1,7 @@ +[*] +indent_size = 2 +indent_style = space +charset = utf-8 +trim_trailing_whitespace = false +end_of_line = lf +insert_final_newline = true diff --git a/dev/k8s-get-deploy-name/.gitignore b/dev/k8s-get-deploy-name/.gitignore new file mode 100644 index 0000000000..07e6e472cc --- /dev/null +++ b/dev/k8s-get-deploy-name/.gitignore @@ -0,0 +1 @@ +/node_modules diff --git a/dev/k8s-get-deploy-name/.npmrc b/dev/k8s-get-deploy-name/.npmrc new file mode 100644 index 0000000000..580a68c499 --- /dev/null +++ b/dev/k8s-get-deploy-name/.npmrc @@ -0,0 +1,3 @@ +audit = false +fund = false +save-exact = true diff --git a/dev/k8s-get-deploy-name/README.md b/dev/k8s-get-deploy-name/README.md new file mode 100644 index 0000000000..a6605e4dd2 --- /dev/null +++ b/dev/k8s-get-deploy-name/README.md @@ -0,0 +1,16 @@ +# Datatracker Get Deploy Name + +This tool process and slugify a git branch into an appropriate subdomain name. + +## Usage + +1. From the `dev/k8s-get-deploy-name` directory, install the dependencies: +```sh +npm install +``` +2. Run the command: (replacing the `branch` argument) +```sh +node /cli.js --branch feat/fooBar-123 +``` + +The subdomain name will be output. It can then be used in a workflow as a namespace name and subdomain value. diff --git a/dev/k8s-get-deploy-name/cli.js b/dev/k8s-get-deploy-name/cli.js new file mode 100644 index 0000000000..b6c3b5119e --- /dev/null +++ b/dev/k8s-get-deploy-name/cli.js @@ -0,0 +1,22 @@ +#!/usr/bin/env node + +import yargs from 'yargs/yargs' +import { hideBin } from 'yargs/helpers' +import slugify from 'slugify' + +const argv = yargs(hideBin(process.argv)).argv + +let branch = argv.branch +if (!branch) { + throw new Error('Missing --branch argument!') +} +if (branch.indexOf('/') >= 0) { + branch = branch.split('/').slice(1).join('-') +} +branch = slugify(branch, { lower: true, strict: true }) +if (branch.length < 1) { + throw new Error('Branch name is empty!') +} +process.stdout.write(`dt-${branch}`) + +process.exit(0) diff --git a/dev/k8s-get-deploy-name/package-lock.json b/dev/k8s-get-deploy-name/package-lock.json new file mode 100644 index 0000000000..e492a4cd38 --- /dev/null +++ b/dev/k8s-get-deploy-name/package-lock.json @@ -0,0 +1,303 @@ +{ + "name": "k8s-get-deploy-name", + "lockfileVersion": 2, + "requires": true, + "packages": { + "": { + "name": "k8s-get-deploy-name", + "dependencies": { + "slugify": "1.6.6", + "yargs": "17.7.2" + } + }, + "node_modules/ansi-regex": { + "version": "5.0.1", + "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", + "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==", + "engines": { + "node": ">=8" + } + }, + "node_modules/ansi-styles": { + "version": "4.3.0", + "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz", + "integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==", + "dependencies": { + "color-convert": "^2.0.1" + }, + "engines": { + "node": ">=8" + }, + "funding": { + "url": "https://github.com/chalk/ansi-styles?sponsor=1" + } + }, + "node_modules/cliui": { + "version": "8.0.1", + "resolved": "https://registry.npmjs.org/cliui/-/cliui-8.0.1.tgz", + "integrity": "sha512-BSeNnyus75C4//NQ9gQt1/csTXyo/8Sb+afLAkzAptFuMsod9HFokGNudZpi/oQV73hnVK+sR+5PVRMd+Dr7YQ==", + "dependencies": { + "string-width": "^4.2.0", + "strip-ansi": "^6.0.1", + "wrap-ansi": "^7.0.0" + }, + "engines": { + "node": ">=12" + } + }, + "node_modules/color-convert": { + "version": "2.0.1", + "resolved": "https://registry.npmjs.org/color-convert/-/color-convert-2.0.1.tgz", + "integrity": "sha512-RRECPsj7iu/xb5oKYcsFHSppFNnsj/52OVTRKb4zP5onXwVF3zVmmToNcOfGC+CRDpfK/U584fMg38ZHCaElKQ==", + "dependencies": { + "color-name": "~1.1.4" + }, + "engines": { + "node": ">=7.0.0" + } + }, + "node_modules/color-name": { + "version": "1.1.4", + "resolved": "https://registry.npmjs.org/color-name/-/color-name-1.1.4.tgz", + "integrity": "sha512-dOy+3AuW3a2wNbZHIuMZpTcgjGuLU/uBL/ubcZF9OXbDo8ff4O8yVp5Bf0efS8uEoYo5q4Fx7dY9OgQGXgAsQA==" + }, + "node_modules/emoji-regex": { + "version": "8.0.0", + "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", + "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" + }, + "node_modules/escalade": { + "version": "3.1.1", + "resolved": "https://registry.npmjs.org/escalade/-/escalade-3.1.1.tgz", + "integrity": "sha512-k0er2gUkLf8O0zKJiAhmkTnJlTvINGv7ygDNPbeIsX/TJjGJZHuh9B2UxbsaEkmlEo9MfhrSzmhIlhRlI2GXnw==", + "engines": { + "node": ">=6" + } + }, + "node_modules/get-caller-file": { + "version": "2.0.5", + "resolved": "https://registry.npmjs.org/get-caller-file/-/get-caller-file-2.0.5.tgz", + "integrity": "sha512-DyFP3BM/3YHTQOCUL/w0OZHR0lpKeGrxotcHWcqNEdnltqFwXVfhEBQ94eIo34AfQpo0rGki4cyIiftY06h2Fg==", + "engines": { + "node": "6.* || 8.* || >= 10.*" + } + }, + "node_modules/is-fullwidth-code-point": { + "version": "3.0.0", + "resolved": "https://registry.npmjs.org/is-fullwidth-code-point/-/is-fullwidth-code-point-3.0.0.tgz", + "integrity": "sha512-zymm5+u+sCsSWyD9qNaejV3DFvhCKclKdizYaJUuHA83RLjb7nSuGnddCHGv0hk+KY7BMAlsWeK4Ueg6EV6XQg==", + "engines": { + "node": ">=8" + } + }, + "node_modules/require-directory": { + "version": "2.1.1", + "resolved": "https://registry.npmjs.org/require-directory/-/require-directory-2.1.1.tgz", + "integrity": "sha512-fGxEI7+wsG9xrvdjsrlmL22OMTTiHRwAMroiEeMgq8gzoLC/PQr7RsRDSTLUg/bZAZtF+TVIkHc6/4RIKrui+Q==", + "engines": { + "node": ">=0.10.0" + } + }, + "node_modules/slugify": { + "version": "1.6.6", + "resolved": "https://registry.npmjs.org/slugify/-/slugify-1.6.6.tgz", + "integrity": "sha512-h+z7HKHYXj6wJU+AnS/+IH8Uh9fdcX1Lrhg1/VMdf9PwoBQXFcXiAdsy2tSK0P6gKwJLXp02r90ahUCqHk9rrw==", + "engines": { + "node": ">=8.0.0" + } + }, + "node_modules/string-width": { + "version": "4.2.3", + "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", + "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", + "dependencies": { + "emoji-regex": "^8.0.0", + "is-fullwidth-code-point": "^3.0.0", + "strip-ansi": "^6.0.1" + }, + "engines": { + "node": ">=8" + } + }, + "node_modules/strip-ansi": { + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", + "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", + "dependencies": { + "ansi-regex": "^5.0.1" + }, + "engines": { + "node": ">=8" + } + }, + "node_modules/wrap-ansi": { + "version": "7.0.0", + "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-7.0.0.tgz", + "integrity": "sha512-YVGIj2kamLSTxw6NsZjoBxfSwsn0ycdesmc4p+Q21c5zPuZ1pl+NfxVdxPtdHvmNVOQ6XSYG4AUtyt/Fi7D16Q==", + "dependencies": { + "ansi-styles": "^4.0.0", + "string-width": "^4.1.0", + "strip-ansi": "^6.0.0" + }, + "engines": { + "node": ">=10" + }, + "funding": { + "url": "https://github.com/chalk/wrap-ansi?sponsor=1" + } + }, + "node_modules/y18n": { + "version": "5.0.8", + "resolved": "https://registry.npmjs.org/y18n/-/y18n-5.0.8.tgz", + "integrity": "sha512-0pfFzegeDWJHJIAmTLRP2DwHjdF5s7jo9tuztdQxAhINCdvS+3nGINqPd00AphqJR/0LhANUS6/+7SCb98YOfA==", + "engines": { + "node": ">=10" + } + }, + "node_modules/yargs": { + "version": "17.7.2", + "resolved": "https://registry.npmjs.org/yargs/-/yargs-17.7.2.tgz", + "integrity": "sha512-7dSzzRQ++CKnNI/krKnYRV7JKKPUXMEh61soaHKg9mrWEhzFWhFnxPxGl+69cD1Ou63C13NUPCnmIcrvqCuM6w==", + "dependencies": { + "cliui": "^8.0.1", + "escalade": "^3.1.1", + "get-caller-file": "^2.0.5", + "require-directory": "^2.1.1", + "string-width": "^4.2.3", + "y18n": "^5.0.5", + "yargs-parser": "^21.1.1" + }, + "engines": { + "node": ">=12" + } + }, + "node_modules/yargs-parser": { + "version": "21.1.1", + "resolved": "https://registry.npmjs.org/yargs-parser/-/yargs-parser-21.1.1.tgz", + "integrity": "sha512-tVpsJW7DdjecAiFpbIB1e3qxIQsE6NoPc5/eTdrbbIC4h0LVsWhnoa3g+m2HclBIujHzsxZ4VJVA+GUuc2/LBw==", + "engines": { + "node": ">=12" + } + } + }, + "dependencies": { + "ansi-regex": { + "version": "5.0.1", + "resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.1.tgz", + "integrity": "sha512-quJQXlTSUGL2LH9SUXo8VwsY4soanhgo6LNSm84E1LBcE8s3O0wpdiRzyR9z/ZZJMlMWv37qOOb9pdJlMUEKFQ==" + }, + "ansi-styles": { + "version": "4.3.0", + "resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz", + "integrity": "sha512-zbB9rCJAT1rbjiVDb2hqKFHNYLxgtk8NURxZ3IZwD3F6NtxbXZQCnnSi1Lkx+IDohdPlFp222wVALIheZJQSEg==", + "requires": { + "color-convert": "^2.0.1" + } + }, + "cliui": { + "version": "8.0.1", + "resolved": "https://registry.npmjs.org/cliui/-/cliui-8.0.1.tgz", + "integrity": "sha512-BSeNnyus75C4//NQ9gQt1/csTXyo/8Sb+afLAkzAptFuMsod9HFokGNudZpi/oQV73hnVK+sR+5PVRMd+Dr7YQ==", + "requires": { + "string-width": "^4.2.0", + "strip-ansi": "^6.0.1", + "wrap-ansi": "^7.0.0" + } + }, + "color-convert": { + "version": "2.0.1", + "resolved": "https://registry.npmjs.org/color-convert/-/color-convert-2.0.1.tgz", + "integrity": "sha512-RRECPsj7iu/xb5oKYcsFHSppFNnsj/52OVTRKb4zP5onXwVF3zVmmToNcOfGC+CRDpfK/U584fMg38ZHCaElKQ==", + "requires": { + "color-name": "~1.1.4" + } + }, + "color-name": { + "version": "1.1.4", + "resolved": "https://registry.npmjs.org/color-name/-/color-name-1.1.4.tgz", + "integrity": "sha512-dOy+3AuW3a2wNbZHIuMZpTcgjGuLU/uBL/ubcZF9OXbDo8ff4O8yVp5Bf0efS8uEoYo5q4Fx7dY9OgQGXgAsQA==" + }, + "emoji-regex": { + "version": "8.0.0", + "resolved": "https://registry.npmjs.org/emoji-regex/-/emoji-regex-8.0.0.tgz", + "integrity": "sha512-MSjYzcWNOA0ewAHpz0MxpYFvwg6yjy1NG3xteoqz644VCo/RPgnr1/GGt+ic3iJTzQ8Eu3TdM14SawnVUmGE6A==" + }, + "escalade": { + "version": "3.1.1", + "resolved": "https://registry.npmjs.org/escalade/-/escalade-3.1.1.tgz", + "integrity": "sha512-k0er2gUkLf8O0zKJiAhmkTnJlTvINGv7ygDNPbeIsX/TJjGJZHuh9B2UxbsaEkmlEo9MfhrSzmhIlhRlI2GXnw==" + }, + "get-caller-file": { + "version": "2.0.5", + "resolved": "https://registry.npmjs.org/get-caller-file/-/get-caller-file-2.0.5.tgz", + "integrity": "sha512-DyFP3BM/3YHTQOCUL/w0OZHR0lpKeGrxotcHWcqNEdnltqFwXVfhEBQ94eIo34AfQpo0rGki4cyIiftY06h2Fg==" + }, + "is-fullwidth-code-point": { + "version": "3.0.0", + "resolved": "https://registry.npmjs.org/is-fullwidth-code-point/-/is-fullwidth-code-point-3.0.0.tgz", + "integrity": "sha512-zymm5+u+sCsSWyD9qNaejV3DFvhCKclKdizYaJUuHA83RLjb7nSuGnddCHGv0hk+KY7BMAlsWeK4Ueg6EV6XQg==" + }, + "require-directory": { + "version": "2.1.1", + "resolved": "https://registry.npmjs.org/require-directory/-/require-directory-2.1.1.tgz", + "integrity": "sha512-fGxEI7+wsG9xrvdjsrlmL22OMTTiHRwAMroiEeMgq8gzoLC/PQr7RsRDSTLUg/bZAZtF+TVIkHc6/4RIKrui+Q==" + }, + "slugify": { + "version": "1.6.6", + "resolved": "https://registry.npmjs.org/slugify/-/slugify-1.6.6.tgz", + "integrity": "sha512-h+z7HKHYXj6wJU+AnS/+IH8Uh9fdcX1Lrhg1/VMdf9PwoBQXFcXiAdsy2tSK0P6gKwJLXp02r90ahUCqHk9rrw==" + }, + "string-width": { + "version": "4.2.3", + "resolved": "https://registry.npmjs.org/string-width/-/string-width-4.2.3.tgz", + "integrity": "sha512-wKyQRQpjJ0sIp62ErSZdGsjMJWsap5oRNihHhu6G7JVO/9jIB6UyevL+tXuOqrng8j/cxKTWyWUwvSTriiZz/g==", + "requires": { + "emoji-regex": "^8.0.0", + "is-fullwidth-code-point": "^3.0.0", + "strip-ansi": "^6.0.1" + } + }, + "strip-ansi": { + "version": "6.0.1", + "resolved": "https://registry.npmjs.org/strip-ansi/-/strip-ansi-6.0.1.tgz", + "integrity": "sha512-Y38VPSHcqkFrCpFnQ9vuSXmquuv5oXOKpGeT6aGrr3o3Gc9AlVa6JBfUSOCnbxGGZF+/0ooI7KrPuUSztUdU5A==", + "requires": { + "ansi-regex": "^5.0.1" + } + }, + "wrap-ansi": { + "version": "7.0.0", + "resolved": "https://registry.npmjs.org/wrap-ansi/-/wrap-ansi-7.0.0.tgz", + "integrity": "sha512-YVGIj2kamLSTxw6NsZjoBxfSwsn0ycdesmc4p+Q21c5zPuZ1pl+NfxVdxPtdHvmNVOQ6XSYG4AUtyt/Fi7D16Q==", + "requires": { + "ansi-styles": "^4.0.0", + "string-width": "^4.1.0", + "strip-ansi": "^6.0.0" + } + }, + "y18n": { + "version": "5.0.8", + "resolved": "https://registry.npmjs.org/y18n/-/y18n-5.0.8.tgz", + "integrity": "sha512-0pfFzegeDWJHJIAmTLRP2DwHjdF5s7jo9tuztdQxAhINCdvS+3nGINqPd00AphqJR/0LhANUS6/+7SCb98YOfA==" + }, + "yargs": { + "version": "17.7.2", + "resolved": "https://registry.npmjs.org/yargs/-/yargs-17.7.2.tgz", + "integrity": "sha512-7dSzzRQ++CKnNI/krKnYRV7JKKPUXMEh61soaHKg9mrWEhzFWhFnxPxGl+69cD1Ou63C13NUPCnmIcrvqCuM6w==", + "requires": { + "cliui": "^8.0.1", + "escalade": "^3.1.1", + "get-caller-file": "^2.0.5", + "require-directory": "^2.1.1", + "string-width": "^4.2.3", + "y18n": "^5.0.5", + "yargs-parser": "^21.1.1" + } + }, + "yargs-parser": { + "version": "21.1.1", + "resolved": "https://registry.npmjs.org/yargs-parser/-/yargs-parser-21.1.1.tgz", + "integrity": "sha512-tVpsJW7DdjecAiFpbIB1e3qxIQsE6NoPc5/eTdrbbIC4h0LVsWhnoa3g+m2HclBIujHzsxZ4VJVA+GUuc2/LBw==" + } + } +} diff --git a/dev/k8s-get-deploy-name/package.json b/dev/k8s-get-deploy-name/package.json new file mode 100644 index 0000000000..849f5d9b8d --- /dev/null +++ b/dev/k8s-get-deploy-name/package.json @@ -0,0 +1,8 @@ +{ + "name": "k8s-get-deploy-name", + "type": "module", + "dependencies": { + "slugify": "1.6.6", + "yargs": "17.7.2" + } +} diff --git a/dev/legacy/add-old-drafts-from-archive.py b/dev/legacy/add-old-drafts-from-archive.py new file mode 100644 index 0000000000..f09c3b4558 --- /dev/null +++ b/dev/legacy/add-old-drafts-from-archive.py @@ -0,0 +1,157 @@ +#!/usr/bin/env python +import sys + +print("This is only here as documention - please read the file") +sys.exit(0) + +# #!/usr/bin/env python +# # Copyright The IETF Trust 2017-2019, All Rights Reserved + +# import datetime +# import os +# import sys +# from pathlib import Path +# from contextlib import closing + +# os.environ["DJANGO_SETTINGS_MODULE"] = "ietf.settings" + +# import django +# django.setup() + +# from django.conf import settings +# from django.core.validators import validate_email, ValidationError +# from ietf.utils.draft import PlaintextDraft +# from ietf.submit.utils import update_authors +# from ietf.utils.timezone import date_today + +# import debug # pyflakes:ignore + +# from ietf.doc.models import Document, NewRevisionDocEvent, DocEvent, State +# from ietf.person.models import Person + +# system = Person.objects.get(name="(System)") +# expired = State.objects.get(type='draft',slug='expired') + +# names = set() +# print 'collecting draft names ...' +# versions = 0 +# for p in Path(settings.INTERNET_DRAFT_PATH).glob('draft*.txt'): +# n = str(p).split('/')[-1].split('-') +# if n[-1][:2].isdigit(): +# name = '-'.join(n[:-1]) +# if '--' in name or '.txt' in name or '[' in name or '=' in name or '&' in name: +# continue +# if name.startswith('draft-draft-'): +# continue +# if name == 'draft-ietf-trade-iotp-v1_0-dsig': +# continue +# if len(n[-1]) != 6: +# continue +# if name.startswith('draft-mlee-'): +# continue +# names.add('-'.join(n[:-1])) + +# count=0 +# print 'iterating through names ...' +# for name in sorted(names): +# if not Document.objects.filter(name=name).exists(): +# paths = list(Path(settings.INTERNET_DRAFT_PATH).glob('%s-??.txt'%name)) +# paths.sort() +# doc = None +# for p in paths: +# n = str(p).split('/')[-1].split('-') +# rev = n[-1][:2] +# with open(str(p)) as txt_file: +# raw = txt_file.read() +# try: +# text = raw.decode('utf8') +# except UnicodeDecodeError: +# text = raw.decode('latin1') +# try: +# draft = PlaintextDraft(text, txt_file.name, name_from_source=True) +# except Exception as e: +# print name, rev, "Can't parse", p,":",e +# continue +# if draft.errors and draft.errors.keys()!=['draftname',]: +# print "Errors - could not process", name, rev, datetime.datetime.fromtimestamp(p.stat().st_mtime, datetime.timezone.utc), draft.errors, draft.get_title().encode('utf8') +# else: +# time = datetime.datetime.fromtimestamp(p.stat().st_mtime, datetime.timezone.utc) +# if not doc: +# doc = Document.objects.create(name=name, +# time=time, +# type_id='draft', +# title=draft.get_title(), +# abstract=draft.get_abstract(), +# rev = rev, +# pages=draft.get_pagecount(), +# words=draft.get_wordcount(), +# expires=time+datetime.timedelta(settings.INTERNET_DRAFT_DAYS_TO_EXPIRE), +# ) +# DocAlias.objects.create(name=doc.name).docs.add(doc) +# doc.states.add(expired) +# # update authors +# authors = [] +# for author in draft.get_author_list(): +# full_name, first_name, middle_initial, last_name, name_suffix, email, country, company = author + +# author_name = full_name.replace("\n", "").replace("\r", "").replace("<", "").replace(">", "").strip() + +# if email: +# try: +# validate_email(email) +# except ValidationError: +# email = "" + +# def turn_into_unicode(s): +# if s is None: +# return u"" + +# if isinstance(s, unicode): +# return s +# else: +# try: +# return s.decode("utf-8") +# except UnicodeDecodeError: +# try: +# return s.decode("latin-1") +# except UnicodeDecodeError: +# return "" + +# author_name = turn_into_unicode(author_name) +# email = turn_into_unicode(email) +# company = turn_into_unicode(company) + +# authors.append({ +# "name": author_name, +# "email": email, +# "affiliation": company, +# "country": country +# }) +# dummysubmission=type('', (), {})() #https://stackoverflow.com/questions/19476816/creating-an-empty-object-in-python +# dummysubmission.authors = authors +# update_authors(doc,dummysubmission) + +# # add a docevent with words explaining where this came from +# events = [] +# e = NewRevisionDocEvent.objects.create( +# type="new_revision", +# doc=doc, +# rev=rev, +# by=system, +# desc="New version available: %s-%s.txt" % (doc.name, doc.rev), +# time=time, +# ) +# events.append(e) +# e = DocEvent.objects.create( +# type="comment", +# doc = doc, +# rev = rev, +# by = system, +# desc = "Revision added from id-archive on %s by %s"%(date_today(),sys.argv[0]), +# time=time, +# ) +# events.append(e) +# doc.time = time +# doc.rev = rev +# doc.save_with_history(events) +# print "Added",name, rev diff --git a/dev/legacy/notes/notes.html b/dev/legacy/notes/notes.html index 85980a5b1b..cb10a18689 100644 --- a/dev/legacy/notes/notes.html +++ b/dev/legacy/notes/notes.html @@ -355,7 +355,7 @@

Introduction

in one place.

With my recent investigations of code analysis tools, I thought it might be a good idea to start collecting these in one place for the project.

-
+
Henrik <henrik@levkowetz.com>, 23 Mar 2014
@@ -398,8 +398,9 @@

PyChecker

do the right thing, but once it was made to run on the datatracker code, and ignore the django code, it didn't report anything that PyFlakes hadn't already caught.

-
-Henrik <henrik@levkowetz.com>, 23 Mar 2014
+
+ Henrik <henrik@levkowetz.com>, 23 Mar 2014 +
diff --git a/dev/legacy/recalculate-rfc-authors-snapshot b/dev/legacy/recalculate-rfc-authors-snapshot new file mode 100644 index 0000000000..cbe7a7a2f3 --- /dev/null +++ b/dev/legacy/recalculate-rfc-authors-snapshot @@ -0,0 +1,8346 @@ +#!/bin/sh + +echo "This is only here as documentation. Please read the file" +exit + +# #!/usr/bin/env python + +# """DANGER, WILL ROBINSON + +# This code was used in the construction of person migrations 0016 through 0019 +# and doc migration 0029 The original intent was to provide a utility that could +# be run periodically, but the need for manual inspection of the results was too +# great. It is here as a _starting point_ for future exploration of updating rfc +# documentauthor sets. Be careful to check that assumptions haven't changed in +# the interim. + +# """ + +# import os, sys +# import django +# import argparse +# from collections import namedtuple +# import subprocess +# from tempfile import mkstemp + +# basedir = os.path.abspath(os.path.join(os.path.dirname(os.path.abspath(__file__)), '../..')) +# sys.path.insert(0, basedir) +# os.environ.setdefault("DJANGO_SETTINGS_MODULE", "ietf.settings") + +# django.setup() + +# from ietf.person.models import Email, Person +# from ietf.doc.models import Document +# from ietf import settings + +# import debug + +# # This is a snapshot dump from the RFC Editor in late April 2017 +# rfced_data = """RFC1 || S. Crocker || +# RFC2 || B. Duvall || +# RFC3 || S.D. Crocker || +# RFC4 || E.B. Shapiro || +# RFC5 || J. Rulifson || +# RFC6 || S.D. Crocker || +# RFC7 || G. Deloche || +# RFC8 || G. Deloche || +# RFC9 || G. Deloche || +# RFC10 || S.D. Crocker || +# RFC11 || G. Deloche || +# RFC12 || M. Wingfield || +# RFC13 || V. Cerf || +# RFC14 || || +# RFC15 || C.S. Carr || +# RFC16 || S. Crocker || +# RFC17 || J.E. Kreznar || +# RFC18 || V. Cerf || +# RFC19 || J.E. Kreznar || +# RFC20 || V.G. Cerf || +# RFC21 || V.G. Cerf || +# RFC22 || V.G. Cerf || +# RFC23 || G. Gregg || +# RFC24 || S.D. Crocker || +# RFC25 || S.D. Crocker || +# RFC26 || || +# RFC27 || S.D. Crocker || +# RFC28 || W.K. English || +# RFC29 || R.E. Kahn || +# RFC30 || S.D. Crocker || +# RFC31 || D. Bobrow, W.R. Sutherland || +# RFC32 || J. Cole || +# RFC33 || S.D. Crocker || +# RFC34 || W.K. English || +# RFC35 || S.D. Crocker || +# RFC36 || S.D. Crocker || +# RFC37 || S.D. Crocker || +# RFC38 || S.M. Wolfe || +# RFC39 || E. Harslem, J.F. Heafner || +# RFC40 || E. Harslem, J.F. Heafner || +# RFC41 || J.T. Melvin || +# RFC42 || E. Ancona || +# RFC43 || A.G. Nemeth || +# RFC44 || A. Shoshani, R. Long, A. Landsberg || +# RFC45 || J. Postel, S.D. Crocker || +# RFC46 || E. Meyer || +# RFC47 || J. Postel, S. Crocker || +# RFC48 || J. Postel, S.D. Crocker || +# RFC49 || E. Meyer || +# RFC50 || E. Harslen, J. Heafner || +# RFC51 || M. Elie || +# RFC52 || J. Postel, S.D. Crocker || +# RFC53 || S.D. Crocker || +# RFC54 || S.D. Crocker, J. Postel, J. Newkirk, M. Kraley || +# RFC55 || J. Newkirk, M. Kraley, J. Postel, S.D. Crocker || +# RFC56 || E. Belove, D. Black, R. Flegal, L.G. Farquar || +# RFC57 || M. Kraley, J. Newkirk || +# RFC58 || T.P. Skinner || +# RFC59 || E. Meyer || +# RFC60 || R. Kalin || +# RFC61 || D.C. Walden || +# RFC62 || D.C. Walden || +# RFC63 || V.G. Cerf || +# RFC64 || M. Elie || +# RFC65 || D.C. Walden || +# RFC66 || S.D. Crocker || +# RFC67 || W.R. Crowther || +# RFC68 || M. Elie || +# RFC69 || A.K. Bhushan || +# RFC70 || S.D. Crocker || +# RFC71 || T. Schipper || +# RFC72 || R.D. Bressler || +# RFC73 || S.D. Crocker || +# RFC74 || J.E. White || +# RFC75 || S.D. Crocker || +# RFC76 || J. Bouknight, J. Madden, G.R. Grossman || +# RFC77 || J. Postel || +# RFC78 || E. Harslem, J.F. Heafner, J.E. White || +# RFC79 || E. Meyer || +# RFC80 || E. Harslem, J.F. Heafner || +# RFC81 || J. Bouknight || +# RFC82 || E. Meyer || +# RFC83 || R.H. Anderson, E. Harslem, J.F. Heafner || +# RFC84 || J.B. North || +# RFC85 || S.D. Crocker || +# RFC86 || S.D. Crocker || +# RFC87 || A. Vezza || +# RFC88 || R.T. Braden, S.M. Wolfe || +# RFC89 || R.M. Metcalfe || +# RFC90 || R.T. Braden || +# RFC91 || G.H. Mealy || +# RFC92 || || +# RFC93 || A.M. McKenzie || +# RFC94 || E. Harslem, J.F. Heafner || +# RFC95 || S. Crocker || +# RFC96 || R.W. Watson || +# RFC97 || J.T. Melvin, R.W. Watson || +# RFC98 || E. Meyer, T. Skinner || +# RFC99 || P.M. Karp || +# RFC100 || P.M. Karp || +# RFC101 || R.W. Watson || +# RFC102 || S.D. Crocker || +# RFC103 || R.B. Kalin || +# RFC104 || J.B. Postel, S.D. Crocker || +# RFC105 || J.E. White || +# RFC106 || T.C. O'Sullivan || +# RFC107 || R.D. Bressler, S.D. Crocker, W.R. Crowther, G.R. Grossman, R.S. Tomlinson, J.E. White || +# RFC108 || R.W. Watson || +# RFC109 || J. Winett || +# RFC110 || J. Winett || +# RFC111 || S.D. Crocker || +# RFC112 || T.C. O'Sullivan || +# RFC113 || E. Harslem, J.F. Heafner, J.E. White || +# RFC114 || A.K. Bhushan || +# RFC115 || R.W. Watson, J.B. North || +# RFC116 || S.D. Crocker || +# RFC117 || J. Wong || +# RFC118 || R.W. Watson || +# RFC119 || M. Krilanovich || +# RFC120 || M. Krilanovich || +# RFC121 || M. Krilanovich || +# RFC122 || J.E. White || +# RFC123 || S.D. Crocker || +# RFC124 || J.T. Melvin || +# RFC125 || J. McConnell || +# RFC126 || J. McConnell || +# RFC127 || J. Postel || +# RFC128 || J. Postel || +# RFC129 || E. Harslem, J. Heafner, E. Meyer || +# RFC130 || J.F. Heafner || +# RFC131 || E. Harslem, J.F. Heafner || +# RFC132 || J.E. White || +# RFC133 || R.L. Sunberg || +# RFC134 || A. Vezza || +# RFC135 || W. Hathaway || +# RFC136 || R.E. Kahn || +# RFC137 || T.C. O'Sullivan || +# RFC138 || R.H. Anderson, V.G. Cerf, E. Harslem, J.F. Heafner, J. Madden, R.M. Metcalfe, A. Shoshani, J.E. White, D.C.M. Wood || +# RFC139 || T.C. O'Sullivan || +# RFC140 || S.D. Crocker || +# RFC141 || E. Harslem, J.F. Heafner || +# RFC142 || C. Kline, J. Wong || +# RFC143 || W. Naylor, J. Wong, C. Kline, J. Postel || +# RFC144 || A. Shoshani || +# RFC145 || J. Postel || +# RFC146 || P.M. Karp, D.B. McKay, D.C.M. Wood || +# RFC147 || J.M. Winett || +# RFC148 || A.K. Bhushan || +# RFC149 || S.D. Crocker || +# RFC150 || R.B. Kalin || +# RFC151 || A. Shoshani || +# RFC152 || M. Wilber || +# RFC153 || J.T. Melvin, R.W. Watson || +# RFC154 || S.D. Crocker || +# RFC155 || J.B. North || +# RFC156 || J. Bouknight || +# RFC157 || V.G. Cerf || +# RFC158 || T.C. O'Sullivan || +# RFC159 || || +# RFC160 || Network Information Center. Stanford Research Institute || +# RFC161 || A. Shoshani || +# RFC162 || M. Kampe || +# RFC163 || V.G. Cerf || +# RFC164 || J.F. Heafner || +# RFC165 || J. Postel || +# RFC166 || R.H. Anderson, V.G. Cerf, E. Harslem, J.F. Heafner, J. Madden, R.M. Metcalfe, A. Shoshani, J.E. White, D.C.M. Wood || +# RFC167 || A.K. Bhushan, R.M. Metcalfe, J.M. Winett || +# RFC168 || J.B. North || +# RFC169 || S.D. Crocker || +# RFC170 || Network Information Center. Stanford Research Institute || +# RFC171 || A. Bhushan, B. Braden, W. Crowther, E. Harslem, J. Heafner, A. McKenize, J. Melvin, B. Sundberg, D. Watson, J. White || +# RFC172 || A. Bhushan, B. Braden, W. Crowther, E. Harslem, J. Heafner, A. McKenzie, J. Melvin, B. Sundberg, D. Watson, J. White || +# RFC173 || P.M. Karp, D.B. McKay || +# RFC174 || J. Postel, V.G. Cerf || +# RFC175 || E. Harslem, J.F. Heafner || +# RFC176 || A.K. Bhushan, R. Kanodia, R.M. Metcalfe, J. Postel || +# RFC177 || J. McConnell || +# RFC178 || I.W. Cotton || +# RFC179 || A.M. McKenzie || +# RFC180 || A.M. McKenzie || +# RFC181 || J. McConnell || +# RFC182 || J.B. North || +# RFC183 || J.M. Winett || +# RFC184 || K.C. Kelley || +# RFC185 || J.B. North || +# RFC186 || J.C. Michener || +# RFC187 || D.B. McKay, D.P. Karp || +# RFC188 || P.M. Karp, D.B. McKay || +# RFC189 || R.T. Braden || +# RFC190 || L.P. Deutsch || +# RFC191 || C.H. Irby || +# RFC192 || R.W. Watson || +# RFC193 || E. Harslem, J.F. Heafner || +# RFC194 || V. Cerf, E. Harslem, J. Heafner, B. Metcalfe, J. White || +# RFC195 || G.H. Mealy || +# RFC196 || R.W. Watson || +# RFC197 || A. Shoshani, E. Harslem || +# RFC198 || J.F. Heafner || +# RFC199 || T. Williams || +# RFC200 || J.B. North || +# RFC201 || || +# RFC202 || S.M. Wolfe, J. Postel || +# RFC203 || R.B. Kalin || +# RFC204 || J. Postel || +# RFC205 || R.T. Braden || +# RFC206 || J. White || +# RFC207 || A. Vezza || +# RFC208 || A.M. McKenzie || +# RFC209 || B. Cosell || +# RFC210 || W. Conrad || +# RFC211 || J.B. North || +# RFC212 || Information Sciences Institute University of Southern California || +# RFC213 || B. Cosell || +# RFC214 || E. Harslem || +# RFC215 || A.M. McKenzie || +# RFC216 || J.E. White || +# RFC217 || J.E. White || +# RFC218 || B. Cosell || +# RFC219 || R. Winter || +# RFC220 || || +# RFC221 || R.W. Watson || +# RFC222 || R.M. Metcalfe || +# RFC223 || J.T. Melvin, R.W. Watson || +# RFC224 || A.M. McKenzie || +# RFC225 || E. Harslem, R. Stoughton || +# RFC226 || P.M. Karp || +# RFC227 || J.F. Heafner, E. Harslem || +# RFC228 || D.C. Walden || +# RFC229 || J. Postel || +# RFC230 || T. Pyke || +# RFC231 || J.F. Heafner, E. Harslem || +# RFC232 || A. Vezza || +# RFC233 || A. Bhushan, R. Metcalfe || +# RFC234 || A. Vezza || +# RFC235 || E. Westheimer || +# RFC236 || J. Postel || +# RFC237 || R.W. Watson || +# RFC238 || R.T. Braden || +# RFC239 || R.T. Braden || +# RFC240 || A.M. McKenzie || +# RFC241 || A.M. McKenzie || +# RFC242 || L. Haibt, A.P. Mullery || +# RFC243 || A.P. Mullery || +# RFC244 || || +# RFC245 || C. Falls || +# RFC246 || A. Vezza || +# RFC247 || P.M. Karp || +# RFC248 || || +# RFC249 || R.F. Borelli || +# RFC250 || H. Brodie || +# RFC251 || D. Stern || +# RFC252 || E. Westheimer || +# RFC253 || J.A. Moorer || +# RFC254 || A. Bhushan || +# RFC255 || E. Westheimer || +# RFC256 || B. Cosell || +# RFC257 || || +# RFC258 || || +# RFC259 || || +# RFC260 || || +# RFC261 || || +# RFC262 || || +# RFC263 || A.M. McKenzie || +# RFC264 || A. Bhushan, B. Braden, W. Crowther, E. Harslem, J. Heafner, A. McKenize, B. Sundberg, D. Watson, J. White || +# RFC265 || A. Bhushan, B. Braden, W. Crowther, E. Harslem, J. Heafner, A. McKenzie, J. Melvin, B. Sundberg, D. Watson, J. White || +# RFC266 || E. Westheimer || +# RFC267 || E. Westheimer || +# RFC268 || J. Postel || +# RFC269 || H. Brodie || +# RFC270 || A.M. McKenzie || +# RFC271 || B. Cosell || +# RFC272 || || +# RFC273 || R.W. Watson || +# RFC274 || E. Forman || +# RFC275 || || +# RFC276 || R.W. Watson || +# RFC277 || || +# RFC278 || A.K. Bhushan, R.T. Braden, E. Harslem, J.F. Heafner, A.M. McKenzie, J.T. Melvin, R.L. Sundberg, R.W. Watson, J.E. White || +# RFC279 || || +# RFC280 || R.W. Watson || +# RFC281 || A.M. McKenzie || +# RFC282 || M.A. Padlipsky || +# RFC283 || R.T. Braden || +# RFC284 || || +# RFC285 || D. Huff || +# RFC286 || E. Forman || +# RFC287 || E. Westheimer || +# RFC288 || E. Westheimer || +# RFC289 || R.W. Watson || +# RFC290 || A.P. Mullery || +# RFC291 || D.B. McKay || +# RFC292 || J.C. Michener, I.W. Cotton, K.C. Kelley, D.E. Liddle, E. Meyer || +# RFC293 || E. Westheimer || +# RFC294 || A.K. Bhushan || +# RFC295 || J. Postel || +# RFC296 || D.E. Liddle || +# RFC297 || D.C. Walden || +# RFC298 || E. Westheimer || +# RFC299 || D. Hopkin || +# RFC300 || J.B. North || +# RFC301 || R. Alter || +# RFC302 || R.F. Bryan || +# RFC303 || Network Information Center. Stanford Research Institute || +# RFC304 || D.B. McKay || +# RFC305 || R. Alter || +# RFC306 || E. Westheimer || +# RFC307 || E. Harslem || +# RFC308 || M. Seriff || +# RFC309 || A.K. Bhushan || +# RFC310 || A.K. Bhushan || +# RFC311 || R.F. Bryan || +# RFC312 || A.M. McKenzie || +# RFC313 || T.C. O'Sullivan || +# RFC314 || I.W. Cotton || +# RFC315 || E. Westheimer || +# RFC316 || D.B. McKay, A.P. Mullery || +# RFC317 || J. Postel || +# RFC318 || J. Postel || +# RFC319 || E. Westheimer || +# RFC320 || R. Reddy || +# RFC321 || P.M. Karp || +# RFC322 || V. Cerf, J. Postel || +# RFC323 || V. Cerf || +# RFC324 || J. Postel || +# RFC325 || G. Hicks || +# RFC326 || E. Westheimer || +# RFC327 || A.K. Bhushan || +# RFC328 || J. Postel || +# RFC329 || Network Information Center. Stanford Research Institute || +# RFC330 || E. Westheimer || +# RFC331 || J.M. McQuillan || +# RFC332 || E. Westheimer || +# RFC333 || R.D. Bressler, D. Murphy, D.C. Walden || +# RFC334 || A.M. McKenzie || +# RFC335 || R.F. Bryan || +# RFC336 || I.W. Cotton || +# RFC337 || || +# RFC338 || R.T. Braden || +# RFC339 || R. Thomas || +# RFC340 || T.C. O'Sullivan || +# RFC341 || || +# RFC342 || E. Westheimer || +# RFC343 || A.M. McKenzie || +# RFC344 || E. Westheimer || +# RFC345 || K.C. Kelley || +# RFC346 || J. Postel || +# RFC347 || J. Postel || +# RFC348 || J. Postel || +# RFC349 || J. Postel || +# RFC350 || R. Stoughton || +# RFC351 || D. Crocker || +# RFC352 || D. Crocker || +# RFC353 || E. Westheimer || +# RFC354 || A.K. Bhushan || +# RFC355 || J. Davidson || +# RFC356 || R. Alter || +# RFC357 || J. Davidson || +# RFC358 || || +# RFC359 || D.C. Walden || +# RFC360 || C. Holland || +# RFC361 || R.D. Bressler || +# RFC362 || E. Westheimer || +# RFC363 || Network Information Center. Stanford Research Institute || +# RFC364 || M.D. Abrams || +# RFC365 || D.C. Walden || +# RFC366 || E. Westheimer || +# RFC367 || E. Westheimer || +# RFC368 || R.T. Braden || +# RFC369 || J.R. Pickens || +# RFC370 || E. Westheimer || +# RFC371 || R.E. Kahn || +# RFC372 || R.W. Watson || +# RFC373 || J. McCarthy || +# RFC374 || A.M. McKenzie || +# RFC375 || || +# RFC376 || E. Westheimer || +# RFC377 || R.T. Braden || +# RFC378 || A.M. McKenzie || +# RFC379 || R. Braden || +# RFC380 || || +# RFC381 || J.M. McQuillan || +# RFC382 || L. McDaniel || +# RFC383 || || +# RFC384 || J.B. North || +# RFC385 || A.K. Bhushan || +# RFC386 || B. Cosell, D.C. Walden || +# RFC387 || K.C. Kelley, J. Meir || +# RFC388 || V. Cerf || +# RFC389 || B. Noble || +# RFC390 || R.T. Braden || +# RFC391 || A.M. McKenzie || +# RFC392 || G. Hicks, B.D. Wessler || +# RFC393 || J.M. Winett || +# RFC394 || J.M. McQuillan || +# RFC395 || J.M. McQuillan || +# RFC396 || S. Bunch || +# RFC397 || || +# RFC398 || J.R. Pickens, E. Faeh || +# RFC399 || M. Krilanovich || +# RFC400 || A.M. McKenzie || +# RFC401 || J. Hansen || +# RFC402 || J.B. North || +# RFC403 || G. Hicks || +# RFC404 || A.M. McKenzie || +# RFC405 || A.M. McKenzie || +# RFC406 || J.M. McQuillan || +# RFC407 || R.D. Bressler, R. Guida, A.M. McKenzie || +# RFC408 || A.D. Owen, J. Postel || +# RFC409 || J.E. White || +# RFC410 || J.M. McQuillan || +# RFC411 || M.A. Padlipsky || +# RFC412 || G. Hicks || +# RFC413 || A.M. McKenzie || +# RFC414 || A.K. Bhushan || +# RFC415 || H. Murray || +# RFC416 || J.C. Norton || +# RFC417 || J. Postel, C. Kline || +# RFC418 || W. Hathaway || +# RFC419 || A. Vezza || +# RFC420 || H. Murray || +# RFC421 || A.M. McKenzie || +# RFC422 || A.M. McKenzie || +# RFC423 || B. Noble || +# RFC424 || || +# RFC425 || R.D. Bressler || +# RFC426 || R. Thomas || +# RFC427 || || +# RFC428 || || +# RFC429 || J. Postel || +# RFC430 || R.T. Braden || +# RFC431 || M. Krilanovich || +# RFC432 || N. Neigus || +# RFC433 || J. Postel || +# RFC434 || A.M. McKenzie || +# RFC435 || B. Cosell, D.C. Walden || +# RFC436 || M. Krilanovich || +# RFC437 || E. Faeh || +# RFC438 || R. Thomas, R. Clements || +# RFC439 || V. Cerf || +# RFC440 || D.C. Walden || +# RFC441 || R.D. Bressler, R. Thomas || +# RFC442 || V. Cerf || +# RFC443 || A.M. McKenzie || +# RFC444 || || +# RFC445 || A.M. McKenzie || +# RFC446 || L.P. Deutsch || +# RFC447 || A.M. McKenzie || +# RFC448 || R.T. Braden || +# RFC449 || D.C. Walden || +# RFC450 || M.A. Padlipsky || +# RFC451 || M.A. Padlipsky || +# RFC452 || J. Winett || +# RFC453 || M.D. Kudlick || +# RFC454 || A.M. McKenzie || +# RFC455 || A.M. McKenzie || +# RFC456 || M.D. Kudlick || +# RFC457 || D.C. Walden || +# RFC458 || R.D. Bressler, R. Thomas || +# RFC459 || W. Kantrowitz || +# RFC460 || C. Kline || +# RFC461 || A.M. McKenzie || +# RFC462 || J. Iseli, D. Crocker || +# RFC463 || A.K. Bhushan || +# RFC464 || M.D. Kudlick || +# RFC465 || || +# RFC466 || J.M. Winett || +# RFC467 || J.D. Burchfiel, R.S. Tomlinson || +# RFC468 || R.T. Braden || +# RFC469 || M.D. Kudlick || +# RFC470 || R. Thomas || +# RFC471 || R. Thomas || +# RFC472 || S. Bunch || +# RFC473 || D.C. Walden || +# RFC474 || S. Bunch || +# RFC475 || A.K. Bhushan || +# RFC476 || A.M. McKenzie || +# RFC477 || M. Krilanovich || +# RFC478 || R.D. Bressler, R. Thomas || +# RFC479 || J.E. White || +# RFC480 || J.E. White || +# RFC481 || || +# RFC482 || A.M. McKenzie || +# RFC483 || M.D. Kudlick || +# RFC484 || || +# RFC485 || J.R. Pickens || +# RFC486 || R.D. Bressler || +# RFC487 || R.D. Bressler || +# RFC488 || M.F. Auerbach || +# RFC489 || J. Postel || +# RFC490 || J.R. Pickens || +# RFC491 || M.A. Padlipsky || +# RFC492 || E. Meyer || +# RFC493 || J.C. Michener, I.W. Cotton, K.C. Kelley, D.E. Liddle, E. Meyer || +# RFC494 || D.C. Walden || +# RFC495 || A.M. McKenzie || +# RFC496 || M.F. Auerbach || +# RFC497 || A.M. McKenzie || +# RFC498 || R.T. Braden || +# RFC499 || B.R. Reussow || +# RFC500 || A. Shoshani, I. Spiegler || +# RFC501 || K.T. Pogran || +# RFC502 || || +# RFC503 || N. Neigus, J. Postel || +# RFC504 || R. Thomas || +# RFC505 || M.A. Padlipsky || +# RFC506 || M.A. Padlipsky || +# RFC507 || || +# RFC508 || L. Pfeifer, J. McAfee || +# RFC509 || A.M. McKenzie || +# RFC510 || J.E. White || +# RFC511 || J.B. North || +# RFC512 || W. Hathaway || +# RFC513 || W. Hathaway || +# RFC514 || W. Kantrowitz || +# RFC515 || R. Winter || +# RFC516 || J. Postel || +# RFC517 || || +# RFC518 || N. Vaughan, E.J. Feinler || +# RFC519 || J.R. Pickens || +# RFC520 || J.D. Day || +# RFC521 || A.M. McKenzie || +# RFC522 || A.M. McKenzie || +# RFC523 || A.K. Bhushan || +# RFC524 || J.E. White || +# RFC525 || W. Parrish, J.R. Pickens || +# RFC526 || W.K. Pratt || +# RFC527 || R. Merryman || +# RFC528 || J.M. McQuillan || +# RFC529 || A.M. McKenzie, R. Thomas, R.S. Tomlinson, K.T. Pogran || +# RFC530 || A.K. Bhushan || +# RFC531 || M.A. Padlipsky || +# RFC532 || R.G. Merryman || +# RFC533 || D.C. Walden || +# RFC534 || D.C. Walden || +# RFC535 || R. Thomas || +# RFC536 || || +# RFC537 || S. Bunch || +# RFC538 || A.M. McKenzie || +# RFC539 || D. Crocker, J. Postel || +# RFC540 || || +# RFC541 || || +# RFC542 || N. Neigus || +# RFC543 || N.D. Meyer || +# RFC544 || N.D. Meyer, K. Kelley || +# RFC545 || J.R. Pickens || +# RFC546 || R. Thomas || +# RFC547 || D.C. Walden || +# RFC548 || D.C. Walden || +# RFC549 || J.C. Michener || +# RFC550 || L.P. Deutsch || +# RFC551 || Y. Feinroth, R. Fink || +# RFC552 || A.D. Owen || +# RFC553 || C.H. Irby, K. Victor || +# RFC554 || || +# RFC555 || J.E. White || +# RFC556 || A.M. McKenzie || +# RFC557 || B.D. Wessler || +# RFC558 || || +# RFC559 || A.K. Bushan || +# RFC560 || D. Crocker, J. Postel || +# RFC561 || A.K. Bhushan, K.T. Pogran, R.S. Tomlinson, J.E. White || +# RFC562 || A.M. McKenzie || +# RFC563 || J. Davidson || +# RFC564 || || +# RFC565 || D. Cantor || +# RFC566 || A.M. McKenzie || +# RFC567 || L.P. Deutsch || +# RFC568 || J.M. McQuillan || +# RFC569 || M.A. Padlipsky || +# RFC570 || J.R. Pickens || +# RFC571 || R. Braden || +# RFC572 || || +# RFC573 || A. Bhushan || +# RFC574 || M. Krilanovich || +# RFC575 || || +# RFC576 || K. Victor || +# RFC577 || D. Crocker || +# RFC578 || A.K. Bhushan, N.D. Ryan || +# RFC579 || A.M. McKenzie || +# RFC580 || J. Postel || +# RFC581 || D. Crocker, J. Postel || +# RFC582 || R. Clements || +# RFC583 || || +# RFC584 || J. Iseli, D. Crocker, N. Neigus || +# RFC585 || D. Crocker, N. Neigus, E.J. Feinler, J. Iseli || +# RFC586 || A.M. McKenzie || +# RFC587 || J. Postel || +# RFC588 || A. Stokes || +# RFC589 || R.T. Braden || +# RFC590 || M.A. Padlipsky || +# RFC591 || D.C. Walden || +# RFC592 || R.W. Watson || +# RFC593 || A.M. McKenzie, J. Postel || +# RFC594 || J.D. Burchfiel || +# RFC595 || W. Hathaway || +# RFC596 || E.A. Taft || +# RFC597 || N. Neigus, E.J. Feinler || +# RFC598 || Network Information Center. Stanford Research Institute || +# RFC599 || R.T. Braden || +# RFC600 || A. Berggreen || +# RFC601 || A.M. McKenzie || +# RFC602 || R.M. Metcalfe || +# RFC603 || J.D. Burchfiel || +# RFC604 || J. Postel || +# RFC605 || || +# RFC606 || L.P. Deutsch || +# RFC607 || M. Krilanovich, G. Gregg || +# RFC608 || M.D. Kudlick || +# RFC609 || B. Ferguson || +# RFC610 || R. Winter, J. Hill, W. Greiff || +# RFC611 || D.C. Walden || +# RFC612 || A.M. McKenzie || +# RFC613 || A.M. McKenzie || +# RFC614 || K.T. Pogran, N. Neigus || +# RFC615 || D. Crocker || +# RFC616 || D. Walden || +# RFC617 || E.A. Taft || +# RFC618 || E.A. Taft || +# RFC619 || W. Naylor, H. Opderbeck || +# RFC620 || B. Ferguson || +# RFC621 || M.D. Kudlick || +# RFC622 || A.M. McKenzie || +# RFC623 || M. Krilanovich || +# RFC624 || M. Krilanovich, G. Gregg, W. Hathaway, J.E. White || +# RFC625 || M.D. Kudlick, E.J. Feinler || +# RFC626 || L. Kleinrock, H. Opderbeck || +# RFC627 || M.D. Kudlick, E.J. Feinler || +# RFC628 || M.L. Keeney || +# RFC629 || J.B. North || +# RFC630 || J. Sussman || +# RFC631 || A. Danthine || +# RFC632 || H. Opderbeck || +# RFC633 || A.M. McKenzie || +# RFC634 || A.M. McKenzie || +# RFC635 || V. Cerf || +# RFC636 || J.D. Burchfiel, B. Cosell, R.S. Tomlinson, D.C. Walden || +# RFC637 || A.M. McKenzie || +# RFC638 || A.M. McKenzie || +# RFC639 || || +# RFC640 || J. Postel || +# RFC641 || || +# RFC642 || J.D. Burchfiel || +# RFC643 || E. Mader || +# RFC644 || R. Thomas || +# RFC645 || D. Crocker || +# RFC646 || || +# RFC647 || M.A. Padlipsky || +# RFC648 || || +# RFC649 || || +# RFC650 || || +# RFC651 || D. Crocker || +# RFC652 || D. Crocker || +# RFC653 || D. Crocker || +# RFC654 || D. Crocker || +# RFC655 || D. Crocker || +# RFC656 || D. Crocker || +# RFC657 || D. Crocker || +# RFC658 || D. Crocker || +# RFC659 || J. Postel || +# RFC660 || D.C. Walden || +# RFC661 || J. Postel || +# RFC662 || R. Kanodia || +# RFC663 || R. Kanodia || +# RFC664 || || +# RFC665 || || +# RFC666 || M.A. Padlipsky || +# RFC667 || S.G. Chipman || +# RFC668 || || +# RFC669 || D.W. Dodds || +# RFC670 || || +# RFC671 || R. Schantz || +# RFC672 || R. Schantz || +# RFC673 || || +# RFC674 || J. Postel, J.E. White || +# RFC675 || V. Cerf, Y. Dalal, C. Sunshine || +# RFC676 || || +# RFC677 || P.R. Johnson, R. Thomas || +# RFC678 || J. Postel || +# RFC679 || D.W. Dodds || +# RFC680 || T.H. Myer, D.A. Henderson || +# RFC681 || S. Holmgren || +# RFC682 || || +# RFC683 || R. Clements || +# RFC684 || R. Schantz || +# RFC685 || M. Beeler || +# RFC686 || B. Harvey || +# RFC687 || D.C. Walden || +# RFC688 || D.C. Walden || +# RFC689 || R. Clements || +# RFC690 || J. Postel || +# RFC691 || B. Harvey || +# RFC692 || S.M. Wolfe || +# RFC693 || || +# RFC694 || J. Postel || +# RFC695 || M. Krilanovich || +# RFC696 || V.G. Cerf || +# RFC697 || J. Lieb || +# RFC698 || T. Mock || +# RFC699 || J. Postel, J. Vernon || +# RFC700 || E. Mader, W.W. Plummer, R.S. Tomlinson || +# RFC701 || D.W. Dodds || +# RFC702 || D.W. Dodds || +# RFC703 || D.W. Dodds || +# RFC704 || P.J. Santos || +# RFC705 || R.F. Bryan || +# RFC706 || J. Postel || +# RFC707 || J.E. White || +# RFC708 || J.E. White || +# RFC709 || || +# RFC710 || || +# RFC711 || || +# RFC712 || J.E. Donnelley || +# RFC713 || J. Haverty || +# RFC714 || A.M. McKenzie || +# RFC715 || || +# RFC716 || D.C. Walden, J. Levin || +# RFC717 || J. Postel || +# RFC718 || J. Postel || +# RFC719 || J. Postel || +# RFC720 || D. Crocker || +# RFC721 || L.L. Garlick || +# RFC722 || J. Haverty || +# RFC723 || || +# RFC724 || D. Crocker, K.T. Pogran, J. Vittal, D.A. Henderson || +# RFC725 || J.D. Day, G.R. Grossman || +# RFC726 || J. Postel, D. Crocker || +# RFC727 || M.R. Crispin || +# RFC728 || J.D. Day || +# RFC729 || D. Crocker || +# RFC730 || J. Postel || +# RFC731 || J.D. Day || +# RFC732 || J.D. Day || +# RFC733 || D. Crocker, J. Vittal, K.T. Pogran, D.A. Henderson || +# RFC734 || M.R. Crispin || +# RFC735 || D. Crocker, R.H. Gumpertz || +# RFC736 || M.R. Crispin || +# RFC737 || K. Harrenstien || +# RFC738 || K. Harrenstien || +# RFC739 || J. Postel || +# RFC740 || R.T. Braden || +# RFC741 || D. Cohen || +# RFC742 || K. Harrenstien || +# RFC743 || K. Harrenstien || +# RFC744 || J. Sattley || +# RFC745 || M. Beeler || +# RFC746 || R. Stallman || +# RFC747 || M.R. Crispin || +# RFC748 || M.R. Crispin || +# RFC749 || B. Greenberg || +# RFC750 || J. Postel || +# RFC751 || P.D. Lebling || +# RFC752 || M.R. Crispin || +# RFC753 || J. Postel || +# RFC754 || J. Postel || +# RFC755 || J. Postel || +# RFC756 || J.R. Pickens, E.J. Feinler, J.E. Mathis || +# RFC757 || D.P. Deutsch || +# RFC758 || J. Postel || +# RFC759 || J. Postel || +# RFC760 || J. Postel || +# RFC761 || J. Postel || +# RFC762 || J. Postel || +# RFC763 || M.D. Abrams || +# RFC764 || J. Postel || +# RFC765 || J. Postel || +# RFC766 || J. Postel || +# RFC767 || J. Postel || +# RFC768 || J. Postel || +# RFC769 || J. Postel || +# RFC770 || J. Postel || +# RFC771 || V.G. Cerf, J. Postel || +# RFC772 || S. Sluizer, J. Postel || +# RFC773 || V.G. Cerf || +# RFC774 || J. Postel || +# RFC775 || D. Mankins, D. Franklin, A.D. Owen || +# RFC776 || J. Postel || +# RFC777 || J. Postel || +# RFC778 || D.L. Mills || +# RFC779 || E. Killian || +# RFC780 || S. Sluizer, J. Postel || +# RFC781 || Z. Su || +# RFC782 || J. Nabielsky, A.P. Skelton || +# RFC783 || K.R. Sollins || +# RFC784 || S. Sluizer, J. Postel || +# RFC785 || S. Sluizer, J. Postel || +# RFC786 || S. Sluizer, J. Postel || +# RFC787 || A.L. Chapin || +# RFC788 || J. Postel || +# RFC789 || E.C. Rosen || +# RFC790 || J. Postel || +# RFC791 || J. Postel || +# RFC792 || J. Postel || +# RFC793 || J. Postel || +# RFC794 || V.G. Cerf || +# RFC795 || J. Postel || +# RFC796 || J. Postel || +# RFC797 || A.R. Katz || +# RFC798 || A.R. Katz || +# RFC799 || D.L. Mills || +# RFC800 || J. Postel, J. Vernon || +# RFC801 || J. Postel || +# RFC802 || A.G. Malis || +# RFC803 || A. Agarwal, M.J. O'Connor, D.L. Mills || +# RFC804 || International Telegraph and Telephone Consultative Committee of the International Telecommunication Union || +# RFC805 || J. Postel || +# RFC806 || National Bureau of Standards || +# RFC807 || J. Postel || +# RFC808 || J. Postel || +# RFC809 || T. Chang || +# RFC810 || E.J. Feinler, K. Harrenstien, Z. Su, V. White || +# RFC811 || K. Harrenstien, V. White, E.J. Feinler || +# RFC812 || K. Harrenstien, V. White || +# RFC813 || D.D. Clark || +# RFC814 || D.D. Clark || +# RFC815 || D.D. Clark || +# RFC816 || D.D. Clark || +# RFC817 || D.D. Clark || +# RFC818 || J. Postel || +# RFC819 || Z. Su, J. Postel || +# RFC820 || J. Postel || +# RFC821 || J. Postel || +# RFC822 || D. Crocker || +# RFC823 || R.M. Hinden, A. Sheltzer || bob.hinden@gmail.com +# RFC824 || W.I. MacGregor, D.C. Tappan || +# RFC825 || J. Postel || +# RFC826 || D. Plummer || +# RFC827 || E.C. Rosen || +# RFC828 || K. Owen || +# RFC829 || V.G. Cerf || +# RFC830 || Z. Su || +# RFC831 || R.T. Braden || +# RFC832 || D. Smallberg || +# RFC833 || D. Smallberg || +# RFC834 || D. Smallberg || +# RFC835 || D. Smallberg || +# RFC836 || D. Smallberg || +# RFC837 || D. Smallberg || +# RFC838 || D. Smallberg || +# RFC839 || D. Smallberg || +# RFC840 || J. Postel || +# RFC841 || National Bureau of Standards || +# RFC842 || D. Smallberg || +# RFC843 || D. Smallberg || +# RFC844 || R. Clements || +# RFC845 || D. Smallberg || +# RFC846 || D. Smallberg || +# RFC847 || A. Westine, D. Smallberg, J. Postel || +# RFC848 || D. Smallberg || +# RFC849 || M.R. Crispin || +# RFC850 || M.R. Horton || +# RFC851 || A.G. Malis || +# RFC852 || A.G. Malis || +# RFC853 || || +# RFC854 || J. Postel, J.K. Reynolds || +# RFC855 || J. Postel, J.K. Reynolds || +# RFC856 || J. Postel, J. Reynolds || +# RFC857 || J. Postel, J. Reynolds || +# RFC858 || J. Postel, J. Reynolds || +# RFC859 || J. Postel, J. Reynolds || +# RFC860 || J. Postel, J. Reynolds || +# RFC861 || J. Postel, J. Reynolds || +# RFC862 || J. Postel || +# RFC863 || J. Postel || +# RFC864 || J. Postel || +# RFC865 || J. Postel || +# RFC866 || J. Postel || +# RFC867 || J. Postel || +# RFC868 || J. Postel, K. Harrenstien || +# RFC869 || R. Hinden || bob.hinden@gmail.com +# RFC870 || J.K. Reynolds, J. Postel || +# RFC871 || M.A. Padlipsky || +# RFC872 || M.A. Padlipsky || +# RFC873 || M.A. Padlipsky || +# RFC874 || M.A. Padlipsky || +# RFC875 || M.A. Padlipsky || +# RFC876 || D. Smallberg || +# RFC877 || J.T. Korb || +# RFC878 || A.G. Malis || +# RFC879 || J. Postel || +# RFC880 || J.K. Reynolds, J. Postel || +# RFC881 || J. Postel || +# RFC882 || P.V. Mockapetris || +# RFC883 || P.V. Mockapetris || +# RFC884 || M. Solomon, E. Wimmers || +# RFC885 || J. Postel || +# RFC886 || M.T. Rose || +# RFC887 || M. Accetta || +# RFC888 || L. Seamonson, E.C. Rosen || +# RFC889 || D.L. Mills || +# RFC890 || J. Postel || +# RFC891 || D.L. Mills || +# RFC892 || International Organization for Standardization || +# RFC893 || S. Leffler, M.J. Karels || +# RFC894 || C. Hornig || +# RFC895 || J. Postel || +# RFC896 || J. Nagle || +# RFC897 || J. Postel || +# RFC898 || R.M. Hinden, J. Postel, M. Muuss, J.K. Reynolds || bob.hinden@gmail.com +# RFC899 || J. Postel, A. Westine || +# RFC900 || J.K. Reynolds, J. Postel || +# RFC901 || J.K. Reynolds, J. Postel || +# RFC902 || J.K. Reynolds, J. Postel || +# RFC903 || R. Finlayson, T. Mann, J.C. Mogul, M. Theimer || +# RFC904 || D.L. Mills || +# RFC905 || ISO || +# RFC906 || R. Finlayson || +# RFC907 || Bolt Beranek and Newman Laboratories || +# RFC908 || D. Velten, R.M. Hinden, J. Sax || bob.hinden@gmail.com +# RFC909 || C. Welles, W. Milliken || +# RFC910 || H.C. Forsdick || +# RFC911 || P. Kirton || +# RFC912 || M. St. Johns || +# RFC913 || M. Lottor || +# RFC914 || D.J. Farber, G. Delp, T.M. Conte || +# RFC915 || M.A. Elvy, R. Nedved || +# RFC916 || G.G. Finn || +# RFC917 || J.C. Mogul || +# RFC918 || J.K. Reynolds || +# RFC919 || J.C. Mogul || +# RFC920 || J. Postel, J.K. Reynolds || +# RFC921 || J. Postel || +# RFC922 || J.C. Mogul || +# RFC923 || J.K. Reynolds, J. Postel || +# RFC924 || J.K. Reynolds, J. Postel || +# RFC925 || J. Postel || +# RFC926 || International Organization for Standardization || +# RFC927 || B.A. Anderson || +# RFC928 || M.A. Padlipsky || +# RFC929 || J. Lilienkamp, R. Mandell, M.A. Padlipsky || +# RFC930 || M. Solomon, E. Wimmers || +# RFC931 || M. St. Johns || +# RFC932 || D.D. Clark || +# RFC933 || S. Silverman || +# RFC934 || M.T. Rose, E.A. Stefferud || +# RFC935 || J.G. Robinson || +# RFC936 || M.J. Karels || +# RFC937 || M. Butler, J. Postel, D. Chase, J. Goldberger, J.K. Reynolds || +# RFC938 || T. Miller || +# RFC939 || National Research Council || +# RFC940 || Gateway Algorithms and Data Structures Task Force || +# RFC941 || International Organization for Standardization || +# RFC942 || National Research Council || +# RFC943 || J.K. Reynolds, J. Postel || +# RFC944 || J.K. Reynolds, J. Postel || +# RFC945 || J. Postel || +# RFC946 || R. Nedved || +# RFC947 || K. Lebowitz, D. Mankins || +# RFC948 || I. Winston || +# RFC949 || M.A. Padlipsky || +# RFC950 || J.C. Mogul, J. Postel || +# RFC951 || W.J. Croft, J. Gilmore || +# RFC952 || K. Harrenstien, M.K. Stahl, E.J. Feinler || +# RFC953 || K. Harrenstien, M.K. Stahl, E.J. Feinler || +# RFC954 || K. Harrenstien, M.K. Stahl, E.J. Feinler || +# RFC955 || R.T. Braden || +# RFC956 || D.L. Mills || +# RFC957 || D.L. Mills || +# RFC958 || D.L. Mills || +# RFC959 || J. Postel, J. Reynolds || +# RFC960 || J.K. Reynolds, J. Postel || +# RFC961 || J.K. Reynolds, J. Postel || +# RFC962 || M.A. Padlipsky || +# RFC963 || D.P. Sidhu || +# RFC964 || D.P. Sidhu, T. Blumer || +# RFC965 || L. Aguilar || +# RFC966 || S.E. Deering, D.R. Cheriton || +# RFC967 || M.A. Padlipsky || +# RFC968 || V.G. Cerf || +# RFC969 || D.D. Clark, M.L. Lambert, L. Zhang || +# RFC970 || J. Nagle || +# RFC971 || A.L. DeSchon || +# RFC972 || F.J. Wancho || +# RFC973 || P.V. Mockapetris || +# RFC974 || C. Partridge || +# RFC975 || D.L. Mills || +# RFC976 || M.R. Horton || +# RFC977 || B. Kantor, P. Lapsley || +# RFC978 || J.K. Reynolds, R. Gillman, W.A. Brackenridge, A. Witkowski, J. Postel || +# RFC979 || A.G. Malis || +# RFC980 || O.J. Jacobsen, J. Postel || +# RFC981 || D.L. Mills || +# RFC982 || H.W. Braun || +# RFC983 || D.E. Cass, M.T. Rose || +# RFC984 || D.D. Clark, M.L. Lambert || +# RFC985 || National Science Foundation, Network Technical Advisory Group || +# RFC986 || R.W. Callon, H.W. Braun || +# RFC987 || S.E. Kille || +# RFC988 || S.E. Deering || +# RFC989 || J. Linn || +# RFC990 || J.K. Reynolds, J. Postel || +# RFC991 || J.K. Reynolds, J. Postel || +# RFC992 || K.P. Birman, T.A. Joseph || +# RFC993 || D.D. Clark, M.L. Lambert || +# RFC994 || International Organization for Standardization || +# RFC995 || International Organization for Standardization || +# RFC996 || D.L. Mills || +# RFC997 || J.K. Reynolds, J. Postel || +# RFC998 || D.D. Clark, M.L. Lambert, L. Zhang || +# RFC999 || A. Westine, J. Postel || +# RFC1000 || J.K. Reynolds, J. Postel || +# RFC1001 || NetBIOS Working Group in the Defense Advanced Research Projects Agency, Internet Activities Board, End-to-End Services Task Force || +# RFC1002 || NetBIOS Working Group in the Defense Advanced Research Projects Agency, Internet Activities Board, End-to-End Services Task Force || +# RFC1003 || A.R. Katz || +# RFC1004 || D.L. Mills || +# RFC1005 || A. Khanna, A.G. Malis || +# RFC1006 || M.T. Rose, D.E. Cass || +# RFC1007 || W. McCoy || +# RFC1008 || W. McCoy || +# RFC1009 || R.T. Braden, J. Postel || +# RFC1010 || J.K. Reynolds, J. Postel || +# RFC1011 || J.K. Reynolds, J. Postel || +# RFC1012 || J.K. Reynolds, J. Postel || +# RFC1013 || R.W. Scheifler || +# RFC1014 || Sun Microsystems || +# RFC1015 || B.M. Leiner || +# RFC1016 || W. Prue, J. Postel || +# RFC1017 || B.M. Leiner || +# RFC1018 || A.M. McKenzie || +# RFC1019 || D. Arnon || +# RFC1020 || S. Romano, M.K. Stahl || +# RFC1021 || C. Partridge, G. Trewitt || +# RFC1022 || C. Partridge, G. Trewitt || +# RFC1023 || G. Trewitt, C. Partridge || +# RFC1024 || C. Partridge, G. Trewitt || +# RFC1025 || J. Postel || +# RFC1026 || S.E. Kille || +# RFC1027 || S. Carl-Mitchell, J.S. Quarterman || +# RFC1028 || J. Davin, J.D. Case, M. Fedor, M.L. Schoffstall || +# RFC1029 || G. Parr || +# RFC1030 || M.L. Lambert || +# RFC1031 || W.D. Lazear || +# RFC1032 || M.K. Stahl || +# RFC1033 || M. Lottor || +# RFC1034 || P.V. Mockapetris || +# RFC1035 || P.V. Mockapetris || +# RFC1036 || M.R. Horton, R. Adams || +# RFC1037 || B. Greenberg, S. Keene || +# RFC1038 || M. St. Johns || +# RFC1039 || D. Latham || +# RFC1040 || J. Linn || +# RFC1041 || Y. Rekhter || +# RFC1042 || J. Postel, J.K. Reynolds || +# RFC1043 || A. Yasuda, T. Thompson || +# RFC1044 || K. Hardwick, J. Lekashman || +# RFC1045 || D.R. Cheriton || +# RFC1046 || W. Prue, J. Postel || +# RFC1047 || C. Partridge || +# RFC1048 || P.A. Prindeville || +# RFC1049 || M.A. Sirbu || +# RFC1050 || Sun Microsystems || +# RFC1051 || P.A. Prindeville || +# RFC1052 || V.G. Cerf || +# RFC1053 || S. Levy, T. Jacobson || +# RFC1054 || S.E. Deering || +# RFC1055 || J.L. Romkey || +# RFC1056 || M.L. Lambert || +# RFC1057 || Sun Microsystems || +# RFC1058 || C.L. Hedrick || +# RFC1059 || D.L. Mills || +# RFC1060 || J.K. Reynolds, J. Postel || +# RFC1061 || || +# RFC1062 || S. Romano, M.K. Stahl, M. Recker || +# RFC1063 || J.C. Mogul, C.A. Kent, C. Partridge, K. McCloghrie || +# RFC1064 || M.R. Crispin || +# RFC1065 || K. McCloghrie, M.T. Rose || +# RFC1066 || K. McCloghrie, M.T. Rose || +# RFC1067 || J.D. Case, M. Fedor, M.L. Schoffstall, J. Davin || +# RFC1068 || A.L. DeSchon, R.T. Braden || +# RFC1069 || R.W. Callon, H.W. Braun || +# RFC1070 || R.A. Hagens, N.E. Hall, M.T. Rose || hagens@cs.wisc.edu, nhall@cs.wisc.edu +# RFC1071 || R.T. Braden, D.A. Borman, C. Partridge || +# RFC1072 || V. Jacobson, R.T. Braden || +# RFC1073 || D. Waitzman || +# RFC1074 || J. Rekhter || +# RFC1075 || D. Waitzman, C. Partridge, S.E. Deering || +# RFC1076 || G. Trewitt, C. Partridge || +# RFC1077 || B.M. Leiner || +# RFC1078 || M. Lottor || +# RFC1079 || C.L. Hedrick || +# RFC1080 || C.L. Hedrick || +# RFC1081 || M.T. Rose || +# RFC1082 || M.T. Rose || +# RFC1083 || Defense Advanced Research Projects Agency, Internet Activities Board || +# RFC1084 || J.K. Reynolds || JKREYNOLDS@ISI.EDU +# RFC1085 || M.T. Rose || mrose17@gmail.com +# RFC1086 || J.P. Onions, M.T. Rose || JPO@CS.NOTT.AC.UK, mrose17@gmail.com +# RFC1087 || Defense Advanced Research Projects Agency, Internet Activities Board || +# RFC1088 || L.J. McLaughlin || ljm@TWG.COM +# RFC1089 || M. Schoffstall, C. Davin, M. Fedor, J. Case || schoff@stonewall.nyser.net, jrd@ptt.lcs.mit.edu, fedor@patton.NYSER.NET, case@UTKUX1.UTK.EDU +# RFC1090 || R. Ullmann || +# RFC1091 || J. VanBokkelen || +# RFC1092 || J. Rekhter || +# RFC1093 || H.W. Braun || +# RFC1094 || B. Nowicki || +# RFC1095 || U.S. Warrier, L. Besaw || +# RFC1096 || G.A. Marcy || +# RFC1097 || B. Miller || +# RFC1098 || J.D. Case, M. Fedor, M.L. Schoffstall, J. Davin || jrd@ptt.lcs.mit.edu +# RFC1099 || J. Reynolds || JKREY@ISI.EDU +# RFC1100 || Defense Advanced Research Projects Agency, Internet Activities Board || +# RFC1101 || P.V. Mockapetris || +# RFC1102 || D.D. Clark || +# RFC1103 || D. Katz || +# RFC1104 || H.W. Braun || hwb@merit.edu +# RFC1105 || K. Lougheed, Y. Rekhter || +# RFC1106 || R. Fox || rfox@tandem.com +# RFC1107 || K.R. Sollins || SOLLINS@XX.LCS.MIT.EDU +# RFC1108 || S. Kent || +# RFC1109 || V.G. Cerf || CERF@A.ISI.EDU +# RFC1110 || A.M. McKenzie || MCKENZIE@BBN.COM +# RFC1111 || J. Postel || POSTEL@ISI.EDU +# RFC1112 || S.E. Deering || deering@PESCADERO.STANFORD.EDU +# RFC1113 || J. Linn || Linn@ultra.enet.dec.com +# RFC1114 || S.T. Kent, J. Linn || kent@BBN.COM, Linn@ultra.enet.dec.com +# RFC1115 || J. Linn || Linn@ultra.enet.dec.com +# RFC1116 || D.A. Borman || dab@CRAY.COM +# RFC1117 || S. Romano, M.K. Stahl, M. Recker || +# RFC1118 || E. Krol || Krol@UXC.CSO.UIUC.EDU +# RFC1119 || D.L. Mills || +# RFC1120 || V. Cerf || VCERF@NRI.RESTON.VA.US +# RFC1121 || J. Postel, L. Kleinrock, V.G. Cerf, B. Boehm || Postel@ISI.EDU, lk@CS.UCLA.EDU, VCerf@NRI.RESTON.VA.US, boehm@CS.UCLA.EDU +# RFC1122 || R. Braden, Ed. || Braden@ISI.EDU +# RFC1123 || R. Braden, Ed. || Braden@ISI.EDU +# RFC1124 || B.M. Leiner || +# RFC1125 || D. Estrin || +# RFC1126 || M. Little || little@SAIC.COM +# RFC1127 || R.T. Braden || Braden@ISI.EDU +# RFC1128 || D.L. Mills || +# RFC1129 || D.L. Mills || +# RFC1130 || Defense Advanced Research Projects Agency, Internet Activities Board || +# RFC1131 || J. Moy || +# RFC1132 || L.J. McLaughlin || ljm@TWG.COM +# RFC1133 || J.Y. Yu, H.W. Braun || jyy@merit.edu, hwb@merit.edu +# RFC1134 || D. Perkins || rdhobby@ucdavis.edu +# RFC1135 || J.K. Reynolds || JKREY@ISI.EDU +# RFC1136 || S. Hares, D. Katz || +# RFC1137 || S. Kille || S.Kille@Cs.Ucl.AC.UK +# RFC1138 || S.E. Kille || S.Kille@Cs.Ucl.AC.UK +# RFC1139 || R.A. Hagens || hagens@CS.WISC.EDU +# RFC1140 || Defense Advanced Research Projects Agency, Internet Activities Board || +# RFC1141 || T. Mallory, A. Kullberg || tmallory@CCV.BBN.COM, akullberg@BBN.COM +# RFC1142 || D. Oran, Ed. || +# RFC1143 || D.J. Bernstein || +# RFC1144 || V. Jacobson || +# RFC1145 || J. Zweig, C. Partridge || zweig@CS.UIUC.EDU, craig@BBN.COM +# RFC1146 || J. Zweig, C. Partridge || zweig@CS.UIUC.EDU, craig@BBN.COM +# RFC1147 || R.H. Stine || STINE@SPARTA.COM +# RFC1148 || S.E. Kille || S.Kille@Cs.Ucl.AC.UK +# RFC1149 || D. Waitzman || dwaitzman@BBN.COM +# RFC1150 || G.S. Malkin, J.K. Reynolds || gmalkin@proteon.com, jkrey@isi.edu +# RFC1151 || C. Partridge, R.M. Hinden || craig@BBN.COM, bob.hinden@gmail.com +# RFC1152 || C. Partridge || craig@BBN.COM +# RFC1153 || F.J. Wancho || +# RFC1154 || D. Robinson, R. Ullmann || +# RFC1155 || M.T. Rose, K. McCloghrie || mrose17@gmail.com +# RFC1156 || K. McCloghrie, M.T. Rose || mrose17@gmail.com +# RFC1157 || J.D. Case, M. Fedor, M.L. Schoffstall, J. Davin || jrd@ptt.lcs.mit.edu +# RFC1158 || M.T. Rose || +# RFC1159 || R. Nelson || nelson@sun.soe.clarkson.edu +# RFC1160 || V. Cerf || VCERF@NRI.RESTON.VA.US +# RFC1161 || M.T. Rose || +# RFC1162 || G. Satz || +# RFC1163 || K. Lougheed, Y. Rekhter || +# RFC1164 || J.C. Honig, D. Katz, M. Mathis, Y. Rekhter, J.Y. Yu || +# RFC1165 || J. Crowcroft, J.P. Onions || JON@CS.UCL.AC.UK, JPO@CS.NOTT.AC.UK +# RFC1166 || S. Kirkpatrick, M.K. Stahl, M. Recker || +# RFC1167 || V.G. Cerf || vcerf@NRI.Reston.VA.US +# RFC1168 || A. Westine, A.L. DeSchon, J. Postel, C.E. Ward || +# RFC1169 || V.G. Cerf, K.L. Mills || vcerf@nri.reston.va.us, MILLS@ECF.NCSL.NIST.GOV +# RFC1170 || R.B. Fougner || +# RFC1171 || D. Perkins || ddp@andrew.cmu.edu +# RFC1172 || D. Perkins, R. Hobby || rdhobby@ucdavis.edu, ddp@andrew.cmu.edu +# RFC1173 || J. VanBokkelen || jbvb@ftp.com +# RFC1174 || V.G. Cerf || vcerf@nri.reston.va.us +# RFC1175 || K.L. Bowers, T.L. LaQuey, J.K. Reynolds, K. Roubicek, M.K. Stahl, A. Yuan || +# RFC1176 || M.R. Crispin || mrc@Tomobiki-Cho.CAC.Washington.EDU +# RFC1177 || G.S. Malkin, A.N. Marine, J.K. Reynolds || gmalkin@ftp.com, APRIL@NIC.DDN.MIL, jkrey@isi.edu +# RFC1178 || D. Libes || libes@cme.nist.gov +# RFC1179 || L. McLaughlin || ljm@twg.com +# RFC1180 || T.J. Socolofsky, C.J. Kale || TEDS@SPIDER.CO.UK, CLAUDIAK@SPIDER.CO.UK +# RFC1181 || R. Blokzijl || k13@nikhef.nl +# RFC1182 || || +# RFC1183 || C.F. Everhart, L.A. Mamakos, R. Ullmann, P.V. Mockapetris || Craig_Everhart@transarc.com, pvm@isi.edu +# RFC1184 || D.A. Borman || dab@CRAY.COM +# RFC1185 || V. Jacobson, R.T. Braden, L. Zhang || van@CSAM.LBL.GOV, Braden@ISI.EDU, lixia@PARC.XEROX.COM +# RFC1186 || R.L. Rivest || rivest@theory.lcs.mit.edu +# RFC1187 || M.T. Rose, K. McCloghrie, J.R. Davin || mrose17@gmail.com, KZM@HLS.COM, jrd@ptt.lcs.mit.edu +# RFC1188 || D. Katz || dkatz@merit.edu +# RFC1189 || U.S. Warrier, L. Besaw, L. LaBarre, B.D. Handspicker || +# RFC1190 || C. Topolcic || Casner@ISI.Edu, CLynn@BBN.Com, ppark@BBN.COM, Schroder@BBN.Com, Topolcic@BBN.Com +# RFC1191 || J.C. Mogul, S.E. Deering || mogul@decwrl.dec.com, deering@xerox.com +# RFC1192 || B. Kahin || kahin@hulaw.harvard.edu +# RFC1193 || D. Ferrari || ferrari@UCBVAX.BERKELEY.EDU +# RFC1194 || D.P. Zimmerman || dpz@dimacs.rutgers.edu +# RFC1195 || R.W. Callon || +# RFC1196 || D.P. Zimmerman || dpz@dimacs.rutgers.edu +# RFC1197 || M. Sherman || +# RFC1198 || R.W. Scheifler || rws@expo.lcs.mit.edu +# RFC1199 || J. Reynolds || JKREY@ISI.EDU +# RFC1200 || Defense Advanced Research Projects Agency, Internet Activities Board || +# RFC1201 || D. Provan || donp@Novell.Com +# RFC1202 || M.T. Rose || mrose17@gmail.com +# RFC1203 || J. Rice || RICE@SUMEX-AIM.STANFORD.EDU +# RFC1204 || S. Yeh, D. Lee || dlee@netix.com +# RFC1205 || P. Chmielewski || paulc@rchland.iinus1.ibm.com +# RFC1206 || G.S. Malkin, A.N. Marine || gmalkin@ftp.com, APRIL@nic.ddn.mil +# RFC1207 || G.S. Malkin, A.N. Marine, J.K. Reynolds || gmalkin@ftp.com, APRIL@nic.ddn.mil, jkrey@isi.edu +# RFC1208 || O.J. Jacobsen, D.C. Lynch || OLE@CSLI.STANFORD.EDU, Lynch@ISI.EDU +# RFC1209 || D. Piscitello, J. Lawrence || dave@sabre.bellcore.com, jcl@sabre.bellcore.com +# RFC1210 || V.G. Cerf, P.T. Kirstein, B. Randell || +# RFC1211 || A. Westine, J. Postel || Westine@ISI.EDU, Postel@ISI.EDU +# RFC1212 || M.T. Rose, K. McCloghrie || mrose17@gmail.com, kzm@hls.com +# RFC1213 || K. McCloghrie, M. Rose || kzm@hls.com, mrose17@gmail.com +# RFC1214 || L. LaBarre || cel@mbunix.mitre.org +# RFC1215 || M.T. Rose || mrose17@gmail.com +# RFC1216 || P. Richard, P. Kynikos || +# RFC1217 || V.G. Cerf || CERF@NRI.RESTON.VA.US +# RFC1218 || North American Directory Forum || +# RFC1219 || P.F. Tsuchiya || tsuchiya@thumper.bellcore.com +# RFC1220 || F. Baker || fbaker@ACC.COM +# RFC1221 || W. Edmond || wbe@bbn.com +# RFC1222 || H.W. Braun, Y. Rekhter || HWB@SDSC.EDU, Yakov@Watson.IBM.COM +# RFC1223 || J.M. Halpern || +# RFC1224 || L. Steinberg || LOUISS@IBM.COM +# RFC1225 || M.T. Rose || mrose17@gmail.com +# RFC1226 || B. Kantor || brian@UCSD.EDU +# RFC1227 || M.T. Rose || mrose17@gmail.com +# RFC1228 || G. Carpenter, B. Wijnen || +# RFC1229 || K. McCloghrie || kzm@hls.com +# RFC1230 || K. McCloghrie, R. Fox || kzm@hls.com, rfox@synoptics.com +# RFC1231 || K. McCloghrie, R. Fox, E. Decker || kzm@hls.com, rfox@synoptics.com, cire@cisco.com +# RFC1232 || F. Baker, C.P. Kolb || fbaker@acc.com, kolb@psi.com +# RFC1233 || T.A. Cox, K. Tesink || tacox@sabre.bellcore.com, kaj@nvuxr.cc.bellcore.com +# RFC1234 || D. Provan || donp@Novell.Com +# RFC1235 || J. Ioannidis, G. Maguire || ji@cs.columbia.edu, maguire@cs.columbia.edu +# RFC1236 || L. Morales, P. Hasse || lmorales@huachuca-emh8.army.mil, phasse@huachuca-emh8.army.mil +# RFC1237 || R. Colella, E. Gardner, R. Callon || colella@osi3.ncsl.nist.gov, epg@gateway.mitre.org +# RFC1238 || G. Satz || +# RFC1239 || J.K. Reynolds || jkrey@isi.edu +# RFC1240 || C. Shue, W. Haggerty, K. Dobbins || chi@osf.org, bill@comm.wang.com +# RFC1241 || R.A. Woodburn, D.L. Mills || woody@cseic.saic.com, mills@udel.edu +# RFC1242 || S. Bradner || SOB@HARVARD.HARVARD.EDU +# RFC1243 || S. Waldbusser || waldbusser@andrew.cmu.edu +# RFC1244 || J.P. Holbrook, J.K. Reynolds || holbrook@cic.net, JKREY@ISI.EDU +# RFC1245 || J. Moy || +# RFC1246 || J. Moy || +# RFC1247 || J. Moy || jmoy@proteon.com +# RFC1248 || F. Baker, R. Coltun || fbaker@acc.com, rcoltun@ni.umd.edu +# RFC1249 || T. Howes, M. Smith, B. Beecher || tim@umich.edu, mcs@umich.edu, bryan@umich.edu +# RFC1250 || J. Postel || +# RFC1251 || G. Malkin || gmalkin@ftp.com +# RFC1252 || F. Baker, R. Coltun || fbaker@acc.com, rcoltun@ni.umd.edu +# RFC1253 || F. Baker, R. Coltun || fbaker@acc.com, rcoltun@ni.umd.edu +# RFC1254 || A. Mankin, K. Ramakrishnan || +# RFC1255 || The North American Directory Forum || +# RFC1256 || S. Deering, Ed. || deering@xerox.com +# RFC1257 || C. Partridge || craig@SICS.SE +# RFC1258 || B. Kantor || brian@UCSD.EDU +# RFC1259 || M. Kapor || mkapor@eff.org +# RFC1260 || || +# RFC1261 || S. Williamson, L. Nobile || scottw@DIIS.DDN.MIL +# RFC1262 || V.G. Cerf || +# RFC1263 || S. O'Malley, L.L. Peterson || llp@cs.arizona.edu, sean@cs.arizona.edu +# RFC1264 || R.M. Hinden || bob.hinden@gmail.com +# RFC1265 || Y. Rekhter || yakov@watson.ibm.com +# RFC1266 || Y. Rekhter || yakov@watson.ibm.com +# RFC1267 || K. Lougheed, Y. Rekhter || +# RFC1268 || Y. Rekhter, P. Gross || yakov@watson.ibm.com +# RFC1269 || S. Willis, J.W. Burruss || +# RFC1270 || F. Kastenholz || +# RFC1271 || S. Waldbusser || waldbusser@andrew.cmu.edu +# RFC1272 || C. Mills, D. Hirsh, G.R. Ruth || +# RFC1273 || M.F. Schwartz || schwartz@cs.colorado.edu +# RFC1274 || P. Barker, S. Kille || P.Barker@cs.ucl.ac.uk, S.Kille@cs.ucl.ac.uk +# RFC1275 || S.E. Hardcastle-Kille || S.Kille@CS.UCL.AC.UK +# RFC1276 || S.E. Hardcastle-Kille || S.Kille@CS.UCL.AC.UK +# RFC1277 || S.E. Hardcastle-Kille || S.Kille@CS.UCL.AC.UK +# RFC1278 || S.E. Hardcastle-Kille || S.Kille@CS.UCL.AC.UK +# RFC1279 || S.E. Hardcastle-Kille || S.Kille@CS.UCL.AC.UK +# RFC1280 || J. Postel || +# RFC1281 || R. Pethia, S. Crocker, B. Fraser || rdp@cert.sei.cmu.edu, crocker@tis.com, byf@cert.sei.cmu.edu +# RFC1282 || B. Kantor || brian@UCSD.EDU +# RFC1283 || M. Rose || +# RFC1284 || J. Cook, Ed. || kasten@europa.clearpoint.com +# RFC1285 || J. Case || case@CS.UTK.EDU +# RFC1286 || E. Decker, P. Langille, A. Rijsinghani, K. McCloghrie || langille@edwin.enet.dec.com, anil@levers.enet.dec.com, kzm@hls.com +# RFC1287 || D. Clark, L. Chapin, V. Cerf, R. Braden, R. Hobby || ddc@LCS.MIT.EDU, vcerf@nri.reston.va.us, lyman@BBN.COM, braden@isi.edu, rdhobby@ucdavis.edu +# RFC1288 || D. Zimmerman || dpz@dimacs.rutgers.edu +# RFC1289 || J. Saperia || saperia@enet.dec.com +# RFC1290 || J. Martin || jmartin@magnus.acs.ohio-state.edu +# RFC1291 || V. Aggarwal || +# RFC1292 || R. Lang, R. Wright || +# RFC1293 || T. Bradley, C. Brown || +# RFC1294 || T. Bradley, C. Brown, A. Malis || +# RFC1295 || The North American Directory Forum || 0004454742@mcimail.com +# RFC1296 || M. Lottor || mkl@nisc.sri.com +# RFC1297 || D. Johnson || +# RFC1298 || R. Wormley, S. Bostock || bwormley@novell.com, steveb@novell.com +# RFC1299 || M. Kennedy || MKENNEDY@ISI.EDU +# RFC1300 || S. Greenfield || 0004689513@mcimail.com +# RFC1301 || S. Armstrong, A. Freier, K. Marzullo || armstrong@wrc.xerox.com, freier@apple.com, marzullo@cs.cornell.edu +# RFC1302 || D. Sitzler, P. Smith, A. Marine || +# RFC1303 || K. McCloghrie, M. Rose || kzm@hls.com, mrose17@gmail.com +# RFC1304 || T. Cox, Ed., K. Tesink, Ed. || tacox@sabre.bellcore.com, kaj@nvuxr.cc.bellcore.com +# RFC1305 || D. Mills || +# RFC1306 || A. Nicholson, J. Young || droid@cray.com, jsy@cray.com +# RFC1307 || J. Young, A. Nicholson || jsy@cray.com, droid@cray.com +# RFC1308 || C. Weider, J. Reynolds || +# RFC1309 || C. Weider, J. Reynolds, S. Heker || jkrey@isi.edu +# RFC1310 || L. Chapin || +# RFC1311 || J. Postel || +# RFC1312 || R. Nelson, G. Arnold || nelson@crynwr.com, geoff@east.sun.com +# RFC1313 || C. Partridge || craig@aland.bbn.com +# RFC1314 || A. Katz, D. Cohen || Katz@ISI.Edu, Cohen@ISI.Edu +# RFC1315 || C. Brown, F. Baker, C. Carvalho || cbrown@wellfleet.com, fbaker@acc.com, charles@acc.com +# RFC1316 || B. Stewart || rlstewart@eng.xyplex.com +# RFC1317 || B. Stewart || rlstewart@eng.xyplex.com +# RFC1318 || B. Stewart || rlstewart@eng.xyplex.com +# RFC1319 || B. Kaliski || burt@rsa.com +# RFC1320 || R. Rivest || rivest@theory.lcs.mit.edu +# RFC1321 || R. Rivest || rivest@theory.lcs.mit.edu +# RFC1322 || D. Estrin, Y. Rekhter, S. Hotz || estrin@usc.edu, yakov@ibm.com, hotz@usc.edu +# RFC1323 || V. Jacobson, R. Braden, D. Borman || van@CSAM.LBL.GOV, Braden@ISI.EDU +# RFC1324 || D. Reed || +# RFC1325 || G. Malkin, A. Marine || gmalkin@Xylogics.COM, april@nisc.sri.com +# RFC1326 || P. Tsuchiya || tsuchiya@thumper.bellcore.com +# RFC1327 || S. Hardcastle-Kille || +# RFC1328 || S. Hardcastle-Kille || S.Kille@CS.UCL.AC.UK +# RFC1329 || P. Kuehn || thimmela@sniabg.wa.sni.de +# RFC1330 || ESCC X.500/X.400 Task Force, ESnet Site Coordinating Comittee (ESCC), Energy Sciences Network (ESnet) || +# RFC1331 || W. Simpson || bsimpson@ray.lloyd.com +# RFC1332 || G. McGregor || Glenn.McGregor@Merit.edu +# RFC1333 || W. Simpson || bsimpson@ray.lloyd.com +# RFC1334 || B. Lloyd, W. Simpson || Bill.Simpson@um.cc.umich.edu +# RFC1335 || Z. Wang, J. Crowcroft || z.wang@cs.ucl.ac.uk, j.crowcroft@cs.ucl.ac.uk +# RFC1336 || G. Malkin || gmalkin@Xylogics.COM +# RFC1337 || R. Braden || Braden@ISI.EDU +# RFC1338 || V. Fuller, T. Li, J. Yu, K. Varadhan || +# RFC1339 || S. Dorner, P. Resnick || s-dorner@uiuc.edu, presnick@qti.qualcomm.com +# RFC1340 || J. Reynolds, J. Postel || +# RFC1341 || N. Borenstein, N. Freed || +# RFC1342 || K. Moore || moore@cs.utk.edu +# RFC1343 || N. Borenstein || +# RFC1344 || N. Borenstein || +# RFC1345 || K. Simonsen || +# RFC1346 || P. Jones || +# RFC1347 || R. Callon || +# RFC1348 || B. Manning || bmanning@rice.edu +# RFC1349 || P. Almquist || +# RFC1350 || K. Sollins || SOLLINS@LCS.MIT.EDU +# RFC1351 || J. Davin, J. Galvin, K. McCloghrie || jrd@ptt.lcs.mit.edu, galvin@tis.com, kzm@hls.com +# RFC1352 || J. Galvin, K. McCloghrie, J. Davin || galvin@tis.com, kzm@hls.com, jrd@ptt.lcs.mit.edu +# RFC1353 || K. McCloghrie, J. Davin, J. Galvin || kzm@hls.com, jrd@ptt.lcs.mit.edu, galvin@tis.com +# RFC1354 || F. Baker || fbaker@acc.com +# RFC1355 || J. Curran, A. Marine || jcurran@nnsc.nsf.net, april@nisc.sri.com +# RFC1356 || A. Malis, D. Robinson, R. Ullmann || +# RFC1357 || D. Cohen || Cohen@ISI.EDU +# RFC1358 || L. Chapin || lyman@BBN.COM +# RFC1359 || ACM SIGUCCS || martyne@nr-tech.cit.cornell.edu +# RFC1360 || J. Postel || +# RFC1361 || D. Mills || mills@udel.edu +# RFC1362 || M. Allen || MALLEN@NOVELL.COM, brian@ray.lloyd.com +# RFC1363 || C. Partridge || craig@aland.bbn.com +# RFC1364 || K. Varadhan || kannan@oar.net +# RFC1365 || K. Siyan || 72550.1634@compuserve.com +# RFC1366 || E. Gerich || epg@MERIT.EDU +# RFC1367 || C. Topolcic || topolcic@NRI.Reston.VA.US +# RFC1368 || D. McMaster, K. McCloghrie || mcmaster@synoptics.com, kzm@hls.com +# RFC1369 || F. Kastenholz || kasten@ftp.com +# RFC1370 || Internet Architecture Board, L. Chapin || +# RFC1371 || P. Gross || pgross@ans.net +# RFC1372 || C. Hedrick, D. Borman || +# RFC1373 || T. Tignor || tpt2@isi.edu +# RFC1374 || J. Renwick, A. Nicholson || jkr@CRAY.COM, droid@CRAY.COM +# RFC1375 || P. Robinson || +# RFC1376 || S. Senum || sjs@network.com +# RFC1377 || D. Katz || dkatz@cisco.com +# RFC1378 || B. Parker || brad@cayman.com +# RFC1379 || R. Braden || Braden@ISI.EDU +# RFC1380 || P. Gross, P. Almquist || pgross@ans.net, Almquist@JESSICA.STANFORD.EDU +# RFC1381 || D. Throop, F. Baker || throop@dg-rtp.dg.com, fbaker@acc.com +# RFC1382 || D. Throop, Ed. || throop@dg-rtp.dg.com +# RFC1383 || C. Huitema || Christian.Huitema@MIRSA.INRIA.FR +# RFC1384 || P. Barker, S.E. Hardcastle-Kille || P.Barker@CS.UCL.AC.UK, S.Kille@ISODE.COM +# RFC1385 || Z. Wang || z.wang@cs.ucl.ac.uk +# RFC1386 || A. Cooper, J. Postel || +# RFC1387 || G. Malkin || gmalkin@Xylogics.COM +# RFC1388 || G. Malkin || gmalkin@Xylogics.COM +# RFC1389 || G. Malkin, F. Baker || gmalkin@Xylogics.COM, fbaker@acc.com +# RFC1390 || D. Katz || dkatz@cisco.com +# RFC1391 || G. Malkin || gmalkin@Xylogics.COM +# RFC1392 || G. Malkin, T. LaQuey Parker || gmalkin@Xylogics.COM, tracy@utexas.edu +# RFC1393 || G. Malkin || gmalkin@Xylogics.COM +# RFC1394 || P. Robinson || +# RFC1395 || J. Reynolds || jkrey@isi.edu +# RFC1396 || S. Crocker || +# RFC1397 || D. Haskin || +# RFC1398 || F. Kastenholz || kasten@ftp.com +# RFC1399 || J. Elliott || elliott@isi.edu +# RFC1400 || S. Williamson || scottw@internic.net +# RFC1401 || Internet Architecture Board || +# RFC1402 || J. Martin || nic@osu.edu +# RFC1403 || K. Varadhan || +# RFC1404 || B. Stockman || +# RFC1405 || C. Allocchio || Claudio.Allocchio@elettra.Trieste.it +# RFC1406 || F. Baker, Ed., J. Watt, Ed. || fbaker@acc.com, james@newbridge.com +# RFC1407 || T. Cox, K. Tesink || tacox@mail.bellcore.com, kaj@cc.bellcore.com +# RFC1408 || D. Borman, Ed. || dab@CRAY.COM, stevea@isc.com +# RFC1409 || D. Borman, Ed. || dab@CRAY.COM, stevea@isc.com +# RFC1410 || J. Postel, Ed. || +# RFC1411 || D. Borman, Ed. || dab@CRAY.COM, stevea@isc.com +# RFC1412 || K. Alagappan || kannan@sejour.lkg.dec.com, stevea@isc.com +# RFC1413 || M. St. Johns || stjohns@DARPA.MIL +# RFC1414 || M. St. Johns, M. Rose || stjohns@DARPA.MIL, mrose17@gmail.com +# RFC1415 || J. Mindel, R. Slaski || +# RFC1416 || D. Borman, Ed. || dab@CRAY.COM, stevea@isc.com +# RFC1417 || The North American Directory Forum || 0004454742@mcimail.com +# RFC1418 || M. Rose || mrose17@gmail.com +# RFC1419 || G. Minshall, M. Ritter || minshall@wc.novell.com, MWRITTER@applelink.apple.com +# RFC1420 || S. Bostock || +# RFC1421 || J. Linn || 104-8456@mcimail.com +# RFC1422 || S. Kent || kent@BBN.COM +# RFC1423 || D. Balenson || balenson@tis.com +# RFC1424 || B. Kaliski || burt@rsa.com +# RFC1425 || J. Klensin, WG Chair, N. Freed, Ed., M. Rose, E. Stefferud, D. Crocker || +# RFC1426 || J. Klensin, WG Chair, N. Freed, Ed., M. Rose, E. Stefferud, D. Crocker || +# RFC1427 || J. Klensin, WG Chair, N. Freed, Ed., K. Moore || +# RFC1428 || G. Vaudreuil || GVaudre@CNRI.Reston.VA.US +# RFC1429 || E. Thomas || +# RFC1430 || S. Hardcastle-Kille, E. Huizer, V. Cerf, R. Hobby, S. Kent || S.Kille@isode.com, vcerf@cnri.reston.va.us, rdhobby@ucdavis.edu, skent@bbn.com +# RFC1431 || P. Barker || +# RFC1432 || J. Quarterman || jsq@tic.com, mids@tic.com +# RFC1433 || J. Garrett, J. Hagan, J. Wong || jwg@garage.att.com, Hagan@UPENN.EDU, jwong@garage.att.com +# RFC1434 || R. Dixon, D. Kushi || rcdixon@ralvmg.vnet.ibm.com, kushi@watson.ibm.com +# RFC1435 || S. Knowles || stev@ftp.com +# RFC1436 || F. Anklesaria, M. McCahill, P. Lindner, D. Johnson, D. Torrey, B. Albert || fxa@boombox.micro.umn.edu, mpm@boombox.micro.umn.edu, lindner@boombox.micro.umn.edu, dmj@boombox.micro.umn.edu, daniel@boombox.micro.umn.edu, alberti@boombox.micro.umn.edu +# RFC1437 || N. Borenstein, M. Linimon || nsb@bellcore.com, linimon@LONESOME.COM +# RFC1438 || A. Lyman Chapin, C. Huitema || Lyman@BBN.COM, Christian.Huitema@MIRSA.INRIA.FR +# RFC1439 || C. Finseth || Craig.A.Finseth-1@umn.edu +# RFC1440 || R. Troth || troth@rice.edu +# RFC1441 || J. Case, K. McCloghrie, M. Rose, S. Waldbusser || +# RFC1442 || J. Case, K. McCloghrie, M. Rose, S. Waldbusser || +# RFC1443 || J. Case, K. McCloghrie, M. Rose, S. Waldbusser || +# RFC1444 || J. Case, K. McCloghrie, M. Rose, S. Waldbusser || +# RFC1445 || J. Galvin, K. McCloghrie || galvin@tis.com +# RFC1446 || J. Galvin, K. McCloghrie || galvin@tis.com +# RFC1447 || K. McCloghrie, J. Galvin || galvin@tis.com +# RFC1448 || J. Case, K. McCloghrie, M. Rose, S. Waldbusser || +# RFC1449 || J. Case, K. McCloghrie, M. Rose, S. Waldbusser || +# RFC1450 || J. Case, K. McCloghrie, M. Rose, S. Waldbusser || +# RFC1451 || J. Case, K. McCloghrie, M. Rose, S. Waldbusser || +# RFC1452 || J. Case, K. McCloghrie, M. Rose, S. Waldbusser || +# RFC1453 || W. Chimiak || chim@relito.medeng.wfu.edu +# RFC1454 || T. Dixon || dixon@rare.nl +# RFC1455 || D. Eastlake 3rd || +# RFC1456 || Vietnamese Standardization Working Group || +# RFC1457 || R. Housley || Housley.McLean_CSD@Xerox.COM +# RFC1458 || R. Braudes, S. Zabele || rebraudes@tasc.com, gszabele@tasc.com +# RFC1459 || J. Oikarinen, D. Reed || +# RFC1460 || M. Rose || mrose17@gmail.com +# RFC1461 || D. Throop || throop@dg-rtp.dg.com +# RFC1462 || E. Krol, E. Hoffman || e-krol@uiuc.edu, ellen@merit.edu +# RFC1463 || E. Hoffman, L. Jackson || +# RFC1464 || R. Rosenbaum || +# RFC1465 || D. Eppenberger || Eppenberger@switch.ch +# RFC1466 || E. Gerich || epg@MERIT.EDU +# RFC1467 || C. Topolcic || topolcic@CNRI.Reston.VA.US +# RFC1468 || J. Murai, M. Crispin, E. van der Poel || jun@wide.ad.jp, MRC@PANDA.COM, erik@poel.juice.or.jp +# RFC1469 || T. Pusateri || pusateri@cs.duke.edu +# RFC1470 || R. Enger, J. Reynolds || enger@reston.ans.net +# RFC1471 || F. Kastenholz || kasten@ftp.com +# RFC1472 || F. Kastenholz || kasten@ftp.com +# RFC1473 || F. Kastenholz || kasten@ftp.com +# RFC1474 || F. Kastenholz || kasten@ftp.com +# RFC1475 || R. Ullmann || +# RFC1476 || R. Ullmann || +# RFC1477 || M. Steenstrup || +# RFC1478 || M. Steenstrup || +# RFC1479 || M. Steenstrup || +# RFC1480 || A. Cooper, J. Postel || +# RFC1481 || C. Huitema || Christian.Huitema@MIRSA.INRIA.FR +# RFC1482 || M. Knopper, S. Richardson || +# RFC1483 || Juha Heinanen || +# RFC1484 || S. Hardcastle-Kille || S.Kille@ISODE.COM +# RFC1485 || S. Hardcastle-Kille || S.Kille@ISODE.COM +# RFC1486 || M. Rose, C. Malamud || mrose17@gmail.com, carl@malamud.com +# RFC1487 || W. Yeong, T. Howes, S. Kille || yeongw@psilink.com, tim@umich.edu, S.Kille@isode.com +# RFC1488 || T. Howes, S. Kille, W. Yeong, C. Robbins || tim@umich.edu, S.Kille@isode.com, yeongw@psilink.com +# RFC1489 || A. Chernov || ache@astral.msk.su +# RFC1490 || T. Bradley, C. Brown, A. Malis || +# RFC1491 || C. Weider, R. Wright || clw@merit.edu, wright@lbl.gov +# RFC1492 || C. Finseth || Craig.A.Finseth-1@umn.edu +# RFC1493 || E. Decker, P. Langille, A. Rijsinghani, K. McCloghrie || langille@edwin.enet.dec.com, anil@levers.enet.dec.com, kzm@hls.com +# RFC1494 || H. Alvestrand, S. Thompson || Harald.Alvestrand@delab.sintef.no, sjt@gateway.ssw.com +# RFC1495 || H. Alvestrand, S. Kille, R. Miles, M. Rose, S. Thompson || Harald.Alvestrand@delab.sintef.no, S.Kille@ISODE.COM, rsm@spyder.ssw.com, mrose17@gmail.com, sjt@gateway.ssw.com +# RFC1496 || H. Alvestrand, J. Romaguera, K. Jordan || Harald.T.Alvestrand@delab.sintef.no, Kevin.E.Jordan@mercury.oss.arh.cpg.cdc.com, Romaguera@netconsult.ch +# RFC1497 || J. Reynolds || jkrey@isi.edu +# RFC1498 || J. Saltzer || Saltzer@MIT.EDU +# RFC1499 || J. Elliott || elliott@isi.edu +# RFC1500 || J. Postel || +# RFC1501 || E. Brunsen || BRUNSENE@EMAIL.ENMU.EDU +# RFC1502 || H. Alvestrand || Harald.Alvestrand@delab.sintef.no +# RFC1503 || K. McCloghrie, M. Rose || kzm@hls.com, mrose17@gmail.com +# RFC1504 || A. Oppenheimer || Oppenheime1@applelink.apple.com +# RFC1505 || A. Costanzo, D. Robinson, R. Ullmann || +# RFC1506 || J. Houttuin || +# RFC1507 || C. Kaufman || +# RFC1508 || J. Linn || +# RFC1509 || J. Wray || Wray@tuxedo.enet.dec.com +# RFC1510 || J. Kohl, C. Neuman || jtkohl@zk3.dec.com, bcn@isi.edu +# RFC1511 || J. Linn || +# RFC1512 || J. Case, A. Rijsinghani || case@CS.UTK.EDU, anil@levers.enet.dec.com +# RFC1513 || S. Waldbusser || waldbusser@cmu.edu +# RFC1514 || P. Grillo, S. Waldbusser || pl0143@mail.psi.net, waldbusser@cmu.edu +# RFC1515 || D. McMaster, K. McCloghrie, S. Roberts || mcmaster@synoptics.com, kzm@hls.com, sroberts@farallon.com +# RFC1516 || D. McMaster, K. McCloghrie || mcmaster@synoptics.com, kzm@hls.com +# RFC1517 || Internet Engineering Steering Group, R. Hinden || bob.hinden@gmail.com +# RFC1518 || Y. Rekhter, T. Li || yakov@watson.ibm.com, tli@cisco.com +# RFC1519 || V. Fuller, T. Li, J. Yu, K. Varadhan || vaf@Stanford.EDU, tli@cisco.com, jyy@merit.edu, kannan@oar.net +# RFC1520 || Y. Rekhter, C. Topolcic || yakov@watson.ibm.com, topolcic@CNRI.Reston.VA.US +# RFC1521 || N. Borenstein, N. Freed || gvaudre@cnri.reston.va.us +# RFC1522 || K. Moore || moore@cs.utk.edu +# RFC1523 || N. Borenstein || nsb@bellcore.com +# RFC1524 || N. Borenstein || nsb@bellcore.com +# RFC1525 || E. Decker, K. McCloghrie, P. Langille, A. Rijsinghani || kzm@hls.com, langille@edwin.enet.dec.com, anil@levers.enet.dec.com +# RFC1526 || D. Piscitello || dave@mail.bellcore.com +# RFC1527 || G. Cook || cook@path.net +# RFC1528 || C. Malamud, M. Rose || +# RFC1529 || C. Malamud, M. Rose || +# RFC1530 || C. Malamud, M. Rose || +# RFC1531 || R. Droms || droms@bucknell.edu +# RFC1532 || W. Wimer || Walter.Wimer@CMU.EDU +# RFC1533 || S. Alexander, R. Droms || stevea@lachman.com, droms@bucknell.edu +# RFC1534 || R. Droms || droms@bucknell.edu +# RFC1535 || E. Gavron || gavron@aces.com +# RFC1536 || A. Kumar, J. Postel, C. Neuman, P. Danzig, S. Miller || anant@isi.edu, postel@isi.edu, bcn@isi.edu, danzig@caldera.usc.edu, smiller@caldera.usc.edu +# RFC1537 || P. Beertema || Piet.Beertema@cwi.nl, anant@isi.edu +# RFC1538 || W. Behl, B. Sterling, W. Teskey || +# RFC1539 || G. Malkin || gmalkin@Xylogics.COM +# RFC1540 || J. Postel || +# RFC1541 || R. Droms || droms@bucknell.edu +# RFC1542 || W. Wimer || Walter.Wimer@CMU.EDU +# RFC1543 || J. Postel || Postel@ISI.EDU, dwaitzman@BBN.COM +# RFC1544 || M. Rose || mrose17@gmail.com +# RFC1545 || D. Piscitello || dave@mail.bellcore.com +# RFC1546 || C. Partridge, T. Mendez, W. Milliken || craig@bbn.com, tmendez@bbn.com, milliken@bbn.com +# RFC1547 || D. Perkins || Bill.Simpson@um.cc.umich.edu +# RFC1548 || W. Simpson || +# RFC1549 || W. Simpson, Ed. || +# RFC1550 || S. Bradner, A. Mankin || sob@harvard.edu, mankin@cmf.nrl.navy.mil +# RFC1551 || M. Allen || mallen@novell.com, fbaker@acc.com +# RFC1552 || W. Simpson || Bill.Simpson@um.cc.umich.edu +# RFC1553 || S. Mathur, M. Lewis || mathur@telebit.com, Mark.S.Lewis@telebit.com +# RFC1554 || M. Ohta, K. Handa || mohta@cc.titech.ac.jp, handa@etl.go.jp +# RFC1555 || H. Nussbacher, Y. Bourvine || hank@vm.tau.ac.il, yehavi@vms.huji.ac.il +# RFC1556 || H. Nussbacher || hank@vm.tau.ac.il +# RFC1557 || U. Choi, K. Chon, H. Park || uhhyung@kaist.ac.kr, chon@cosmos.kaist.ac.kr, hjpark@dino.media.co.kr +# RFC1558 || T. Howes || tim@umich.edu +# RFC1559 || J. Saperia || saperia@tay.dec.com +# RFC1560 || B. Leiner, Y. Rekhter || leiner@nsipo.nasa.gov, yakov@watson.ibm.com +# RFC1561 || D. Piscitello || wk04464@worldlink.com +# RFC1562 || G. Michaelson, M. Prior || G.Michaelson@cc.uq.oz.au, mrp@itd.adelaide.edu.au +# RFC1563 || N. Borenstein || nsb@bellcore.com +# RFC1564 || P. Barker, R. Hedberg || P.Barker@cs.ucl.ac.uk, Roland.Hedberg@rc.tudelft.nl, Roland.Hedberg@umdac.umu.se +# RFC1565 || S. Kille, N. Freed || S.Kille@isode.com, ned@innosoft.com +# RFC1566 || S. Kille, N. Freed || S.Kille@isode.com, ned@innosoft.com +# RFC1567 || G. Mansfield, S. Kille || glenn@aic.co.jp, S.Kille@isode.com +# RFC1568 || A. Gwinn || allen@mail.cox.smu.edu +# RFC1569 || M. Rose || mrose17@gmail.com +# RFC1570 || W. Simpson, Ed. || +# RFC1571 || D. Borman || dab@CRAY.COM +# RFC1572 || S. Alexander, Ed. || +# RFC1573 || K. McCloghrie, F. Kastenholz || kzm@hls.com, kasten@ftp.com +# RFC1574 || S. Hares, C. Wittbrodt || skh@merit.edu, cjw@magnolia.Stanford.EDU +# RFC1575 || S. Hares, C. Wittbrodt || skh@merit.edu, cjw@magnolia.Stanford.EDU +# RFC1576 || J. Penner || jjp@bscs.com +# RFC1577 || M. Laubach || laubach@hpl.hp.com +# RFC1578 || J. Sellers || sellers@quest.arc.nasa.gov +# RFC1579 || S. Bellovin || smb@research.att.com +# RFC1580 || EARN Staff || earndoc@earncc.earn.net +# RFC1581 || G. Meyer || gerry@spider.co.uk +# RFC1582 || G. Meyer || gerry@spider.co.uk +# RFC1583 || J. Moy || +# RFC1584 || J. Moy || +# RFC1585 || J. Moy || jmoy@proteon.com +# RFC1586 || O. deSouza, M. Rodrigues || osmund.desouza@att.com, manoel.rodrigues@att.com +# RFC1587 || R. Coltun, V. Fuller || rcoltun@rainbow-bridge.com, vaf@Valinor.Stanford.EDU +# RFC1588 || J. Postel, C. Anderson || +# RFC1589 || D. Mills || mills@udel.edu +# RFC1590 || J. Postel || Postel@ISI.EDU +# RFC1591 || J. Postel || Postel@ISI.EDU +# RFC1592 || B. Wijnen, G. Carpenter, K. Curran, A. Sehgal, G. Waters || +# RFC1593 || W. McKenzie, J. Cheng || mckenzie@ralvma.vnet.ibm.com, cheng@ralvm6.vnet.ibm.com +# RFC1594 || A. Marine, J. Reynolds, G. Malkin || amarine@atlas.arc.nasa.gov, jkrey@isi.edu, gmalkin@Xylogics.COM +# RFC1595 || T. Brown, K. Tesink || tacox@mail.bellcore.com, kaj@cc.bellcore.com +# RFC1596 || T. Brown, Ed. || tacox@mail.bellcore.com +# RFC1597 || Y. Rekhter, B. Moskowitz, D. Karrenberg, G. de Groot || yakov@watson.ibm.com, 3858921@mcimail.com, Daniel.Karrenberg@ripe.net, GeertJan.deGroot@ripe.net +# RFC1598 || W. Simpson || Bill.Simpson@um.cc.umich.edu +# RFC1599 || M. Kennedy || MKENNEDY@ISI.EDU +# RFC1600 || J. Postel || +# RFC1601 || C. Huitema || Christian.Huitema@sophia.inria.fr +# RFC1602 || Internet Architecture Board, Internet Engineering Steering Group || Christian.Huitema@MIRSA.INRIA.FR, 0006423401@mcimail.com +# RFC1603 || E. Huizer, D. Crocker || +# RFC1604 || T. Brown, Ed. || tacox@mail.bellcore.com +# RFC1605 || W. Shakespeare || +# RFC1606 || J. Onions || j.onions@nexor.co.uk +# RFC1607 || V. Cerf || vcerf@isoc.org, vinton_cerf@mcimail.com +# RFC1608 || T. Johannsen, G. Mansfield, M. Kosters, S. Sataluri || Thomas.Johannsen@ifn.et.tu-dresden.de, glenn@aic.co.jp, markk@internic.net, sri@qsun.att.com +# RFC1609 || G. Mansfield, T. Johannsen, M. Knopper || glenn@aic.co.jp, Thomas.Johannsen@ifn.et.tu-dresden.de, mak@merit.edu +# RFC1610 || J. Postel || +# RFC1611 || R. Austein, J. Saperia || sra@epilogue.com, saperia@zko.dec.com +# RFC1612 || R. Austein, J. Saperia || sra@epilogue.com, saperia@zko.dec.com +# RFC1613 || J. Forster, G. Satz, G. Glick, R. Day || forster@cisco.com, satz@cisco.com, gglick@cisco.com, R.Day@jnt.ac.uk +# RFC1614 || C. Adie || C.J.Adie@edinburgh.ac.uk +# RFC1615 || J. Houttuin, J. Craigie || +# RFC1616 || RARE WG-MSG Task Force 88, E. Huizer, Ed., J. Romaguera, Ed. || +# RFC1617 || P. Barker, S. Kille, T. Lenggenhager || p.barker@cs.ucl.ac.uk, s.kille@isode.com, lenggenhager@switch.ch +# RFC1618 || W. Simpson || Bill.Simpson@um.cc.umich.edu +# RFC1619 || W. Simpson || Bill.Simpson@um.cc.umich.edu +# RFC1620 || B. Braden, J. Postel, Y. Rekhter || Braden@ISI.EDU, Postel@ISI.EDU, Yakov@WATSON.IBM.COM +# RFC1621 || P. Francis || francis@cactus.ntt.jp +# RFC1622 || P. Francis || francis@cactus.ntt.jp +# RFC1623 || F. Kastenholz || kasten@ftp.com +# RFC1624 || A. Rijsinghani, Ed. || anil@levers.enet.dec.com +# RFC1625 || M. St. Pierre, J. Fullton, K. Gamiel, J. Goldman, B. Kahle, J. Kunze, H. Morris, F. Schiettecatte || saint@wais.com, jim.fullton@cnidr.org, kevin.gamiel@cnidr.org, jonathan@think.com, brewster@wais.com, jak@violet.berkeley.edu, morris@wais.com, francois@wais.com +# RFC1626 || R. Atkinson || atkinson@itd.nrl.navy.mil +# RFC1627 || E. Lear, E. Fair, D. Crocker, T. Kessler || +# RFC1628 || J. Case, Ed. || case@SNMP.COM +# RFC1629 || R. Colella, R. Callon, E. Gardner, Y. Rekhter || colella@nist.gov, callon@wellfleet.com, epg@gateway.mitre.org, yakov@watson.ibm.com +# RFC1630 || T. Berners-Lee || timbl@info.cern.ch +# RFC1631 || K. Egevang, P. Francis || kbe@craycom.dk, francis@cactus.ntt.jp +# RFC1632 || A. Getchell, Ed., S. Sataluri, Ed. || +# RFC1633 || R. Braden, D. Clark, S. Shenker || Braden@ISI.EDU, ddc@LCS.MIT.EDU, Shenker@PARC.XEROX.COM +# RFC1634 || M. Allen || mallen@novell.com, fbaker@acc.com +# RFC1635 || P. Deutsch, A. Emtage, A. Marine || peterd@bunyip.com, bajan@bunyip.com, amarine@atlas.arc.nasa.gov +# RFC1636 || R. Braden, D. Clark, S. Crocker, C. Huitema || Braden@ISI.EDU, ddc@lcs.mit.edu, crocker@tis.com, Christian.Huitema@MIRSA.INRIA.FR +# RFC1637 || B. Manning, R. Colella || bmanning@rice.edu, colella@nist.gov +# RFC1638 || F. Baker, R. Bowen || Rich_Bowen@vnet.ibm.com +# RFC1639 || D. Piscitello || dave@corecom.com +# RFC1640 || S. Crocker || crocker@TIS.COM +# RFC1641 || D. Goldsmith, M. Davis || david_goldsmith@taligent.com, mark_davis@taligent.com +# RFC1642 || D. Goldsmith, M. Davis || david_goldsmith@taligent.com, mark_davis@taligent.com +# RFC1643 || F. Kastenholz || kasten@ftp.com +# RFC1644 || R. Braden || Braden@ISI.EDU +# RFC1645 || A. Gwinn || allen@mail.cox.smu.edu +# RFC1646 || C. Graves, T. Butts, M. Angel || cvg@oc.com, tom@oc.com, angel@oc.com +# RFC1647 || B. Kelly || kellywh@mail.auburn.edu +# RFC1648 || A. Cargille || +# RFC1649 || R. Hagens, A. Hansen || hagens@ans.net, Alf.Hansen@uninett.no +# RFC1650 || F. Kastenholz || kasten@ftp.com +# RFC1651 || J. Klensin, N. Freed, M. Rose, E. Stefferud, D. Crocker || klensin@mci.net, ned@innosoft.com, mrose17@gmail.com, stef@nma.com, dcrocker@sgi.com +# RFC1652 || J. Klensin, N. Freed, M. Rose, E. Stefferud, D. Crocker || klensin@mci.net, ned@innosoft.com, mrose17@gmail.com, stef@nma.com, dcrocker@sgi.com +# RFC1653 || J. Klensin, N. Freed, K. Moore || klensin@mci.net, ned@innosoft.com, moore@cs.utk.edu +# RFC1654 || Y. Rekhter, Ed., T. Li, Ed. || +# RFC1655 || Y. Rekhter, Ed., P. Gross, Ed. || yakov@watson.ibm.com, 0006423401@mcimail.com +# RFC1656 || P. Traina || pst@cisco.com +# RFC1657 || S. Willis, J. Burruss, J. Chu, Ed. || swillis@wellfleet.com, jburruss@wellfleet.com, jychu@watson.ibm.com +# RFC1658 || B. Stewart || rlstewart@eng.xyplex.com +# RFC1659 || B. Stewart || rlstewart@eng.xyplex.com +# RFC1660 || B. Stewart || rlstewart@eng.xyplex.com +# RFC1661 || W. Simpson, Ed. || +# RFC1662 || W. Simpson, Ed. || +# RFC1663 || D. Rand || dave_rand@novell.com +# RFC1664 || C. Allocchio, A. Bonito, B. Cole, S. Giordano, R. Hagens || +# RFC1665 || Z. Kielczewski, Ed., D. Kostick, Ed., K. Shih, Ed. || zbig@eicon.qc.ca, dck2@mail.bellcore.com, kmshih@novell.com +# RFC1666 || Z. Kielczewski, Ed., D. Kostick, Ed., K. Shih, Ed. || zbig@eicon.qc.ca, dck2@mail.bellcore.com, kmshih@novell.com +# RFC1667 || S. Symington, D. Wood, M. Pullen || susan@gateway.mitre.org, wood@mitre.org, mpullen@cs.gmu.edu +# RFC1668 || D. Estrin, T. Li, Y. Rekhter || estrin@usc.edu, tli@cisco.com, yakov@watson.ibm.com +# RFC1669 || J. Curran || jcurran@near.net +# RFC1670 || D. Heagerty || denise@dxcoms.cern.ch +# RFC1671 || B. Carpenter || brian@dxcoms.cern.ch +# RFC1672 || N. Brownlee || n.brownlee@auckland.ac.nz +# RFC1673 || R. Skelton || RSKELTON@msm.epri.com +# RFC1674 || M. Taylor || mark.s.taylor@airdata.com +# RFC1675 || S. Bellovin || smb@research.att.com +# RFC1676 || A. Ghiselli, D. Salomoni, C. Vistoli || Salomoni@infn.it, Vistoli@infn.it, Ghiselli@infn.it +# RFC1677 || B. Adamson || adamson@itd.nrl.navy.mil +# RFC1678 || E. Britton, J. Tavs || brittone@vnet.ibm.com, tavs@vnet.ibm.com +# RFC1679 || D. Green, P. Irey, D. Marlow, K. O'Donoghue || dtgreen@relay.nswc.navy.mil, pirey@relay.nswc.navy.mil, dmarlow@relay.nswc.navy.mil, kodonog@relay.nswc.navy.mil +# RFC1680 || C. Brazdziunas || crb@faline.bellcore.com +# RFC1681 || S. Bellovin || smb@research.att.com +# RFC1682 || J. Bound || bound@zk3.dec.com +# RFC1683 || R. Clark, M. Ammar, K. Calvert || rjc@cc.gatech.edu, ammar@cc.gatech.edu, calvert@cc.gatech.edu +# RFC1684 || P. Jurg || +# RFC1685 || H. Alvestrand || +# RFC1686 || M. Vecchi || mpvecchi@twcable.com +# RFC1687 || E. Fleischman || ericf@atc.boeing.com +# RFC1688 || W. Simpson || Bill.Simpson@um.cc.umich.edu +# RFC1689 || J. Foster, Ed. || +# RFC1690 || G. Huston || Geoff.Huston@aarnet.edu.au +# RFC1691 || W. Turner || wrt1@cornell.edu +# RFC1692 || P. Cameron, D. Crocker, D. Cohen, J. Postel || cameron@xylint.co.uk, dcrocker@sgi.com, Cohen@myricom.com, Postel@ISI.EDU +# RFC1693 || T. Connolly, P. Amer, P. Conrad || connolly@udel.edu, amer@udel.edu, pconrad@udel.edu +# RFC1694 || T. Brown, Ed., K. Tesink, Ed. || tacox@mail.bellcore.com, kaj@cc.bellcore.com +# RFC1695 || M. Ahmed, Ed., K. Tesink, Ed. || mxa@mail.bellcore.com, kaj@cc.bellcore.com +# RFC1696 || J. Barnes, L. Brown, R. Royston, S. Waldbusser || barnes@xylogics.com, brown_l@msm.cdx.mot.com, rroyston@usr.com, swol@andrew.cmu.edu +# RFC1697 || D. Brower, Ed., B. Purvy, A. Daniel, M. Sinykin, J. Smith || daveb@ingres.com, bpurvy@us.oracle.com, anthony@informix.com, msinykin@us.oracle.com, jaysmith@us.oracle.com +# RFC1698 || P. Furniss || P.Furniss@ulcc.ac.uk +# RFC1699 || J. Elliott || elliott@isi.edu +# RFC1700 || J. Reynolds, J. Postel || jkrey@isi.edu, postel@isi.edu +# RFC1701 || S. Hanks, T. Li, D. Farinacci, P. Traina || stan@netsmiths.com, tli@cisco.com, dino@cisco.com, pst@cisco.com +# RFC1702 || S. Hanks, T. Li, D. Farinacci, P. Traina || stan@netsmiths.com, tli@cisco.com, dino@cisco.com, pst@cisco.com +# RFC1703 || M. Rose || mrose17@gmail.com +# RFC1704 || N. Haller, R. Atkinson || +# RFC1705 || R. Carlson, D. Ficarella || RACarlson@anl.gov, ficarell@cpdmfg.cig.mot.com +# RFC1706 || B. Manning, R. Colella || bmanning@isi.edu, colella@nist.gov +# RFC1707 || M. McGovern, R. Ullmann || scrivner@world.std.com, rullmann@crd.lotus.com +# RFC1708 || D. Gowin || drg508@crane-ns.nwscc.sea06.navy.MIL +# RFC1709 || J. Gargano, D. Wasley || jcgargano@ucdavis.edu, dlw@berkeley.edu +# RFC1710 || R. Hinden || bob.hinden@gmail.com +# RFC1711 || J. Houttuin || houttuin@rare.nl +# RFC1712 || C. Farrell, M. Schulze, S. Pleitner, D. Baldoni || craig@cs.curtin.edu.au, mike@cs.curtin.edu.au, pleitner@cs.curtin.edu.au, flint@cs.curtin.edu.au +# RFC1713 || A. Romao || artur@fct.unl.pt +# RFC1714 || S. Williamson, M. Kosters || scottw@internic.net, markk@internic.net +# RFC1715 || C. Huitema || Christian.Huitema@MIRSA.INRIA.FR +# RFC1716 || P. Almquist, F. Kastenholz || +# RFC1717 || K. Sklower, B. Lloyd, G. McGregor, D. Carr || sklower@CS.Berkeley.EDU, brian@lloyd.com, glenn@lloyd.com, dcarr@Newbridge.COM +# RFC1718 || IETF Secretariat, G. Malkin || ietf-info@cnri.reston.va.us, gmalkin@Xylogics.COM +# RFC1719 || P. Gross || phill_gross@mcimail.com +# RFC1720 || J. Postel || +# RFC1721 || G. Malkin || gmalkin@Xylogics.COM +# RFC1722 || G. Malkin || gmalkin@Xylogics.COM +# RFC1723 || G. Malkin || gmalkin@Xylogics.COM +# RFC1724 || G. Malkin, F. Baker || gmalkin@Xylogics.COM, fred@cisco.com +# RFC1725 || J. Myers, M. Rose || mrose17@gmail.com +# RFC1726 || C. Partridge, F. Kastenholz || craig@aland.bbn.com, kasten@ftp.com +# RFC1727 || C. Weider, P. Deutsch || clw@bunyip.com, peterd@bunyip.com +# RFC1728 || C. Weider || clw@bunyip.com +# RFC1729 || C. Lynch || clifford.lynch@ucop.edu +# RFC1730 || M. Crispin || MRC@CAC.Washington.EDU +# RFC1731 || J. Myers || +# RFC1732 || M. Crispin || MRC@CAC.Washington.EDU +# RFC1733 || M. Crispin || MRC@CAC.Washington.EDU +# RFC1734 || J. Myers || +# RFC1735 || J. Heinanen, R. Govindan || Juha.Heinanen@datanet.tele.fi, govindan@isi.edu +# RFC1736 || J. Kunze || jak@violet.berkeley.edu +# RFC1737 || K. Sollins, L. Masinter || masinter@parc.xerox.com, sollins@lcs.mit.edu +# RFC1738 || T. Berners-Lee, L. Masinter, M. McCahill || +# RFC1739 || G. Kessler, S. Shepard || kumquat@hill.com, sds@hill.com +# RFC1740 || P. Faltstrom, D. Crocker, E. Fair || paf@nada.kth.se, dcrocker@mordor.stanford.edu, fair@apple.com +# RFC1741 || P. Faltstrom, D. Crocker, E. Fair || paf@nada.kth.se, dcrocker@mordor.stanford.edu, fair@apple.com +# RFC1742 || S. Waldbusser, K. Frisa || waldbusser@cmu.edu, kfrisa@fore.com +# RFC1743 || K. McCloghrie, E. Decker || kzm@cisco.com, cire@cisco.com +# RFC1744 || G. Huston || Geoff.Huston@aarnet.edu.au +# RFC1745 || K. Varadhan, S. Hares, Y. Rekhter || kannan@isi.edu, skh@merit.edu, yakov@watson.ibm.com +# RFC1746 || B. Manning, D. Perkins || bmanning@isi.edu, dperkins@tenet.edu +# RFC1747 || J. Hilgeman, Chair, S. Nix, A. Bartky, W. Clark, Ed. || jeffh@apertus.com, snix@metaplex.com, alan@sync.com, wclark@cisco.com +# RFC1748 || K. McCloghrie, E. Decker || kzm@cisco.com, cire@cisco.com +# RFC1749 || K. McCloghrie, F. Baker, E. Decker || kzm@cisco.com, fred@cisco.com, cire@cisco.com +# RFC1750 || D. Eastlake 3rd, S. Crocker, J. Schiller || dee@lkg.dec.com, crocker@cybercash.com, jis@mit.edu +# RFC1751 || D. McDonald || danmcd@itd.nrl.navy.mil +# RFC1752 || S. Bradner, A. Mankin || sob@harvard.edu, mankin@isi.edu +# RFC1753 || N. Chiappa || jnc@lcs.mit.edu +# RFC1754 || M. Laubach || laubach@com21.com +# RFC1755 || M. Perez, F. Liaw, A. Mankin, E. Hoffman, D. Grossman, A. Malis || perez@isi.edu, fong@fore.com, mankin@isi.edu, hoffman@isi.edu, dan@merlin.dev.cdx.mot.com, malis@maelstrom.timeplex.com +# RFC1756 || T. Rinne || Timo.Rinne@hut.fi +# RFC1757 || S. Waldbusser || waldbusser@cmu.edu +# RFC1758 || The North American Directory Forum || 0004454742@mcimail.com +# RFC1759 || R. Smith, F. Wright, T. Hastings, S. Zilles, J. Gyllenskog || rlsmith@nb.ppd.ti.com, don@lexmark.com, tom.hastings@alum.mit.edu, szilles@mv.us.adobe.com, jgyllens@hpdmd48.boi.hp.com +# RFC1760 || N. Haller || nmh@bellcore.com +# RFC1761 || B. Callaghan, R. Gilligan || brent.callaghan@eng.sun.com, bob.gilligan@eng.sun.com +# RFC1762 || S. Senum || sjs@digibd.com +# RFC1763 || S. Senum || sjs@digibd.com +# RFC1764 || S. Senum || sjs@digibd.com +# RFC1765 || J. Moy || jmoy@casc.com +# RFC1766 || H. Alvestrand || Harald.T.Alvestrand@uninett.no +# RFC1767 || D. Crocker || +# RFC1768 || D. Marlow || dmarlow@relay.nswc.navy.mil +# RFC1769 || D. Mills || mills@udel.edu +# RFC1770 || C. Graff || bud@fotlan5.fotlan.army.mil +# RFC1771 || Y. Rekhter, T. Li || +# RFC1772 || Y. Rekhter, P. Gross || yakov@watson.ibm.com, 0006423401@mcimail.com +# RFC1773 || P. Traina || pst@cisco.com +# RFC1774 || P. Traina, Ed. || +# RFC1775 || D. Crocker || dcrocker@mordor.stanford.edu +# RFC1776 || S. Crocker || crocker@cybercash.com +# RFC1777 || W. Yeong, T. Howes, S. Kille || yeongw@psilink.com, tim@umich.edu, S.Kille@isode.com +# RFC1778 || T. Howes, S. Kille, W. Yeong, C. Robbins || tim@umich.edu, S.Kille@isode.com, yeongw@psilink.com +# RFC1779 || S. Kille || S.Kille@ISODE.COM +# RFC1780 || J. Postel, Ed. || +# RFC1781 || S. Kille || S.Kille@ISODE.COM +# RFC1782 || G. Malkin, A. Harkin || gmalkin@xylogics.com, ash@cup.hp.com +# RFC1783 || G. Malkin, A. Harkin || gmalkin@xylogics.com, ash@cup.hp.com +# RFC1784 || G. Malkin, A. Harkin || gmalkin@xylogics.com, ash@cup.hp.com +# RFC1785 || G. Malkin, A. Harkin || gmalkin@xylogics.com, ash@cup.hp.com +# RFC1786 || T. Bates, E. Gerich, L. Joncheray, J-M. Jouanigot, D. Karrenberg, M. Terpstra, J. Yu || +# RFC1787 || Y. Rekhter || +# RFC1788 || W. Simpson || +# RFC1789 || C. Yang || cqyang@cs.unt.edu +# RFC1790 || V. Cerf || vcerf@isoc.org +# RFC1791 || T. Sung || tae@novell.Com +# RFC1792 || T. Sung || tae@novell.Com +# RFC1793 || J. Moy || jmoy@casc.com +# RFC1794 || T. Brisco || brisco@rutgers.edu +# RFC1795 || L. Wells, Chair, A. Bartky, Ed. || +# RFC1796 || C. Huitema, J. Postel, S. Crocker || Christian.Huitema@MIRSA.INRIA.FR, Postel@ISI.EDU, crocker@cybercash.com +# RFC1797 || Internet Assigned Numbers Authority (IANA) || iana@isi.edu +# RFC1798 || A. Young || A.Young@isode.com +# RFC1799 || M. Kennedy || MKENNEDY@ISI.EDU +# RFC1800 || J. Postel, Ed. || +# RFC1801 || S. Kille || S.Kille@ISODE.COM +# RFC1802 || H. Alvestrand, K. Jordan, S. Langlois, J. Romaguera || Harald.T.Alvestrand@uninett.no, Kevin.E.Jordan@cdc.com, Sylvain.Langlois@der.edf.fr, Romaguera@NetConsult.ch +# RFC1803 || R. Wright, A. Getchell, T. Howes, S. Sataluri, P. Yee, W. Yeong || +# RFC1804 || G. Mansfield, P. Rajeev, S. Raghavan, T. Howes || glenn@aic.co.jp, rajeev%hss@lando.hns.com, svr@iitm.ernet.in, tim@umich.edu +# RFC1805 || A. Rubin || rubin@bellcore.com +# RFC1806 || R. Troost, S. Dorner || rens@century.com, sdorner@qualcomm.com +# RFC1807 || R. Lasher, D. Cohen || rlasher@forsythe.stanford.edu, Cohen@myri.com +# RFC1808 || R. Fielding || fielding@ics.uci.edu +# RFC1809 || C. Partridge || craig@aland.bbn.com +# RFC1810 || J. Touch || touch@isi.edu +# RFC1811 || Federal Networking Council || execdir@fnc.gov +# RFC1812 || F. Baker, Ed. || +# RFC1813 || B. Callaghan, B. Pawlowski, P. Staubach || brent.callaghan@eng.sun.com, beepy@netapp.com, peter.staubach@eng.sun.com +# RFC1814 || E. Gerich || epg@merit.edu +# RFC1815 || M. Ohta || mohta@cc.titech.ac.jp +# RFC1816 || Federal Networking Council || execdir@fnc.gov +# RFC1817 || Y. Rekhter || yakov@cisco.com +# RFC1818 || J. Postel, T. Li, Y. Rekhter || postel@isi.edu, yakov@cisco.com, tli@cisco.com +# RFC1819 || L. Delgrossi, Ed., L. Berger, Ed. || lberger@bbn.com, luca@andersen.fr, dat@bbn.com, stevej@syzygycomm.com, schaller@heidelbg.ibm.com +# RFC1820 || E. Huizer || Erik.Huizer@SURFnet.nl +# RFC1821 || M. Borden, E. Crawley, B. Davie, S. Batsell || +# RFC1822 || J. Lowe || +# RFC1823 || T. Howes, M. Smith || tim@umich.edu, mcs@umich.edu +# RFC1824 || H. Danisch || danisch@ira.uka.de +# RFC1825 || R. Atkinson || +# RFC1826 || R. Atkinson || +# RFC1827 || R. Atkinson || +# RFC1828 || P. Metzger, W. Simpson || +# RFC1829 || P. Karn, P. Metzger, W. Simpson || +# RFC1830 || G. Vaudreuil || Greg.Vaudreuil@Octel.com +# RFC1831 || R. Srinivasan || raj@eng.sun.com +# RFC1832 || R. Srinivasan || raj@eng.sun.com +# RFC1833 || R. Srinivasan || raj@eng.sun.com +# RFC1834 || J. Gargano, K. Weiss || jcgargano@ucdavis.edu, krweiss@ucdavis.edu +# RFC1835 || P. Deutsch, R. Schoultz, P. Faltstrom, C. Weider || peterd@bunyip.com, schoultz@sunet.se, paf@bunyip.com, clw@bunyip.com +# RFC1836 || S. Kille || S.Kille@ISODE.COM +# RFC1837 || S. Kille || S.Kille@ISODE.COM +# RFC1838 || S. Kille || S.Kille@ISODE.COM +# RFC1839 || || +# RFC1840 || || +# RFC1841 || J. Chapman, D. Coli, A. Harvey, B. Jensen, K. Rowett || joelle@cisco.com, dcoli@cisco.com, agh@cisco.com, bent@cisco.com, krowett@cisco.com +# RFC1842 || Y. Wei, Y. Zhang, J. Li, J. Ding, Y. Jiang || HZRFC@usai.asiainfo.com, zhang@orion.harvard.edu, jian@is.rice.edu, ding@Beijing.AsiaInfo.com, yjj@eng.umd.edu +# RFC1843 || F. Lee || lee@csl.stanford.edu +# RFC1844 || E. Huizer || Erik.Huizer@SURFnet.nl +# RFC1845 || D. Crocker, N. Freed, A. Cargille || dcrocker@mordor.stanford.edu, ned@innosoft.com +# RFC1846 || A. Durand, F. Dupont || Alain.Durand@imag.fr, Francis.Dupont@inria.fr +# RFC1847 || J. Galvin, S. Murphy, S. Crocker, N. Freed || galvin@tis.com, sandy@tis.com, crocker@cybercash.com, ned@innosoft.com +# RFC1848 || S. Crocker, N. Freed, J. Galvin, S. Murphy || crocker@cybercash.com, galvin@tis.com, murphy@tis.com, ned@innosoft.com +# RFC1849 || H. Spencer || henry@zoo.utoronto.ca +# RFC1850 || F. Baker, R. Coltun || fred@cisco.com, rcoltun@rainbow-bridge.com +# RFC1851 || P. Karn, P. Metzger, W. Simpson || +# RFC1852 || P. Metzger, W. Simpson || +# RFC1853 || W. Simpson || +# RFC1854 || N. Freed || ned@innosoft.com +# RFC1855 || S. Hambridge || sallyh@ludwig.sc.intel.com +# RFC1856 || H. Clark || henryc@bbnplanet.com +# RFC1857 || M. Lambert || lambert@psc.edu +# RFC1858 || G. Ziemba, D. Reed, P. Traina || paul@alantec.com, darrenr@cyber.com.au, pst@cisco.com +# RFC1859 || Y. Pouffary || pouffary@taec.enet.dec.com +# RFC1860 || T. Pummill, B. Manning || trop@alantec.com, bmanning@isi.edu +# RFC1861 || A. Gwinn || allen@mail.cox.smu.edu +# RFC1862 || M. McCahill, J. Romkey, M. Schwartz, K. Sollins, T. Verschuren, C. Weider || mpm@boombox.micro.umn.edu, romkey@apocalypse.org, schwartz@cs.colorado.edu, sollins@lcs.mit.edu, Ton.Verschuren@surfnet.nl, clw@bunyip.com +# RFC1863 || D. Haskin || dhaskin@baynetworks.com +# RFC1864 || J. Myers, M. Rose || mrose17@gmail.com +# RFC1865 || W. Houser, J. Griffin, C. Hage || houser.walt@forum.va.gov, agriffin@cpcug.org, carl@chage.com +# RFC1866 || T. Berners-Lee, D. Connolly || timbl@w3.org, connolly@w3.org +# RFC1867 || E. Nebel, L. Masinter || masinter@parc.xerox.com, nebel@xsoft.sd.xerox.com +# RFC1868 || G. Malkin || gmalkin@xylogics.com +# RFC1869 || J. Klensin, N. Freed, M. Rose, E. Stefferud, D. Crocker || klensin@mci.net, ned@innosoft.com, mrose17@gmail.com, stef@nma.com, dcrocker@mordor.stanford.edu +# RFC1870 || J. Klensin, N. Freed, K. Moore || klensin@mci.net, ned@innosoft.com, moore@cs.utk.edu +# RFC1871 || J. Postel || postel@isi.edu +# RFC1872 || E. Levinson || ELevinson@Accurate.com +# RFC1873 || E. Levinson || +# RFC1874 || E. Levinson || ELevinson@Accurate.com +# RFC1875 || N. Berge || Nils.Harald.Berge@nr.no +# RFC1876 || C. Davis, P. Vixie, T. Goodwin, I. Dickinson || ckd@kei.com, paul@vix.com, tim@pipex.net, idickins@fore.co.uk +# RFC1877 || S. Cobb || stevec@microsoft.com +# RFC1878 || T. Pummill, B. Manning || trop@alantec.com, bmanning@isi.edu +# RFC1879 || B. Manning, Ed. || +# RFC1880 || J. Postel, Ed. || +# RFC1881 || IAB, IESG || +# RFC1882 || B. Hancock || hancock@network-1.com +# RFC1883 || S. Deering, R. Hinden || bob.hinden@gmail.com +# RFC1884 || R. Hinden, Ed., S. Deering, Ed. || bob.hinden@gmail.com +# RFC1885 || A. Conta, S. Deering || deering@parc.xerox.com +# RFC1886 || S. Thomson, C. Huitema || set@thumper.bellcore.com, Christian.Huitema@MIRSA.INRIA.FR +# RFC1887 || Y. Rekhter, Ed., T. Li, Ed. || +# RFC1888 || J. Bound, B. Carpenter, D. Harrington, J. Houldsworth, A. Lloyd || +# RFC1889 || Audio-Video Transport Working Group, H. Schulzrinne, S. Casner, R. Frederick, V. Jacobson || schulzrinne@fokus.gmd.de, casner@precept.com, frederic@parc.xerox.com, van@ee.lbl.gov +# RFC1890 || Audio-Video Transport Working Group, H. Schulzrinne || schulzrinne@fokus.gmd.de +# RFC1891 || K. Moore || moore@cs.utk.edu +# RFC1892 || G. Vaudreuil || Greg.Vaudreuil@Octel.com +# RFC1893 || G. Vaudreuil || Greg.Vaudreuil@Octel.com +# RFC1894 || K. Moore, G. Vaudreuil || moore@cs.utk.edu, Greg.Vaudreuil@Octel.Com +# RFC1895 || E. Levinson || ELevinson@Accurate.com +# RFC1896 || P. Resnick, A. Walker || presnick@qti.qualcomm.com, amanda@intercon.com +# RFC1897 || R. Hinden, J. Postel || bob.hinden@gmail.com, postel@isi.edu +# RFC1898 || D. Eastlake 3rd, B. Boesch, S. Crocker, M. Yesil || dee@cybercash.com, boesch@cybercash.com, crocker@cybercash.com, magdalen@cybercash.com +# RFC1899 || J. Elliott || elliott@isi.edu +# RFC1900 || B. Carpenter, Y. Rekhter || brian@dxcoms.cern.ch, yakov@cisco.com +# RFC1901 || J. Case, K. McCloghrie, M. Rose, S. Waldbusser || +# RFC1902 || J. Case, K. McCloghrie, M. Rose, S. Waldbusser || +# RFC1903 || J. Case, K. McCloghrie, M. Rose, S. Waldbusser || +# RFC1904 || J. Case, K. McCloghrie, M. Rose, S. Waldbusser || +# RFC1905 || J. Case, K. McCloghrie, M. Rose, S. Waldbusser || +# RFC1906 || J. Case, K. McCloghrie, M. Rose, S. Waldbusser || +# RFC1907 || J. Case, K. McCloghrie, M. Rose, S. Waldbusser || +# RFC1908 || J. Case, K. McCloghrie, M. Rose, S. Waldbusser || +# RFC1909 || K. McCloghrie, Ed. || +# RFC1910 || G. Waters, Ed. || +# RFC1911 || G. Vaudreuil || Greg.Vaudreuil@Octel.Com +# RFC1912 || D. Barr || barr@math.psu.edu +# RFC1913 || C. Weider, J. Fullton, S. Spero || clw@bunyip.com, fullton@cnidr.org, ses@eit.com +# RFC1914 || P. Faltstrom, R. Schoultz, C. Weider || paf@bunyip.com, schoultz@sunet.se, clw@bunyip.com +# RFC1915 || F. Kastenholz || kasten@ftp.com +# RFC1916 || H. Berkowitz, P. Ferguson, W. Leland, P. Nesser || hcb@clark.net, pferguso@cisco.com, wel@bellcore.com, pjnesser@rocket.com +# RFC1917 || P. Nesser II || pjnesser@martigny.ai.mit.edu +# RFC1918 || Y. Rekhter, B. Moskowitz, D. Karrenberg, G. J. de Groot, E. Lear || yakov@cisco.com, rgm3@is.chrysler.com, Daniel.Karrenberg@ripe.net, GeertJan.deGroot@ripe.net, lear@sgi.com +# RFC1919 || M. Chatel || mchatel@pax.eunet.ch +# RFC1920 || J. Postel || +# RFC1921 || J. Dujonc || J.Y.Dujonc@frcl.bull.fr +# RFC1922 || HF. Zhu, DY. Hu, ZG. Wang, TC. Kao, WCH. Chang, M. Crispin || zhf@net.tsinghua.edu.cn, hdy@tsinghua.edu.cn, tckao@iiidns.iii.org.tw, chung@iiidns.iii.org.tw, MRC@CAC.Washington.EDU +# RFC1923 || J. Halpern, S. Bradner || jhalpern@newbridge.com, sob@harvard.edu +# RFC1924 || R. Elz || kre@munnari.OZ.AU +# RFC1925 || R. Callon || rcallon@baynetworks.com +# RFC1926 || J. Eriksson || bygg@sunet.se +# RFC1927 || C. Rogers || rogers@isi.edu +# RFC1928 || M. Leech, M. Ganis, Y. Lee, R. Kuris, D. Koblas, L. Jones || mleech@bnr.ca +# RFC1929 || M. Leech || mleech@bnr.ca +# RFC1930 || J. Hawkinson, T. Bates || jhawk@bbnplanet.com, Tony.Bates@mci.net +# RFC1931 || D. Brownell || dbrownell@sun.com +# RFC1932 || R. Cole, D. Shur, C. Villamizar || rgc@qsun.att.com, d.shur@att.com, curtis@ans.net +# RFC1933 || R. Gilligan, E. Nordmark || Bob.Gilligan@Eng.Sun.COM, Erik.Nordmark@Eng.Sun.COM +# RFC1934 || K. Smith || ksmith@ascend.com +# RFC1935 || J. Quarterman, S. Carl-Mitchell || tic@tic.com +# RFC1936 || J. Touch, B. Parham || touch@isi.edu, bparham@isi.edu +# RFC1937 || Y. Rekhter, D. Kandlur || yakov@cisco.com, kandlur@watson.ibm.com +# RFC1938 || N. Haller, C. Metz || +# RFC1939 || J. Myers, M. Rose || mrose17@gmail.com +# RFC1940 || D. Estrin, T. Li, Y. Rekhter, K. Varadhan, D. Zappala || estrin@isi.edu, tli@cisco.com, yakov@cisco.com, kannan@isi.edu, daniel@isi.edu +# RFC1941 || J. Sellers, J. Robichaux || julier@internic.net, sellers@quest.arc.nasa.gov +# RFC1942 || D. Raggett || dsr@w3.org +# RFC1943 || B. Jennings || jennings@sandia.gov +# RFC1944 || S. Bradner, J. McQuaid || +# RFC1945 || T. Berners-Lee, R. Fielding, H. Frystyk || timbl@w3.org, fielding@ics.uci.edu, frystyk@w3.org +# RFC1946 || S. Jackowski || Stevej@NetManage.com +# RFC1947 || D. Spinellis || D.Spinellis@senanet.com +# RFC1948 || S. Bellovin || smb@research.att.com +# RFC1949 || A. Ballardie || A.Ballardie@cs.ucl.ac.uk +# RFC1950 || P. Deutsch, J-L. Gailly || +# RFC1951 || P. Deutsch || +# RFC1952 || P. Deutsch || +# RFC1953 || P. Newman, W. Edwards, R. Hinden, E. Hoffman, F. Ching Liaw, T. Lyon, G. Minshall || bob.hinden@gmail.com +# RFC1954 || P. Newman, W. Edwards, R. Hinden, E. Hoffman, F. Ching Liaw, T. Lyon, G. Minshall || bob.hinden@gmail.com +# RFC1955 || R. Hinden || bob.hinden@gmail.com +# RFC1956 || D. Engebretson, R. Plzak || engebred@ncr.disa.mil, plzak@nic.ddn.mil +# RFC1957 || R. Nelson || nelson@crynwr.com +# RFC1958 || B. Carpenter, Ed. || +# RFC1959 || T. Howes, M. Smith || tim@umich.edu, mcs@umich.edu +# RFC1960 || T. Howes || tim@umich.edu +# RFC1961 || P. McMahon || p.v.mcmahon@rea0803.wins.icl.co.uk +# RFC1962 || D. Rand || dlr@daver.bungi.com +# RFC1963 || K. Schneider, S. Venters || kevin@adtran.com, sventers@adtran.com +# RFC1964 || J. Linn || +# RFC1965 || P. Traina || pst@cisco.com +# RFC1966 || T. Bates, R. Chandra || tbates@cisco.com, rchandra@cisco.com +# RFC1967 || K. Schneider, R. Friend || kschneider@adtran.com, rfriend@stac.com +# RFC1968 || G. Meyer || gerry@spider.co.uk +# RFC1969 || K. Sklower, G. Meyer || sklower@CS.Berkeley.EDU, gerry@spider.co.uk +# RFC1970 || T. Narten, E. Nordmark, W. Simpson || +# RFC1971 || S. Thomson, T. Narten || +# RFC1972 || M. Crawford || crawdad@fnal.gov +# RFC1973 || W. Simpson || +# RFC1974 || R. Friend, W. Simpson || rfriend@stac.com +# RFC1975 || D. Schremp, J. Black, J. Weiss || dhs@magna.telco.com, jtb@magna.telco.com, jaw@magna.telco.com +# RFC1976 || K. Schneider, S. Venters || kevin@adtran.com, sventers@adtran.com +# RFC1977 || V. Schryver || vjs@rhyolite.com +# RFC1978 || D. Rand || dave_rand@novell.com +# RFC1979 || J. Woods || jfw@funhouse.com +# RFC1980 || J. Seidman || jim@spyglass.com +# RFC1981 || J. McCann, S. Deering, J. Mogul || deering@parc.xerox.com, mogul@pa.dec.com +# RFC1982 || R. Elz, R. Bush || randy@psg.com +# RFC1983 || G. Malkin, Ed. || +# RFC1984 || IAB, IESG || brian@dxcoms.cern.ch, fred@cisco.com +# RFC1985 || J. De Winter || jack@wildbear.on.ca +# RFC1986 || W. Polites, W. Wollman, D. Woo, R. Langan || +# RFC1987 || P. Newman, W. Edwards, R. Hinden, E. Hoffman, F. Ching Liaw, T. Lyon, G. Minshall || bob.hinden@gmail.com +# RFC1988 || G. McAnally, D. Gilbert, J. Flick || johnf@hprnd.rose.hp.com +# RFC1989 || W. Simpson || +# RFC1990 || K. Sklower, B. Lloyd, G. McGregor, D. Carr, T. Coradetti || sklower@CS.Berkeley.EDU, brian@lloyd.com, glenn@lloyd.com, dcarr@Newbridge.COM, 70761.1664@compuserve.com +# RFC1991 || D. Atkins, W. Stallings, P. Zimmermann || warlord@MIT.EDU, stallings@ACM.org, prz@acm.org +# RFC1992 || I. Castineyra, N. Chiappa, M. Steenstrup || isidro@bbn.com, gnc@ginger.lcs.mit.edu, msteenst@bbn.com +# RFC1993 || A. Barbir, D. Carr, W. Simpson || +# RFC1994 || W. Simpson || +# RFC1995 || M. Ohta || mohta@necom830.hpcl.titech.ac.jp +# RFC1996 || P. Vixie || paul@vix.com +# RFC1997 || R. Chandra, P. Traina, T. Li || pst@cisco.com, rchandra@cisco.com, tli@skat.usc.edu +# RFC1998 || E. Chen, T. Bates || enke@mci.net, tbates@cisco.com +# RFC1999 || J. Elliott || elliott@isi.edu +# RFC2000 || J. Postel, Ed. || +# RFC2001 || W. Stevens || rstevens@noao.edu +# RFC2002 || C. Perkins, Ed. || +# RFC2003 || C. Perkins || perk@watson.ibm.com, solomon@comm.mot.com +# RFC2004 || C. Perkins || perk@watson.ibm.com, solomon@comm.mot.com +# RFC2005 || J. Solomon || solomon@comm.mot.com +# RFC2006 || D. Cong, M. Hamlen, C. Perkins || +# RFC2007 || J. Foster, M. Isaacs, M. Prior || Jill.Foster@newcastle.ac.uk, mmi@dcs.gla.ac.uk, mrp@connect.com.au +# RFC2008 || Y. Rekhter, T. Li || yakov@cisco.com, tli@cisco.com +# RFC2009 || T. Imielinski, J. Navas || +# RFC2010 || B. Manning, P. Vixie || bmanning@isi.edu, paul@vix.com +# RFC2011 || K. McCloghrie, Ed. || +# RFC2012 || K. McCloghrie, Ed. || +# RFC2013 || K. McCloghrie, Ed. || +# RFC2014 || A. Weinrib, J. Postel || +# RFC2015 || M. Elkins || +# RFC2016 || L. Daigle, P. Deutsch, B. Heelan, C. Alpaugh, M. Maclachlan || ura-bunyip@bunyip.com +# RFC2017 || N. Freed, K. Moore, A. Cargille || ned@innosoft.com, moore@cs.utk.edu +# RFC2018 || M. Mathis, J. Mahdavi, S. Floyd, A. Romanow || +# RFC2019 || M. Crawford || crawdad@fnal.gov +# RFC2020 || J. Flick || +# RFC2021 || S. Waldbusser || waldbusser@ins.com +# RFC2022 || G. Armitage || gja@thumper.bellcore.com +# RFC2023 || D. Haskin, E. Allen || +# RFC2024 || D. Chen, Ed., P. Gayek, S. Nix || dchen@vnet.ibm.com, gayek@vnet.ibm.com, snix@metaplex.com +# RFC2025 || C. Adams || cadams@bnr.ca +# RFC2026 || S. Bradner || +# RFC2027 || J. Galvin || +# RFC2028 || R. Hovey, S. Bradner || hovey@wnpv01.enet.dec.com, sob@harvard.edu +# RFC2029 || M. Speer, D. Hoffman || michael.speer@eng.sun.com, don.hoffman@eng.sun.com +# RFC2030 || D. Mills || +# RFC2031 || E. Huizer || +# RFC2032 || T. Turletti, C. Huitema || turletti@sophia.inria.fr, huitema@bellcore.com +# RFC2033 || J. Myers || +# RFC2034 || N. Freed || +# RFC2035 || L. Berc, W. Fenner, R. Frederick, S. McCanne || berc@pa.dec.com, fenner@cmf.nrl.navy.mil, frederick@parc.xerox.com, mccanne@ee.lbl.gov +# RFC2036 || G. Huston || +# RFC2037 || K. McCloghrie, A. Bierman || kzm@cisco.com, andy@yumaworks.com +# RFC2038 || D. Hoffman, G. Fernando, V. Goyal || gerard.fernando@eng.sun.com, goyal@precept.com, don.hoffman@eng.sun.com +# RFC2039 || C. Kalbfleisch || +# RFC2040 || R. Baldwin, R. Rivest || baldwin@rsa.com, rivest@theory.lcs.mit.edu +# RFC2041 || B. Noble, G. Nguyen, M. Satyanarayanan, R. Katz || bnoble@cs.cmu.edu, gnguyen@cs.berkeley.edu, satya@cs.cmu.edu, randy@cs.berkeley.edu +# RFC2042 || B. Manning || bmanning@isi.edu +# RFC2043 || A. Fuqua || fuqua@vnet.ibm.com +# RFC2044 || F. Yergeau || fyergeau@alis.com +# RFC2045 || N. Freed, N. Borenstein || ned@innosoft.com, nsb@nsb.fv.com, Greg.Vaudreuil@Octel.Com +# RFC2046 || N. Freed, N. Borenstein || ned@innosoft.com, nsb@nsb.fv.com, Greg.Vaudreuil@Octel.Com +# RFC2047 || K. Moore || moore@cs.utk.edu +# RFC2048 || N. Freed, J. Klensin, J. Postel || ned@innosoft.com, klensin@mci.net, Postel@ISI.EDU +# RFC2049 || N. Freed, N. Borenstein || ned@innosoft.com, nsb@nsb.fv.com, Greg.Vaudreuil@Octel.Com +# RFC2050 || K. Hubbard, M. Kosters, D. Conrad, D. Karrenberg, J. Postel || kimh@internic.net, markk@internic.net, davidc@APNIC.NET, dfk@RIPE.NET, Postel@ISI.EDU +# RFC2051 || M. Allen, B. Clouston, Z. Kielczewski, W. Kwan, B. Moore || mallen@hq.walldata.com, clouston@cisco.com, zbig@cisco.com, billk@jti.com, remoore@ralvm6.vnet.ibm.com +# RFC2052 || A. Gulbrandsen, P. Vixie || agulbra@troll.no, paul@vix.com +# RFC2053 || E. Der-Danieliantz || edd@acm.org +# RFC2054 || B. Callaghan || brent.callaghan@eng.sun.com +# RFC2055 || B. Callaghan || brent.callaghan@eng.sun.com +# RFC2056 || R. Denenberg, J. Kunze, D. Lynch || +# RFC2057 || S. Bradner || sob@harvard.edu +# RFC2058 || C. Rigney, A. Rubens, W. Simpson, S. Willens || cdr@livingston.com, acr@merit.edu, wsimpson@greendragon.com, steve@livingston.com +# RFC2059 || C. Rigney || cdr@livingston.com +# RFC2060 || M. Crispin || MRC@CAC.Washington.EDU +# RFC2061 || M. Crispin || MRC@CAC.Washington.EDU +# RFC2062 || M. Crispin || MRC@CAC.Washington.EDU +# RFC2063 || N. Brownlee, C. Mills, G. Ruth || cmills@bbn.com, gruth@gte.com +# RFC2064 || N. Brownlee || +# RFC2065 || D. Eastlake 3rd, C. Kaufman || dee@cybercash.com, charlie_kaufman@iris.com +# RFC2066 || R. Gellens || Randy@MV.Unisys.Com +# RFC2067 || J. Renwick || jkr@NetStar.com +# RFC2068 || R. Fielding, J. Gettys, J. Mogul, H. Frystyk, T. Berners-Lee || fielding@ics.uci.edu, jg@w3.org, mogul@wrl.dec.com, frystyk@w3.org, timbl@w3.org +# RFC2069 || J. Franks, P. Hallam-Baker, J. Hostetler, P. Leach, A. Luotonen, E. Sink, L. Stewart || john@math.nwu.edu, hallam@w3.org, jeff@spyglass.com, paulle@microsoft.com, luotonen@netscape.com, eric@spyglass.com, stewart@OpenMarket.com +# RFC2070 || F. Yergeau, G. Nicol, G. Adams, M. Duerst || fyergeau@alis.com, gtn@ebt.com, glenn@spyglass.com, mduerst@ifi.unizh.ch +# RFC2071 || P. Ferguson, H. Berkowitz || pferguso@cisco.com, hcb@clark.net +# RFC2072 || H. Berkowitz || hcb@clark.net +# RFC2073 || Y. Rekhter, P. Lothberg, R. Hinden, S. Deering, J. Postel || bob.hinden@gmail.com +# RFC2074 || A. Bierman, R. Iddon || andy@yumaworks.com, robin_iddon@3mail.3com.com +# RFC2075 || C. Partridge || craig@bbn.com +# RFC2076 || J. Palme || +# RFC2077 || S. Nelson, C. Parks, Mitra || +# RFC2078 || J. Linn || John.Linn@ov.com +# RFC2079 || M. Smith || mcs@netscape.com +# RFC2080 || G. Malkin, R. Minnear || gmalkin@Xylogics.COM, minnear@ipsilon.com +# RFC2081 || G. Malkin || gmalkin@xylogics.com +# RFC2082 || F. Baker, R. Atkinson || rja@cisco.com +# RFC2083 || T. Boutell || boutell@boutell.com +# RFC2084 || G. Bossert, S. Cooper, W. Drummond || bossert@corp.sgi.com, sc@corp.sgi.com, drummond@ieee.org +# RFC2085 || M. Oehler, R. Glenn || mjo@tycho.ncsc.mil, rob.glenn@nist.gov +# RFC2086 || J. Myers || +# RFC2087 || J. Myers || +# RFC2088 || J. Myers || +# RFC2089 || B. Wijnen, D. Levi || +# RFC2090 || A. Emberson || tom@lanworks.com +# RFC2091 || G. Meyer, S. Sherry || +# RFC2092 || S. Sherry, G. Meyer || +# RFC2093 || H. Harney, C. Muckenhirn || +# RFC2094 || H. Harney, C. Muckenhirn || +# RFC2095 || J. Klensin, R. Catoe, P. Krumviede || klensin@mci.net, randy@mci.net, paul@mci.net +# RFC2096 || F. Baker || fred@cisco.com +# RFC2097 || G. Pall || gurdeep@microsoft.com +# RFC2098 || Y. Katsube, K. Nagami, H. Esaki || +# RFC2099 || J. Elliott || elliott@isi.edu +# RFC2100 || J. Ashworth || jra@scfn.thpl.lib.fl.us +# RFC2101 || B. Carpenter, J. Crowcroft, Y. Rekhter || brian@dxcoms.cern.ch, j.crowcroft@cs.ucl.ac.uk, yakov@cisco.com +# RFC2102 || R. Ramanathan || ramanath@bbn.com +# RFC2103 || R. Ramanathan || +# RFC2104 || H. Krawczyk, M. Bellare, R. Canetti || hugo@watson.ibm.com, mihir@cs.ucsd.edu, canetti@watson.ibm.com +# RFC2105 || Y. Rekhter, B. Davie, D. Katz, E. Rosen, G. Swallow || yakov@cisco.com, bsd@cisco.com, dkatz@cisco.com, erosen@cisco.com, swallow@cisco.com +# RFC2106 || S. Chiang, J. Lee, H. Yasuda || schiang@cisco.com, jolee@cisco.com, yasuda@eme068.cow.melco.co.jp +# RFC2107 || K. Hamzeh || kory@ascend.com +# RFC2108 || K. de Graaf, D. Romascanu, D. McMaster, K. McCloghrie || kdegraaf@isd.3com.com, dromasca@gmail.com , mcmaster@cisco.com, kzm@cisco.com +# RFC2109 || D. Kristol, L. Montulli || +# RFC2110 || J. Palme, A. Hopmann || +# RFC2111 || E. Levinson || +# RFC2112 || E. Levinson || +# RFC2113 || D. Katz || +# RFC2114 || S. Chiang, J. Lee, H. Yasuda || schiang@cisco.com, jolee@cisco.com, yasuda@eme068.cow.melco.co.jp +# RFC2115 || C. Brown, F. Baker || +# RFC2116 || C. Apple, K. Rossen || +# RFC2117 || D. Estrin, D. Farinacci, A. Helmy, D. Thaler, S. Deering, M. Handley, V. Jacobson, C. Liu, P. Sharma, L. Wei || +# RFC2118 || G. Pall || +# RFC2119 || S. Bradner || +# RFC2120 || D. Chadwick || +# RFC2121 || G. Armitage || gja@thumper.bellcore.com +# RFC2122 || D. Mavrakis, H. Layec, K. Kartmann || +# RFC2123 || N. Brownlee || +# RFC2124 || P. Amsden, J. Amweg, P. Calato, S. Bensley, G. Lyons || amsden@ctron.com, amsden@ctron.com, amsden@ctron.com, amsden@ctron.com, amsden@ctron.com +# RFC2125 || C. Richards, K. Smith || +# RFC2126 || Y. Pouffary, A. Young || pouffary@taec.enet.dec.com, A.Young@isode.com +# RFC2127 || G. Roeck, Ed. || groeck@cisco.com +# RFC2128 || G. Roeck, Ed. || groeck@cisco.com +# RFC2129 || K. Nagami, Y. Katsube, Y. Shobatake, A. Mogi, S. Matsuzawa, T. Jinmei, H. Esaki || +# RFC2130 || C. Weider, C. Preston, K. Simonsen, H. Alvestrand, R. Atkinson, M. Crispin, P. Svanberg || cweider@microsoft.com, cecilia@well.com, Keld@dkuug.dk, Harald.T.Alvestrand@uninett.no, rja@cisco.com, mrc@cac.washington.edu, psv@nada.kth.se +# RFC2131 || R. Droms || droms@bucknell.edu +# RFC2132 || S. Alexander, R. Droms || sca@engr.sgi.com, droms@bucknell.edu +# RFC2133 || R. Gilligan, S. Thomson, J. Bound, W. Stevens || gilligan@freegate.net, set@thumper.bellcore.com, rstevens@kohala.com +# RFC2134 || ISOC Board of Trustees || +# RFC2135 || ISOC Board of Trustees || +# RFC2136 || P. Vixie, Ed., S. Thomson, Y. Rekhter, J. Bound || yakov@cisco.com, set@thumper.bellcore.com, bound@zk3.dec.com, paul@vix.com +# RFC2137 || D. Eastlake 3rd || dee@cybercash.com +# RFC2138 || C. Rigney, A. Rubens, W. Simpson, S. Willens || cdr@livingston.com, acr@merit.edu, wsimpson@greendragon.com, steve@livingston.com +# RFC2139 || C. Rigney || cdr@livingston.com +# RFC2140 || J. Touch || +# RFC2141 || R. Moats || +# RFC2142 || D. Crocker || +# RFC2143 || B. Elliston || +# RFC2144 || C. Adams || +# RFC2145 || J. C. Mogul, R. Fielding, J. Gettys, H. Frystyk || +# RFC2146 || Federal Networking Council || execdir@fnc.gov +# RFC2147 || D. Borman || +# RFC2148 || H. Alvestrand, P. Jurg || +# RFC2149 || R. Talpade, M. Ammar || +# RFC2150 || J. Max, W. Stickle || jlm@rainfarm.com, wls@rainfarm.com +# RFC2151 || G. Kessler, S. Shepard || +# RFC2152 || D. Goldsmith, M. Davis || goldsmith@apple.com, mark_davis@taligent.com +# RFC2153 || W. Simpson || +# RFC2154 || S. Murphy, M. Badger, B. Wellington || +# RFC2155 || B. Clouston, B. Moore || +# RFC2156 || S. Kille || +# RFC2157 || H. Alvestrand || Harald.T.Alvestrand@uninett.no +# RFC2158 || H. Alvestrand || Harald.T.Alvestrand@uninett.no +# RFC2159 || H. Alvestrand || Harald.T.Alvestrand@uninett.no +# RFC2160 || H. Alvestrand || Harald.T.Alvestrand@uninett.no +# RFC2161 || H. Alvestrand || Harald.T.Alvestrand@uninett.no +# RFC2162 || C. Allocchio || Claudio.Allocchio@elettra.Trieste.it +# RFC2163 || C. Allocchio || +# RFC2164 || S. Kille || S.Kille@ISODE.COM +# RFC2165 || J. Veizades, E. Guttman, C. Perkins, S. Kaplan || cperkins@Corp.sun.com +# RFC2166 || D. Bryant, P. Brittain || +# RFC2167 || S. Williamson, M. Kosters, D. Blacka, J. Singh, K. Zeilstra || +# RFC2168 || R. Daniel, M. Mealling || +# RFC2169 || R. Daniel || +# RFC2170 || W. Almesberger, J. Le Boudec, P. Oechslin || +# RFC2171 || K. Murakami, M. Maruyama || +# RFC2172 || M. Maruyama, K. Murakami || +# RFC2173 || K. Murakami, M. Maruyama || +# RFC2174 || K. Murakami, M. Maruyama || +# RFC2175 || K. Murakami, M. Maruyama || +# RFC2176 || K. Murakami, M. Maruyama || +# RFC2177 || B. Leiba || +# RFC2178 || J. Moy || jmoy@casc.com +# RFC2179 || A. Gwinn || allen@mail.cox.smu.edu, ssh@wwsi.com +# RFC2180 || M. Gahrns || mikega@microsoft.com +# RFC2181 || R. Elz, R. Bush || kre@munnari.OZ.AU, randy@psg.com +# RFC2182 || R. Elz, R. Bush, S. Bradner, M. Patton || kre@munnari.OZ.AU, randy@psg.com, sob@harvard.edu, MAP@POBOX.COM +# RFC2183 || R. Troost, S. Dorner, K. Moore, Ed. || rens@century.com, sdorner@qualcomm.com +# RFC2184 || N. Freed, K. Moore || +# RFC2185 || R. Callon, D. Haskin || +# RFC2186 || D. Wessels, K. Claffy || wessels@nlanr.net, kc@nlanr.net +# RFC2187 || D. Wessels, K. Claffy || wessels@nlanr.net, kc@nlanr.net +# RFC2188 || M. Banan, M. Taylor, J. Cheng || +# RFC2189 || A. Ballardie || +# RFC2190 || C. Zhu || czhu@ibeam.intel.com +# RFC2191 || G. Armitage || gja@dnrc.bell-labs.com +# RFC2192 || C. Newman || chris.newman@innosoft.com +# RFC2193 || M. Gahrns || mikega@microsoft.com +# RFC2194 || B. Aboba, J. Lu, J. Alsop, J. Ding, W. Wang || bernarda@microsoft.com, juanlu@aimnet.net, jalsop@ipass.com, ding@bjai.asiainfo.com, weiwang@merit.edu +# RFC2195 || J. Klensin, R. Catoe, P. Krumviede || klensin@mci.net, randy@mci.net, paul@mci.net +# RFC2196 || B. Fraser || +# RFC2197 || N. Freed || ned.freed@innosoft.com +# RFC2198 || C. Perkins, I. Kouvelas, O. Hodson, V. Hardman, M. Handley, J.C. Bolot, A. Vega-Garcia, S. Fosse-Parisis || mjh@isi.edu +# RFC2199 || A. Ramos || ramos@isi.edu +# RFC2200 || J. Postel || +# RFC2201 || A. Ballardie || +# RFC2202 || P. Cheng, R. Glenn || pau@watson.ibm.com, rob.glenn@nist.gov +# RFC2203 || M. Eisler, A. Chiu, L. Ling || mre@eng.sun.com, hacker@eng.sun.com, lling@eng.sun.com +# RFC2204 || D. Nash || dnash@ford.com +# RFC2205 || R. Braden, Ed., L. Zhang, S. Berson, S. Herzog, S. Jamin || Braden@ISI.EDU, lixia@cs.ucla.edu, Berson@ISI.EDU, Herzog@WATSON.IBM.COM, jamin@EECS.UMICH.EDU +# RFC2206 || F. Baker, J. Krawczyk, A. Sastry || fred@cisco.com, jjk@tiac.net, arun@cisco.com +# RFC2207 || L. Berger, T. O'Malley || timo@bbn.com +# RFC2208 || A. Mankin, Ed., F. Baker, B. Braden, S. Bradner, M. O'Dell, A. Romanow, A. Weinrib, L. Zhang || aweinrib@ibeam.intel.com, braden@isi.edu, lixia@cs.ucla.edu, allyn@eng.sun.com, mo@uu.net +# RFC2209 || R. Braden, L. Zhang || Braden@ISI.EDU, lixia@cs.ucla.edu +# RFC2210 || J. Wroclawski || jtw@lcs.mit.edu +# RFC2211 || J. Wroclawski || jtw@lcs.mit.edu +# RFC2212 || S. Shenker, C. Partridge, R. Guerin || shenker@parc.xerox.com, craig@bbn.com, guerin@watson.ibm.com +# RFC2213 || F. Baker, J. Krawczyk, A. Sastry || fred@cisco.com, jjk@tiac.net, arun@cisco.com +# RFC2214 || F. Baker, J. Krawczyk, A. Sastry || fred@cisco.com, jjk@tiac.net, arun@cisco.com +# RFC2215 || S. Shenker, J. Wroclawski || shenker@parc.xerox.com, jtw@lcs.mit.edu +# RFC2216 || S. Shenker, J. Wroclawski || shenker@parc.xerox.com, jtw@lcs.mit.edu +# RFC2217 || G. Clark || glenc@cisco.com +# RFC2218 || T. Genovese, B. Jennings || TonyG@Microsoft.com, jennings@sandia.gov +# RFC2219 || M. Hamilton, R. Wright || m.t.hamilton@lut.ac.uk, wright@lbl.gov +# RFC2220 || R. Guenther || rgue@loc.gov +# RFC2221 || M. Gahrns || mikega@microsoft.com +# RFC2222 || J. Myers || jgmyers@netscape.com +# RFC2223 || J. Postel, J. Reynolds || Postel@ISI.EDU, jkrey@isi.edu, dwaitzman@BBN.COM +# RFC2224 || B. Callaghan || brent.callaghan@eng.sun.com +# RFC2225 || M. Laubach, J. Halpern || +# RFC2226 || T. Smith, G. Armitage || tjsmith@vnet.ibm.com, gja@lucent.com +# RFC2227 || J. Mogul, P. Leach || mogul@wrl.dec.com, paulle@microsoft.com +# RFC2228 || M. Horowitz, S. Lunt || marc@cygnus.com +# RFC2229 || R. Faith, B. Martin || faith@cs.unc.edu, bamartin@miranda.org +# RFC2230 || R. Atkinson || +# RFC2231 || N. Freed, K. Moore || ned.freed@innosoft.com, moore@cs.utk.edu +# RFC2232 || B. Clouston, Ed., B. Moore, Ed. || clouston@cisco.com, remoore@ralvm6.vnet.ibm.com +# RFC2233 || K. McCloghrie, F. Kastenholz || kzm@cisco.com, kasten@ftp.com +# RFC2234 || D. Crocker, Ed., P. Overell || dcrocker@bbiw.net, paul@bayleaf.org.uk +# RFC2235 || R. Zakon || zakon@info.isoc.org +# RFC2236 || W. Fenner || fenner@parc.xerox.com +# RFC2237 || K. Tamaru || kenzat@microsoft.com +# RFC2238 || B. Clouston, Ed., B. Moore, Ed. || clouston@cisco.com, remoore@ralvm6.vnet.ibm.com +# RFC2239 || K. de Graaf, D. Romascanu, D. McMaster, K. McCloghrie, S. Roberts || kdegraaf@isd.3com.com, dromasca@gmail.com , mcmaster@cisco.com, kzm@cisco.com, sroberts@farallon.com +# RFC2240 || O. Vaughan || owain@vaughan.com +# RFC2241 || D. Provan || donp@Novell.Com +# RFC2242 || R. Droms, K. Fong || +# RFC2243 || C. Metz || +# RFC2244 || C. Newman, J. G. Myers || +# RFC2245 || C. Newman || +# RFC2246 || T. Dierks, C. Allen || tdierks@certicom.com, pck@netcom.com, relyea@netscape.com, jar@netscape.com, msabin@netcom.com, dansimon@microsoft.com, tomw@netscape.com, hugo@watson.ibm.com +# RFC2247 || S. Kille, M. Wahl, A. Grimstad, R. Huber, S. Sataluri || S.Kille@ISODE.COM, M.Wahl@critical-angle.com, alg@att.com, rvh@att.com, sri@att.com +# RFC2248 || N. Freed, S. Kille || ned.freed@innosoft.com, S.Kille@isode.com +# RFC2249 || N. Freed, S. Kille || ned.freed@innosoft.com, S.Kille@isode.com +# RFC2250 || D. Hoffman, G. Fernando, V. Goyal, M. Civanlar || gerard.fernando@eng.sun.com, goyal@precept.com, don.hoffman@eng.sun.com, civanlar@research.att.com +# RFC2251 || M. Wahl, T. Howes, S. Kille || M.Wahl@critical-angle.com, howes@netscape.com, S.Kille@isode.com +# RFC2252 || M. Wahl, A. Coulbeck, T. Howes, S. Kille || M.Wahl@critical-angle.com, A.Coulbeck@isode.com, howes@netscape.com, S.Kille@isode.com +# RFC2253 || M. Wahl, S. Kille, T. Howes || M.Wahl@critical-angle.com, S.Kille@ISODE.COM, howes@netscape.com +# RFC2254 || T. Howes || howes@netscape.com +# RFC2255 || T. Howes, M. Smith || howes@netscape.com, mcs@netscape.com +# RFC2256 || M. Wahl || M.Wahl@critical-angle.com +# RFC2257 || M. Daniele, B. Wijnen, D. Francisco || daniele@zk3.dec.com, wijnen@vnet.ibm.com, dfrancis@cisco.com +# RFC2258 || J. Ordille || joann@bell-labs.com +# RFC2259 || J. Elliott, J. Ordille || jim@apocalypse.org, joann@bell-labs.com +# RFC2260 || T. Bates, Y. Rekhter || tbates@cisco.com, yakov@cisco.com +# RFC2261 || D. Harrington, R. Presuhn, B. Wijnen || +# RFC2262 || J. Case, D. Harrington, R. Presuhn, B. Wijnen || +# RFC2263 || D. Levi, P. Meyer, B. Stewart || +# RFC2264 || U. Blumenthal, B. Wijnen || +# RFC2265 || B. Wijnen, R. Presuhn, K. McCloghrie || +# RFC2266 || J. Flick || +# RFC2267 || P. Ferguson, D. Senie || ferguson@cisco.com, dts@senie.com +# RFC2268 || R. Rivest || rsa-labs@rsa.com +# RFC2269 || G. Armitage || gja@lucent.com +# RFC2270 || J. Stewart, T. Bates, R. Chandra, E. Chen || jstewart@isi.edu, tbates@cisco.com, rchandra@cisco.com, enkechen@cisco.com +# RFC2271 || D. Harrington, R. Presuhn, B. Wijnen || +# RFC2272 || J. Case, D. Harrington, R. Presuhn, B. Wijnen || +# RFC2273 || D. Levi, P. Meyer, B. Stewart || +# RFC2274 || U. Blumenthal, B. Wijnen || +# RFC2275 || B. Wijnen, R. Presuhn, K. McCloghrie || +# RFC2276 || K. Sollins || sollins@lcs.mit.edu +# RFC2277 || H. Alvestrand || Harald.T.Alvestrand@uninett.no +# RFC2278 || N. Freed, J. Postel || ned.freed@innosoft.com, Postel@ISI.EDU +# RFC2279 || F. Yergeau || fyergeau@alis.com +# RFC2280 || C. Alaettinoglu, T. Bates, E. Gerich, D. Karrenberg, D. Meyer, M. Terpstra, C. Villamizar || cengiz@isi.edu, tbates@cisco.com, epg@home.net, dfk@ripe.net, meyer@antc.uoregon.edu, marten@BayNetworks.com, curtis@ans.net +# RFC2281 || T. Li, B. Cole, P. Morton, D. Li || tli@juniper.net, cole@juniper.net, pmorton@cisco.com, dawnli@cisco.com +# RFC2282 || J. Galvin || +# RFC2283 || T. Bates, R. Chandra, D. Katz, Y. Rekhter || +# RFC2284 || L. Blunk, J. Vollbrecht || ljb@merit.edu, jrv@merit.edu +# RFC2285 || R. Mandeville || bob.mandeville@eunet.fr +# RFC2286 || J. Kapp || skapp@reapertech.com +# RFC2287 || C. Krupczak, J. Saperia || cheryl@empiretech.com +# RFC2288 || C. Lynch, C. Preston, R. Daniel || cliff@cni.org, cecilia@well.com, rdaniel@acl.lanl.gov +# RFC2289 || N. Haller, C. Metz, P. Nesser, M. Straw || +# RFC2290 || J. Solomon, S. Glass || solomon@comm.mot.com, glass@ftp.com +# RFC2291 || J. Slein, F. Vitali, E. Whitehead, D. Durand || slein@wrc.xerox.com, fabio@cs.unibo.it, ejw@ics.uci.edu, dgd@cs.bu.edu +# RFC2292 || W. Stevens, M. Thomas || rstevens@kohala.com, matt.thomas@altavista-software.com +# RFC2293 || S. Kille || S.Kille@ISODE.COM +# RFC2294 || S. Kille || S.Kille@ISODE.COM +# RFC2295 || K. Holtman, A. Mutz || koen@win.tue.nl, mutz@hpl.hp.com +# RFC2296 || K. Holtman, A. Mutz || koen@win.tue.nl, mutz@hpl.hp.com +# RFC2297 || P. Newman, W. Edwards, R. Hinden, E. Hoffman, F. Ching Liaw, T. Lyon, G. Minshall || bob.hinden@gmail.com +# RFC2298 || R. Fajman || raf@cu.nih.gov +# RFC2299 || A. Ramos || ramos@isi.edu +# RFC2300 || J. Postel || Postel@ISI.EDU +# RFC2301 || L. McIntyre, S. Zilles, R. Buckley, D. Venable, G. Parsons, J. Rafferty || +# RFC2302 || G. Parsons, J. Rafferty, S. Zilles || +# RFC2303 || C. Allocchio || +# RFC2304 || C. Allocchio || +# RFC2305 || K. Toyoda, H. Ohno, J. Murai, D. Wing || +# RFC2306 || G. Parsons, J. Rafferty || +# RFC2307 || L. Howard || lukeh@xedoc.com +# RFC2308 || M. Andrews || Mark.Andrews@cmis.csiro.au +# RFC2309 || B. Braden, D. Clark, J. Crowcroft, B. Davie, S. Deering, D. Estrin, S. Floyd, V. Jacobson, G. Minshall, C. Partridge, L. Peterson, K. Ramakrishnan, S. Shenker, J. Wroclawski, L. Zhang || Braden@ISI.EDU, DDC@lcs.mit.edu, Jon.Crowcroft@cs.ucl.ac.uk, bdavie@cisco.com, deering@cisco.com, Estrin@usc.edu, Floyd@ee.lbl.gov, Van@ee.lbl.gov, Minshall@fiberlane.com, craig@bbn.com, LLP@cs.arizona.edu, KKRama@research.att.com, Shenker@parc.xerox.com, JTW@lcs.mit.edu, Lixia@cs.ucla.edu +# RFC2310 || K. Holtman || koen@win.tue.nl +# RFC2311 || S. Dusse, P. Hoffman, B. Ramsdell, L. Lundblade, L. Repka || spock@rsa.com, phoffman@imc.org, blaker@deming.com, lgl@qualcomm.com, repka@netscape.com +# RFC2312 || S. Dusse, P. Hoffman, B. Ramsdell, J. Weinstein || spock@rsa.com, phoffman@imc.org, blaker@deming.com, jsw@netscape.com +# RFC2313 || B. Kaliski || burt@rsa.com +# RFC2314 || B. Kaliski || burt@rsa.com +# RFC2315 || B. Kaliski || burt@rsa.com +# RFC2316 || S. Bellovin || +# RFC2317 || H. Eidnes, G. de Groot, P. Vixie || Havard.Eidnes@runit.sintef.no, GeertJan.deGroot@bsdi.com, paul@vix.com +# RFC2318 || H. Lie, B. Bos, C. Lilley || howcome@w3.org, bert@w3.org, chris@w3.org +# RFC2319 || KOI8-U Working Group || +# RFC2320 || M. Greene, J. Luciani, K. White, T. Kuo || maria@xedia.com, luciani@baynetworks.com, kennethw@vnet.ibm.com, ted_kuo@Baynetworks.com +# RFC2321 || A. Bressen || bressen@leftbank.com +# RFC2322 || K. van den Hout, A. Koopal, R. van Mook || koos@cetis.hvu.nl, andre@NL.net, remco@sateh.com +# RFC2323 || A. Ramos || ramos@isi.edu +# RFC2324 || L. Masinter || masinter@parc.xerox.com +# RFC2325 || M. Slavitch || slavitch@loran.com +# RFC2326 || H. Schulzrinne, A. Rao, R. Lanphier || schulzrinne@cs.columbia.edu, anup@netscape.com, robla@real.com +# RFC2327 || M. Handley, V. Jacobson || +# RFC2328 || J. Moy || jmoy@casc.com +# RFC2329 || J. Moy || jmoy@casc.com +# RFC2330 || V. Paxson, G. Almes, J. Mahdavi, M. Mathis || vern@ee.lbl.gov, almes@advanced.org, mahdavi@psc.edu, mathis@psc.edu +# RFC2331 || M. Maher || maher@isi.edu +# RFC2332 || J. Luciani, D. Katz, D. Piscitello, B. Cole, N. Doraswamy || dkatz@cisco.com, luciani@baynetworks.com, bcole@jnx.com, naganand@baynetworks.com +# RFC2333 || D. Cansever || dcansever@gte.com +# RFC2334 || J. Luciani, G. Armitage, J. Halpern, N. Doraswamy || luciani@baynetworks.com, gja@lucent.com, jhalpern@Newbridge.COM, naganand@baynetworks.com +# RFC2335 || J. Luciani || luciani@baynetworks.com +# RFC2336 || J. Luciani || +# RFC2337 || D. Farinacci, D. Meyer, Y. Rekhter || +# RFC2338 || S. Knight, D. Weaver, D. Whipple, R. Hinden, D. Mitzel, P. Hunt, P. Higginson, M. Shand, A. Lindem || Steven.Knight@ascend.com, Doug.Weaver@ascend.com, dwhipple@microsoft.com, bob.hinden@gmail.com, mitzel@iprg.nokia.com, hunt@iprg.nokia.com, higginson@mail.dec.com, shand@mail.dec.com +# RFC2339 || The Internet Society, Sun Microsystems || +# RFC2340 || B. Jamoussi, D. Jamieson, D. Williston, S. Gabe || jamoussi@Nortel.ca, djamies@Nortel.ca, danwil@Nortel.ca, spgabe@Nortel.ca +# RFC2341 || A. Valencia, M. Littlewood, T. Kolar || tkolar@cisco.com, littlewo@cisco.com, valencia@cisco.com +# RFC2342 || M. Gahrns, C. Newman || mikega@microsoft.com, chris.newman@innosoft.com +# RFC2343 || M. Civanlar, G. Cash, B. Haskell || civanlar@research.att.com, glenn@research.att.com, bgh@research.att.com +# RFC2344 || G. Montenegro, Ed. || +# RFC2345 || J. Klensin, T. Wolf, G. Oglesby || klensin@mci.net, ted@usa.net, gary@mci.net +# RFC2346 || J. Palme || jpalme@dsv.su.se +# RFC2347 || G. Malkin, A. Harkin || gmalkin@baynetworks.com, ash@cup.hp.com +# RFC2348 || G. Malkin, A. Harkin || gmalkin@baynetworks.com, ash@cup.hp.com +# RFC2349 || G. Malkin, A. Harkin || gmalkin@baynetworks.com, ash@cup.hp.com +# RFC2350 || N. Brownlee, E. Guttman || n.brownlee@auckland.ac.nz, Erik.Guttman@sun.com +# RFC2351 || A. Robert || arobert@par1.par.sita.int +# RFC2352 || O. Vaughan || owain@vaughan.com +# RFC2353 || G. Dudley || dudleyg@us.ibm.com +# RFC2354 || C. Perkins, O. Hodson || c.perkins@cs.ucl.ac.uk, o.hodson@cs.ucl.ac.uk +# RFC2355 || B. Kelly || kellywh@mail.auburn.edu +# RFC2356 || G. Montenegro, V. Gupta || gabriel.montenegro@Eng.Sun.COM, vipul.gupta@Eng.Sun.COM +# RFC2357 || A. Mankin, A. Romanow, S. Bradner, V. Paxson || mankin@east.isi.edu, allyn@mci.net, sob@harvard.edu, vern@ee.lbl.gov +# RFC2358 || J. Flick, J. Johnson || johnf@hprnd.rose.hp.com, jeff@redbacknetworks.com +# RFC2359 || J. Myers || jgmyers@netscape.com +# RFC2360 || G. Scott || +# RFC2361 || E. Fleischman || ericfl@microsoft.com, Eric.Fleischman@PSS.Boeing.com +# RFC2362 || D. Estrin, D. Farinacci, A. Helmy, D. Thaler, S. Deering, M. Handley, V. Jacobson, C. Liu, P. Sharma, L. Wei || estrin@usc.edu, dino@cisco.com, ahelmy@catarina.usc.edu, thalerd@eecs.umich.edu, deering@parc.xerox.com, m.handley@cs.ucl.ac.uk, van@ee.lbl.gov, charley@catarina.usc.edu, puneet@catarina.usc.edu, lwei@cisco.com +# RFC2363 || G. Gross, M. Kaycee, A. Li, A. Malis, J. Stephens || gmgross@lucent.com, mjk@nj.paradyne.com, alin@shastanets.com, malis@ascend.com, john@cayman.com +# RFC2364 || G. Gross, M. Kaycee, A. Li, A. Malis, J. Stephens || gmgross@lucent.com, mjk@nj.paradyne.com, alin@shastanets.com, malis@ascend.com, john@cayman.com +# RFC2365 || D. Meyer || dmm@cisco.com +# RFC2366 || C. Chung, M. Greene || cchung@tieo.saic.com +# RFC2367 || D. McDonald, C. Metz, B. Phan || danmcd@eng.sun.com, cmetz@inner.net, phan@itd.nrl.navy.mil +# RFC2368 || P. Hoffman, L. Masinter, J. Zawinski || +# RFC2369 || G. Neufeld, J. Baer || +# RFC2370 || R. Coltun || +# RFC2371 || J. Lyon, K. Evans, J. Klein || JimLyon@Microsoft.Com, Keith.Evans@Tandem.Com, Johannes.Klein@Tandem.Com +# RFC2372 || K. Evans, J. Klein, J. Lyon || Keith.Evans@Tandem.Com, Johannes.Klein@Tandem.Com, JimLyon@Microsoft.Com +# RFC2373 || R. Hinden, S. Deering || bob.hinden@gmail.com +# RFC2374 || R. Hinden, M. O'Dell, S. Deering || bob.hinden@gmail.com, mo@uunet.uu.net, deering@cisco.com +# RFC2375 || R. Hinden, S. Deering || bob.hinden@gmail.com, deering@cisco.com +# RFC2376 || E. Whitehead, M. Murata || +# RFC2377 || A. Grimstad, R. Huber, S. Sataluri, M. Wahl || alg@att.com, rvh@att.com, srs@lucent.com, M.Wahl@critical-angle.com +# RFC2378 || R. Hedberg, P. Pomes || Roland.Hedberg@umdac.umu.se, ppomes@qualcomm.com +# RFC2379 || L. Berger || lberger@fore.com +# RFC2380 || L. Berger || lberger@fore.com +# RFC2381 || M. Garrett, M. Borden || mwg@bellcore.com, mborden@baynetworks.com +# RFC2382 || E. Crawley, Ed., L. Berger, S. Berson, F. Baker, M. Borden, J. Krawczyk || esc@argon.com, lberger@fore.com, berson@isi.edu, fred@cisco.com, mborden@baynetworks.com, jj@arrowpoint.com +# RFC2383 || M. Suzuki || suzuki@nal.ecl.net +# RFC2384 || R. Gellens || Randy@Qualcomm.Com +# RFC2385 || A. Heffernan || ahh@cisco.com +# RFC2386 || E. Crawley, R. Nair, B. Rajagopalan, H. Sandick || +# RFC2387 || E. Levinson || XIson@cnj.digex.com +# RFC2388 || L. Masinter || masinter@parc.xerox.com +# RFC2389 || P. Hethmon, R. Elz || +# RFC2390 || T. Bradley, C. Brown, A. Malis || tbradley@avici.com, cbrown@juno.com, malis@ascend.com +# RFC2391 || P. Srisuresh, D. Gan || suresh@ra.lucent.com, dhg@juniper.net +# RFC2392 || E. Levinson || XIson@cnj.digex.net +# RFC2393 || A. Shacham, R. Monsour, R. Pereira, M. Thomas || shacham@cisco.com, rmonsour@hifn.com, rpereira@timestep.com, matt.thomas@altavista-software.com, naganand@baynetworks.com +# RFC2394 || R. Pereira || +# RFC2395 || R. Friend, R. Monsour || rfriend@hifn.com, rmonsour@hifn.com +# RFC2396 || T. Berners-Lee, R. Fielding, L. Masinter || timbl@w3.org, fielding@ics.uci.edu, masinter@parc.xerox.com +# RFC2397 || L. Masinter || +# RFC2398 || S. Parker, C. Schmechel || sparker@eng.sun.com, cschmec@eng.sun.com +# RFC2399 || A. Ramos || ramos@isi.edu +# RFC2400 || J. Postel, J. Reynolds || Postel@ISI.EDU, JKRey@ISI.EDU +# RFC2401 || S. Kent, R. Atkinson || +# RFC2402 || S. Kent, R. Atkinson || +# RFC2403 || C. Madson, R. Glenn || +# RFC2404 || C. Madson, R. Glenn || +# RFC2405 || C. Madson, N. Doraswamy || +# RFC2406 || S. Kent, R. Atkinson || +# RFC2407 || D. Piper || ddp@network-alchemy.com +# RFC2408 || D. Maughan, M. Schertler, M. Schneider, J. Turner || wdm@tycho.ncsc.mil, mss@tycho.ncsc.mil, mjs@securify.com, jeff.turner@raba.com +# RFC2409 || D. Harkins, D. Carrel || dharkins@cisco.com, carrel@ipsec.org +# RFC2410 || R. Glenn, S. Kent || +# RFC2411 || R. Thayer, N. Doraswamy, R. Glenn || naganand@baynetworks.com, rob.glenn@nist.gov +# RFC2412 || H. Orman || ho@darpa.mil +# RFC2413 || S. Weibel, J. Kunze, C. Lagoze, M. Wolf || weibel@oclc.org, jak@ckm.ucsf.edu, lagoze@cs.cornell.edu, misha.wolf@reuters.com +# RFC2414 || M. Allman, S. Floyd, C. Partridge || mallman@lerc.nasa.gov, floyd@ee.lbl.gov, craig@bbn.com +# RFC2415 || K. Poduri, K. Nichols || kpoduri@Baynetworks.com, knichols@baynetworks.com +# RFC2416 || T. Shepard, C. Partridge || shep@alum.mit.edu, craig@bbn.com +# RFC2417 || C. Chung, M. Greene || chihschung@aol.com, maria@xedia.com +# RFC2418 || S. Bradner || +# RFC2419 || K. Sklower, G. Meyer || sklower@CS.Berkeley.EDU +# RFC2420 || H. Kummert || kummert@nentec.de +# RFC2421 || G. Vaudreuil, G. Parsons || Glenn.Parsons@Nortel.ca, GregV@Lucent.Com +# RFC2422 || G. Vaudreuil, G. Parsons || Glenn.Parsons@Nortel.ca, GregV@Lucent.Com +# RFC2423 || G. Vaudreuil, G. Parsons || Glenn.Parsons@Nortel.ca, GregV@Lucent.Com +# RFC2424 || G. Vaudreuil, G. Parsons || Glenn.Parsons@Nortel.ca, GregV@Lucent.Com +# RFC2425 || T. Howes, M. Smith, F. Dawson || howes@netscape.com, mcs@netscape.com, frank_dawson@lotus.com +# RFC2426 || F. Dawson, T. Howes || +# RFC2427 || C. Brown, A. Malis || cbrown@juno.com, malis@ascend.com +# RFC2428 || M. Allman, S. Ostermann, C. Metz || mallman@lerc.nasa.gov, ostermann@cs.ohiou.edu, cmetz@inner.net +# RFC2429 || C. Bormann, L. Cline, G. Deisher, T. Gardos, C. Maciocco, D. Newell, J. Ott, G. Sullivan, S. Wenger, C. Zhu || +# RFC2430 || T. Li, Y. Rekhter || tli@juniper.net, yakov@cisco.com +# RFC2431 || D. Tynan || dtynan@claddagh.ie +# RFC2432 || K. Dubray || kdubray@ironbridgenetworks.com +# RFC2433 || G. Zorn, S. Cobb || glennz@microsoft.com, stevec@microsoft.com +# RFC2434 || T. Narten, H. Alvestrand || narten@raleigh.ibm.com, Harald@Alvestrand.no +# RFC2435 || L. Berc, W. Fenner, R. Frederick, S. McCanne, P. Stewart || berc@pa.dec.com, fenner@parc.xerox.com, frederick@parc.xerox.com, mccanne@cs.berkeley.edu, stewart@parc.xerox.com +# RFC2436 || R. Brett, S. Bradner, G. Parsons || rfbrett@nortel.ca, sob@harvard.edu, Glenn.Parsons@Nortel.ca +# RFC2437 || B. Kaliski, J. Staddon || burt@rsa.com, jstaddon@rsa.com +# RFC2438 || M. O'Dell, H. Alvestrand, B. Wijnen, S. Bradner || mo@uu.net, Harald.Alvestrand@maxware.no, wijnen@vnet.ibm.com, sob@harvard.edu +# RFC2439 || C. Villamizar, R. Chandra, R. Govindan || curtis@ans.net, rchandra@cisco.com, govindan@isi.edu +# RFC2440 || J. Callas, L. Donnerhacke, H. Finney, R. Thayer || +# RFC2441 || D. Cohen || cohen@myri.com +# RFC2442 || N. Freed, D. Newman, J. Belissent, M. Hoy || ned.freed@innosoft.com, dan.newman@innosoft.com, jacques.belissent@eng.sun.com +# RFC2443 || J. Luciani, A. Gallo || luciani@baynetworks.com, gallo@raleigh.ibm.com +# RFC2444 || C. Newman || chris.newman@innosoft.com +# RFC2445 || F. Dawson, D. Stenerson || +# RFC2446 || S. Silverberg, S. Mansour, F. Dawson, R. Hopson || +# RFC2447 || F. Dawson, S. Mansour, S. Silverberg || +# RFC2448 || M. Civanlar, G. Cash, B. Haskell || civanlar@research.att.com, glenn@research.att.com, bgh@research.att.com +# RFC2449 || R. Gellens, C. Newman, L. Lundblade || randy@qualcomm.com, chris.newman@innosoft.com, lgl@qualcomm.com +# RFC2450 || R. Hinden || bob.hinden@gmail.com +# RFC2451 || R. Pereira, R. Adams || +# RFC2452 || M. Daniele || daniele@zk3.dec.com +# RFC2453 || G. Malkin || gmalkin@baynetworks.com +# RFC2454 || M. Daniele || daniele@zk3.dec.com +# RFC2455 || B. Clouston, B. Moore || clouston@cisco.com, remoore@us.ibm.com +# RFC2456 || B. Clouston, B. Moore || clouston@cisco.com, remoore@us.ibm.com +# RFC2457 || B. Clouston, B. Moore || clouston@cisco.com, remoore@us.ibm.com +# RFC2458 || H. Lu, M. Krishnaswamy, L. Conroy, S. Bellovin, F. Burg, A. DeSimone, K. Tewani, P. Davidson, H. Schulzrinne, K. Vishwanathan || smb@research.att.com, fburg@hogpb.att.com, lwc@roke.co.uk, pauldav@nortel.ca, murali@bell-labs.com, hui-lan.lu@bell-labs.com, schulzrinne@cs.columbia.edu, tewani@att.com, kumar@isochrone.com +# RFC2459 || R. Housley, W. Ford, W. Polk, D. Solo || housley@spyrus.com, wford@verisign.com, wpolk@nist.gov, david.solo@citicorp.com +# RFC2460 || S. Deering, R. Hinden || deering@cisco.com, bob.hinden@gmail.com +# RFC2461 || T. Narten, E. Nordmark, W. Simpson || narten@raleigh.ibm.com, nordmark@sun.com, Bill.Simpson@um.cc.umich.edu +# RFC2462 || S. Thomson, T. Narten || +# RFC2463 || A. Conta, S. Deering || aconta@lucent.com, deering@cisco.com +# RFC2464 || M. Crawford || crawdad@fnal.gov +# RFC2465 || D. Haskin, S. Onishi || dhaskin@baynetworks.com, sonishi@baynetworks.com +# RFC2466 || D. Haskin, S. Onishi || dhaskin@baynetworks.com, sonishi@baynetworks.com +# RFC2467 || M. Crawford || crawdad@fnal.gov +# RFC2468 || V. Cerf || vcerf@mci.net +# RFC2469 || T. Narten, C. Burton || narten@raleigh.ibm.com, burton@rtp.vnet.ibm.com +# RFC2470 || M. Crawford, T. Narten, S. Thomas || crawdad@fnal.gov, narten@raleigh.ibm.com, stephen.thomas@transnexus.com +# RFC2471 || R. Hinden, R. Fink, J. Postel || bob.hinden@gmail.com, rlfink@lbl.gov +# RFC2472 || D. Haskin, E. Allen || dhaskin@baynetworks.com, eallen@baynetworks.com +# RFC2473 || A. Conta, S. Deering || aconta@lucent.com, deering@cisco.com +# RFC2474 || K. Nichols, S. Blake, F. Baker, D. Black || kmn@cisco.com, slblake@torrentnet.com, fred@cisco.com, black_david@emc.com +# RFC2475 || S. Blake, D. Black, M. Carlson, E. Davies, Z. Wang, W. Weiss || slblake@torrentnet.com, black_david@emc.com, mark.carlson@sun.com, elwynd@nortel.co.uk, zhwang@bell-labs.com, wweiss@lucent.com +# RFC2476 || R. Gellens, J. Klensin || Randy@Qualcomm.Com, klensin@mci.net +# RFC2477 || B. Aboba, G. Zorn || bernarda@microsoft.com, glennz@microsoft.com +# RFC2478 || E. Baize, D. Pinkas || +# RFC2479 || C. Adams || cadams@entrust.com +# RFC2480 || N. Freed || ned.freed@innosoft.com +# RFC2481 || K. Ramakrishnan, S. Floyd || +# RFC2482 || K. Whistler, G. Adams || kenw@sybase.com, glenn@spyglass.com +# RFC2483 || M. Mealling, R. Daniel || michaelm@rwhois.net, rdaniel@lanl.gov +# RFC2484 || G. Zorn || glennz@microsoft.com +# RFC2485 || S. Drach || drach@sun.com +# RFC2486 || B. Aboba, M. Beadles || bernarda@microsoft.com, mbeadles@wcom.net +# RFC2487 || P. Hoffman || phoffman@imc.org +# RFC2488 || M. Allman, D. Glover, L. Sanchez || mallman@lerc.nasa.gov, Daniel.R.Glover@lerc.nasa.gov, lsanchez@ir.bbn.com +# RFC2489 || R. Droms || droms@bucknell.edu +# RFC2490 || M. Pullen, R. Malghan, L. Lavu, G. Duan, J. Ma, H. Nah || mpullen@gmu.edu, rmalghan@bacon.gmu.edu, llavu@bacon.gmu.edu, gduan@us.oracle.com, jma@newbridge.com, hnah@bacon.gmu.edu +# RFC2491 || G. Armitage, P. Schulter, M. Jork, G. Harter || gja@lucent.com, paschulter@acm.org, jork@kar.dec.com, harter@zk3.dec.com +# RFC2492 || G. Armitage, P. Schulter, M. Jork || gja@lucent.com, paschulter@acm.org, jork@kar.dec.com +# RFC2493 || K. Tesink, Ed. || kaj@bellcore.com +# RFC2494 || D. Fowler, Ed. || davef@newbridge.com +# RFC2495 || D. Fowler, Ed. || davef@newbridge.com +# RFC2496 || D. Fowler, Ed. || davef@newbridge.com +# RFC2497 || I. Souvatzis || is@netbsd.org +# RFC2498 || J. Mahdavi, V. Paxson || mahdavi@psc.edu, vern@ee.lbl.gov +# RFC2499 || A. Ramos || ramos@isi.edu +# RFC2500 || J. Reynolds, R. Braden || +# RFC2501 || S. Corson, J. Macker || corson@isr.umd.edu, macker@itd.nrl.navy.mil +# RFC2502 || M. Pullen, M. Myjak, C. Bouwens || mpullen@gmu.edu, mmyjak@virtualworkshop.com, christina.bouwens@cpmx.mail.saic.com +# RFC2503 || R. Moulton, M. Needleman || ruth@muswell.demon.co.uk +# RFC2504 || E. Guttman, L. Leong, G. Malkin || erik.guttman@sun.com, lorna@colt.net, gmalkin@baynetworks.com +# RFC2505 || G. Lindberg || +# RFC2506 || K. Holtman, A. Mutz, T. Hardie || koen@win.tue.nl, andy_mutz@hp.com, hardie@equinix.com +# RFC2507 || M. Degermark, B. Nordgren, S. Pink || micke@sm.luth.se, bcn@lulea.trab.se, steve@sm.luth.se +# RFC2508 || S. Casner, V. Jacobson || casner@cisco.com, van@cisco.com +# RFC2509 || M. Engan, S. Casner, C. Bormann || engan@effnet.com, casner@cisco.com, cabo@tzi.org +# RFC2510 || C. Adams, S. Farrell || cadams@entrust.com, stephen.farrell@sse.ie +# RFC2511 || M. Myers, C. Adams, D. Solo, D. Kemp || mmyers@verisign.com, cadams@entrust.com, david.solo@citicorp.com, dpkemp@missi.ncsc.mil +# RFC2512 || K. McCloghrie, J. Heinanen, W. Greene, A. Prasad || kzm@cisco.com, jh@telia.fi, wedge.greene@mci.com, aprasad@cisco.com +# RFC2513 || K. McCloghrie, J. Heinanen, W. Greene, A. Prasad || kzm@cisco.com, jh@telia.fi, wedge.greene@mci.com, aprasad@cisco.com +# RFC2514 || M. Noto, E. Spiegel, K. Tesink || mspiegel@cisco.com, kaj@bellcore.com +# RFC2515 || K. Tesink, Ed || kaj@bellcore.com +# RFC2516 || L. Mamakos, K. Lidl, J. Evarts, D. Carrel, D. Simone, R. Wheeler || louie@uu.net, lidl@uu.net, jde@uu.net, carrel@RedBack.net, dan@RedBack.net, ross@routerware.com +# RFC2517 || R. Moats, R. Huber || jayhawk@att.com, rvh@att.com +# RFC2518 || Y. Goland, E. Whitehead, A. Faizi, S. Carter, D. Jensen || yarong@microsoft.com, ejw@ics.uci.edu, asad@netscape.com, srcarter@novell.com, dcjensen@novell.com +# RFC2519 || E. Chen, J. Stewart || enkechen@cisco.com, jstewart@juniper.net +# RFC2520 || J. Luciani, H. Suzuki, N. Doraswamy, D. Horton || luciani@baynetworks.com, hsuzuki@cisco.com, naganand@baynetworks.com, d.horton@citr.com.au +# RFC2521 || P. Karn, W. Simpson || +# RFC2522 || P. Karn, W. Simpson || +# RFC2523 || P. Karn, W. Simpson || +# RFC2524 || M. Banan || +# RFC2525 || V. Paxson, M. Allman, S. Dawson, W. Fenner, J. Griner, I. Heavens, K. Lahey, J. Semke, B. Volz || vern@aciri.org, sdawson@eecs.umich.edu, fenner@parc.xerox.com, jgriner@grc.nasa.gov, ian@spider.com, kml@nas.nasa.gov, semke@psc.edu, volz@process.com +# RFC2526 || D. Johnson, S. Deering || dbj@cs.cmu.edu, deering@cisco.com +# RFC2527 || S. Chokhani, W. Ford || +# RFC2528 || R. Housley, W. Polk || housley@spyrus.com, wpolk@nist.gov +# RFC2529 || B. Carpenter, C. Jung || brian@hursley.ibm.com, cmj@3Com.com +# RFC2530 || D. Wing || dwing-ietf@fuggles.com +# RFC2531 || G. Klyne, L. McIntyre || GK@ACM.ORG, Lloyd.McIntyre@pahv.xerox.com +# RFC2532 || L. Masinter, D. Wing || masinter@parc.xerox.com, dwing-ietf@fuggles.com +# RFC2533 || G. Klyne || GK@ACM.ORG +# RFC2534 || L. Masinter, D. Wing, A. Mutz, K. Holtman || masinter@parc.xerox.com, dwing-ietf@fuggles.com, koen@win.tue.nl +# RFC2535 || D. Eastlake 3rd || dee3@us.ibm.com +# RFC2536 || D. Eastlake 3rd || dee3@us.ibm.com +# RFC2537 || D. Eastlake 3rd || dee3@us.ibm.com +# RFC2538 || D. Eastlake 3rd, O. Gudmundsson || dee3@us.ibm.com, ogud@tislabs.com +# RFC2539 || D. Eastlake 3rd || dee3@us.ibm.com +# RFC2540 || D. Eastlake 3rd || dee3@us.ibm.com +# RFC2541 || D. Eastlake 3rd || dee3@us.ibm.com +# RFC2542 || L. Masinter || masinter@parc.xerox.com +# RFC2543 || M. Handley, H. Schulzrinne, E. Schooler, J. Rosenberg || +# RFC2544 || S. Bradner, J. McQuaid || +# RFC2545 || P. Marques, F. Dupont || +# RFC2546 || A. Durand, B. Buclin || Alain.Durand@imag.fr, Bertrand.Buclin@ch.att.com +# RFC2547 || E. Rosen, Y. Rekhter || erosen@cisco.com, yakov@cisco.com +# RFC2548 || G. Zorn || +# RFC2549 || D. Waitzman || djw@vineyard.net +# RFC2550 || S. Glassman, M. Manasse, J. Mogul || steveg@pa.dec.com, msm@pa.dec.com, mogul@pa.dec.com +# RFC2551 || S. Bradner || +# RFC2552 || M. Blinov, M. Bessonov, C. Clissmann || mch@net-cs.ucd.ie, mikeb@net-cs.ucd.ie, ciaranc@net-cs.ucd.ie +# RFC2553 || R. Gilligan, S. Thomson, J. Bound, W. Stevens || gilligan@freegate.com, set@thumper.bellcore.com, bound@zk3.dec.com, rstevens@kohala.com +# RFC2554 || J. Myers || jgmyers@netscape.com +# RFC2555 || RFC Editor, et al. || braden@isi.edu, jkrey@isi.edu, crocker@mbl.edu, vcerf@mci.net, feinler@juno.com, celeste@isi.edu +# RFC2556 || S. Bradner || sob@harvard.edu +# RFC2557 || J. Palme, A. Hopmann, N. Shelness || jpalme@dsv.su.se, alexhop@microsoft.com, Shelness@lotus.com, stef@nma.com +# RFC2558 || K. Tesink || kaj@research.telcordia.com +# RFC2559 || S. Boeyen, T. Howes, P. Richard || sharon.boeyen@entrust.com, howes@netscape.com, patr@xcert.com +# RFC2560 || M. Myers, R. Ankney, A. Malpani, S. Galperin, C. Adams || mmyers@verisign.com, rankney@erols.com, ambarish@valicert.com, galperin@mycfo.com, cadams@entrust.com +# RFC2561 || K. White, R. Moore || kennethw@vnet.ibm.com, remoore@us.ibm.com +# RFC2562 || K. White, R. Moore || kennethw@vnet.ibm.com, remoore@us.ibm.com +# RFC2563 || R. Troll || rtroll@corp.home.net +# RFC2564 || C. Kalbfleisch, C. Krupczak, R. Presuhn, J. Saperia || cwk@verio.net, cheryl@empiretech.com, randy_presuhn@bmc.com, saperia@mediaone.net +# RFC2565 || R. Herriot, Ed., S. Butler, P. Moore, R. Turner || rherriot@pahv.xerox.com, sbutler@boi.hp.com, paulmo@microsoft.com, rturner@sharplabs.com, rherriot@pahv.xerox.com +# RFC2566 || R. deBry, T. Hastings, R. Herriot, S. Isaacson, P. Powell || sisaacson@novell.com, tom.hastings@alum.mit.edu, robert.herriot@pahv.xerox.com, debryro@uvsc.edu, papowell@astart.com +# RFC2567 || F. Wright || +# RFC2568 || S. Zilles || +# RFC2569 || R. Herriot, Ed., T. Hastings, N. Jacobs, J. Martin || rherriot@pahv.xerox.com, Norm.Jacobs@Central.sun.com, tom.hastings@alum.mit.edu, jkm@underscore.com +# RFC2570 || J. Case, R. Mundy, D. Partain, B. Stewart || +# RFC2571 || B. Wijnen, D. Harrington, R. Presuhn || +# RFC2572 || J. Case, D. Harrington, R. Presuhn, B. Wijnen || +# RFC2573 || D. Levi, P. Meyer, B. Stewart || +# RFC2574 || U. Blumenthal, B. Wijnen || +# RFC2575 || B. Wijnen, R. Presuhn, K. McCloghrie || +# RFC2576 || R. Frye, D. Levi, S. Routhier, B. Wijnen || +# RFC2577 || M. Allman, S. Ostermann || mallman@grc.nasa.gov, ostermann@cs.ohiou.edu +# RFC2578 || K. McCloghrie, Ed., D. Perkins, Ed., J. Schoenwaelder, Ed. || +# RFC2579 || K. McCloghrie, Ed., D. Perkins, Ed., J. Schoenwaelder, Ed. || +# RFC2580 || K. McCloghrie, Ed., D. Perkins, Ed., J. Schoenwaelder, Ed. || +# RFC2581 || M. Allman, V. Paxson, W. Stevens || mallman@grc.nasa.gov, vern@aciri.org, rstevens@kohala.com +# RFC2582 || S. Floyd, T. Henderson || +# RFC2583 || R. Carlson, L. Winkler || RACarlson@anl.gov, lwinkler@anl.gov +# RFC2584 || B. Clouston, B. Moore || clouston@cisco.com, remoore@us.ibm.com +# RFC2585 || R. Housley, P. Hoffman || housley@spyrus.com, phoffman@imc.org +# RFC2586 || J. Salsman, H. Alvestrand || James@bovik.org, Harald.T.Alvestrand@uninett.no +# RFC2587 || S. Boeyen, T. Howes, P. Richard || sharon.boeyen@entrust.com, howes@netscape.com, patr@xcert.com +# RFC2588 || R. Finlayson || finlayson@live.com +# RFC2589 || Y. Yaacovi, M. Wahl, T. Genovese || yoramy@microsoft.com, tonyg@microsoft.com +# RFC2590 || A. Conta, A. Malis, M. Mueller || aconta@lucent.com, malis@ascend.com, memueller@lucent.com +# RFC2591 || D. Levi, J. Schoenwaelder || +# RFC2592 || D. Levi, J. Schoenwaelder || +# RFC2593 || J. Schoenwaelder, J. Quittek || schoenw@ibr.cs.tu-bs.de, quittek@ccrle.nec.de +# RFC2594 || H. Hazewinkel, C. Kalbfleisch, J. Schoenwaelder || +# RFC2595 || C. Newman || chris.newman@innosoft.com +# RFC2596 || M. Wahl, T. Howes || M.Wahl@innosoft.com, howes@netscape.com +# RFC2597 || J. Heinanen, F. Baker, W. Weiss, J. Wroclawski || jh@telia.fi, fred@cisco.com, wweiss@lucent.com, jtw@lcs.mit.edu +# RFC2598 || V. Jacobson, K. Nichols, K. Poduri || van@cisco.com, kmn@cisco.com, kpoduri@baynetworks.com +# RFC2599 || A. DeLaCruz || delacruz@isi.edu +# RFC2600 || J. Reynolds, R. Braden || +# RFC2601 || M. Davison || mike.davison@cisco.com +# RFC2602 || M. Davison || mike.davison@cisco.com +# RFC2603 || M. Davison || mike.davison@cisco.com +# RFC2604 || R. Gellens || randy@qualcomm.com +# RFC2605 || G. Mansfield, S. Kille || glenn@cysols.com, Steve.Kille@MessagingDirect.com +# RFC2606 || D. Eastlake 3rd, A. Panitz || dee3@us.ibm.com, buglady@fuschia.net +# RFC2607 || B. Aboba, J. Vollbrecht || bernarda@microsoft.com, jrv@merit.edu +# RFC2608 || E. Guttman, C. Perkins, J. Veizades, M. Day || Erik.Guttman@sun.com, cperkins@sun.com, veizades@home.net, mday@vinca.com +# RFC2609 || E. Guttman, C. Perkins, J. Kempf || erik.guttman@sun.com, cperkins@sun.com, james.kempf@sun.com +# RFC2610 || C. Perkins, E. Guttman || Charles.Perkins@Sun.Com, Erik.Guttman@Sun.Com +# RFC2611 || L. Daigle, D. van Gulik, R. Iannella, P. Faltstrom || leslie@thinkingcat.com, Dirk.vanGulik@jrc.it, renato@dstc.edu.au, paf@swip.net +# RFC2612 || C. Adams, J. Gilchrist || carlisle.adams@entrust.com, jeff.gilchrist@entrust.com +# RFC2613 || R. Waterman, B. Lahaye, D. Romascanu, S. Waldbusser || rich@allot.com, dromasca@gmail.com , waldbusser@ins.com +# RFC2614 || J. Kempf, E. Guttman || james.kempf@sun.com, erik.guttman@sun.com +# RFC2615 || A. Malis, W. Simpson || malis@ascend.com, wsimpson@GreenDragon.com +# RFC2616 || R. Fielding, J. Gettys, J. Mogul, H. Frystyk, L. Masinter, P. Leach, T. Berners-Lee || fielding@ics.uci.edu, jg@w3.org, mogul@wrl.dec.com, frystyk@w3.org, masinter@parc.xerox.com, paulle@microsoft.com, timbl@w3.org +# RFC2617 || J. Franks, P. Hallam-Baker, J. Hostetler, S. Lawrence, P. Leach, A. Luotonen, L. Stewart || john@math.nwu.edu, pbaker@verisign.com, jeff@AbiSource.com, lawrence@agranat.com, paulle@microsoft.com, stewart@OpenMarket.com +# RFC2618 || B. Aboba, G. Zorn || bernarda@microsoft.com, glennz@microsoft.com +# RFC2619 || G. Zorn, B. Aboba || bernarda@microsoft.com, glennz@microsoft.com +# RFC2620 || B. Aboba, G. Zorn || bernarda@microsoft.com, glennz@microsoft.com +# RFC2621 || G. Zorn, B. Aboba || bernarda@microsoft.com, glennz@microsoft.com +# RFC2622 || C. Alaettinoglu, C. Villamizar, E. Gerich, D. Kessens, D. Meyer, T. Bates, D. Karrenberg, M. Terpstra || cengiz@isi.edu, curtis@avici.com, epg@home.net, David.Kessens@qwest.net, meyer@antc.uoregon.edu, tbates@cisco.com, dfk@ripe.net, marten@BayNetworks.com +# RFC2623 || M. Eisler || mre@eng.sun.com +# RFC2624 || S. Shepler || spencer.shepler@eng.sun.com +# RFC2625 || M. Rajagopal, R. Bhagwat, W. Rickard || murali@gadzoox.com, raj@gadzoox.com, wayne@gadzoox.com +# RFC2626 || P. Nesser II || +# RFC2627 || D. Wallner, E. Harder, R. Agee || dmwalln@orion.ncsc.mil, ejh@tycho.ncsc.mil +# RFC2628 || V. Smyslov || svan@trustworks.com +# RFC2629 || M. Rose || mrose17@gmail.com +# RFC2630 || R. Housley || housley@spyrus.com +# RFC2631 || E. Rescorla || ekr@rtfm.com +# RFC2632 || B. Ramsdell, Ed. || +# RFC2633 || B. Ramsdell, Ed. || +# RFC2634 || P. Hoffman, Ed. || phoffman@imc.org +# RFC2635 || S. Hambridge, A. Lunde || +# RFC2636 || R. Gellens || randy@qualcomm.com +# RFC2637 || K. Hamzeh, G. Pall, W. Verthein, J. Taarud, W. Little, G. Zorn || kory@ascend.com, gurdeep@microsoft.com, glennz@microsoft.com +# RFC2638 || K. Nichols, V. Jacobson, L. Zhang || kmn@cisco.com, van@cisco.com, lixia@cs.ucla.edu +# RFC2639 || T. Hastings, C. Manros || tom.hastings@alum.mit.edu, manros@cp10.es.xerox.com +# RFC2640 || B. Curtin || curtinw@ftm.disa.mil +# RFC2641 || D. Hamilton, D. Ruffen || daveh@ctron.com, ruffen@ctron.com +# RFC2642 || L. Kane || lkane@ctron.com +# RFC2643 || D. Ruffen, T. Len, J. Yanacek || ruffen@ctron.com, len@ctron.com, jyanacek@ctron.com +# RFC2644 || D. Senie || dts@senie.com +# RFC2645 || R. Gellens || randy@qualcomm.com +# RFC2646 || R. Gellens, Ed. || +# RFC2647 || D. Newman || +# RFC2648 || R. Moats || jayhawk@att.com +# RFC2649 || B. Greenblatt, P. Richard || bgreenblatt@directory-applications.com, patr@xcert.com +# RFC2650 || D. Meyer, J. Schmitz, C. Orange, M. Prior, C. Alaettinoglu || dmm@cisco.com, SchmitzJo@aol.com, orange@spiritone.com, mrp@connect.com.au, cengiz@isi.edu +# RFC2651 || J. Allen, M. Mealling || jeff.allen@acm.org, michael.mealling@RWhois.net +# RFC2652 || J. Allen, M. Mealling || jeff.allen@acm.org, michael.mealling@RWhois.net +# RFC2653 || J. Allen, P. Leach, R. Hedberg || jeff.allen@acm.org, paulle@microsoft.com, roland@catalogix.ac.se +# RFC2654 || R. Hedberg, B. Greenblatt, R. Moats, M. Wahl || roland@catalogix.ac.se, bgreenblatt@directory-applications.com, jayhawk@att.com +# RFC2655 || T. Hardie, M. Bowman, D. Hardy, M. Schwartz, D. Wessels || hardie@equinix.com, mic@transarc.com, dhardy@netscape.com, wessels@nlanr.net +# RFC2656 || T. Hardie || hardie@equinix.com +# RFC2657 || R. Hedberg || roland@catalogix.ac.se +# RFC2658 || K. McKay || kylem@qualcomm.com +# RFC2659 || E. Rescorla, A. Schiffman || ekr@rtfm.com, ams@terisa.com +# RFC2660 || E. Rescorla, A. Schiffman || ekr@rtfm.com, ams@terisa.com +# RFC2661 || W. Townsley, A. Valencia, A. Rubens, G. Pall, G. Zorn, B. Palter || gurdeep@microsoft.com, palter@zev.net, acr@del.com, townsley@cisco.com, vandys@cisco.com, gwz@acm.org +# RFC2662 || G. Bathrick, F. Ly || bathricg@agcs.com, faye@coppermountain.com +# RFC2663 || P. Srisuresh, M. Holdrege || srisuresh@lucent.com, holdrege@lucent.com +# RFC2664 || R. Plzak, A. Wells, E. Krol || plzakr@saic.com, awel@cs.wisc.edu, krol@uiuc.edu +# RFC2665 || J. Flick, J. Johnson || johnf@rose.hp.com, jeff@redbacknetworks.com +# RFC2666 || J. Flick || johnf@rose.hp.com +# RFC2667 || D. Thaler || dthaler@microsoft.com +# RFC2668 || A. Smith, J. Flick, K. de Graaf, D. Romascanu, D. McMaster, K. McCloghrie, S. Roberts || andrew@extremenetworks.com, johnf@rose.hp.com, kdegraaf@argon.com, dromasca@gmail.com , mcmaster@cisco.com, kzm@cisco.com, sroberts@farallon.com +# RFC2669 || M. St. Johns, Ed. || stjohns@corp.home.net +# RFC2670 || M. St. Johns, Ed. || stjohns@corp.home.net +# RFC2671 || P. Vixie || vixie@isc.org +# RFC2672 || M. Crawford || crawdad@fnal.gov +# RFC2673 || M. Crawford || crawdad@fnal.gov +# RFC2674 || E. Bell, A. Smith, P. Langille, A. Rijhsinghani, K. McCloghrie || Les_Bell@3Com.com, andrew@extremenetworks.com, langille@newbridge.com, anil@cabletron.com, kzm@cisco.com +# RFC2675 || D. Borman, S. Deering, R. Hinden || dab@bsdi.com, deering@cisco.com, bob.hinden@gmail.com +# RFC2676 || G. Apostolopoulos, S. Kama, D. Williams, R. Guerin, A. Orda, T. Przygienda || georgeap@watson.ibm.com, guerin@ee.upenn.edu, ariel@ee.technion.ac.il, dougw@watson.ibm.com +# RFC2677 || M. Greene, J. Cucchiara, J. Luciani || luciani@baynetworks.com, maria@xedia.com, joan@ironbridgenetworks.com +# RFC2678 || J. Mahdavi, V. Paxson || mahdavi@psc.edu, vern@ee.lbl.gov +# RFC2679 || G. Almes, S. Kalidindi, M. Zekauskas || almes@advanced.org, kalidindi@advanced.org, matt@advanced.org +# RFC2680 || G. Almes, S. Kalidindi, M. Zekauskas || almes@advanced.org, kalidindi@advanced.org, matt@advanced.org +# RFC2681 || G. Almes, S. Kalidindi, M. Zekauskas || almes@advanced.org, kalidindi@advanced.org, matt@advanced.org +# RFC2682 || I. Widjaja, A. Elwalid || indra.widjaja@fnc.fujitsu.com, anwar@lucent.com +# RFC2683 || B. Leiba || leiba@watson.ibm.com +# RFC2684 || D. Grossman, J. Heinanen || dan@dma.isg.mot.com, jh@telia.fi +# RFC2685 || B. Fox, B. Gleeson || barbarafox@lucent.com, bgleeson@shastanets.com +# RFC2686 || C. Bormann || cabo@tzi.org +# RFC2687 || C. Bormann || cabo@tzi.org +# RFC2688 || S. Jackowski, D. Putzolu, E. Crawley, B. Davie || stevej@DeterministicNetworks.com, David.Putzolu@intel.com, esc@argon.com, bdavie@cisco.com +# RFC2689 || C. Bormann || cabo@tzi.org +# RFC2690 || S. Bradner || +# RFC2691 || S. Bradner || +# RFC2692 || C. Ellison || carl.m.ellison@intel.com +# RFC2693 || C. Ellison, B. Frantz, B. Lampson, R. Rivest, B. Thomas, T. Ylonen || carl.m.ellison@intel.com, frantz@netcom.com, blampson@microsoft.com, rivest@theory.lcs.mit.edu, bt0008@sbc.com, ylo@ssh.fi +# RFC2694 || P. Srisuresh, G. Tsirtsis, P. Akkiraju, A. Heffernan || srisuresh@yahoo.com, george@gideon.bt.co.uk, spa@cisco.com, ahh@juniper.net +# RFC2695 || A. Chiu || +# RFC2696 || C. Weider, A. Herron, A. Anantha, T. Howes || cweider@microsoft.com, andyhe@microsoft.com, anoopa@microsoft.com, howes@netscape.com +# RFC2697 || J. Heinanen, R. Guerin || jh@telia.fi, guerin@ee.upenn.edu +# RFC2698 || J. Heinanen, R. Guerin || jh@telia.fi, guerin@ee.upenn.edu +# RFC2699 || S. Ginoza || ginoza@isi.edu +# RFC2700 || J. Reynolds, R. Braden || +# RFC2701 || G. Malkin || +# RFC2702 || D. Awduche, J. Malcolm, J. Agogbua, M. O'Dell, J. McManus || awduche@uu.net, jmalcolm@uu.net, ja@uu.net, mo@uu.net, jmcmanus@uu.net +# RFC2703 || G. Klyne || GK@ACM.ORG +# RFC2704 || M. Blaze, J. Feigenbaum, J. Ioannidis, A. Keromytis || mab@research.att.com, jf@research.att.com, ji@research.att.com, angelos@dsl.cis.upenn.edu +# RFC2705 || M. Arango, A. Dugan, I. Elliott, C. Huitema, S. Pickett || marango@rslcom.com, andrew.dugan@l3.com, ike.elliott@l3.com, huitema@research.telcordia.com, ScottP@vertical.com +# RFC2706 || D. Eastlake 3rd, T. Goldstein || dee3@us.ibm.com, tgoldstein@brodia.com +# RFC2707 || R. Bergman, T. Hastings, S. Isaacson, H. Lewis || rbergma@dpc.com, tom.hastings@alum.mit.edu, scott_isaacson@novell.com, harryl@us.ibm.com +# RFC2708 || R. Bergman || rbergman@dpc.com, tom.hastings@alum.mit.edu, scott_isaacson@novell.com, harryl@us.ibm.com, bpenteco@boi.hp.com +# RFC2709 || P. Srisuresh || srisuresh@lucent.com +# RFC2710 || S. Deering, W. Fenner, B. Haberman || deering@cisco.com, fenner@research.att.com, haberman@raleigh.ibm.com +# RFC2711 || C. Partridge, A. Jackson || craig@bbn.com, awjacks@bbn.com +# RFC2712 || A. Medvinsky, M. Hur || amedvins@excitecorp.com, matt.hur@cybersafe.com +# RFC2713 || V. Ryan, S. Seligman, R. Lee || vincent.ryan@ireland.sun.com, scott.seligman@eng.sun.com, rosanna.lee@eng.sun.com +# RFC2714 || V. Ryan, R. Lee, S. Seligman || vincent.ryan@ireland.sun.com, rosanna.lee@eng.sun.com, scott.seligman@eng.sun.com +# RFC2715 || D. Thaler || dthaler@microsoft.com +# RFC2716 || B. Aboba, D. Simon || bernarda@microsoft.com, dansimon@microsoft.com +# RFC2717 || R. Petke, I. King || rpetke@wcom.net, iking@microsoft.com +# RFC2718 || L. Masinter, H. Alvestrand, D. Zigmond, R. Petke || masinter@parc.xerox.com, harald.alvestrand@maxware.no, djz@corp.webtv.net, rpetke@wcom.net +# RFC2719 || L. Ong, I. Rytina, M. Garcia, H. Schwarzbauer, L. Coene, H. Lin, I. Juhasz, M. Holdrege, C. Sharp || long@nortelnetworks.com, ian.rytina@ericsson.com, holdrege@lucent.com, lode.coene@siemens.atea.be, Miguel.A.Garcia@ericsson.com, chsharp@cisco.com, imre.i.juhasz@telia.se, hlin@research.telcordia.com, HannsJuergen.Schwarzbauer@icn.siemens.de +# RFC2720 || N. Brownlee || n.brownlee@auckland.ac.nz +# RFC2721 || N. Brownlee || n.brownlee@auckland.ac.nz +# RFC2722 || N. Brownlee, C. Mills, G. Ruth || n.brownlee@auckland.ac.nz, cmills@gte.com, gruth@bbn.com +# RFC2723 || N. Brownlee || n.brownlee@auckland.ac.nz +# RFC2724 || S. Handelman, S. Stibler, N. Brownlee, G. Ruth || swhandel@us.ibm.com, stibler@us.ibm.com, n.brownlee@auckland.ac.nz, gruth@bbn.com +# RFC2725 || C. Villamizar, C. Alaettinoglu, D. Meyer, S. Murphy || curtis@avici.com, cengiz@ISI.EDU, dmm@cisco.com, sandy@tis.com +# RFC2726 || J. Zsako || zsako@banknet.net +# RFC2727 || J. Galvin || +# RFC2728 || R. Panabaker, S. Wegerif, D. Zigmond || +# RFC2729 || P. Bagnall, R. Briscoe, A. Poppitt || pete@surfaceeffect.com, bob.briscoe@bt.com, apoppitt@jungle.bt.co.uk +# RFC2730 || S. Hanna, B. Patel, M. Shah || steve.hanna@sun.com, baiju.v.patel@intel.com, munils@microsoft.com +# RFC2731 || J. Kunze || jak@ckm.ucsf.edu +# RFC2732 || R. Hinden, B. Carpenter, L. Masinter || bob.hinden@gmail.com, brian@icair.org, LMM@acm.org +# RFC2733 || J. Rosenberg, H. Schulzrinne || schulzrinne@cs.columbia.edu +# RFC2734 || P. Johansson || +# RFC2735 || B. Fox, B. Petri || bfox@equipecom.com, bernhard.petri@icn.siemens.de +# RFC2736 || M. Handley, C. Perkins || mjh@aciri.org, C.Perkins@cs.ucl.ac.uk +# RFC2737 || K. McCloghrie, A. Bierman || kzm@cisco.com, andy@yumaworks.com +# RFC2738 || G. Klyne || GK@ACM.ORG +# RFC2739 || T. Small, D. Hennessy, F. Dawson || +# RFC2740 || R. Coltun, D. Ferguson, J. Moy || rcoltun@siara.com, dennis@juniper.com, jmoy@sycamorenet.com +# RFC2741 || M. Daniele, B. Wijnen, M. Ellison, D. Francisco || daniele@zk3.dec.com, wijnen@vnet.ibm.com, ellison@world.std.com, dfrancis@cisco.com +# RFC2742 || L. Heintz, S. Gudur, M. Ellison || lheintz@cisco.com, sgudur@hotmail.com +# RFC2743 || J. Linn || +# RFC2744 || J. Wray || John_Wray@Iris.com +# RFC2745 || A. Terzis, B. Braden, S. Vincent, L. Zhang || terzis@cs.ucla.edu, braden@isi.edu, svincent@cisco.com, lixia@cs.ucla.edu +# RFC2746 || A. Terzis, J. Krawczyk, J. Wroclawski, L. Zhang || jj@arrowpoint.com, jtw@lcs.mit.edu, lixia@cs.ucla.edu, terzis@cs.ucla.edu +# RFC2747 || F. Baker, B. Lindell, M. Talwar || fred@cisco.com, lindell@ISI.EDU, mohitt@microsoft.com +# RFC2748 || D. Durham, Ed., J. Boyle, R. Cohen, S. Herzog, R. Rajan, A. Sastry || +# RFC2749 || S. Herzog, Ed., J. Boyle, R. Cohen, D. Durham, R. Rajan, A. Sastry || +# RFC2750 || S. Herzog || +# RFC2751 || S. Herzog || +# RFC2752 || S. Yadav, R. Yavatkar, R. Pabbati, P. Ford, T. Moore, S. Herzog || +# RFC2753 || R. Yavatkar, D. Pendarakis, R. Guerin || raj.yavatkar@intel.com, dimitris@watson.ibm.com, guerin@ee.upenn.edu +# RFC2754 || C. Alaettinoglu, C. Villamizar, R. Govindan || cengiz@isi.edu, curtis@avici.com, govindan@isi.edu +# RFC2755 || A. Chiu, M. Eisler, B. Callaghan || alex.chiu@Eng.sun.com, michael.eisler@Eng.sun.com, brent.callaghan@Eng.sun.com +# RFC2756 || P. Vixie, D. Wessels || vixie@isc.org, wessels@nlanr.net +# RFC2757 || G. Montenegro, S. Dawkins, M. Kojo, V. Magret, N. Vaidya || gab@sun.com, sdawkins@nortel.com, kojo@cs.helsinki.fi, vincent.magret@aud.alcatel.com, vaidya@cs.tamu.edu +# RFC2758 || K. White || wkenneth@us.ibm.com +# RFC2759 || G. Zorn || gwz@acm.org +# RFC2760 || M. Allman, Ed., S. Dawkins, D. Glover, J. Griner, D. Tran, T. Henderson, J. Heidemann, J. Touch, H. Kruse, S. Ostermann, K. Scott, J. Semke || mallman@grc.nasa.gov, Spencer.Dawkins.sdawkins@nt.com, Daniel.R.Glover@grc.nasa.gov, jgriner@grc.nasa.gov, dtran@grc.nasa.gov, tomh@cs.berkeley.edu, johnh@isi.edu, touch@isi.edu, hkruse1@ohiou.edu, ostermann@cs.ohiou.edu, kscott@mitre.org, semke@psc.edu +# RFC2761 || J. Dunn, C. Martin || +# RFC2762 || J. Rosenberg, H. Schulzrinne || jdrosen@dynamicsoft.com, schulzrinne@cs.columbia.edu +# RFC2763 || N. Shen, H. Smit || naiming@siara.com, hsmit@cisco.com +# RFC2764 || B. Gleeson, A. Lin, J. Heinanen, G. Armitage, A. Malis || +# RFC2765 || E. Nordmark || nordmark@sun.com +# RFC2766 || G. Tsirtsis, P. Srisuresh || george.tsirtsis@bt.com, srisuresh@yahoo.com +# RFC2767 || K. Tsuchiya, H. Higuchi, Y. Atarashi || tsuchi@ebina.hitachi.co.jp, h-higuti@ebina.hitachi.co.jp, atarashi@ebina.hitachi.co.jp +# RFC2768 || B. Aiken, J. Strassner, B. Carpenter, I. Foster, C. Lynch, J. Mambretti, R. Moore, B. Teitelbaum || raiken@cisco.com, raiken@cisco.com, johns@cisco.com, brian@hursley.ibm.com, foster@mcs.anl.gov, cliff@cni.org, j-mambretti@nwu.edu, moore@sdsc.edu, ben@internet2.edu +# RFC2769 || C. Villamizar, C. Alaettinoglu, R. Govindan, D. Meyer || curtis@avici.com, cengiz@ISI.EDU, govindan@ISI.EDU, dmm@cisco.com +# RFC2770 || D. Meyer, P. Lothberg || dmm@cisco.com, roll@sprint.net +# RFC2771 || R. Finlayson || finlayson@live.com +# RFC2772 || R. Rockell, R. Fink || rrockell@sprint.net, fink@es.net +# RFC2773 || R. Housley, P. Yee, W. Nace || housley@spyrus.com, yee@spyrus.com +# RFC2774 || H. Nielsen, P. Leach, S. Lawrence || frystyk@microsoft.com, paulle@microsoft.com, lawrence@agranat.com +# RFC2775 || B. Carpenter || brian@icair.org +# RFC2776 || M. Handley, D. Thaler, R. Kermode || mjh@aciri.org, dthaler@microsoft.com, Roger.Kermode@motorola.com +# RFC2777 || D. Eastlake 3rd || Donald.Eastlake@motorola.com +# RFC2778 || M. Day, J. Rosenberg, H. Sugano || mday@alum.mit.edu, suga@flab.fujitsu.co.jp +# RFC2779 || M. Day, S. Aggarwal, G. Mohr, J. Vincent || mday@alum.mit.edu, sonuag@microsoft.com, gojomo@usa.net, jesse@intonet.com +# RFC2780 || S. Bradner, V. Paxson || sob@harvard.edu, vern@aciri.org +# RFC2781 || P. Hoffman, F. Yergeau || phoffman@imc.org, fyergeau@alis.com +# RFC2782 || A. Gulbrandsen, P. Vixie, L. Esibov || arnt@troll.no, levone@microsoft.com +# RFC2783 || J. Mogul, D. Mills, J. Brittenson, J. Stone, U. Windl || mogul@wrl.dec.com, mills@udel.edu, jonathan@dsg.stanford.edu, ulrich.windl@rz.uni-regensburg.de +# RFC2784 || D. Farinacci, T. Li, S. Hanks, D. Meyer, P. Traina || dino@procket.com, tony1@home.net, stan_hanks@enron.net, dmm@cisco.com, pst@juniper.net +# RFC2785 || R. Zuccherato || robert.zuccherato@entrust.com +# RFC2786 || M. St. Johns || stjohns@corp.home.net +# RFC2787 || B. Jewell, D. Chuang || bjewell@coppermountain.com, david_chuang@cosinecom.com +# RFC2788 || N. Freed, S. Kille || ned.freed@innosoft.com, Steve.Kille@MessagingDirect.com +# RFC2789 || N. Freed, S. Kille || ned.freed@innosoft.com, Steve.Kille@MessagingDirect.com +# RFC2790 || S. Waldbusser, P. Grillo || waldbusser@ins.com +# RFC2791 || J. Yu || jyy@cosinecom.com +# RFC2792 || M. Blaze, J. Ioannidis, A. Keromytis || +# RFC2793 || G. Hellstrom || gunnar.hellstrom@omnitor.se +# RFC2794 || P. Calhoun, C. Perkins || +# RFC2795 || S. Christey || steqve@shore.net +# RFC2796 || T. Bates, R. Chandra, E. Chen || tbates@cisco.com, rchandra@redback.com, enke@redback.com +# RFC2797 || M. Myers, X. Liu, J. Schaad, J. Weinstein || mmyers@verisign.com, xliu@cisco.com, jimsch@nwlink.com, jsw@meer.net +# RFC2798 || M. Smith || mcs@netscape.com +# RFC2799 || S. Ginoza || ginoza@isi.edu +# RFC2800 || J. Reynolds, R. Braden, S. Ginoza || +# RFC2801 || D. Burdett || david.burdett@commerceone.com +# RFC2802 || K. Davidson, Y. Kawatsura || kent@differential.com, kawatura@bisd.hitachi.co.jp +# RFC2803 || H. Maruyama, K. Tamura, N. Uramoto || maruyama@jp.ibm.com, kent@trl.ibm.co.jp, uramoto@jp.ibm.com +# RFC2804 || IAB, IESG || fred@cisco.com, brian@icair.org +# RFC2805 || N. Greene, M. Ramalho, B. Rosen || ngreene@nortelnetworks.com, mramalho@cisco.com, brosen@eng.fore.com +# RFC2806 || A. Vaha-Sipila || avs@iki.fi +# RFC2807 || J. Reagle || reagle@w3.org +# RFC2808 || M. Nystrom || magnus@rsasecurity.com +# RFC2809 || B. Aboba, G. Zorn || bernarda@microsoft.com, gwz@cisco.com +# RFC2810 || C. Kalt || Christophe.Kalt@gmail.com +# RFC2811 || C. Kalt || Christophe.Kalt@gmail.com +# RFC2812 || C. Kalt || Christophe.Kalt@gmail.com +# RFC2813 || C. Kalt || Christophe.Kalt@gmail.com +# RFC2814 || R. Yavatkar, D. Hoffman, Y. Bernet, F. Baker, M. Speer || yavatkar@ibeam.intel.com, yoramb@microsoft.com, fred@cisco.com, speer@Eng.Sun.COM +# RFC2815 || M. Seaman, A. Smith, E. Crawley, J. Wroclawski || andrew@extremenetworks.com, jtw@lcs.mit.edu +# RFC2816 || A. Ghanwani, J. Pace, V. Srinivasan, A. Smith, M. Seaman || aghanwan@nortelnetworks.com, pacew@us.ibm.com, vijay@cosinecom.com, andrew@extremenetworks.com +# RFC2817 || R. Khare, S. Lawrence || rohit@4K-associates.com, lawrence@agranat.com +# RFC2818 || E. Rescorla || ekr@rtfm.com +# RFC2819 || S. Waldbusser || +# RFC2820 || E. Stokes, D. Byrne, B. Blakley, P. Behera || blakley@dascom.com, stokes@austin.ibm.com, djbyrne@us.ibm.com, prasanta@netscape.com +# RFC2821 || J. Klensin, Ed. || +# RFC2822 || P. Resnick, Ed. || presnick@qti.qualcomm.com +# RFC2823 || J. Carlson, P. Langner, E. Hernandez-Valencia, J. Manchester || james.d.carlson@sun.com, plangner@lucent.com, enrique@lucent.com, sterling@hotair.hobl.lucent.com +# RFC2824 || J. Lennox, H. Schulzrinne || lennox@cs.columbia.edu, schulzrinne@cs.columbia.edu +# RFC2825 || IAB, L. Daigle, Ed. || iab@iab.org +# RFC2826 || Internet Architecture Board || iab@iab.org +# RFC2827 || P. Ferguson, D. Senie || ferguson@cisco.com, dts@senie.com +# RFC2828 || R. Shirey || rshirey@bbn.com +# RFC2829 || M. Wahl, H. Alvestrand, J. Hodges, R. Morgan || M.Wahl@innosoft.com, Harald@Alvestrand.no, JHodges@oblix.com, rlmorgan@washington.edu +# RFC2830 || J. Hodges, R. Morgan, M. Wahl || JHodges@oblix.com, rlmorgan@washington.edu, M.Wahl@innosoft.com +# RFC2831 || P. Leach, C. Newman || paulle@microsoft.com, chris.newman@innosoft.com +# RFC2832 || S. Hollenbeck, M. Srivastava || shollenb@netsol.com, manojs@netsol.com +# RFC2833 || H. Schulzrinne, S. Petrack || schulzrinne@cs.columbia.edu, scott.petrack@metatel.com +# RFC2834 || J.-M. Pittet || jmp@sgi.com +# RFC2835 || J.-M. Pittet || jmp@sgi.com +# RFC2836 || S. Brim, B. Carpenter, F. Le Faucheur || sbrim@cisco.com, brian@icair.org, flefauch@cisco.com +# RFC2837 || K. Teow || +# RFC2838 || D. Zigmond, M. Vickers || djz@corp.webtv.net, mav@liberate.com +# RFC2839 || F. da Cruz, J. Altman || +# RFC2840 || J. Altman, F. da Cruz || +# RFC2841 || P. Metzger, W. Simpson || +# RFC2842 || R. Chandra, J. Scudder || rchandra@redback.com, jgs@cisco.com +# RFC2843 || P. Droz, T. Przygienda || dro@zurich.ibm.com, prz@siara.com +# RFC2844 || T. Przygienda, P. Droz, R. Haas || prz@siara.com, dro@zurich.ibm.com, rha@zurich.ibm.com +# RFC2845 || P. Vixie, O. Gudmundsson, D. Eastlake 3rd, B. Wellington || vixie@isc.org, ogud@tislabs.com, dee3@torque.pothole.com, Brian.Wellington@nominum.com +# RFC2846 || C. Allocchio || +# RFC2847 || M. Eisler || mike@eisler.com +# RFC2848 || S. Petrack, L. Conroy || scott.petrack@metatel.com, lwc@roke.co.uk +# RFC2849 || G. Good || ggood@netscape.com +# RFC2850 || Internet Architecture Board, B. Carpenter, Ed. || brian@icair.org +# RFC2851 || M. Daniele, B. Haberman, S. Routhier, J. Schoenwaelder || daniele@zk3.dec.com, haberman@nortelnetworks.com, sar@epilogue.com, schoenw@ibr.cs.tu-bs.de +# RFC2852 || D. Newman || dan.newman@sun.com +# RFC2853 || J. Kabat, M. Upadhyay || jackk@valicert.com, mdu@eng.sun.com +# RFC2854 || D. Connolly, L. Masinter || connolly@w3.org, LM@att.com +# RFC2855 || K. Fujisawa || fujisawa@sm.sony.co.jp +# RFC2856 || A. Bierman, K. McCloghrie, R. Presuhn || andy@yumaworks.com, kzm@cisco.com, rpresuhn@bmc.com +# RFC2857 || A. Keromytis, N. Provos || angelos@dsl.cis.upenn.edu, provos@citi.umich.edu, rgm@icsa.net, tytso@valinux.com +# RFC2858 || T. Bates, Y. Rekhter, R. Chandra, D. Katz || tbates@cisco.com, rchandra@redback.com, dkatz@jnx.com, yakov@cisco.com +# RFC2859 || W. Fang, N. Seddigh, B. Nandy || wfang@cs.princeton.edu, nseddigh@nortelnetworks.com, bnandy@nortelnetworks.com +# RFC2860 || B. Carpenter, F. Baker, M. Roberts || brian@icair.org, fred@cisco.com, roberts@icann.org +# RFC2861 || M. Handley, J. Padhye, S. Floyd || mjh@aciri.org, padhye@aciri.org, floyd@aciri.org +# RFC2862 || M. Civanlar, G. Cash || civanlar@research.att.com, glenn@research.att.com +# RFC2863 || K. McCloghrie, F. Kastenholz || kzm@cisco.com, kasten@argon.com +# RFC2864 || K. McCloghrie, G. Hanson || kzm@cisco.com, gary_hanson@adc.com +# RFC2865 || C. Rigney, S. Willens, A. Rubens, W. Simpson || cdr@telemancy.com, acr@merit.edu, wsimpson@greendragon.com, steve@livingston.com +# RFC2866 || C. Rigney || cdr@telemancy.com +# RFC2867 || G. Zorn, B. Aboba, D. Mitton || gwz@cisco.com, dmitton@nortelnetworks.com, aboba@internaut.com +# RFC2868 || G. Zorn, D. Leifer, A. Rubens, J. Shriver, M. Holdrege, I. Goyret || gwz@cisco.com, leifer@del.com, John.Shriver@intel.com, acr@del.com, matt@ipverse.com, igoyret@lucent.com +# RFC2869 || C. Rigney, W. Willats, P. Calhoun || cdr@telemancy.com, ward@cyno.com, pcalhoun@eng.sun.com, arubens@tutsys.com, bernarda@microsoft.com +# RFC2870 || R. Bush, D. Karrenberg, M. Kosters, R. Plzak || randy@psg.com, daniel.karrenberg@ripe.net, markk@netsol.com, plzakr@saic.com +# RFC2871 || J. Rosenberg, H. Schulzrinne || +# RFC2872 || Y. Bernet, R. Pabbati || yoramb@microsoft.com, rameshpa@microsoft.com +# RFC2873 || X. Xiao, A. Hannan, V. Paxson, E. Crabbe || xipeng@gblx.net, alan@ivmg.net, edc@explosive.net, vern@aciri.org +# RFC2874 || M. Crawford, C. Huitema || crawdad@fnal.gov, huitema@microsoft.com +# RFC2875 || H. Prafullchandra, J. Schaad || hemma@cp.net, jimsch@exmsft.com +# RFC2876 || J. Pawling || john.pawling@wang.com +# RFC2877 || T. Murphy Jr., P. Rieth, J. Stevens || murphyte@us.ibm.com, rieth@us.ibm.com, jssteven@us.ibm.com +# RFC2878 || M. Higashiyama, F. Baker || Mitsuru.Higashiyama@yy.anritsu.co.jp, fred.baker@cisco.com +# RFC2879 || G. Klyne, L. McIntyre || GK@ACM.ORG, Lloyd.McIntyre@pahv.xerox.com +# RFC2880 || L. McIntyre, G. Klyne || Lloyd.McIntyre@pahv.xerox.com, GK@ACM.ORG +# RFC2881 || D. Mitton, M. Beadles || dmitton@nortelnetworks.com, mbeadles@smartpipes.com +# RFC2882 || D. Mitton || dmitton@nortelnetworks.com +# RFC2883 || S. Floyd, J. Mahdavi, M. Mathis, M. Podolsky || floyd@aciri.org, mahdavi@novell.com, mathis@psc.edu, podolsky@eecs.berkeley.edu +# RFC2884 || J. Hadi Salim, U. Ahmed || hadi@nortelnetworks.com, ahmed@sce.carleton.ca +# RFC2885 || F. Cuervo, N. Greene, C. Huitema, A. Rayhan, B. Rosen, J. Segers || +# RFC2886 || T. Taylor || tom.taylor.stds@gmail.com +# RFC2887 || M. Handley, S. Floyd, B. Whetten, R. Kermode, L. Vicisano, M. Luby || mjh@aciri.org, floyd@aciri.org, whetten@talarian.com, Roger.Kermode@motorola.com, lorenzo@cisco.com, luby@digitalfountain.com +# RFC2888 || P. Srisuresh || srisuresh@yahoo.com +# RFC2889 || R. Mandeville, J. Perser || bob@cqos.com, jerry_perser@netcomsystems.com +# RFC2890 || G. Dommety || gdommety@cisco.com +# RFC2891 || T. Howes, M. Wahl, A. Anantha || anoopa@microsoft.com, howes@loudcloud.com, Mark.Wahl@sun.com +# RFC2892 || D. Tsiang, G. Suwala || tsiang@cisco.com, gsuwala@cisco.com +# RFC2893 || R. Gilligan, E. Nordmark || gilligan@freegate.com, nordmark@eng.sun.com +# RFC2894 || M. Crawford || crawdad@fnal.gov +# RFC2895 || A. Bierman, C. Bucci, R. Iddon || andy@yumaworks.com, cbucci@cisco.com +# RFC2896 || A. Bierman, C. Bucci, R. Iddon || andy@yumaworks.com, cbucci@cisco.com +# RFC2897 || D. Cromwell || cromwell@nortelnetworks.com +# RFC2898 || B. Kaliski || bkaliski@rsasecurity.com +# RFC2899 || S. Ginoza || ginoza@isi.edu +# RFC2900 || J. Reynolds, R. Braden, S. Ginoza || +# RFC2901 || Z. Wenzel, J. Klensin, R. Bush, S. Huter || zita@nsrc.org, klensin@nsrc.org, randy@nsrc.org, sghuter@nsrc.org +# RFC2902 || S. Deering, S. Hares, C. Perkins, R. Perlman || deering@cisco.com, skh@nexthop.com, Radia.Perlman@sun.com, Charles.Perkins@nokia.com +# RFC2903 || C. de Laat, G. Gross, L. Gommans, J. Vollbrecht, D. Spence || delaat@phys.uu.nl, gmgross@lucent.com, jrv@interlinknetworks.com, dspence@interlinknetworks.com +# RFC2904 || J. Vollbrecht, P. Calhoun, S. Farrell, L. Gommans, G. Gross, B. de Bruijn, C. de Laat, M. Holdrege, D. Spence || pcalhoun@eng.sun.com, stephen.farrell@baltimore.ie, betty@euronet.nl, delaat@phys.uu.nl, matt@ipverse.com, dspence@interlinknetworks.com +# RFC2905 || J. Vollbrecht, P. Calhoun, S. Farrell, L. Gommans, G. Gross, B. de Bruijn, C. de Laat, M. Holdrege, D. Spence || jrv@interlinknetworks.com, pcalhoun@eng.sun.com, stephen.farrell@baltimore.ie, gmgross@lucent.com, betty@euronet.nl, delaat@phys.uu.nl, matt@ipverse.com, dspence@interlinknetworks.com +# RFC2906 || S. Farrell, J. Vollbrecht, P. Calhoun, L. Gommans, G. Gross, B. de Bruijn, C. de Laat, M. Holdrege, D. Spence || stephen.farrell@baltimore.ie, jrv@interlinknetworks.com, pcalhoun@eng.sun.com, gmgross@lucent.com, betty@euronet.nl, delaat@phys.uu.nl, matt@ipverse.com, dspence@interlinknetworks.com +# RFC2907 || R. Kermode || Roger.Kermode@motorola.com +# RFC2908 || D. Thaler, M. Handley, D. Estrin || dthaler@microsoft.com, mjh@aciri.org, estrin@usc.edu +# RFC2909 || P. Radoslavov, D. Estrin, R. Govindan, M. Handley, S. Kumar, D. Thaler || pavlin@catarina.usc.edu, estrin@isi.edu, govindan@isi.edu, mjh@aciri.org, kkumar@usc.edu, dthaler@microsoft.com +# RFC2910 || R. Herriot, Ed., S. Butler, P. Moore, R. Turner, J. Wenn || robert.herriot@pahv.xerox.com, sbutler@boi.hp.com, pmoore@peerless.com, jwenn@cp10.es.xerox.com, tom.hastings@alum.mit.edu, robert.herriot@pahv.xerox.com +# RFC2911 || T. Hastings, Ed., R. Herriot, R. deBry, S. Isaacson, P. Powell || sisaacson@novell.com, tom.hastings@alum.mit.edu, robert.herriot@pahv.xerox.com, debryro@uvsc.edu, papowell@astart.com +# RFC2912 || G. Klyne || GK@ACM.ORG +# RFC2913 || G. Klyne || GK@ACM.ORG +# RFC2914 || S. Floyd || +# RFC2915 || M. Mealling, R. Daniel || michaelm@netsol.com, rdaniel@datafusion.net +# RFC2916 || P. Faltstrom || paf@cisco.com +# RFC2917 || K. Muthukrishnan, A. Malis || mkarthik@lucent.com, Andy.Malis@vivacenetworks.com +# RFC2918 || E. Chen || enke@redback.com +# RFC2919 || R. Chandhok, G. Wenger || chandhok@qualcomm.com, gwenger@qualcomm.com +# RFC2920 || N. Freed || ned.freed@innosoft.com +# RFC2921 || B. Fink || fink@es.net +# RFC2922 || A. Bierman, K. Jones || andy@yumaworks.com, kejones@nortelnetworks.com +# RFC2923 || K. Lahey || +# RFC2924 || N. Brownlee, A. Blount || n.brownlee@auckland.ac.nz, blount@alum.mit.edu +# RFC2925 || K. White || wkenneth@us.ibm.com +# RFC2926 || J. Kempf, R. Moats, P. St. Pierre || james.kempf@sun.com, rmoats@coreon.net, Pete.StPierre@Eng.Sun.COM +# RFC2927 || M. Wahl || Mark.Wahl@sun.com +# RFC2928 || R. Hinden, S. Deering, R. Fink, T. Hain || bob.hinden@gmail.com, deering@cisco.com, rlfink@lbl.gov, tonyhain@microsoft.com +# RFC2929 || D. Eastlake 3rd, E. Brunner-Williams, B. Manning || Donald.Eastlake@motorola.com, brunner@engage.com, bmanning@isi.edu +# RFC2930 || D. Eastlake 3rd || Donald.Eastlake@motorola.com +# RFC2931 || D. Eastlake 3rd || Donald.Eastlake@motorola.com +# RFC2932 || K. McCloghrie, D. Farinacci, D. Thaler || kzm@cisco.com, dthaler@microsoft.com +# RFC2933 || K. McCloghrie, D. Farinacci, D. Thaler || kzm@cisco.com, dthaler@microsoft.com +# RFC2934 || K. McCloghrie, D. Farinacci, D. Thaler, B. Fenner || kzm@cisco.com, dthaler@microsoft.com, fenner@research.att.com +# RFC2935 || D. Eastlake 3rd, C. Smith || Donald.Eastlake@motorola.com, chris.smith@royalbank.com +# RFC2936 || D. Eastlake 3rd, C. Smith, D. Soroka || Donald.Eastlake@motorola.com, chris.smith@royalbank.com, dsoroka@us.ibm.com +# RFC2937 || C. Smith || cs@Eng.Sun.COM +# RFC2938 || G. Klyne, L. Masinter || GK@ACM.ORG, LMM@acm.org +# RFC2939 || R. Droms || droms@bucknell.edu +# RFC2940 || A. Smith, D. Partain, J. Seligson || David.Partain@ericsson.com, jseligso@nortelnetworks.com +# RFC2941 || T. Ts'o, Ed., J. Altman || tytso@mit.edu, jaltman@columbia.edu +# RFC2942 || T. Ts'o || +# RFC2943 || R. Housley, T. Horting, P. Yee || housley@spyrus.com, thorting@spyrus.com, yee@spyrus.com +# RFC2944 || T. Wu || tjw@cs.Stanford.EDU +# RFC2945 || T. Wu || tjw@cs.Stanford.EDU +# RFC2946 || T. Ts'o || tytso@mit.edu +# RFC2947 || J. Altman || jaltman@columbia.edu +# RFC2948 || J. Altman || jaltman@columbia.edu +# RFC2949 || J. Altman || jaltman@columbia.edu +# RFC2950 || J. Altman || jaltman@columbia.edu +# RFC2951 || R. Housley, T. Horting, P. Yee || housley@spyrus.com, thorting@spyrus.com, yee@spyrus.com +# RFC2952 || T. Ts'o || tytso@mit.edu +# RFC2953 || T. Ts'o || tytso@mit.edu +# RFC2954 || K. Rehbehn, D. Fowler || krehbehn@megisto.com, fowler@syndesis.com +# RFC2955 || K. Rehbehn, O. Nicklass, G. Mouradian || krehbehn@megisto.com, orly_n@rad.co.il, gvm@att.com +# RFC2956 || M. Kaat || Marijke.Kaat@surfnet.nl +# RFC2957 || L. Daigle, P. Faltstrom || paf@cisco.com +# RFC2958 || L. Daigle, P. Faltstrom || paf@cisco.com +# RFC2959 || M. Baugher, B. Strahm, I. Suconick || mbaugher@passedge.com, Bill.Strahm@intel.com, irina@ennovatenetworks.com +# RFC2960 || R. Stewart, Q. Xie, K. Morneault, C. Sharp, H. Schwarzbauer, T. Taylor, I. Rytina, M. Kalla, L. Zhang, V. Paxson || randall@lakerest.net, qxie1@email.mot.com, kmorneau@cisco.com, chsharp@cisco.com, HannsJuergen.Schwarzbauer@icn.siemens.de, tom.taylor.stds@gmail.com, ian.rytina@ericsson.com, mkalla@telcordia.com, lixia@cs.ucla.edu, vern@aciri.org +# RFC2961 || L. Berger, D. Gan, G. Swallow, P. Pan, F. Tommasi, S. Molendini || lberger@labn.net, swallow@cisco.com, franco.tommasi@unile.it, molendini@ultra5.unile.it +# RFC2962 || D. Raz, J. Schoenwaelder, B. Sugla || raz@lucent.com, schoenw@ibr.cs.tu-bs.de, sugla@ispsoft.com +# RFC2963 || O. Bonaventure, S. De Cnodder || Olivier.Bonaventure@info.fundp.ac.be, stefaan.de_cnodder@alcatel.be +# RFC2964 || K. Moore, N. Freed || moore@cs.utk.edu, ned.freed@innosoft.com +# RFC2965 || D. Kristol, L. Montulli || +# RFC2966 || T. Li, T. Przygienda, H. Smit || tli@procket.com, prz@redback.com, henk@procket.com +# RFC2967 || L. Daigle, R. Hedberg || leslie@thinkingcat.com, Roland@catalogix.se +# RFC2968 || L. Daigle, T. Eklof || leslie@thinkingcat.com, thommy.eklof@hotsip.com +# RFC2969 || T. Eklof, L. Daigle || thommy.eklof@hotsip.com, leslie@thinkingcat.com +# RFC2970 || L. Daigle, T. Eklof || leslie@thinkingcat.com, thommy.eklof@hotsip.com +# RFC2971 || T. Showalter || tjs@mirapoint.com +# RFC2972 || N. Popp, M. Mealling, L. Masinter, K. Sollins || LMM@acm.org, michaelm@netsol.com, nico@realnames.com, sollins@lcs.mit.edu +# RFC2973 || R. Balay, D. Katz, J. Parker || Rajesh.Balay@cosinecom.com, dkatz@juniper.net, jparker@axiowave.com +# RFC2974 || M. Handley, C. Perkins, E. Whelan || mjh@aciri.org, csp@isi.edu, e.whelan@cs.ucl.ac.uk +# RFC2975 || B. Aboba, J. Arkko, D. Harrington || bernarda@microsoft.com, Jari.Arkko@ericsson.com, dbh@cabletron.com +# RFC2976 || S. Donovan || +# RFC2977 || S. Glass, T. Hiller, S. Jacobs, C. Perkins || +# RFC2978 || N. Freed, J. Postel || ned.freed@innosoft.com +# RFC2979 || N. Freed || ned.freed@innosoft.com +# RFC2980 || S. Barber || sob@academ.com +# RFC2981 || R. Kavasseri, Ed. || ramk@cisco.com +# RFC2982 || R. Kavasseri, Ed. || ramk@cisco.com +# RFC2983 || D. Black || black_david@emc.com +# RFC2984 || C. Adams || cadams@entrust.com +# RFC2985 || M. Nystrom, B. Kaliski || magnus@rsasecurity.com, bkaliski@rsasecurity.com +# RFC2986 || M. Nystrom, B. Kaliski || magnus@rsasecurity.com, bkaliski@rsasecurity.com +# RFC2987 || P. Hoffman || phoffman@imc.org +# RFC2988 || V. Paxson, M. Allman || vern@aciri.org, mallman@grc.nasa.gov +# RFC2989 || B. Aboba, P. Calhoun, S. Glass, T. Hiller, P. McCann, H. Shiino, P. Walsh, G. Zorn, G. Dommety, C. Perkins, B. Patil, D. Mitton, S. Manning, M. Beadles, X. Chen, S. Sivalingham, A. Hameed, M. Munson, S. Jacobs, B. Lim, B. Hirschman, R. Hsu, H. Koo, M. Lipford, E. Campbell, Y. Xu, S. Baba, E. Jaques || bernarda@microsoft.com, pcalhoun@eng.sun.com, steven.glass@sun.com, tom.hiller@lucent.com, mccap@lucent.com, hshiino@lucent.com, walshp@lucent.com, gwz@cisco.com, gdommety@cisco.com, charliep@iprg.nokia.com, Basavaraj.Patil@nokia.com, dmitton@nortelnetworks.com, smanning@nortelnetworks.com, mbeadles@smartpipes.com, xing.chen@usa.alcatel.com, s.sivalingham@ericsson.com, none, mmunson@mobilnet.gte.com, sjacobs@gte.com, bklim@lgic.co.kr, qa4053@email.mot.com, rhsu@qualcomm.com, hskoo@sta.samsung.com, mlipfo01@sprintspectrum.com, ed_campbell@3com.com, yxu@watercove.com, sbaba@tari.toshiba.com, ejaques@akamail.com +# RFC2990 || G. Huston || gih@telstra.net +# RFC2991 || D. Thaler, C. Hopps || dthaler@dthaler.microsoft.com, chopps@nexthop.com +# RFC2992 || C. Hopps || chopps@nexthop.com +# RFC2993 || T. Hain || tonyhain@microsoft.com +# RFC2994 || H. Ohta, M. Matsui || hidenori@iss.isl.melco.co.jp, matsui@iss.isl.melco.co.jp +# RFC2995 || H. Lu, Ed., I. Faynberg, J. Voelker, M. Weissman, W. Zhang, S. Rhim, J. Hwang, S. Ago, S. Moeenuddin, S. Hadvani, S. Nyckelgard, J. Yoakum, L. Robart || faynberg@lucent.com, huilanlu@lucent.com, jvoelker@lucent.com, maw1@lucent.com, wzz@lucent.com, syrhim@kt.co.kr, jkhwang@kt.co.kr, ago@ssf.abk.nec.co.jp, moeen@asl.dl.nec.com, hadvani@asl.dl.nec.com, soren.m.nyckelgard@telia.se, yoakum@nortelnetworks.com, robart@nortelnetworks.com +# RFC2996 || Y. Bernet || yoramb@microsoft.com +# RFC2997 || Y. Bernet, A. Smith, B. Davie || Yoramb@microsoft.com, bsd@cisco.com +# RFC2998 || Y. Bernet, P. Ford, R. Yavatkar, F. Baker, L. Zhang, M. Speer, R. Braden, B. Davie, J. Wroclawski, E. Felstaine || yoramb@microsoft.com, raj.yavatkar@intel.com, peterf@microsoft.com, fred@cisco.com, lixia@cs.ucla.edu, speer@Eng.Sun.COM, braden@isi.edu, bsd@cisco.com, jtw@lcs.mit.edu +# RFC2999 || S. Ginoza || ginoza@isi.edu +# RFC3000 || J. Reynolds, R. Braden, S. Ginoza, L. Shiota || +# RFC3001 || M. Mealling || michaelm@netsol.com +# RFC3002 || D. Mitzel || mitzel@iprg.nokia.com +# RFC3003 || M. Nilsson || nilsson@id3.org +# RFC3004 || G. Stump, R. Droms, Y. Gu, R. Vyaghrapuri, A. Demirtjis, B. Beser, J. Privat || stumpga@us.ibm.com, rdroms@cisco.com, yegu@microsoft.com, rameshv@microsoft.com, annd@microsoft.com +# RFC3005 || S. Harris || srh@merit.edu +# RFC3006 || B. Davie, C. Iturralde, D. Oran, S. Casner, J. Wroclawski || bsd@cisco.com, cei@cisco.com, oran@cisco.com, casner@acm.org, jtw@lcs.mit.edu +# RFC3007 || B. Wellington || Brian.Wellington@nominum.com +# RFC3008 || B. Wellington || Brian.Wellington@nominum.com +# RFC3009 || J. Rosenberg, H. Schulzrinne || jdrosen@dynamicsoft.com, schulzrinne@cs.columbia.edu +# RFC3010 || S. Shepler, B. Callaghan, D. Robinson, R. Thurlow, C. Beame, M. Eisler, D. Noveck || beame@bws.com, brent.callaghan@sun.com, mike@eisler.com, david.robinson@sun.com, robert.thurlow@sun.com +# RFC3011 || G. Waters || +# RFC3012 || C. Perkins, P. Calhoun || +# RFC3013 || T. Killalea || tomk@neart.org +# RFC3014 || R. Kavasseri || ramk@cisco.com +# RFC3015 || F. Cuervo, N. Greene, A. Rayhan, C. Huitema, B. Rosen, J. Segers || +# RFC3016 || Y. Kikuchi, T. Nomura, S. Fukunaga, Y. Matsui, H. Kimata || yoshihiro.kikuchi@toshiba.co.jp, matsui@drl.mei.co.jp, t-nomura@ccm.cl.nec.co.jp, fukunaga444@oki.co.jp, kimata@nttvdt.hil.ntt.co.jp +# RFC3017 || M. Riegel, G. Zorn || maximilian.riegel@icn.siemens.de, gwz@cisco.com +# RFC3018 || A. Bogdanov || a_bogdanov@iname.ru +# RFC3019 || B. Haberman, R. Worzella || haberman@nortelnetworks.com, worzella@us.ibm.com +# RFC3020 || P. Pate, B. Lynch, K. Rehbehn || prayson.pate@overturenetworks.com, bob.lynch@overturenetworks.com, krehbehn@megisto.com +# RFC3021 || A. Retana, R. White, V. Fuller, D. McPherson || aretana@cisco.com, riw@cisco.com, vaf@valinor.barrnet.net, danny@ambernetworks.com +# RFC3022 || P. Srisuresh, K. Egevang || srisuresh@yahoo.com, kjeld.egevang@intel.com +# RFC3023 || M. Murata, S. St. Laurent, D. Kohn || mmurata@trl.ibm.co.jp, simonstl@simonstl.com, dan@dankohn.com +# RFC3024 || G. Montenegro, Ed. || +# RFC3025 || G. Dommety, K. Leung || gdommety@cisco.com, kleung@cisco.com +# RFC3026 || R. Blane || Roy_Blane@inmarsat.com +# RFC3027 || M. Holdrege, P. Srisuresh || matt@ipverse.com, srisuresh@yahoo.com +# RFC3028 || T. Showalter || tjs@mirapoint.com +# RFC3029 || C. Adams, P. Sylvester, M. Zolotarev, R. Zuccherato || cadams@entrust.com, mzolotarev@baltimore.com, peter.sylvester@edelweb.fr, robert.zuccherato@entrust.com +# RFC3030 || G. Vaudreuil || GregV@ieee.org +# RFC3031 || E. Rosen, A. Viswanathan, R. Callon || erosen@cisco.com, arun@force10networks.com, rcallon@juniper.net +# RFC3032 || E. Rosen, D. Tappan, G. Fedorkow, Y. Rekhter, D. Farinacci, T. Li, A. Conta || erosen@cisco.com, tappan@cisco.com, yakov@juniper.net, fedorkow@cisco.com, dino@procket.com, tli@procket.com, aconta@txc.com +# RFC3033 || M. Suzuki || suzuki.muneyoshi@lab.ntt.co.jp +# RFC3034 || A. Conta, P. Doolan, A. Malis || aconta@txc.com, pdoolan@ennovatenetworks.com, Andy.Malis@vivacenetworks.com +# RFC3035 || B. Davie, J. Lawrence, K. McCloghrie, E. Rosen, G. Swallow, Y. Rekhter, P. Doolan || bsd@cisco.com, pdoolan@ennovatenetworks.com, jlawrenc@cisco.com, kzm@cisco.com, yakov@juniper.net, erosen@cisco.com, swallow@cisco.com +# RFC3036 || L. Andersson, P. Doolan, N. Feldman, A. Fredette, B. Thomas || loa.andersson@nortelnetworks.com, pdoolan@ennovatenetworks.com, nkf@us.ibm.com, fredette@photonex.com, rhthomas@cisco.com +# RFC3037 || B. Thomas, E. Gray || ewgray@mindspring.com, rhthomas@cisco.com +# RFC3038 || K. Nagami, Y. Katsube, N. Demizu, H. Esaki, P. Doolan || ken.nagami@toshiba.co.jp, demizu@dd.iij4u.or.jp, hiroshi@wide.ad.jp, yasuhiro.katsube@toshiba.co.jp, pdoolan@ennovatenetworks.com +# RFC3039 || S. Santesson, W. Polk, P. Barzin, M. Nystrom || stefan@addtrust.com, wpolk@nist.gov, barzin@secude.com, magnus@rsasecurity.com +# RFC3040 || I. Cooper, I. Melve, G. Tomlinson || icooper@equinix.com, Ingrid.Melve@uninett.no, gary.tomlinson@cacheflow.com +# RFC3041 || T. Narten, R. Draves || narten@raleigh.ibm.com, richdr@microsoft.com +# RFC3042 || M. Allman, H. Balakrishnan, S. Floyd || mallman@grc.nasa.gov, hari@lcs.mit.edu, floyd@aciri.org +# RFC3043 || M. Mealling || michaelm@netsol.com +# RFC3044 || S. Rozenfeld || +# RFC3045 || M. Meredith || mark_meredith@novell.com +# RFC3046 || M. Patrick || michael.patrick@motorola.com +# RFC3047 || P. Luthi || luthip@pictel.com +# RFC3048 || B. Whetten, L. Vicisano, R. Kermode, M. Handley, S. Floyd, M. Luby || whetten@talarian.com, lorenzo@cisco.com, Roger.Kermode@motorola.com, mjh@aciri.org, luby@digitalfountain.com +# RFC3049 || J. Naugle, K. Kasthurirangan, G. Ledford || jnaugle@us.ibm.com, kasthuri@us.ibm.com, gledford@zephyrcorp.com +# RFC3050 || J. Lennox, H. Schulzrinne, J. Rosenberg || lennox@cs.columbia.edu, jdrosen@dynamicsoft.com, schulzrinne@cs.columbia.edu +# RFC3051 || J. Heath, J. Border || jheath@hns.com, border@hns.com +# RFC3052 || M. Eder, S. Nag || michael.eder@nokia.com, thinker@monmouth.com +# RFC3053 || A. Durand, P. Fasano, I. Guardini, D. Lento || Alain.Durand@sun.com, paolo.fasano@cselt.it, ivano.guardini@cselt.it, dlento@mail.tim.it +# RFC3054 || P. Blatherwick, R. Bell, P. Holland || blather@nortelnetworks.com, rtbell@cisco.com, phil.holland@circa.ca +# RFC3055 || M. Krishnaswamy, D. Romascanu || murali@lucent.com, dromasca@gmail.com +# RFC3056 || B. Carpenter, K. Moore || brian@icair.org, moore@cs.utk.edu +# RFC3057 || K. Morneault, S. Rengasami, M. Kalla, G. Sidebottom || kmorneau@cisco.com, mkalla@telcordia.com, srengasa@telcordia.com, gregside@nortelnetworks.com +# RFC3058 || S. Teiwes, P. Hartmann, D. Kuenzi || stephan.teiwes@it-sec.com, peter.hartmann@it-sec.com, dkuenzi@724.com +# RFC3059 || E. Guttman || Erik.Guttman@sun.com +# RFC3060 || B. Moore, E. Ellesson, J. Strassner, A. Westerinen || eellesson@lboard.com, remoore@us.ibm.com, johns@cisco.com, andreaw@cisco.com +# RFC3061 || M. Mealling || michaelm@netsol.com +# RFC3062 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC3063 || Y. Ohba, Y. Katsube, E. Rosen, P. Doolan || yoshihiro.ohba@toshiba.co.jp, yasuhiro.katsube@toshiba.co.jp, erosen@cisco.com, pdoolan@ennovatenetworks.com +# RFC3064 || B. Foster || bfoster@cisco.com +# RFC3065 || P. Traina, D. McPherson, J. Scudder || danny@ambernetworks.com, jgs@cisco.com +# RFC3066 || H. Alvestrand || Harald@Alvestrand.no +# RFC3067 || J. Arvidsson, A. Cormack, Y. Demchenko, J. Meijer || Jimmy.J.Arvidsson@telia.se, Andrew.Cormack@ukerna.ac.uk, demch@terena.nl, jan.meijer@surfnet.nl +# RFC3068 || C. Huitema || huitema@microsoft.com +# RFC3069 || D. McPherson, B. Dykes || danny@ambernetworks.com, bdykes@onesecure.com +# RFC3070 || V. Rawat, R. Tio, S. Nanji, R. Verma || vrawat@oni.com, tor@redback.com, rverma@dc.com, suhail@redback.com +# RFC3071 || J. Klensin || klensin@jck.com +# RFC3072 || M. Wildgrube || max@wildgrube.com +# RFC3073 || J. Collins || jcollins@bitstream.com +# RFC3074 || B. Volz, S. Gonczi, T. Lemon, R. Stevens || bernie.volz@ericsson.com, steve.gonczi@networkengines.com, ted.lemon@nominum.com, robs@join.com +# RFC3075 || D. Eastlake 3rd, J. Reagle, D. Solo || Donald.Eastlake@motorola.com, reagle@w3.org, dsolo@alum.mit.edu +# RFC3076 || J. Boyer || jboyer@PureEdge.com +# RFC3077 || E. Duros, W. Dabbous, H. Izumiyama, N. Fujii, Y. Zhang || +# RFC3078 || G. Pall, G. Zorn || gurdeep@microsoft.com, gwz@cisco.com +# RFC3079 || G. Zorn || gwz@cisco.com +# RFC3080 || M. Rose || mrose17@gmail.com +# RFC3081 || M. Rose || mrose17@gmail.com +# RFC3082 || J. Kempf, J. Goldschmidt || james.kempf@sun.com, jason.goldschmidt@sun.com +# RFC3083 || R. Woundy || rwoundy@cisco.com +# RFC3084 || K. Chan, J. Seligson, D. Durham, S. Gai, K. McCloghrie, S. Herzog, F. Reichmeyer, R. Yavatkar, A. Smith || khchan@nortelnetworks.com, sgai@cisco.com, Herzog@iphighway.com, kzm@cisco.com, franr@pfn.com, raj.yavatkar@intel.com, andrew@allegronetworks.com +# RFC3085 || A. Coates, D. Allen, D. Rivers-Moore || tony.coates@reuters.com, ho73@dial.pipex.com, daniel.rivers-moore@rivcom.com +# RFC3086 || K. Nichols, B. Carpenter || nichols@packetdesign.com, brian@icair.org +# RFC3087 || B. Campbell, R. Sparks || bcampbell@dynamicsoft.com, rsparks@dynamicsoft.com +# RFC3088 || K. Zeilenga || kurt@openldap.org +# RFC3089 || H. Kitamura || kitamura@da.jp.nec.com +# RFC3090 || E. Lewis || lewis@tislabs.com +# RFC3091 || H. Kennedy || kennedyh@engin.umich.edu +# RFC3092 || D. Eastlake 3rd, C. Manros, E. Raymond || Donald.Eastlake@motorola.com, manros@cp10.es.xerox.com, esr@thyrsus.com +# RFC3093 || M. Gaynor, S. Bradner || +# RFC3094 || D. Sprague, R. Benedyk, D. Brendes, J. Keller || david.sprague@tekelec.com, dan.brendes@tekelec.com, robby.benedyk@tekelec.com, joe.keller@tekelec.com +# RFC3095 || C. Bormann, C. Burmeister, M. Degermark, H. Fukushima, H. Hannu, L-E. Jonsson, R. Hakenberg, T. Koren, K. Le, Z. Liu, A. Martensson, A. Miyazaki, K. Svanbro, T. Wiebke, T. Yoshimura, H. Zheng || cabo@tzi.org, burmeister@panasonic.de, micke@cs.arizona.edu, fukusima@isl.mei.co.jp, hans.hannu@ericsson.com, lars-erik.jonsson@ericsson.com, hakenberg@panasonic.de, tmima@cisco.com, khiem.le@nokia.com, zhigang.liu@nokia.com, anton.martensson@era.ericsson.se, akihiro@isl.mei.co.jp, krister.svanbro@ericsson.com, wiebke@panasonic.de, yoshi@spg.yrp.nttdocomo.co.jp, haihong.zheng@nokia.com +# RFC3096 || M. Degermark, Ed. || +# RFC3097 || R. Braden, L. Zhang || Braden@ISI.EDU, lixia@cs.ucla.edu +# RFC3098 || T. Gavin, D. Eastlake 3rd, S. Hambridge || tedgavin@newsguy.com, Donald.Eastlake@motorola.com, sallyh@ludwig.sc.intel.com +# RFC3099 || S. Ginoza || ginoza@isi.edu +# RFC3100 || || +# RFC3101 || P. Murphy || pmurphy@noc.usgs.net +# RFC3102 || M. Borella, J. Lo, D. Grabelsky, G. Montenegro || mike_borella@commworks.com, yidarlo@yahoo.com, david_grabelsky@commworks.com, gab@sun.com +# RFC3103 || M. Borella, D. Grabelsky, J. Lo, K. Taniguchi || mike_borella@commworks.com, david_grabelsky@commworks.com, yidarlo@yahoo.com, taniguti@ccrl.sj.nec.com +# RFC3104 || G. Montenegro, M. Borella || gab@sun.com, mike_borella@commworks.com +# RFC3105 || J. Kempf, G. Montenegro || gab@sun.com +# RFC3106 || D. Eastlake 3rd, T. Goldstein || Donald.Eastlake@motorola.com, tgoldstein@brodia.com +# RFC3107 || Y. Rekhter, E. Rosen || yakov@juniper.net, erosen@cisco.com +# RFC3108 || R. Kumar, M. Mostafa || rkumar@cisco.com, mmostafa@cisco.com +# RFC3109 || R. Braden, R. Bush, J. Klensin || braden@isi.edu, randy@psg.com, klensin@jck.com +# RFC3110 || D. Eastlake 3rd || Donald.Eastlake@motorola.com +# RFC3111 || E. Guttman || Erik.Guttman@germany.sun.com +# RFC3112 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC3113 || K. Rosenbrock, R. Sanmugam, S. Bradner, J. Klensin || rosenbrock@etsi.fr, sob@harvard.edu, 3GPPContact@etsi.fr +# RFC3114 || W. Nicolls || wnicolls@forsythesolutions.com +# RFC3115 || G. Dommety, K. Leung || gdommety@cisco.com, kleung@cisco.com +# RFC3116 || J. Dunn, C. Martin || Jeffrey.Dunn@worldnet.att.net, Cynthia.E.Martin@worldnet.att.net +# RFC3117 || M. Rose || mrose17@gmail.com +# RFC3118 || R. Droms, Ed., W. Arbaugh, Ed. || +# RFC3119 || R. Finlayson || finlayson@live.com +# RFC3120 || K. Best, N. Walsh || karl.best@oasis-open.org, Norman.Walsh@East.Sun.COM +# RFC3121 || K. Best, N. Walsh || karl.best@oasis-open.org, Norman.Walsh@East.Sun.COM +# RFC3122 || A. Conta || aconta@txc.com +# RFC3123 || P. Koch || pk@TechFak.Uni-Bielefeld.DE +# RFC3124 || H. Balakrishnan, S. Seshan || hari@lcs.mit.edu, srini@cmu.edu +# RFC3125 || J. Ross, D. Pinkas, N. Pope || harri.rasilainen@etsi.fr, ross@secstan.com, Denis.Pinkas@bull.net, pope@secstan.com +# RFC3126 || D. Pinkas, J. Ross, N. Pope || harri.rasilainen@etsi.fr, Denis.Pinkas@bull.net, ross@secstan.com, pope@secstan.com +# RFC3127 || D. Mitton, M. St.Johns, S. Barkley, D. Nelson, B. Patil, M. Stevens, B. Wolff || dmitton@nortelnetworks.com, stjohns@rainmakertechnologies.com, stuartb@uu.net, dnelson@enterasys.com, Basavaraj.Patil@nokia.com, mstevens@ellacoya.com, barney@databus.com +# RFC3128 || I. Miller || Ian_Miller@singularis.ltd.uk +# RFC3129 || M. Thomas || mat@cisco.com +# RFC3130 || E. Lewis || +# RFC3131 || S. Bradner, P. Calhoun, H. Cuschieri, S. Dennett, G. Flynn, M. Lipford, M. McPheters || sob@harvard.edu, pcalhoun@eng.sun.com, hcuschie@tia.eia.org, S.Dennett@motorola.com, gerry.flynn@verizonwireless.com, mjmcpheters@lucent.com +# RFC3132 || J. Kempf || +# RFC3133 || J. Dunn, C. Martin || +# RFC3134 || J. Dunn, C. Martin || +# RFC3135 || J. Border, M. Kojo, J. Griner, G. Montenegro, Z. Shelby || border@hns.com, kojo@cs.helsinki.fi, jgriner@grc.nasa.gov, gab@sun.com, zach.shelby@ee.oulu.fi +# RFC3136 || L. Slutsman, Ed., I. Faynberg, H. Lu, M. Weissman || faynberg@lucent.com, huilanlu@lucent.com, maw1@lucent.com, lslutsman@att.com +# RFC3137 || A. Retana, L. Nguyen, R. White, A. Zinin, D. McPherson || aretana@cisco.com, lhnguyen@cisco.com, riw@cisco.com, azinin@nexsi.com, danny@ambernetworks.com +# RFC3138 || D. Meyer || dmm@sprint.net +# RFC3139 || L. Sanchez, K. McCloghrie, J. Saperia || kzm@cisco.com, lsanchez@megisto.com, saperia@jdscons.com +# RFC3140 || D. Black, S. Brim, B. Carpenter, F. Le Faucheur || black_david@emc.com, sbrim@cisco.com, brian@icair.org, flefauch@cisco.com +# RFC3141 || T. Hiller, P. Walsh, X. Chen, M. Munson, G. Dommety, S. Sivalingham, B. Lim, P. McCann, H. Shiino, B. Hirschman, S. Manning, R. Hsu, H. Koo, M. Lipford, P. Calhoun, C. Lo, E. Jaques, E. Campbell, Y.Xu,S.Baba,T.Ayaki,T.Seki,A.Hameed || pcalhoun@eng.sun.com, gdommety@cisco.com, tom.hiller@lucent.com, rhsu@qualcomm.com, mlipfo01@sprintspectrum.com, serge@awardsolutions.com, mccap@lucent.com, mmunson@gte.net, hskoo@sta.samsung.com, walshp@lucent.com, yxu@watercove.com, qa4053@email.mot.com, ejaques@akamail.com, s.sivalingham@ericsson.com, xing.chen@usa.alcatel.com, bklim@lge.com, hshiino@lucent.com, sbaba@tari.toshiba.com, ayaki@ddi.co.jp, Charles.Lo@vodafone-us.com, t-seki@kddi.com +# RFC3142 || J. Hagino, K. Yamamoto || itojun@iijlab.net, kazu@iijlab.net +# RFC3143 || I. Cooper, J. Dilley || icooper@equinix.com, jad@akamai.com +# RFC3144 || D. Romascanu || dromasca@gmail.com +# RFC3145 || R. Verma, M. Verma, J. Carlson || rverma@dc.com, Madhvi_Verma@3com.com, james.d.carlson@sun.com +# RFC3146 || K. Fujisawa, A. Onoe || fujisawa@sm.sony.co.jp, onoe@sm.sony.co.jp +# RFC3147 || P. Christian || christi@nortelnetworks.com +# RFC3148 || M. Mathis, M. Allman || mathis@psc.edu, mallman@bbn.com +# RFC3149 || A. Srinath, G. Levendel, K. Fritz, R. Kalyanaram || Ashok.Srinath@sylantro.com, Gil.Levendel@sylantro.com, Kent.Fritz@sylantro.com, Raghuraman.Kal@wipro.com +# RFC3150 || S. Dawkins, G. Montenegro, M. Kojo, V. Magret || spencer.dawkins@fnc.fujitsu.com, gab@sun.com, kojo@cs.helsinki.fi, vincent.magret@alcatel.com +# RFC3151 || N. Walsh, J. Cowan, P. Grosso || Norman.Walsh@East.Sun.COM, jcowan@reutershealth.com, pgrosso@arbortext.com +# RFC3152 || R. Bush || randy@psg.com +# RFC3153 || R. Pazhyannur, I. Ali, C. Fox || pazhynnr@cig.mot.com, fia225@email.mot.com, fox@cisco.com +# RFC3154 || J. Kempf, C. Castelluccia, P. Mutaf, N. Nakajima, Y. Ohba, R. Ramjee, Y. Saifullah, B. Sarikaya, X. Xu || James.Kempf@Sun.COM, pars.mutaf@inria.fr, claude.castelluccia@inria.fr, nnakajima@tari.toshiba.com, yohba@tari.toshiba.com, ramjee@bell-labs.com, Yousuf.Saifullah@nokia.com, Behcet.Sarikaya@usa.alcatel.com +# RFC3155 || S. Dawkins, G. Montenegro, M. Kojo, V. Magret, N. Vaidya || spencer.dawkins@fnc.fujitsu.com, gab@sun.com, kojo@cs.helsinki.fi, vincent.magret@alcatel.com +# RFC3156 || M. Elkins, D. Del Torto, R. Levien, T. Roessler || +# RFC3157 || A. Arsenault, S. Farrell || aarsenault@dvnet.com, stephen.farrell@baltimore.ie +# RFC3158 || C. Perkins, J. Rosenberg, H. Schulzrinne || csp@isi.edu, jdrosen@dynamicsoft.com, schulzrinne@cs.columbia.edu +# RFC3159 || K. McCloghrie, M. Fine, J. Seligson, K. Chan, S. Hahn, R. Sahita, A. Smith, F. Reichmeyer || mfine@cisco.com, jseligso@nortelnetworks.com, khchan@nortelnetworks.com, scott.hahn@intel.com, ravi.sahita@intel.com, andrew@allegronetworks.com, franr@pfn.com +# RFC3160 || S. Harris || +# RFC3161 || C. Adams, P. Cain, D. Pinkas, R. Zuccherato || cadams@entrust.com, pcain@bbn.com, Denis.Pinkas@bull.net, robert.zuccherato@entrust.com +# RFC3162 || B. Aboba, G. Zorn, D. Mitton || bernarda@microsoft.com, gwz@cisco.com +# RFC3163 || R. Zuccherato, M. Nystrom || robert.zuccherato@entrust.com, magnus@rsasecurity.com +# RFC3164 || C. Lonvick || clonvick@cisco.com +# RFC3165 || D. Levi, J. Schoenwaelder || +# RFC3166 || D. Meyer, J. Scudder || dmm@sprint.net, jgs@cisco.com +# RFC3167 || D. Meyer, J. Scudder || dmm@sprint.net, jgs@cisco.com +# RFC3168 || K. Ramakrishnan, S. Floyd, D. Black || kk@teraoptic.com, floyd@aciri.org, black_david@emc.com +# RFC3169 || M. Beadles, D. Mitton || dmitton@nortelnetworks.com +# RFC3170 || B. Quinn, K. Almeroth || bquinn@celoxnetworks.com, almeroth@cs.ucsb.edu +# RFC3171 || Z. Albanna, K. Almeroth, D. Meyer, M. Schipper || zaid@juniper.net, almeroth@cs.ucsb.edu, dmm@sprint.net, iana@iana.org +# RFC3172 || G. Huston, Ed. || +# RFC3173 || A. Shacham, B. Monsour, R. Pereira, M. Thomas || shacham@shacham.net, bob@bobmonsour.com, royp@cisco.com, matt@3am-software.com +# RFC3174 || D. Eastlake 3rd, P. Jones || Donald.Eastlake@motorola.com, paulej@packetizer.com +# RFC3175 || F. Baker, C. Iturralde, F. Le Faucheur, B. Davie || fred@cisco.com, cei@cisco.com, flefauch@cisco.com, bdavie@cisco.com +# RFC3176 || P. Phaal, S. Panchen, N. McKee || peter_phaal@INMON.COM, sonia_panchen@INMON.COM, neil_mckee@INMON.COM +# RFC3177 || IAB, IESG || +# RFC3178 || J. Hagino, H. Snyder || itojun@iijlab.net, hal@vailsys.com +# RFC3179 || J. Schoenwaelder, J. Quittek || schoenw@ibr.cs.tu-bs.de, quittek@ccrle.nec.de +# RFC3180 || D. Meyer, P. Lothberg || dmm@sprint.net, roll@sprint.net +# RFC3181 || S. Herzog || herzog@policyconsulting.com +# RFC3182 || S. Yadav, R. Yavatkar, R. Pabbati, P. Ford, T. Moore, S. Herzog, R. Hess || Satyendra.Yadav@intel.com, Raj.Yavatkar@intel.com, rameshpa@microsoft.com, peterf@microsoft.com, timmoore@microsoft.com, herzog@policyconsulting.com, rodney.hess@intel.com +# RFC3183 || T. Dean, W. Ottaway || tbdean@QinetiQ.com, wjottaway@QinetiQ.com +# RFC3184 || S. Harris || srh@merit.edu +# RFC3185 || S. Farrell, S. Turner || stephen.farrell@baltimore.ie, turners@ieca.com +# RFC3186 || S. Shimizu, T. Kawano, K. Murakami, E. Beier || shimizu@ntt-20.ecl.net, kawano@core.ecl.net, murakami@ntt-20.ecl.net, Beier@bina.de +# RFC3187 || J. Hakala, H. Walravens || juha.hakala@helsinki.fi, hartmut.walravens@sbb.spk-berlin.de +# RFC3188 || J. Hakala || juha.hakala@helsinki.fi +# RFC3189 || K. Kobayashi, A. Ogawa, S. Casner, C. Bormann || ikob@koganei.wide.ad.jp, akimichi@sfc.wide.ad.jp, casner@acm.org, cabo@tzi.org +# RFC3190 || K. Kobayashi, A. Ogawa, S. Casner, C. Bormann || ikob@koganei.wide.ad.jp, akimichi@sfc.wide.ad.jp, casner@acm.org, cabo@tzi.org +# RFC3191 || C. Allocchio || +# RFC3192 || C. Allocchio || +# RFC3193 || B. Patel, B. Aboba, W. Dixon, G. Zorn, S. Booth || baiju.v.patel@intel.com, bernarda@microsoft.com, wdixon@microsoft.com, gwz@cisco.com, ebooth@cisco.com +# RFC3194 || A. Durand, C. Huitema || +# RFC3195 || D. New, M. Rose || dnew@san.rr.com, mrose17@gmail.com +# RFC3196 || T. Hastings, C. Manros, P. Zehler, C. Kugler, H. Holst || tom.hastings@alum.mit.edu, Kugler@us.ibm.com, hh@I-data.com, Peter.Zehler@xerox.com +# RFC3197 || R. Austein || sra@hactrn.net +# RFC3198 || A. Westerinen, J. Schnizlein, J. Strassner, M. Scherling, B. Quinn, S. Herzog, A. Huynh, M. Carlson, J. Perry, S. Waldbusser || andreaw@cisco.com, john.schnizlein@cisco.com, john.strassner@intelliden.com, mscherling@xcert.com, bquinn@celoxnetworks.com, jay.perry@netapp.com, herzog@PolicyConsulting.com, mark.carlson@sun.com, waldbusser@nextbeacon.com +# RFC3199 || S. Ginoza || ginoza@isi.edu +# RFC3200 || || +# RFC3201 || R. Steinberger, O. Nicklass || robert.steinberger@fnc.fujitsu.com, Orly_n@rad.co.il +# RFC3202 || R. Steinberger, O. Nicklass || robert.steinberger@fnc.fujitsu.com, Orly_n@rad.co.il +# RFC3203 || Y. T'Joens, C. Hublet, P. De Schrijver || yves.tjoens@alcatel.be, p2@mind.be, Christian.Hublet@alcatel.be +# RFC3204 || E. Zimmerer, J. Peterson, A. Vemuri, L. Ong, F. Audet, M. Watson, M. Zonoun || eric_zimmerer@yahoo.com, Aparna.Vemuri@Qwest.com, jon.peterson@neustar.com, lyndon_ong@yahoo.com, mzonoun@nortelnetworks.com, audet@nortelnetworks.com, mwatson@nortelnetworks.com +# RFC3205 || K. Moore || moore@cs.utk.edu +# RFC3206 || R. Gellens || randy@qualcomm.com +# RFC3207 || P. Hoffman || phoffman@imc.org +# RFC3208 || T. Speakman, J. Crowcroft, J. Gemmell, D. Farinacci, S. Lin, D. Leshchiner, M. Luby, T. Montgomery, L. Rizzo, A. Tweedly, N. Bhaskar, R. Edmonstone, R. Sumanasekera, L. Vicisano || speakman@cisco.com, dino@procket.com, steven@juniper.net, agt@cisco.com, nbhaskar@cisco.com, redmonst@cisco.com, rajitha@cisco.com, lorenzo@cisco.com, j.crowcroft@cs.ucl.ac.uk, jgemmell@microsoft.com, dleshc@tibco.com, luby@digitalfountain.com, todd@talarian.com, luigi@iet.unipi.it +# RFC3209 || D. Awduche, L. Berger, D. Gan, T. Li, V. Srinivasan, G. Swallow || awduche@movaz.com, lberger@movaz.com, dhg@juniper.net, tli@procket.com, vsriniva@cosinecom.com, swallow@cisco.com +# RFC3210 || D. Awduche, A. Hannan, X. Xiao || awduche@movaz.com, alan@routingloop.com, xxiao@photuris.com +# RFC3211 || P. Gutmann || pgut001@cs.auckland.ac.nz +# RFC3212 || B. Jamoussi, Ed., L. Andersson, R. Callon, R. Dantu, L. Wu, P. Doolan, T. Worster, N. Feldman, A. Fredette, M. Girish, E. Gray, J. Heinanen, T. Kilty, A. Malis || loa.andersson@utfors.se, rcallon@juniper.net, rdantu@netrake.com, pdoolan@acm.org, Nkf@us.ibm.com, afredette@charter.net, eric.gray@sandburst.com, jh@song.fi, tim-kilty@mediaone.net, Andy.Malis@vivacenetworks.com, muckai@atoga.com, fsb@thefsb.org, liwwu@cisco.com +# RFC3213 || J. Ash, M. Girish, E. Gray, B. Jamoussi, G. Wright || gash@att.com, eric.gray@sandburst.com, gwright@nortelnetworks.com, muckai@atoga.com, Jamoussi@nortelnetworks.com +# RFC3214 || J. Ash, Y. Lee, P. Ashwood-Smith, B. Jamoussi, D. Fedyk, D. Skalecki, L. Li || gash@att.com, jamoussi@NortelNetworks.com, petera@NortelNetworks.com, dareks@nortelnetworks.com, ylee@ceterusnetworks.com, lili@ss8networks.com, dwfedyk@nortelnetworks.com +# RFC3215 || C. Boscher, P. Cheval, L. Wu, E. Gray || christophe.boscher@alcatel.fr, pierrick.cheval@space.alcatel.fr, liwwu@cisco.com, eric.gray@sandburst.com +# RFC3216 || C. Elliott, D. Harrington, J. Jason, J. Schoenwaelder, F. Strauss, W. Weiss || chelliot@cisco.com, dbh@enterasys.com, jamie.jason@intel.com, schoenw@ibr.cs.tu-bs.de, strauss@ibr.cs.tu-bs.de, wweiss@ellacoya.com +# RFC3217 || R. Housley || rhousley@rsasecurity.com +# RFC3218 || E. Rescorla || ekr@rtfm.com +# RFC3219 || J. Rosenberg, H. Salama, M. Squire || jdrosen@dynamicsoft.com, hsalama@cisco.com, mattsquire@acm.org +# RFC3220 || C. Perkins, Ed. || charliep@iprg.nokia.com +# RFC3221 || G. Huston || +# RFC3222 || G. Trotter || Guy_Trotter@agilent.com +# RFC3223 || || +# RFC3224 || E. Guttman || erik.guttman@sun.com +# RFC3225 || D. Conrad || david.conrad@nominum.com +# RFC3226 || O. Gudmundsson || ogud@ogud.com +# RFC3227 || D. Brezinski, T. Killalea || dbrezinski@In-Q-Tel.org, tomk@neart.org +# RFC3228 || B. Fenner || fenner@research.att.com +# RFC3229 || J. Mogul, B. Krishnamurthy, F. Douglis, A. Feldmann, Y. Goland, A. van Hoff, D. Hellerstein || +# RFC3230 || J. Mogul, A. Van Hoff || JeffMogul@acm.org, avh@marimba.com +# RFC3231 || D. Levi, J. Schoenwaelder || +# RFC3232 || J. Reynolds, Ed. || rfc-editor@rfc-editor.org +# RFC3233 || P. Hoffman, S. Bradner || +# RFC3234 || B. Carpenter, S. Brim || brian@hursley.ibm.com, sbrim@cisco.com +# RFC3235 || D. Senie || dts@senie.com +# RFC3236 || M. Baker, P. Stark || mbaker@planetfred.com, distobj@acm.org, Peter.Stark@ecs.ericsson.com +# RFC3237 || M. Tuexen, Q. Xie, R. Stewart, M. Shore, L. Ong, J. Loughney, M. Stillman || Michael.Tuexen@icn.siemens.de, qxie1@email.mot.com, randall@lakerest.net, mshore@cisco.com, lyong@ciena.com, john.loughney@nokia.com, maureen.stillman@nokia.com +# RFC3238 || S. Floyd, L. Daigle || iab@iab.org +# RFC3239 || C. Kugler, H. Lewis, T. Hastings || kugler@us.ibm.com, tom.hastings@alum.mit.edu, harryl@us.ibm.com +# RFC3240 || D. Clunie, E. Cordonnier || dclunie@dclunie.com, emmanuel.cordonnier@etiam.com +# RFC3241 || C. Bormann || cabo@tzi.org +# RFC3242 || L-E. Jonsson, G. Pelletier || lars-erik.jonsson@ericsson.com, ghyslain.pelletier@epl.ericsson.se +# RFC3243 || L-E. Jonsson || lars-erik.jonsson@ericsson.com +# RFC3244 || M. Swift, J. Trostle, J. Brezak || mikesw@cs.washington.edu, john3725@world.std.com, jbrezak@microsoft.com +# RFC3245 || J. Klensin, Ed., IAB || iab@iab.org, sob@harvard.edu, paf@cisco.com +# RFC3246 || B. Davie, A. Charny, J.C.R. Bennet, K. Benson, J.Y. Le Boudec, W. Courtney, S. Davari, V. Firoiu, D. Stiliadis || bsd@cisco.com, acharny@cisco.com, jcrb@motorola.com, Kent.Benson@tellabs.com, jean-yves.leboudec@epfl.ch, bill.courtney@trw.com, shahram_davari@pmc-sierra.com, vfiroiu@nortelnetworks.com, stiliadi@bell-labs.com +# RFC3247 || A. Charny, J. Bennet, K. Benson, J. Boudec, A. Chiu, W. Courtney, S. Davari, V. Firoiu, C. Kalmanek, K. Ramakrishnan || acharny@cisco.com, jcrb@motorola.com, Kent.Benson@tellabs.com, jean-yves.leboudec@epfl.ch, angela.chiu@celion.com, bill.courtney@trw.com, shahram_davari@pmc-sierra.com, vfiroiu@nortelnetworks.com, crk@research.att.com, kk@teraoptic.com +# RFC3248 || G. Armitage, B. Carpenter, A. Casati, J. Crowcroft, J. Halpern, B. Kumar, J. Schnizlein || +# RFC3249 || V. Cancio, M. Moldovan, H. Tamura, D. Wing || vcancio@pacbell.net, mmoldovan@g3nova.com, tamura@toda.ricoh.co.jp, dwing-ietf@fuggles.com +# RFC3250 || L. McIntyre, G. Parsons, J. Rafferty || lmcintyre@pahv.xerox.com, gparsons@nortelnetworks.com, jraff@brooktrout.com +# RFC3251 || B. Rajagopalan || braja@tellium.com +# RFC3252 || H. Kennedy || kennedyh@engin.umich.edu +# RFC3253 || G. Clemm, J. Amsden, T. Ellison, C. Kaler, J. Whitehead || geoffrey.clemm@rational.com, jamsden@us.ibm.com, tim_ellison@uk.ibm.com, ckaler@microsoft.com, ejw@cse.ucsc.edu +# RFC3254 || H. Alvestrand || Harald@alvestrand.no +# RFC3255 || N. Jones, C. Murton || nrjones@agere.com, murton@nortelnetworks.com +# RFC3256 || D. Jones, R. Woundy || doug@yas.com, rwoundy@broadband.att.com +# RFC3257 || L. Coene || +# RFC3258 || T. Hardie || Ted.Hardie@nominum.com +# RFC3259 || J. Ott, C. Perkins, D. Kutscher || jo@tzi.uni-bremen.de, csp@isi.edu, dku@tzi.uni-bremen.de +# RFC3260 || D. Grossman || dan@dma.isg.mot.com +# RFC3261 || J. Rosenberg, H. Schulzrinne, G. Camarillo, A. Johnston, J. Peterson, R. Sparks, M. Handley, E. Schooler || jdrosen@dynamicsoft.com, schulzrinne@cs.columbia.edu, Gonzalo.Camarillo@ericsson.com, alan.johnston@wcom.com, jon.peterson@neustar.com, rsparks@dynamicsoft.com, mjh@icir.org, schooler@research.att.com +# RFC3262 || J. Rosenberg, H. Schulzrinne || jdrosen@dynamicsoft.com, schulzrinne@cs.columbia.edu +# RFC3263 || J. Rosenberg, H. Schulzrinne || jdrosen@dynamicsoft.com, schulzrinne@cs.columbia.edu +# RFC3264 || J. Rosenberg, H. Schulzrinne || jdrosen@dynamicsoft.com, schulzrinne@cs.columbia.edu +# RFC3265 || A. B. Roach || adam@dynamicsoft.com +# RFC3266 || S. Olson, G. Camarillo, A. B. Roach || seanol@microsoft.com, Gonzalo.Camarillo@ericsson.com, adam@dynamicsoft.com +# RFC3267 || J. Sjoberg, M. Westerlund, A. Lakaniemi, Q. Xie || Johan.Sjoberg@ericsson.com, Magnus.Westerlund@ericsson.com, ari.lakaniemi@nokia.com, qxie1@email.mot.com +# RFC3268 || P. Chown || pc@skygate.co.uk +# RFC3269 || R. Kermode, L. Vicisano || Roger.Kermode@motorola.com, lorenzo@cisco.com +# RFC3270 || F. Le Faucheur, L. Wu, B. Davie, S. Davari, P. Vaananen, R. Krishnan, P. Cheval, J. Heinanen || flefauch@cisco.com, liwwu@cisco.com, bsd@cisco.com, davari@ieee.org, pasi.vaananen@nokia.com, ram@axiowave.com, pierrick.cheval@space.alcatel.fr, jh@song.fi +# RFC3271 || V. Cerf || vinton.g.cerf@wcom.com +# RFC3272 || D. Awduche, A. Chiu, A. Elwalid, I. Widjaja, X. Xiao || awduche@movaz.com, angela.chiu@celion.com, anwar@lucent.com, iwidjaja@research.bell-labs.com, xipeng@redback.com +# RFC3273 || S. Waldbusser || waldbusser@nextbeacon.com +# RFC3274 || P. Gutmann || pgut001@cs.auckland.ac.nz +# RFC3275 || D. Eastlake 3rd, J. Reagle, D. Solo || Donald.Eastlake@motorola.com, reagle@w3.org, dsolo@alum.mit.edu +# RFC3276 || B. Ray, R. Abbi || rray@pesa.com, Rajesh.Abbi@alcatel.com +# RFC3277 || D. McPherson || danny@tcb.net +# RFC3278 || S. Blake-Wilson, D. Brown, P. Lambert || sblakewi@certicom.com, dbrown@certicom.com, plambert@sprintmail.com +# RFC3279 || L. Bassham, W. Polk, R. Housley || tim.polk@nist.gov, rhousley@rsasecurity.com, lbassham@nist.gov +# RFC3280 || R. Housley, W. Polk, W. Ford, D. Solo || rhousley@rsasecurity.com, wford@verisign.com, wpolk@nist.gov, dsolo@alum.mit.edu +# RFC3281 || S. Farrell, R. Housley || stephen.farrell@baltimore.ie, rhousley@rsasecurity.com +# RFC3282 || H. Alvestrand || Harald@Alvestrand.no +# RFC3283 || B. Mahoney, G. Babics, A. Taler || bobmah@mit.edu, georgeb@steltor.com, alex@0--0.org +# RFC3284 || D. Korn, J. MacDonald, J. Mogul, K. Vo || kpv@research.att.com, dgk@research.att.com, JeffMogul@acm.org, jmacd@cs.berkeley.edu +# RFC3285 || M. Gahrns, T. Hain || mikega@microsoft.com, ahain@cisco.com +# RFC3286 || L. Ong, J. Yoakum || lyong@ciena.com, yoakum@nortelnetworks.com +# RFC3287 || A. Bierman || andy@yumaworks.com +# RFC3288 || E. O'Tuathail, M. Rose || eamon.otuathail@clipcode.com, mrose17@gmail.com +# RFC3289 || F. Baker, K. Chan, A. Smith || fred@cisco.com, khchan@nortelnetworks.com, ah_smith@acm.org +# RFC3290 || Y. Bernet, S. Blake, D. Grossman, A. Smith || ybernet@msn.com, steven.blake@ericsson.com, dan@dma.isg.mot.com, ah_smith@acm.org +# RFC3291 || M. Daniele, B. Haberman, S. Routhier, J. Schoenwaelder || md@world.std.com, bkhabs@nc.rr.com, sar@epilogue.com, schoenw@ibr.cs.tu-bs.de +# RFC3292 || A. Doria, F. Hellstrand, K. Sundell, T. Worster || avri@acm.org, fiffi@nortelnetworks.com, ksundell@nortelnetworks.com, fsb@thefsb.org +# RFC3293 || T. Worster, A. Doria, J. Buerkle || fsb@thefsb.org, avri@acm.com, Joachim.Buerkle@nortelnetworks.com +# RFC3294 || A. Doria, K. Sundell || avri@acm.org, sundell@nortelnetworks.com +# RFC3295 || H. Sjostrand, J. Buerkle, B. Srinivasan || hans@ipunplugged.com, joachim.buerkle@nortelnetworks.com, balaji@cplane.com +# RFC3296 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC3297 || G. Klyne, R. Iwazaki, D. Crocker || GK@ACM.ORG, iwa@rdl.toshibatec.co.jp, dcrocker@brandenburg.com +# RFC3298 || I. Faynberg, J. Gato, H. Lu, L. Slutsman || lslutsman@att.com, faynberg@lucent.com, jgato@airtel.es, huilanlu@lucent.com +# RFC3299 || S. Ginoza || ginoza@isi.edu +# RFC3300 || J. Reynolds, R. Braden, S. Ginoza, A. De La Cruz || +# RFC3301 || Y. T'Joens, P. Crivellari, B. Sales || paolo.crivellari@belgacom.be +# RFC3302 || G. Parsons, J. Rafferty || gparsons@nortelnetworks.com, jraff@brooktrout.com +# RFC3303 || P. Srisuresh, J. Kuthan, J. Rosenberg, A. Molitor, A. Rayhan || srisuresh@yahoo.com, kuthan@fokus.fhg.de, jdrosen@dynamicsoft.com, amolitor@visi.com, rayhan@ee.ryerson.ca +# RFC3304 || R. P. Swale, P. A. Mart, P. Sijben, S. Brim, M. Shore || richard.swale@bt.com, paul.sijben@picopoint.com, philip.mart@marconi.com, sbrim@cisco.com, mshore@cisco.com +# RFC3305 || M. Mealling, Ed., R. Denenberg, Ed. || michael@verisignlabs.com, rden@loc.gov +# RFC3306 || B. Haberman, D. Thaler || bkhabs@nc.rr.com, dthaler@microsoft.com +# RFC3307 || B. Haberman || bkhabs@nc.rr.com +# RFC3308 || P. Calhoun, W. Luo, D. McPherson, K. Peirce || pcalhoun@bstormnetworks.com, luo@cisco.com, danny@tcb.net, Ken@malibunetworks.com +# RFC3309 || J. Stone, R. Stewart, D. Otis || jonathan@dsg.stanford.edu, randall@lakerest.net, dotis@sanlight.net +# RFC3310 || A. Niemi, J. Arkko, V. Torvinen || aki.niemi@nokia.com, jari.arkko@ericsson.com, vesa.torvinen@ericsson.fi +# RFC3311 || J. Rosenberg || jdrosen@dynamicsoft.com +# RFC3312 || G. Camarillo, Ed., W. Marshall, Ed., J. Rosenberg || Gonzalo.Camarillo@ericsson.com, wtm@research.att.com, jdrosen@dynamicsoft.com +# RFC3313 || W. Marshall, Ed. || +# RFC3314 || M. Wasserman, Ed. || +# RFC3315 || R. Droms, Ed., J. Bound, B. Volz, T. Lemon, C. Perkins, M. Carney || Jim.Bound@hp.com, volz@metrocast.net, Ted.Lemon@nominum.com, charles.perkins@nokia.com, michael.carney@sun.com +# RFC3316 || J. Arkko, G. Kuijpers, H. Soliman, J. Loughney, J. Wiljakka || jari.arkko@ericsson.com, gerben.a.kuijpers@ted.ericsson.se, john.loughney@nokia.com, hesham.soliman@era.ericsson.se, juha.wiljakka@nokia.com +# RFC3317 || K. Chan, R. Sahita, S. Hahn, K. McCloghrie || khchan@nortelnetworks.com, ravi.sahita@intel.com, scott.hahn@intel.com, kzm@cisco.com +# RFC3318 || R. Sahita, Ed., S. Hahn, K. Chan, K. McCloghrie || ravi.sahita@intel.com, scott.hahn@intel.com, khchan@nortelnetworks.com, kzm@cisco.com +# RFC3319 || H. Schulzrinne, B. Volz || schulzrinne@cs.columbia.edu, volz@metrocast.net +# RFC3320 || R. Price, C. Bormann, J. Christoffersson, H. Hannu, Z. Liu, J. Rosenberg || richard.price@roke.co.uk, cabo@tzi.org, jan.christoffersson@epl.ericsson.se, hans.hannu@epl.ericsson.se, zhigang.c.liu@nokia.com, jdrosen@dynamicsoft.com +# RFC3321 || H. Hannu, J. Christoffersson, S. Forsgren, K.-C. Leung, Z. Liu, R. Price || hans.hannu@epl.ericsson.se, jan.christoffersson@epl.ericsson.se, StefanForsgren@alvishagglunds.se, kcleung@cs.ttu.edu, zhigang.c.liu@nokia.com, richard.price@roke.co.uk +# RFC3322 || H. Hannu || hans.hannu@epl.ericsson.se +# RFC3323 || J. Peterson || jon.peterson@neustar.biz +# RFC3324 || M. Watson || mwatson@nortelnetworks.com +# RFC3325 || C. Jennings, J. Peterson, M. Watson || fluffy@cisco.com, Jon.Peterson@NeuStar.biz, mwatson@nortelnetworks.com +# RFC3326 || H. Schulzrinne, D. Oran, G. Camarillo || schulzrinne@cs.columbia.edu, oran@cisco.com, Gonzalo.Camarillo@ericsson.com +# RFC3327 || D. Willis, B. Hoeneisen || dean.willis@softarmor.com, hoeneisen@switch.ch +# RFC3328 || || +# RFC3329 || J. Arkko, V. Torvinen, G. Camarillo, A. Niemi, T. Haukka || jari.arkko@ericsson.com, vesa.torvinen@ericsson.fi, Gonzalo.Camarillo@ericsson.com, aki.niemi@nokia.com, tao.haukka@nokia.com +# RFC3330 || IANA || iana@iana.org +# RFC3331 || K. Morneault, R. Dantu, G. Sidebottom, B. Bidulock, J. Heitz || kmorneau@cisco.com, rdantu@netrake.com, greg@signatustechnologies.com, bidulock@openss7.org, jheitz@lucent.com +# RFC3332 || G. Sidebottom, Ed., K. Morneault, Ed., J. Pastor-Balbas, Ed. || +# RFC3334 || T. Zseby, S. Zander, C. Carle || zseby@fokus.fhg.de, zander@fokus.fhg.de, carle@fokus.fhg.de +# RFC3335 || T. Harding, R. Drummond, C. Shih || tharding@cyclonecommerce.com, chuck.shih@gartner.com, rik@drummondgroup.com +# RFC3336 || B. Thompson, T. Koren, B. Buffam || brucet@cisco.com, tmima@cisco.com, bruce@seawaynetworks.com +# RFC3337 || B. Thompson, T. Koren, B. Buffam || brucet@cisco.com, bruce@seawaynetworks.com, tmima@cisco.com +# RFC3338 || S. Lee, M-K. Shin, Y-J. Kim, E. Nordmark, A. Durand || syl@pec.etri.re.kr, mkshin@pec.etri.re.kr, yjkim@pec.etri.re.kr, Alain.Durand@sun.com, erik.nordmark@sun.com +# RFC3339 || G. Klyne, C. Newman || chris.newman@sun.com, GK@ACM.ORG +# RFC3340 || M. Rose, G. Klyne, D. Crocker || mrose17@gmail.com, Graham.Klyne@MIMEsweeper.com, dcrocker@brandenburg.com +# RFC3341 || M. Rose, G. Klyne, D. Crocker || mrose17@gmail.com, Graham.Klyne@MIMEsweeper.com, dcrocker@brandenburg.com +# RFC3342 || E. Dixon, H. Franklin, J. Kint, G. Klyne, D. New, S. Pead, M. Rose, M. Schwartz || Graham.Klyne@MIMEsweeper.com, mrose17@gmail.com, schwartz@CodeOnTheRoad.com, edixon@myrealbox.com, huston@franklin.ro, d20@icosahedron.org, dnew@san.rr.com, spead@fiber.net +# RFC3343 || M. Rose, G. Klyne, D. Crocker || mrose17@gmail.com, gk@ninebynine.org, dcrocker@brandenburg.com +# RFC3344 || C. Perkins, Ed. || Basavaraj.Patil@nokia.com, PRoberts@MEGISTO.com, charliep@iprg.nokia.com +# RFC3345 || D. McPherson, V. Gill, D. Walton, A. Retana || danny@tcb.net, vijay@umbc.edu, dwalton@cisco.com, aretana@cisco.com +# RFC3346 || J. Boyle, V. Gill, A. Hannan, D. Cooper, D. Awduche, B. Christian, W.S. Lai || jboyle@pdnets.com, vijay@umbc.edu, alan@routingloop.com, dcooper@gblx.net, awduche@movaz.com, blaine@uu.net, wlai@att.com +# RFC3347 || M. Krueger, R. Haagens || marjorie_krueger@hp.com, Randy_Haagens@hp.com, csapuntz@stanford.edu, mbakke@cisco.com +# RFC3348 || M. Gahrns, R. Cheng || mikega@microsoft.com, raych@microsoft.com +# RFC3349 || M. Rose || mrose17@gmail.com +# RFC3351 || N. Charlton, M. Gasson, G. Gybels, M. Spanner, A. van Wijk || nathan@millpark.com, michael.gasson@korusolutions.com, Guido.Gybels@rnid.org.uk, mike.spanner@rnid.org.uk, Arnoud.van.Wijk@eln.ericsson.se +# RFC3352 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC3353 || D. Ooms, B. Sales, W. Livens, A. Acharya, F. Griffoul, F. Ansari || Dirk.Ooms@alcatel.be, Bernard.Sales@alcatel.be, WLivens@colt-telecom.be, arup@us.ibm.com, griffoul@ulticom.com, furquan@dnrc.bell-labs.com +# RFC3354 || D. Eastlake 3rd || Donald.Eastlake@motorola.com +# RFC3355 || A. Singh, R. Turner, R. Tio, S. Nanji || rturner@eng.paradyne.com, tor@redback.com, asingh1@motorola.com, suhail@redback.com +# RFC3356 || G. Fishman, S. Bradner || +# RFC3357 || R. Koodli, R. Ravikanth || rajeev.koodli@nokia.com, rravikanth@axiowave.com +# RFC3358 || T. Przygienda || prz@xebeo.com +# RFC3359 || T. Przygienda || +# RFC3360 || S. Floyd || floyd@icir.org +# RFC3361 || H. Schulzrinne || schulzrinne@cs.columbia.edu +# RFC3362 || G. Parsons || gparsons@nortelnetworks.com +# RFC3363 || R. Bush, A. Durand, B. Fink, O. Gudmundsson, T. Hain || +# RFC3364 || R. Austein || sra@hactrn.net +# RFC3365 || J. Schiller || jis@mit.edu +# RFC3366 || G. Fairhurst, L. Wood || gorry@erg.abdn.ac.uk, lwood@cisco.com +# RFC3367 || N. Popp, M. Mealling, M. Moseley || npopp@verisign.com, michael@verisignlabs.com, marshall@netword.com +# RFC3368 || M. Mealling || michael@verisignlabs.com +# RFC3369 || R. Housley || rhousley@rsasecurity.com +# RFC3370 || R. Housley || rhousley@rsasecurity.com +# RFC3371 || E. Caves, P. Calhoun, R. Wheeler || evan@occamnetworks.com, pcalhoun@bstormnetworks.com +# RFC3372 || A. Vemuri, J. Peterson || Aparna.Vemuri@Qwest.com, jon.peterson@neustar.biz +# RFC3373 || D. Katz, R. Saluja || dkatz@juniper.net, rajesh.saluja@tenetindia.com +# RFC3374 || J. Kempf, Ed. || henrik@levkowetz.com, pcalhoun@bstormnetworks.com, kempf@docomolabs-usa.com, gkenward@nortelnetworks.com, hmsyed@nortelnetworks.com, jmanner@cs.helsinki.fi, madjid.nakhjiri@motorola.com, govind.krishnamurthi@nokia.com, rajeev.koodli@nokia.com, kulwinder.atwal@zucotto.com, mat@cisco.com, mat.horan@comdev.cc, phil_neumiller@3com.com +# RFC3375 || S. Hollenbeck || +# RFC3376 || B. Cain, S. Deering, I. Kouvelas, B. Fenner, A. Thyagarajan || deering@cisco.com, fenner@research.att.com, kouvelas@cisco.com +# RFC3377 || J. Hodges, R. Morgan || Jeff.Hodges@sun.com, rlmorgan@washington.edu +# RFC3378 || R. Housley, S. Hollenbeck || rhousley@rsasecurity.com, shollenbeck@verisign.com +# RFC3379 || D. Pinkas, R. Housley || Denis.Pinkas@bull.net, rhousley@rsasecurity.com +# RFC3380 || T. Hastings, R. Herriot, C. Kugler, H. Lewis || kugler@us.ibm.com, tom.hastings@alum.mit.edu, bob@Herriot.com, harryl@us.ibm.com +# RFC3381 || T. Hastings, H. Lewis, R. Bergman || tom.hastings@alum.mit.edu, harryl@us.ibm.com, rbergma@hitachi-hkis.com +# RFC3382 || R. deBry, T. Hastings, R. Herriot, K. Ocke, P. Zehler || debryro@uvsc.edu, tom.hastings@alum.mit.edu, bob@herriot.com, KOcke@crt.xerox.com, Peter.Zehler@xerox.com +# RFC3383 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC3384 || E. Stokes, R. Weiser, R. Moats, R. Huber || rweiser@trustdst.com, stokese@us.ibm.com, rmoats@lemurnetworks.net, rvh@att.com +# RFC3385 || D. Sheinwald, J. Satran, P. Thaler, V. Cavanna || julian_satran@il.ibm.com, Dafna_Sheinwald@il.ibm.com, pat_thaler@agilent.com, vince_cavanna@agilent.com +# RFC3386 || W. Lai, Ed., D. McDysan, Ed. || +# RFC3387 || M. Eder, H. Chaskar, S. Nag || Michael.eder@nokia.com, thinker@monmouth.com, hemant.chaskar@nokia.com +# RFC3388 || G. Camarillo, G. Eriksson, J. Holler, H. Schulzrinne || Gonzalo.Camarillo@ericsson.com, Jan.Holler@era.ericsson.se, Goran.AP.Eriksson@era.ericsson.se, schulzrinne@cs.columbia.edu +# RFC3389 || R. Zopf || zopf@lucent.com +# RFC3390 || M. Allman, S. Floyd, C. Partridge || mallman@bbn.com, floyd@icir.org, craig@bbn.com +# RFC3391 || R. Herriot || bob@herriot.com +# RFC3392 || R. Chandra, J. Scudder || rchandra@redback.com, jgs@cisco.com +# RFC3393 || C. Demichelis, P. Chimento || carlo.demichelis@tilab.com, chimento@torrentnet.com +# RFC3394 || J. Schaad, R. Housley || jimsch@exmsft.com, rhousley@rsasecurity.com +# RFC3395 || A. Bierman, C. Bucci, R. Dietz, A. Warth || andy@yumaworks.com, cbucci@cisco.com, rdietz@hifn.com, dahoss@earthlink.net +# RFC3396 || T. Lemon, S. Cheshire || mellon@nominum.com, rfc@stuartcheshire.org +# RFC3397 || B. Aboba, S. Cheshire || bernarda@microsoft.com, rfc@stuartcheshire.org +# RFC3398 || G. Camarillo, A. B. Roach, J. Peterson, L. Ong || Gonzalo.Camarillo@Ericsson.com, adam@dynamicsoft.com, jon.peterson@neustar.biz, lyOng@ciena.com +# RFC3400 || || +# RFC3401 || M. Mealling || michael@neonym.net +# RFC3402 || M. Mealling || michael@neonym.net +# RFC3403 || M. Mealling || michael@neonym.net +# RFC3404 || M. Mealling || michael@neonym.net +# RFC3405 || M. Mealling || michael@neonym.net +# RFC3406 || L. Daigle, D. van Gulik, R. Iannella, P. Faltstrom || leslie@thinkingcat.com, renato@iprsystems.com, paf@cisco.com +# RFC3407 || F. Andreasen || fandreas@cisco.com +# RFC3408 || Z. Liu, K. Le || khiem.le@nokia.com +# RFC3409 || K. Svanbro || krister.svanbro@ericsson.com +# RFC3410 || J. Case, R. Mundy, D. Partain, B. Stewart || +# RFC3411 || D. Harrington, R. Presuhn, B. Wijnen || +# RFC3412 || J. Case, D. Harrington, R. Presuhn, B. Wijnen || +# RFC3413 || D. Levi, P. Meyer, B. Stewart || +# RFC3414 || U. Blumenthal, B. Wijnen || +# RFC3415 || B. Wijnen, R. Presuhn, K. McCloghrie || +# RFC3416 || R. Presuhn, Ed. || +# RFC3417 || R. Presuhn, Ed. || +# RFC3418 || R. Presuhn, Ed. || +# RFC3419 || M. Daniele, J. Schoenwaelder || md@world.std.com, schoenw@ibr.cs.tu-bs.de +# RFC3420 || R. Sparks || rsparks@dynamicsoft.com +# RFC3421 || W. Zhao, H. Schulzrinne, E. Guttman, C. Bisdikian, W. Jerome || zwb@cs.columbia.edu, hgs@cs.columbia.edu, Erik.Guttman@sun.com, bisdik@us.ibm.com, wfj@us.ibm.com +# RFC3422 || O. Okamoto, M. Maruyama, T. Sajima || okamoto.osamu@lab.ntt.co.jp, mitsuru@core.ecl.net, tjs@sun.com +# RFC3423 || K. Zhang, E. Elkin || kevinzhang@ieee.org, eitan@xacct.com +# RFC3424 || L. Daigle, Ed., IAB || iab@iab.org +# RFC3425 || D. Lawrence || tale@nominum.com +# RFC3426 || S. Floyd || iab@iab.org +# RFC3427 || A. Mankin, S. Bradner, R. Mahy, D. Willis, J. Ott, B. Rosen || mankin@psg.com, sob@harvard.edu, rohan@cisco.com, dean.willis@softarmor.com, brian.rosen@marconi.com, jo@ipdialog.com +# RFC3428 || B. Campbell, Ed., J. Rosenberg, H. Schulzrinne, C. Huitema, D. Gurle || bcampbell@dynamicsoft.com, jdrosen@dynamicsoft.com, schulzrinne@cs.columbia.edu, huitema@microsoft.com, dgurle@microsoft.com +# RFC3429 || H. Ohta || ohta.hiroshi@lab.ntt.co.jp +# RFC3430 || J. Schoenwaelder || schoenw@ibr.cs.tu-bs.de +# RFC3431 || W. Segmuller || whs@watson.ibm.com +# RFC3432 || V. Raisanen, G. Grotefeld, A. Morton || Vilho.Raisanen@nokia.com, g.grotefeld@motorola.com, acmorton@att.com +# RFC3433 || A. Bierman, D. Romascanu, K.C. Norseth || andy@yumaworks.com, dromasca@gmail.com , kenyon.c.norseth@L-3com.com +# RFC3434 || A. Bierman, K. McCloghrie || andy@yumaworks.com, kzm@cisco.com +# RFC3435 || F. Andreasen, B. Foster || fandreas@cisco.com, bfoster@cisco.com +# RFC3436 || A. Jungmaier, E. Rescorla, M. Tuexen || ajung@exp-math.uni-essen.de, ekr@rtfm.com, Michael.Tuexen@siemens.com +# RFC3437 || W. Palter, W. Townsley || mark@townsley.net, palter.ietf@zev.net +# RFC3438 || W. Townsley || mark@townsley.net +# RFC3439 || R. Bush, D. Meyer || randy@psg.com, dmm@maoz.com +# RFC3440 || F. Ly, G. Bathrick || faye@pedestalnetworks.com, greg.bathrick@nokia.com +# RFC3441 || R. Kumar || rkumar@cisco.com +# RFC3442 || T. Lemon, S. Cheshire, B. Volz || Ted.Lemon@nominum.com, rfc@stuartcheshire.org, bernie.volz@ericsson.com +# RFC3443 || P. Agarwal, B. Akyol || puneet@acm.org, bora@cisco.com +# RFC3444 || A. Pras, J. Schoenwaelder || pras@ctit.utwente.nl, schoenw@informatik.uni-osnabrueck.de +# RFC3445 || D. Massey, S. Rose || masseyd@isi.edu, scott.rose@nist.gov +# RFC3446 || D. Kim, D. Meyer, H. Kilmer, D. Farinacci || dorian@blackrose.org, hank@rem.com, dino@procket.com, dmm@maoz.com +# RFC3447 || J. Jonsson, B. Kaliski || jonsson@mathematik.uni-marburg.de, bkaliski@rsasecurity.com +# RFC3448 || M. Handley, S. Floyd, J. Padhye, J. Widmer || mjh@icir.org, floyd@icir.org, padhye@microsoft.com, widmer@informatik.uni-mannheim.de +# RFC3449 || H. Balakrishnan, V. Padmanabhan, G. Fairhurst, M. Sooriyabandara || hari@lcs.mit.edu, padmanab@microsoft.com, gorry@erg.abdn.ac.uk, mahesh@erg.abdn.ac.uk +# RFC3450 || M. Luby, J. Gemmell, L. Vicisano, L. Rizzo, J. Crowcroft || luby@digitalfountain.com, jgemmell@microsoft.com, lorenzo@cisco.com, luigi@iet.unipi.it, Jon.Crowcroft@cl.cam.ac.uk +# RFC3451 || M. Luby, J. Gemmell, L. Vicisano, L. Rizzo, M. Handley, J. Crowcroft || luby@digitalfountain.com, jgemmell@microsoft.com, lorenzo@cisco.com, luigi@iet.unipi.it, mjh@icir.org, Jon.Crowcroft@cl.cam.ac.uk +# RFC3452 || M. Luby, L. Vicisano, J. Gemmell, L. Rizzo, M. Handley, J. Crowcroft || luby@digitalfountain.com, lorenzo@cisco.com, jgemmell@microsoft.com, luigi@iet.unipi.it, mjh@icir.org, Jon.Crowcroft@cl.cam.ac.uk +# RFC3453 || M. Luby, L. Vicisano, J. Gemmell, L. Rizzo, M. Handley, J. Crowcroft || luby@digitalfountain.com, lorenzo@cisco.com, jgemmell@microsoft.com, luigi@iet.unipi.it, mjh@icir.org, Jon.Crowcroft@cl.cam.ac.uk +# RFC3454 || P. Hoffman, M. Blanchet || paul.hoffman@imc.org, Marc.Blanchet@viagenie.qc.ca +# RFC3455 || M. Garcia-Martin, E. Henrikson, D. Mills || miguel.a.garcia@ericsson.com, ehenrikson@lucent.com, duncan.mills@vf.vodafone.co.uk +# RFC3456 || B. Patel, B. Aboba, S. Kelly, V. Gupta || baiju.v.patel@intel.com, bernarda@microsoft.com, scott@hyperthought.com, vipul.gupta@sun.com +# RFC3457 || S. Kelly, S. Ramamoorthi || +# RFC3458 || E. Burger, E. Candell, C. Eliot, G. Klyne || e.burger@ieee.org, emily.candell@comverse.com, GK-IETF@ninebynine.org, charle@Microsoft.com +# RFC3459 || E. Burger || e.burger@ieee.org +# RFC3460 || B. Moore, Ed. || remoore@us.ibm.com +# RFC3461 || K. Moore || moore@cs.utk.edu +# RFC3462 || G. Vaudreuil || GregV@ieee.org +# RFC3463 || G. Vaudreuil || GregV@ieee.org +# RFC3464 || K. Moore, G. Vaudreuil || moore@cs.utk.edu, GregV@ieee.org +# RFC3465 || M. Allman || mallman@bbn.com +# RFC3466 || M. Day, B. Cain, G. Tomlinson, P. Rzewski || mday@alum.mit.edu, bcain@storigen.com, gary@tomlinsongroup.net, philrz@yahoo.com +# RFC3467 || J. Klensin || +# RFC3468 || L. Andersson, G. Swallow || loa@pi.se, swallow@cisco.com +# RFC3469 || V. Sharma, Ed., F. Hellstrand, Ed. || +# RFC3470 || S. Hollenbeck, M. Rose, L. Masinter || shollenbeck@verisign.com, mrose17@gmail.com, LMM@acm.org +# RFC3471 || L. Berger, Ed. || lberger@movaz.com +# RFC3472 || P. Ashwood-Smith, Ed., L. Berger, Ed. || petera@nortelnetworks.com, lberger@movaz.com +# RFC3473 || L. Berger, Ed. || lberger@movaz.com +# RFC3474 || Z. Lin, D. Pendarakis || zhiwlin@nyct.com, dpendarakis@tellium.com +# RFC3475 || O. Aboul-Magd || osama@nortelnetworks.com +# RFC3476 || B. Rajagopalan || braja@tellium.com +# RFC3477 || K. Kompella, Y. Rekhter || kireeti@juniper.net, yakov@juniper.net +# RFC3478 || M. Leelanivas, Y. Rekhter, R. Aggarwal || manoj@juniper.net, yakov@juniper.net, rahul@redback.com +# RFC3479 || A. Farrel, Ed. || afarrel@movaz.com, pjb@dataconnection.com, pmatthews@hyperchip.com, ewgray@GraIyMage.com, jack.shaio@vivacenetworks.com, tob@laurelnetworks.com, andy.malis@vivacenetworks.com +# RFC3480 || K. Kompella, Y. Rekhter, A. Kullberg || kireeti@juniper.net, yakov@juniper.net, akullber@netplane.com +# RFC3481 || H. Inamura, Ed., G. Montenegro, Ed., R. Ludwig, A. Gurtov, F. Khafizov || inamura@mml.yrp.nttdocomo.co.jp, gab@sun.com, Reiner.Ludwig@Ericsson.com, andrei.gurtov@sonera.com, faridk@nortelnetworks.com +# RFC3482 || M. Foster, T. McGarry, J. Yu || mark.foster@neustar.biz, tom.mcgarry@neustar.biz, james.yu@neustar.biz +# RFC3483 || D. Rawlins, A. Kulkarni, M. Bokaemper, K. Chan || Diana.Rawlins@wcom.com, amol.kulkarni@intel.com, khchan@nortelnetworks.com, mbokaemper@juniper.net +# RFC3484 || R. Draves || richdr@microsoft.com +# RFC3485 || M. Garcia-Martin, C. Bormann, J. Ott, R. Price, A. B. Roach || miguel.a.garcia@ericsson.com, cabo@tzi.org, jo@tzi.org, richard.price@roke.co.uk, adam@dynamicsoft.com +# RFC3486 || G. Camarillo || Gonzalo.Camarillo@ericsson.com +# RFC3487 || H. Schulzrinne || schulzrinne@cs.columbia.edu +# RFC3488 || I. Wu, T. Eckert || iwu@cisco.com +# RFC3489 || J. Rosenberg, J. Weinberger, C. Huitema, R. Mahy || jdrosen@dynamicsoft.com, jweinberger@dynamicsoft.com, huitema@microsoft.com, rohan@cisco.com +# RFC3490 || P. Faltstrom, P. Hoffman, A. Costello || paf@cisco.com, phoffman@imc.org +# RFC3491 || P. Hoffman, M. Blanchet || paul.hoffman@imc.org, Marc.Blanchet@viagenie.qc.ca +# RFC3492 || A. Costello || +# RFC3493 || R. Gilligan, S. Thomson, J. Bound, J. McCann, W. Stevens || gilligan@intransa.com, sethomso@cisco.com, Jim.Bound@hp.com, Jack.McCann@hp.com +# RFC3494 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC3495 || B. Beser, P. Duffy, Ed. || burcak@juniper.net, paduffy@cisco.com +# RFC3496 || A. G. Malis, T. Hsiao || Andy.Malis@vivacenetworks.com, Tony.Hsiao@VivaceNetworks.com +# RFC3497 || L. Gharai, C. Perkins, G. Goncher, A. Mankin || ladan@isi.edu, csp@csperkins.org, mankin@psg.com, Gary.Goncher@tek.com +# RFC3498 || J. Kuhfeld, J. Johnson, M. Thatcher || +# RFC3499 || S. Ginoza || ginoza@isi.edu +# RFC3500 || || +# RFC3501 || M. Crispin || MRC@CAC.Washington.EDU +# RFC3502 || M. Crispin || MRC@CAC.Washington.EDU +# RFC3503 || A. Melnikov || mel@messagingdirect.com +# RFC3504 || D. Eastlake || Donald.Eastlake@motorola.com +# RFC3505 || D. Eastlake || Donald.Eastlake@motorola.com +# RFC3506 || K. Fujimura, D. Eastlake || fujimura@isl.ntt.co.jp, Donald.Eastlake@motorola.com +# RFC3507 || J. Elson, A. Cerpa || jelson@cs.ucla.edu, cerpa@cs.ucla.edu +# RFC3508 || O. Levin || orit@radvision.com +# RFC3509 || A. Zinin, A. Lindem, D. Yeung || zinin@psg.com, myeung@procket.com, acee@redback.com +# RFC3510 || R. Herriot, I. McDonald || bob@herriot.com, imcdonald@sharplabs.com +# RFC3511 || B. Hickman, D. Newman, S. Tadjudin, T. Martin || brooks.hickman@spirentcom.com, dnewman@networktest.com, Saldju.Tadjudin@spirentcom.com, tmartin@gvnw.com +# RFC3512 || M. MacFaden, D. Partain, J. Saperia, W. Tackabury || +# RFC3513 || R. Hinden, S. Deering || bob.hinden@gmail.com, deering@cisco.com +# RFC3514 || S. Bellovin || bellovin@acm.org +# RFC3515 || R. Sparks || rsparks@dynamicsoft.com +# RFC3516 || L. Nerenberg || lyndon@orthanc.ab.ca +# RFC3517 || E. Blanton, M. Allman, K. Fall, L. Wang || eblanton@cs.purdue.edu, mallman@bbn.com, kfall@intel-research.net, lwang0@uky.edu +# RFC3518 || M. Higashiyama, F. Baker, T. Liao || Mitsuru.Higashiyama@yy.anritsu.co.jp, fred@cisco.com, tawei@cisco.com +# RFC3519 || H. Levkowetz, S. Vaarala || henrik@levkowetz.com, sami.vaarala@iki.fi +# RFC3520 || L-N. Hamer, B. Gage, B. Kosinski, H. Shieh || nhamer@nortelnetworks.com, brettk@invidi.com, gageb@nortelnetworks.com, hugh.shieh@attws.com +# RFC3521 || L-N. Hamer, B. Gage, H. Shieh || nhamer@nortelnetworks.com, gageb@nortelnetworks.com, hugh.shieh@attws.com +# RFC3522 || R. Ludwig, M. Meyer || Reiner.Ludwig@eed.ericsson.se, Michael.Meyer@eed.ericsson.se +# RFC3523 || J. Polk || jmpolk@cisco.com +# RFC3524 || G. Camarillo, A. Monrad || Gonzalo.Camarillo@ericsson.com, atle.monrad@ericsson.com +# RFC3525 || C. Groves, Ed., M. Pantaleo, Ed., T. Anderson, Ed., T. Taylor, Ed. || tlatla@verizon.net, Christian.Groves@ericsson.com.au, Marcello.Pantaleo@eed.ericsson.se, tom.taylor.stds@gmail.com +# RFC3526 || T. Kivinen, M. Kojo || kivinen@ssh.fi, mika.kojo@helsinki.fi +# RFC3527 || K. Kinnear, M. Stapp, R. Johnson, J. Kumarasamy || kkinnear@cisco.com, mjs@cisco.com, jayk@cisco.com, raj@cisco.com +# RFC3528 || W. Zhao, H. Schulzrinne, E. Guttman || zwb@cs.columbia.edu, hgs@cs.columbia.edu, Erik.Guttman@sun.com +# RFC3529 || W. Harold || wharold@us.ibm.com +# RFC3530 || S. Shepler, B. Callaghan, D. Robinson, R. Thurlow, C. Beame, M. Eisler, D. Noveck || beame@bws.com, brent.callaghan@sun.com, mike@eisler.com, dnoveck@netapp.com, david.robinson@sun.com, robert.thurlow@sun.com +# RFC3531 || M. Blanchet || Marc.Blanchet@viagenie.qc.ca +# RFC3532 || T. Anderson, J. Buerkle || todd.a.anderson@intel.com, joachim.buerkle@nortelnetworks.com +# RFC3533 || S. Pfeiffer || Silvia.Pfeiffer@csiro.au +# RFC3534 || L. Walleij || triad@df.lth.se +# RFC3535 || J. Schoenwaelder || j.schoenwaelder@iu-bremen.de +# RFC3536 || P. Hoffman || paul.hoffman@imc.org +# RFC3537 || J. Schaad, R. Housley || jimsch@exmsft.com, housley@vigilsec.com +# RFC3538 || Y. Kawatsura || kawatura@bisd.hitachi.co.jp +# RFC3539 || B. Aboba, J. Wood || bernarda@microsoft.com, jonwood@speakeasy.net +# RFC3540 || N. Spring, D. Wetherall, D. Ely || nspring@cs.washington.edu, djw@cs.washington.edu, ely@cs.washington.edu +# RFC3541 || A. Walsh || aaron@mantiscorp.com +# RFC3542 || W. Stevens, M. Thomas, E. Nordmark, T. Jinmei || matt@3am-software.com, Erik.Nordmark@sun.com, jinmei@isl.rdc.toshiba.co.jp +# RFC3543 || S. Glass, M. Chandra || steven.glass@sun.com, mchandra@cisco.com +# RFC3544 || T. Koren, S. Casner, C. Bormann || tmima@cisco.com, casner@packetdesign.com, cabo@tzi.org +# RFC3545 || T. Koren, S. Casner, J. Geevarghese, B. Thompson, P. Ruddy || tmima@cisco.com, casner@acm.org, geevjohn@hotmail.com, brucet@cisco.com, pruddy@cisco.com +# RFC3546 || S. Blake-Wilson, M. Nystrom, D. Hopwood, J. Mikkelsen, T. Wright || sblakewilson@bcisse.com, magnus@rsasecurity.com, david.hopwood@zetnet.co.uk, janm@transactionware.com, timothy.wright@vodafone.com +# RFC3547 || M. Baugher, B. Weis, T. Hardjono, H. Harney || mbaugher@cisco.com, thardjono@verisign.com, hh@sparta.com, bew@cisco.com +# RFC3548 || S. Josefsson, Ed. || +# RFC3549 || J. Salim, H. Khosravi, A. Kleen, A. Kuznetsov || hadi@znyx.com, hormuzd.m.khosravi@intel.com, ak@suse.de, kuznet@ms2.inr.ac.ru +# RFC3550 || H. Schulzrinne, S. Casner, R. Frederick, V. Jacobson || schulzrinne@cs.columbia.edu, casner@acm.org, ronf@bluecoat.com, van@packetdesign.com +# RFC3551 || H. Schulzrinne, S. Casner || schulzrinne@cs.columbia.edu, casner@acm.org +# RFC3552 || E. Rescorla, B. Korver || ekr@rtfm.com, briank@xythos.com, iab@iab.org +# RFC3553 || M. Mealling, L. Masinter, T. Hardie, G. Klyne || michael@verisignlabs.com, LMM@acm.org, hardie@qualcomm.com, GK-IETF@ninebynine.org +# RFC3554 || S. Bellovin, J. Ioannidis, A. Keromytis, R. Stewart || smb@research.att.com, ji@research.att.com, angelos@cs.columbia.edu, randall@lakerest.net +# RFC3555 || S. Casner, P. Hoschka || casner@acm.org, ph@w3.org +# RFC3556 || S. Casner || casner@acm.org +# RFC3557 || Q. Xie, Ed. || bdp003@motorola.com, Senaka.Balasuriya@motorola.com, yoonie@verbaltek.com, stephane.maes@oracle.com, hgarudad@qualcomm.com, Qiaobing.Xie@motorola.com +# RFC3558 || A. Li || adamli@icsl.ucla.edu +# RFC3559 || D. Thaler || dthaler@microsoft.com +# RFC3560 || R. Housley || housley@vigilsec.com +# RFC3561 || C. Perkins, E. Belding-Royer, S. Das || Charles.Perkins@nokia.com, ebelding@cs.ucsb.edu, sdas@ececs.uc.edu +# RFC3562 || M. Leech || mleech@nortelnetworks.com +# RFC3563 || A. Zinin || zinin@psg.com +# RFC3564 || F. Le Faucheur, W. Lai || +# RFC3565 || J. Schaad || jimsch@exmsft.com +# RFC3566 || S. Frankel, H. Herbert || sheila.frankel@nist.gov, howard.c.herbert@intel.com +# RFC3567 || T. Li, R. Atkinson || tli@procket.net, rja@extremenetworks.com +# RFC3568 || A. Barbir, B. Cain, R. Nair, O. Spatscheck || abbieb@nortelnetworks.com, bcain@storigen.com, nair_raj@yahoo.com, spatsch@research.att.com +# RFC3569 || S. Bhattacharyya, Ed. || +# RFC3570 || P. Rzewski, M. Day, D. Gilletti || mday@alum.mit.edu, dgilletti@yahoo.com, philrz@yahoo.com +# RFC3571 || D. Rawlins, A. Kulkarni, K. Ho Chan, M. Bokaemper, D. Dutt || Diana.Rawlins@mci.com, amol.kulkarni@intel.com, khchan@nortelnetworks.com, mbokaemper@juniper.net, ddutt@cisco.com +# RFC3572 || T. Ogura, M. Maruyama, T. Yoshida || ogura@core.ecl.net, mitsuru@core.ecl.net, yoshida@peta.arch.ecl.net +# RFC3573 || I. Goyret || igoyret@lucent.com +# RFC3574 || J. Soininen, Ed. || +# RFC3575 || B. Aboba || bernarda@microsoft.com +# RFC3576 || M. Chiba, G. Dommety, M. Eklund, D. Mitton, B. Aboba || mchiba@cisco.com, gdommety@cisco.com, meklund@cisco.com, david@mitton.com, bernarda@microsoft.com +# RFC3577 || S. Waldbusser, R. Cole, C. Kalbfleisch, D. Romascanu || waldbusser@nextbeacon.com, cwk@verio.net, rgcole@att.com, dromasca@gmail.com +# RFC3578 || G. Camarillo, A. B. Roach, J. Peterson, L. Ong || Gonzalo.Camarillo@ericsson.com, adam@dynamicsoft.com, jon.peterson@neustar.biz, lyong@ciena.com +# RFC3579 || B. Aboba, P. Calhoun || bernarda@microsoft.com, pcalhoun@airespace.com +# RFC3580 || P. Congdon, B. Aboba, A. Smith, G. Zorn, J. Roese || paul_congdon@hp.com, bernarda@microsoft.com, ah_smith@acm.org, jjr@enterasys.com, gwz@cisco.com +# RFC3581 || J. Rosenberg, H. Schulzrinne || jdrosen@dynamicsoft.com, schulzrinne@cs.columbia.edu +# RFC3582 || J. Abley, B. Black, V. Gill || jabley@isc.org, ben@layer8.net, vijaygill9@aol.com +# RFC3583 || H. Chaskar, Ed. || john.loughney@nokia.com, hemant.chaskar@nokia.com +# RFC3584 || R. Frye, D. Levi, S. Routhier, B. Wijnen || +# RFC3585 || J. Jason, L. Rafalow, E. Vyncke || jamie.jason@intel.com, rafalow@watson.ibm.com, evyncke@cisco.com +# RFC3586 || M. Blaze, A. Keromytis, M. Richardson, L. Sanchez || mab@crypto.com, angelos@cs.columbia.edu, mcr@sandelman.ottawa.on.ca, lsanchez@xapiens.com +# RFC3587 || R. Hinden, S. Deering, E. Nordmark || bob.hinden@gmail.com, erik.nordmark@sun.com +# RFC3588 || P. Calhoun, J. Loughney, E. Guttman, G. Zorn, J. Arkko || pcalhoun@airespace.com, john.Loughney@nokia.com, Jari.Arkko@ericsson.com, erik.guttman@sun.com +# RFC3589 || J. Loughney || john.Loughney@Nokia.com +# RFC3590 || B. Haberman || brian@innovationslab.net +# RFC3591 || H-K. Lam, M. Stewart, A. Huynh || mstewart1@nc.rr.com, a_n_huynh@yahoo.com, hklam@lucent.com +# RFC3592 || K. Tesink || kaj@research.telcordia.com +# RFC3593 || K. Tesink, Ed. || +# RFC3594 || P. Duffy || paduffy@cisco.com +# RFC3595 || B. Wijnen || bwijnen@lucent.com +# RFC3596 || S. Thomson, C. Huitema, V. Ksinant, M. Souissi || sethomso@cisco.com, huitema@microsoft.com, vladimir.ksinant@6wind.com, Mohsen.Souissi@nic.fr +# RFC3597 || A. Gustafsson || gson@nominum.com +# RFC3598 || K. Murchison || ken@oceana.com +# RFC3599 || S. Ginoza || ginoza@isi.edu +# RFC3600 || J. Reynolds, Ed., S. Ginoza, Ed. || +# RFC3601 || C. Allocchio || Claudio.Allocchio@garr.it +# RFC3602 || S. Frankel, R. Glenn, S. Kelly || sheila.frankel@nist.gov, scott@hyperthought.com, rob.glenn@nist.gov +# RFC3603 || W. Marshall, Ed., F. Andreasen, Ed. || +# RFC3604 || H. Khosravi, G. Kullgren, S. Shew, J. Sadler, A. Watanabe || hormuzd.m.khosravi@intel.com, geku@nortelnetworks.com, Jonathan.Sadler@tellabs.com, sdshew@nortelnetworks.com, Shiomoto.Kohei@lab.ntt.co.jp, atsushi@exa.onlab.ntt.co.jp, okamoto@exa.onlab.ntt.co.jp +# RFC3605 || C. Huitema || huitema@microsoft.com +# RFC3606 || F. Ly, M. Noto, A. Smith, E. Spiegel, K. Tesink || faye@pedestalnetworks.com, mnoto@cisco.com, ah_smith@acm.org, mspiegel@cisco.com, kaj@research.telcordia.com +# RFC3607 || M. Leech || mleech@nortelnetworks.com +# RFC3608 || D. Willis, B. Hoeneisen || dean.willis@softarmor.com, hoeneisen@switch.ch +# RFC3609 || R. Bonica, K. Kompella, D. Meyer || ronald.p.bonica@mci.com, kireeti@juniper.net, dmm@maoz.com +# RFC3610 || D. Whiting, R. Housley, N. Ferguson || dwhiting@hifn.com, housley@vigilsec.com, niels@macfergus.com +# RFC3611 || T. Friedman, Ed., R. Caceres, Ed., A. Clark, Ed. || almeroth@cs.ucsb.edu, caceres@watson.ibm.com, alan@telchemy.com, robert.cole@jhuapl.edu, duffield@research.att.com, timur.friedman@lip6.fr, khedayat@brixnet.com, ksarac@utdallas.edu, magnus.westerlund@ericsson.com +# RFC3612 || A. Farrel || adrian@olddog.co.uk +# RFC3613 || R. Morgan, K. Hazelton || rlmorgan@washington.edu, hazelton@doit.wisc.edu +# RFC3614 || J. Smith || jrsmith@watson.ibm.com +# RFC3615 || J. Gustin, A. Goyens || jean-marc.gustin@swift.com, andre.goyens@swift.com +# RFC3616 || F. Bellifemine, I. Constantinescu, S. Willmott || Fabio.Bellifemine@TILAB.COM, ion.constantinescu@epfl.ch, steve@lsi.upc.es +# RFC3617 || E. Lear || lear@cisco.com +# RFC3618 || B. Fenner, Ed., D. Meyer, Ed. || +# RFC3619 || S. Shah, M. Yip || sshah@extremenetworks.com, my@extremenetworks.com +# RFC3620 || D. New || dnew@san.rr.com +# RFC3621 || A. Berger, D. Romascanu || avib@PowerDsine.com, dromasca@gmail.com +# RFC3622 || M. Mealling || michael@neonym.net +# RFC3623 || J. Moy, P. Pillay-Esnault, A. Lindem || jmoy@sycamorenet.com, padma@juniper.net, acee@redback.com +# RFC3624 || B. Foster, D. Auerbach, F. Andreasen || fandreas@cisco.com, dea@cisco.com, bfoster@cisco.com +# RFC3625 || R. Gellens, H. Garudadri || +# RFC3626 || T. Clausen, Ed., P. Jacquet, Ed. || T.Clausen@computer.org, Philippe.Jacquet@inria.fr +# RFC3627 || P. Savola || psavola@funet.fi +# RFC3628 || D. Pinkas, N. Pope, J. Ross || Denis.Pinkas@bull.net, pope@secstan.com, ross@secstan.com, claire.desclercs@etsi.org +# RFC3629 || F. Yergeau || fyergeau@alis.com +# RFC3630 || D. Katz, K. Kompella, D. Yeung || dkatz@juniper.net, myeung@procket.com, kireeti@juniper.net +# RFC3631 || S. Bellovin, Ed., J. Schiller, Ed., C. Kaufman, Ed. || +# RFC3632 || S. Hollenbeck, S. Veeramachaneni, S. Yalamanchilli || shollenbeck@verisign.com, sveerama@verisign.com, syalamanchilli@verisign.com +# RFC3633 || O. Troan, R. Droms || ot@cisco.com, rdroms@cisco.com +# RFC3634 || K. Luehrs, R. Woundy, J. Bevilacqua, N. Davoust || k.luehrs@cablelabs.com, richard_woundy@cable.comcast.com, john@yas.com, nancy@yas.com +# RFC3635 || J. Flick || johnf@rose.hp.com +# RFC3636 || J. Flick || johnf@rose.hp.com +# RFC3637 || C.M. Heard, Ed. || +# RFC3638 || J. Flick, C. M. Heard || johnf@rose.hp.com, heard@pobox.com +# RFC3639 || M. St. Johns, Ed., G. Huston, Ed., IAB || +# RFC3640 || J. van der Meer, D. Mackie, V. Swaminathan, D. Singer, P. Gentric || jan.vandermeer@philips.com, dmackie@apple.com, viswanathan.swaminathan@sun.com, singer@apple.com, philippe.gentric@philips.com +# RFC3641 || S. Legg || steven.legg@adacel.com.au +# RFC3642 || S. Legg || steven.legg@adacel.com.au +# RFC3643 || R. Weber, M. Rajagopal, F. Travostino, M. O'Donnell, C. Monia, M. Merhar || roweber@ieee.org, muralir@broadcom.com, travos@nortelnetworks.com, cmonia@pacbell.net, milan.merhar@sun.com +# RFC3644 || Y. Snir, Y. Ramberg, J. Strassner, R. Cohen, B. Moore || yramberg@cisco.com, ysnir@cisco.com, john.strassner@intelliden.com, ronc@lyciumnetworks.com, remoore@us.ibm.com +# RFC3645 || S. Kwan, P. Garg, J. Gilroy, L. Esibov, J. Westhead, R. Hall || skwan@microsoft.com, praeritg@microsoft.com, jamesg@microsoft.com, levone@microsoft.com, randyhall@lucent.com, jwesth@microsoft.com +# RFC3646 || R. Droms, Ed. || rdroms@cisco.com +# RFC3647 || S. Chokhani, W. Ford, R. Sabett, C. Merrill, S. Wu || chokhani@orionsec.com, wford@verisign.com, rsabett@cooley.com, cmerrill@mccarter.com, swu@infoliance.com +# RFC3648 || J. Whitehead, J. Reschke, Ed. || ejw@cse.ucsc.edu, julian.reschke@greenbytes.de +# RFC3649 || S. Floyd || floyd@acm.org +# RFC3650 || S. Sun, L. Lannom, B. Boesch || ssun@cnri.reston.va.us, llannom@cnri.reston.va.us, bboesch@cnri.reston.va.us +# RFC3651 || S. Sun, S. Reilly, L. Lannom || ssun@cnri.reston.va.us, sreilly@cnri.reston.va.us, llannom@cnri.reston.va.us +# RFC3652 || S. Sun, S. Reilly, L. Lannom, J. Petrone || ssun@cnri.reston.va.us, sreilly@cnri.reston.va.us, llannom@cnri.reston.va.us, jpetrone@cnri.reston.va.us +# RFC3653 || J. Boyer, M. Hughes, J. Reagle || jboyer@PureEdge.com, Merlin.Hughes@betrusted.com, reagle@mit.edu +# RFC3654 || H. Khosravi, Ed., T. Anderson, Ed. || edbowen@us.ibm.com, rdantu@unt.edu, avri@acm.org, ram.gopal@nokia.com, hadi@znyx.com, muneyb@avaya.com, margaret.wasserman@nokia.com, hormuzd.m.khosravi@intel.com, todd.a.anderson@intel.com +# RFC3655 || B. Wellington, O. Gudmundsson || Brian.Wellington@nominum.com, ogud@ogud.com +# RFC3656 || R. Siemborski || +# RFC3657 || S. Moriai, A. Kato || camellia@isl.ntt.co.jp, akato@po.ntts.co.jp +# RFC3658 || O. Gudmundsson || ds-rfc@ogud.com +# RFC3659 || P. Hethmon || phethmon@hethmon.com +# RFC3660 || B. Foster, F. Andreasen || bfoster@cisco.com, fandreas@cisco.com +# RFC3661 || B. Foster, C. Sivachelvan || chelliah@cisco.com, bfoster@cisco.com +# RFC3662 || R. Bless, K. Nichols, K. Wehrle || bless@tm.uka.de, knichols@ieee.org, Klaus.Wehrle@uni-tuebingen.de +# RFC3663 || A. Newton || anewton@verisignlabs.com +# RFC3664 || P. Hoffman || paul.hoffman@vpnc.org +# RFC3665 || A. Johnston, S. Donovan, R. Sparks, C. Cunningham, K. Summers || alan.johnston@mci.com, sdonovan@dynamicsoft.com, rsparks@dynamicsoft.com, ccunningham@dynamicsoft.com, kevin.summers@sonusnet.com +# RFC3666 || A. Johnston, S. Donovan, R. Sparks, C. Cunningham, K. Summers || alan.johnston@mci.com, sdonovan@dynamicsoft.com, rsparks@dynamicsoft.com, ccunningham@dynamicsoft.com, kevin.summers@sonusnet.com +# RFC3667 || S. Bradner || +# RFC3668 || S. Bradner || +# RFC3669 || S. Brim || sbrim@cisco.com +# RFC3670 || B. Moore, D. Durham, J. Strassner, A. Westerinen, W. Weiss || remoore@us.ibm.com, david.durham@intel.com, john.strassner@intelliden.com, andreaw@cisco.com, walterweiss@attbi.com +# RFC3671 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC3672 || K. Zeilenga || Kurt@OpenLDAP.org, steven.legg@adacel.com.au +# RFC3673 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC3674 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC3675 || D. Eastlake 3rd || dee3@torque.pothole.com +# RFC3676 || R. Gellens || randy@qualcomm.com +# RFC3677 || L. Daigle, Ed., Internet Architecture Board || iab@iab.org +# RFC3678 || D. Thaler, B. Fenner, B. Quinn || dthaler@microsoft.com, fenner@research.att.com, rcq@ipmulticast.com +# RFC3679 || R. Droms || rdroms@cisco.com +# RFC3680 || J. Rosenberg || jdrosen@dynamicsoft.com +# RFC3681 || R. Bush, R. Fink || randy@psg.com, bob@thefinks.com +# RFC3682 || V. Gill, J. Heasley, D. Meyer || vijay@umbc.edu, heas@shrubbery.net, dmm@1-4-5.net +# RFC3683 || M. Rose || mrose17@gmail.com +# RFC3684 || R. Ogier, F. Templin, M. Lewis || ogier@erg.sri.com, ftemplin@iprg.nokia.com, lewis@erg.sri.com +# RFC3685 || C. Daboo || daboo@cyrusoft.com +# RFC3686 || R. Housley || housley@vigilsec.com +# RFC3687 || S. Legg || steven.legg@adacel.com.au +# RFC3688 || M. Mealling || michael@verisignlabs.com +# RFC3689 || K. Carlberg, R. Atkinson || k.carlberg@cs.ucl.ac.uk, rja@extremenetworks.com +# RFC3690 || K. Carlberg, R. Atkinson || k.carlberg@cs.ucl.ac.uk, rja@extremenetworks.com +# RFC3691 || A. Melnikov || Alexey.Melnikov@isode.com +# RFC3692 || T. Narten || narten@us.ibm.com +# RFC3693 || J. Cuellar, J. Morris, D. Mulligan, J. Peterson, J. Polk || Jorge.Cuellar@siemens.com, jmorris@cdt.org, dmulligan@law.berkeley.edu, jon.peterson@neustar.biz, jmpolk@cisco.com +# RFC3694 || M. Danley, D. Mulligan, J. Morris, J. Peterson || mre213@nyu.edu, dmulligan@law.berkeley.edu, jmorris@cdt.org, jon.peterson@neustar.biz +# RFC3695 || M. Luby, L. Vicisano || luby@digitalfountain.com, lorenzo@cisco.com +# RFC3696 || J. Klensin || john-ietf@jck.com +# RFC3697 || J. Rajahalme, A. Conta, B. Carpenter, S. Deering || jarno.rajahalme@nokia.com, aconta@txc.com, brc@zurich.ibm.com +# RFC3698 || K. Zeilenga, Ed. || Kurt@OpenLDAP.org +# RFC3700 || J. Reynolds, Ed., S. Ginoza, Ed. || +# RFC3701 || R. Fink, R. Hinden || bob@thefinks.com, bob.hinden@gmail.com +# RFC3702 || J. Loughney, G. Camarillo || John.Loughney@nokia.com, Gonzalo.Camarillo@ericsson.com +# RFC3703 || J. Strassner, B. Moore, R. Moats, E. Ellesson || john.strassner@intelliden.com, remoore@us.ibm.com, rmoats@lemurnetworks.net, ellesson@mindspring.com +# RFC3704 || F. Baker, P. Savola || fred@cisco.com, psavola@funet.fi +# RFC3705 || B. Ray, R. Abbi || rray@pesa.com, Rajesh.Abbi@alcatel.com +# RFC3706 || G. Huang, S. Beaulieu, D. Rochefort || +# RFC3707 || A. Newton || anewton@verisignlabs.com +# RFC3708 || E. Blanton, M. Allman || eblanton@cs.purdue.edu, mallman@icir.org +# RFC3709 || S. Santesson, R. Housley, T. Freeman || stefans@microsoft.com, housley@vigilsec.com, trevorf@microsoft.com +# RFC3710 || H. Alvestrand || harald@alvestrand.no +# RFC3711 || M. Baugher, D. McGrew, M. Naslund, E. Carrara, K. Norrman || mbaugher@cisco.com, elisabetta.carrara@ericsson.com, mcgrew@cisco.com, mats.naslund@ericsson.com, karl.norrman@ericsson.com +# RFC3712 || P. Fleming, I. McDonald || flemingp@us.ibm.com, flemingp@us.ibm.com, flemingp@us.ibm.com +# RFC3713 || M. Matsui, J. Nakajima, S. Moriai || matsui@iss.isl.melco.co.jp, june15@iss.isl.melco.co.jp, shiho@rd.scei.sony.co.jp +# RFC3714 || S. Floyd, Ed., J. Kempf, Ed. || iab@iab.org +# RFC3715 || B. Aboba, W. Dixon || bernarda@microsoft.com, ietf-wd@v6security.com +# RFC3716 || IAB Advisory Committee || iab@iab.org +# RFC3717 || B. Rajagopalan, J. Luciani, D. Awduche || braja@tellium.com, james_luciani@mindspring.com, awduche@awduche.com +# RFC3718 || R. McGowan || +# RFC3719 || J. Parker, Ed. || jparker@axiowave.com +# RFC3720 || J. Satran, K. Meth, C. Sapuntzakis, M. Chadalapaka, E. Zeidner || Julian_Satran@il.ibm.com, meth@il.ibm.com, csapuntz@alum.mit.edu, efri@xiv.co.il, cbm@rose.hp.com +# RFC3721 || M. Bakke, J. Hafner, J. Hufferd, K. Voruganti, M. Krueger || kaladhar@us.ibm.com, mbakke@cisco.com, hafner@almaden.ibm.com, hufferd@us.ibm.com, marjorie_krueger@hp.com +# RFC3722 || M. Bakke || mbakke@cisco.com +# RFC3723 || B. Aboba, J. Tseng, J. Walker, V. Rangan, F. Travostino || bernarda@microsoft.com, joshtseng@yahoo.com, jesse.walker@intel.com, vrangan@brocade.com, travos@nortelnetworks.com +# RFC3724 || J. Kempf, Ed., R. Austein, Ed., IAB || +# RFC3725 || J. Rosenberg, J. Peterson, H. Schulzrinne, G. Camarillo || jdrosen@dynamicsoft.com, jon.peterson@neustar.biz, schulzrinne@cs.columbia.edu, Gonzalo.Camarillo@ericsson.com +# RFC3726 || M. Brunner, Ed. || brunner@netlab.nec.de, robert.hancock@roke.co.uk, eleanor.hepworth@roke.co.uk, cornelia.kappler@siemens.com, Hannes.Tschofenig@mchp.siemens.de +# RFC3727 || S. Legg || steven.legg@adacel.com.au +# RFC3728 || B. Ray, R. Abbi || rray@pesa.com, Rajesh.Abbi@alcatel.com +# RFC3729 || S. Waldbusser || waldbusser@nextbeacon.com +# RFC3730 || S. Hollenbeck || shollenbeck@verisign.com +# RFC3731 || S. Hollenbeck || shollenbeck@verisign.com +# RFC3732 || S. Hollenbeck || shollenbeck@verisign.com +# RFC3733 || S. Hollenbeck || shollenbeck@verisign.com +# RFC3734 || S. Hollenbeck || shollenbeck@verisign.com +# RFC3735 || S. Hollenbeck || shollenbeck@verisign.com +# RFC3736 || R. Droms || rdroms@cisco.com +# RFC3737 || B. Wijnen, A. Bierman || bwijnen@lucent.com, andy@yumaworks.com +# RFC3738 || M. Luby, V. Goyal || luby@digitalfountain.com, v.goyal@ieee.org +# RFC3739 || S. Santesson, M. Nystrom, T. Polk || stefans@microsoft.com, wpolk@nist.gov, magnus@rsasecurity.com +# RFC3740 || T. Hardjono, B. Weis || thardjono@verisign.com, bew@cisco.com +# RFC3741 || J. Boyer, D. Eastlake 3rd, J. Reagle || jboyer@PureEdge.com, Donald.Eastlake@motorola.com, reagle@mit.edu +# RFC3742 || S. Floyd || floyd@icir.org +# RFC3743 || K. Konishi, K. Huang, H. Qian, Y. Ko || konishi@jp.apan.net, huangk@alum.sinica.edu, Hlqian@cnnic.net.cn, yw@mrko.pe.kr, jseng@pobox.org.sg, rickard@rickardgroup.com +# RFC3744 || G. Clemm, J. Reschke, E. Sedlar, J. Whitehead || geoffrey.clemm@us.ibm.com, julian.reschke@greenbytes.de, eric.sedlar@oracle.com, ejw@cse.ucsc.edu +# RFC3745 || D. Singer, R. Clark, D. Lee || singer@apple.com, richard@elysium.ltd.uk, dlee@yahoo-inc.com +# RFC3746 || L. Yang, R. Dantu, T. Anderson, R. Gopal || lily.l.yang@intel.com, rdantu@unt.edu, todd.a.anderson@intel.com, ram.gopal@nokia.com +# RFC3747 || H. Hazewinkel, Ed., D. Partain, Ed. || +# RFC3748 || B. Aboba, L. Blunk, J. Vollbrecht, J. Carlson, H. Levkowetz, Ed. || bernarda@microsoft.com, ljb@merit.edu, jrv@umich.edu, james.d.carlson@sun.com, henrik@levkowetz.com +# RFC3749 || S. Hollenbeck || shollenbeck@verisign.com +# RFC3750 || C. Huitema, R. Austein, S. Satapati, R. van der Pol || huitema@microsoft.com, sra@isc.org, satapati@cisco.com, Ronald.vanderPol@nlnetlabs.nl +# RFC3751 || S. Bradner || sob@harvard.edu +# RFC3752 || A. Barbir, E. Burger, R. Chen, S. McHenry, H. Orman, R. Penno || abbieb@nortelnetworks.com, e.burger@ieee.org, chen@research.att.com, stephen@mchenry.net, ho@alum.mit.edu, rpenno@nortelnetworks.com +# RFC3753 || J. Manner, Ed., M. Kojo, Ed. || jmanner@cs.helsinki.fi, kojo@cs.helsinki.fi +# RFC3754 || R. Bless, K. Wehrle || bless@tm.uka.de, Klaus.Wehrle@uni-tuebingen.de +# RFC3755 || S. Weiler || weiler@tislabs.com +# RFC3756 || P. Nikander, Ed., J. Kempf, E. Nordmark || pekka.nikander@nomadiclab.com, kempf@docomolabs-usa.com, erik.nordmark@sun.com +# RFC3757 || O. Kolkman, J. Schlyter, E. Lewis || olaf@ripe.net, jakob@nic.se, edlewis@arin.net +# RFC3758 || R. Stewart, M. Ramalho, Q. Xie, M. Tuexen, P. Conrad || randall@lakerest.net, mramalho@cisco.com, qxie1@email.mot.com, tuexen@fh-muenster.de, conrad@acm.org +# RFC3759 || L-E. Jonsson || lars-erik.jonsson@ericsson.com +# RFC3760 || D. Gustafson, M. Just, M. Nystrom || degustafson@comcast.net, Just.Mike@tbs-sct.gc.ca, magnus@rsasecurity.com +# RFC3761 || P. Faltstrom, M. Mealling || paf@cisco.com +# RFC3762 || O. Levin || oritl@microsoft.com +# RFC3763 || S. Shalunov, B. Teitelbaum || shalunov@internet2.edu, ben@internet2.edu +# RFC3764 || J. Peterson || jon.peterson@neustar.biz +# RFC3765 || G. Huston || gih@telstra.net +# RFC3766 || H. Orman, P. Hoffman || hilarie@purplestreak.com, paul.hoffman@vpnc.org +# RFC3767 || S. Farrell, Ed. || +# RFC3768 || R. Hinden, Ed. || bob.hinden@gmail.com +# RFC3769 || S. Miyakawa, R. Droms || miyakawa@nttv6.jp, rdroms@cisco.com +# RFC3770 || R. Housley, T. Moore || housley@vigilsec.com, timmoore@microsoft.com +# RFC3771 || R. Harrison, K. Zeilenga || roger_harrison@novell.com, Kurt@OpenLDAP.org +# RFC3772 || J. Carlson, R. Winslow || +# RFC3773 || E. Candell || emily.candell@comverse.com +# RFC3774 || E. Davies, Ed. || +# RFC3775 || D. Johnson, C. Perkins, J. Arkko || dbj@cs.rice.edu, charliep@iprg.nokia.com, jari.arkko@ericsson.com +# RFC3776 || J. Arkko, V. Devarapalli, F. Dupont || jari.arkko@ericsson.com, vijayd@iprg.nokia.com, Francis.Dupont@enst-bretagne.fr +# RFC3777 || J. Galvin, Ed. || galvin+ietf@elistx.com +# RFC3778 || E. Taft, J. Pravetz, S. Zilles, L. Masinter || taft@adobe.com, jpravetz@adobe.com, szilles@adobe.com, LMM@acm.org +# RFC3779 || C. Lynn, S. Kent, K. Seo || CLynn@BBN.Com, Kent@BBN.Com, KSeo@BBN.Com +# RFC3780 || F. Strauss, J. Schoenwaelder || strauss@ibr.cs.tu-bs.de, j.schoenwaelder@iu-bremen.de +# RFC3781 || F. Strauss, J. Schoenwaelder || strauss@ibr.cs.tu-bs.de, j.schoenwaelder@iu-bremen.de +# RFC3782 || S. Floyd, T. Henderson, A. Gurtov || floyd@acm.org, thomas.r.henderson@boeing.com, andrei.gurtov@teliasonera.com +# RFC3783 || M. Chadalapaka, R. Elliott || cbm@rose.hp.com, elliott@hp.com +# RFC3784 || H. Smit, T. Li || hhwsmit@xs4all.nl, tony.li@tony.li +# RFC3785 || F. Le Faucheur, R. Uppili, A. Vedrenne, P. Merckx, T. Telkamp || flefauch@cisco.com, alain.vedrenne@equant.com, pierre.merckx@equant.com, telkamp@gblx.net +# RFC3786 || A. Hermelin, S. Previdi, M. Shand || amir@montilio.com, sprevidi@cisco.com, mshand@cisco.com +# RFC3787 || J. Parker, Ed. || jparker@axiowave.com +# RFC3788 || J. Loughney, M. Tuexen, Ed., J. Pastor-Balbas || john.loughney@nokia.com, tuexen@fh-muenster.de, j.javier.pastor@ericsson.com +# RFC3789 || P. Nesser, II, A. Bergstrom, Ed. || phil@nesser.com, andreas.bergstrom@hiof.no +# RFC3790 || C. Mickles, Ed., P. Nesser, II || cmickles.ee88@gtalumni.org, phil@nesser.com +# RFC3791 || C. Olvera, P. Nesser, II || cesar.olvera@consulintel.es, phil@nesser.com +# RFC3792 || P. Nesser, II, A. Bergstrom, Ed. || phil@nesser.com, andreas.bergstrom@hiof.no +# RFC3793 || P. Nesser, II, A. Bergstrom, Ed. || phil@nesser.com, andreas.bergstrom@hiof.no +# RFC3794 || P. Nesser, II, A. Bergstrom, Ed. || phil@nesser.com, andreas.bergstrom@hiof.no +# RFC3795 || R. Sofia, P. Nesser, II || rsofia@zmail.pt, phil@nesser.com +# RFC3796 || P. Nesser, II, A. Bergstrom, Ed. || phil@nesser.com, andreas.bergstrom@hiof.no +# RFC3797 || D. Eastlake 3rd || Donald.Eastlake@motorola.com +# RFC3798 || T. Hansen, Ed., G. Vaudreuil, Ed. || GregV@ieee.org +# RFC3801 || G. Vaudreuil, G. Parsons || gregv@ieee.org, GParsons@NortelNetworks.com +# RFC3802 || G. Vaudreuil, G. Parsons || gregv@ieee.org, gparsons@nortelnetworks.com +# RFC3803 || G. Vaudreuil, G. Parsons || gregv@ieee.org, gparsons@nortelnetworks.com +# RFC3804 || G. Parsons || gparsons@nortelnetworks.com +# RFC3805 || R. Bergman, H. Lewis, I. McDonald || Ron.Bergman@hitachi-ps.us, harryl@us.ibm.com, imcdonald@sharplabs.com +# RFC3806 || R. Bergman, H. Lewis, I. McDonald || Ron.Bergman@hitachi-ps.us, harryl@us.ibm.com, imcdonald@sharplabs.com +# RFC3807 || E. Weilandt, N. Khanchandani, S. Rao || eva.weilandt@temic.com, rsanjay@nortelnetworks.com, neerajk@nortelnetworks.com +# RFC3808 || I. McDonald || imcdonald@sharplabs.com, iana@iana.org +# RFC3809 || A. Nagarajan, Ed. || +# RFC3810 || R. Vida, Ed., L. Costa, Ed. || Rolland.Vida@lip6.fr, Luis.Costa@lip6.fr, Serge.Fdida@lip6.fr, deering@cisco.com, fenner@research.att.com, kouvelas@cisco.com, brian@innovationslab.net +# RFC3811 || T. Nadeau, Ed., J. Cucchiara, Ed. || tnadeau@cisco.com, jcucchiara@mindspring.com +# RFC3812 || C. Srinivasan, A. Viswanathan, T. Nadeau || cheenu@bloomberg.net, arunv@force10networks.com, tnadeau@cisco.com +# RFC3813 || C. Srinivasan, A. Viswanathan, T. Nadeau || cheenu@bloomberg.net, arunv@force10networks.com, tnadeau@cisco.com +# RFC3814 || T. Nadeau, C. Srinivasan, A. Viswanathan || tnadeau@cisco.com, cheenu@bloomberg.net, arunv@force10networks.com +# RFC3815 || J. Cucchiara, H. Sjostrand, J. Luciani || james_luciani@mindspring.com, hans@ipunplugged.com, jcucchiara@mindspring.com +# RFC3816 || J. Quittek, M. Stiemerling, H. Hartenstein || quittek@netlab.nec.de, stiemerling@netlab.nec.de, hartenstein@rz.uni-karlsruhe.de +# RFC3817 || W. Townsley, R. da Silva || mark@townsley.net, rdasilva@va.rr.com +# RFC3818 || V. Schryver || vjs@rhyolite.com +# RFC3819 || P. Karn, Ed., C. Bormann, G. Fairhurst, D. Grossman, R. Ludwig, J. Mahdavi, G. Montenegro, J. Touch, L. Wood || karn@qualcomm.com, cabo@tzi.org, gorry@erg.abdn.ac.uk, Dan.Grossman@motorola.com, Reiner.Ludwig@ericsson.com, jmahdavi@earthlink.net, gab@sun.com, touch@isi.edu, lwood@cisco.com +# RFC3820 || S. Tuecke, V. Welch, D. Engert, L. Pearlman, M. Thompson || tuecke@mcs.anl.gov, vwelch@ncsa.uiuc.edu, deengert@anl.gov, laura@isi.edu, mrthompson@lbl.gov +# RFC3821 || M. Rajagopal, E. Rodriguez, R. Weber || +# RFC3822 || D. Peterson || dap@cnt.com +# RFC3823 || B. Kovitz || bkovitz@caltech.edu +# RFC3824 || J. Peterson, H. Liu, J. Yu, B. Campbell || jon.peterson@neustar.biz, hong.liu@neustar.biz, james.yu@neustar.biz, bcampbell@dynamicsoft.com +# RFC3825 || J. Polk, J. Schnizlein, M. Linsner || jmpolk@cisco.com, john.schnizlein@cisco.com, marc.linsner@cisco.com +# RFC3826 || U. Blumenthal, F. Maino, K. McCloghrie || uri@bell-labs.com, fmaino@andiamo.com, kzm@cisco.com +# RFC3827 || K. Sarcar || kanoj.sarcar@sun.com +# RFC3828 || L-A. Larzon, M. Degermark, S. Pink, L-E. Jonsson, Ed., G. Fairhurst, Ed. || lln@csee.ltu.se, micke@cs.arizona.edu, steve@cs.arizona.edu, lars-erik.jonsson@ericsson.com, gorry@erg.abdn.ac.uk +# RFC3829 || R. Weltman, M. Smith, M. Wahl || robw@worldspot.com, mcs@pearlcrescent.com +# RFC3830 || J. Arkko, E. Carrara, F. Lindholm, M. Naslund, K. Norrman || jari.arkko@ericsson.com, elisabetta.carrara@ericsson.com, fredrik.lindholm@ericsson.com, mats.naslund@ericsson.com, karl.norrman@ericsson.com +# RFC3831 || C. DeSanti || cds@cisco.com +# RFC3832 || W. Zhao, H. Schulzrinne, E. Guttman, C. Bisdikian, W. Jerome || zwb@cs.columbia.edu, hgs@cs.columbia.edu, Erik.Guttman@sun.com, bisdik@us.ibm.com, wfj@us.ibm.com +# RFC3833 || D. Atkins, R. Austein || derek@ihtfp.com, sra@isc.org +# RFC3834 || K. Moore || moore@cs.utk.edu +# RFC3835 || A. Barbir, R. Penno, R. Chen, M. Hofmann, H. Orman || abbieb@nortelnetworks.com, chen@research.att.com, hofmann@bell-labs.com, ho@alum.mit.edu, rpenno@nortelnetworks.com +# RFC3836 || A. Beck, M. Hofmann, H. Orman, R. Penno, A. Terzis || abeck@bell-labs.com, hofmann@bell-labs.com, ho@alum.mit.edu, rpenno@nortelnetworks.com, terzis@cs.jhu.edu +# RFC3837 || A. Barbir, O. Batuner, B. Srinivas, M. Hofmann, H. Orman || abbieb@nortelnetworks.com, batuner@attbi.com, bindignavile.srinivas@nokia.com, hofmann@bell-labs.com, ho@alum.mit.edu +# RFC3838 || A. Barbir, O. Batuner, A. Beck, T. Chan, H. Orman || abbieb@nortelnetworks.com, batuner@attbi.com, abeck@bell-labs.com, Tat.Chan@nokia.com, ho@alum.mit.edu +# RFC3839 || R. Castagno, D. Singer || +# RFC3840 || J. Rosenberg, H. Schulzrinne, P. Kyzivat || jdrosen@dynamicsoft.com, schulzrinne@cs.columbia.edu, pkyzivat@cisco.com +# RFC3841 || J. Rosenberg, H. Schulzrinne, P. Kyzivat || jdrosen@dynamicsoft.com, schulzrinne@cs.columbia.edu, pkyzivat@cisco.com +# RFC3842 || R. Mahy || rohan@cisco.com +# RFC3843 || L-E. Jonsson, G. Pelletier || lars-erik.jonsson@ericsson.com, ghyslain.pelletier@ericsson.com +# RFC3844 || E. Davies, Ed., J. Hofmann, Ed. || elwynd@nortelnetworks.com, jeanette@wz-berlin.de +# RFC3845 || J. Schlyter, Ed. || jakob@nic.se +# RFC3846 || F. Johansson, T. Johansson || fredrik@ipunplugged.com, tony.johansson@bytemobile.com +# RFC3847 || M. Shand, L. Ginsberg || mshand@cisco.com, ginsberg@cisco.com +# RFC3848 || C. Newman || chris.newman@sun.com +# RFC3849 || G. Huston, A. Lord, P. Smith || gih@apnic.net, anne@apnic.net, pfs@cisco.com +# RFC3850 || B. Ramsdell, Ed. || +# RFC3851 || B. Ramsdell, Ed. || +# RFC3852 || R. Housley || housley@vigilsec.com +# RFC3853 || J. Peterson || jon.peterson@neustar.biz +# RFC3854 || P. Hoffman, C. Bonatti, A. Eggen || +# RFC3855 || P. Hoffman, C. Bonatti || phoffman@imc.org, bonattic@ieca.com +# RFC3856 || J. Rosenberg || jdrosen@dynamicsoft.com +# RFC3857 || J. Rosenberg || jdrosen@dynamicsoft.com +# RFC3858 || J. Rosenberg || jdrosen@dynamicsoft.com +# RFC3859 || J. Peterson || jon.peterson@neustar.biz +# RFC3860 || J. Peterson || jon.peterson@neustar.biz +# RFC3861 || J. Peterson || jon.peterson@neustar.biz +# RFC3862 || G. Klyne, D. Atkins || GK-IETF@ninebynine.org, derek@ihtfp.com +# RFC3863 || H. Sugano, S. Fujimoto, G. Klyne, A. Bateman, W. Carr, J. Peterson || sugano.h@jp.fujitsu.com, shingo_fujimoto@jp.fujitsu.com, GK@ninebynine.org, bateman@acm.org, wayne.carr@intel.com, jon.peterson@neustar.biz +# RFC3864 || G. Klyne, M. Nottingham, J. Mogul || GK-IETF@ninebynine.org, mnot@pobox.com, JeffMogul@acm.org +# RFC3865 || C. Malamud || carl@media.org +# RFC3866 || K. Zeilenga, Ed. || +# RFC3867 || Y. Kawatsura, M. Hiroya, H. Beykirch || ykawatsu@itg.hitachi.co.jp, hiroya@st.rim.or.jp, hbbeykirch@web.de +# RFC3868 || J. Loughney, Ed., G. Sidebottom, L. Coene, G. Verwimp, J. Keller, B. Bidulock || john.Loughney@nokia.com, greg@signatustechnologies.com, lode.coene@siemens.com, gery.verwimp@siemens.com, joe.keller@tekelec.com, bidulock@openss7.org +# RFC3869 || R. Atkinson, Ed., S. Floyd, Ed., Internet Architecture Board || iab@iab.org, iab@iab.org, iab@iab.org +# RFC3870 || A. Swartz || me@aaronsw.com +# RFC3871 || G. Jones, Ed. || gmj3871@pobox.com +# RFC3872 || D. Zinman, D. Walker, J. Jiang || dzinman@rogers.com, david.walker@sedna-wireless.com, jjiang@syndesis.com +# RFC3873 || J. Pastor, M. Belinchon || J.Javier.Pastor@ericsson.com, maria.carmen.belinchon@ericsson.com +# RFC3874 || R. Housley || housley@vigilsec.com +# RFC3875 || D. Robinson, K. Coar || drtr@apache.org, coar@apache.org +# RFC3876 || D. Chadwick, S. Mullan || d.w.chadwick@salford.ac.uk, sean.mullan@sun.com +# RFC3877 || S. Chisholm, D. Romascanu || schishol@nortelnetworks.com, dromasca@gmail.com +# RFC3878 || H. Lam, A. Huynh, D. Perkins || hklam@lucent.com, a_n_huynh@yahoo.com, dperkins@snmpinfo.com +# RFC3879 || C. Huitema, B. Carpenter || huitema@microsoft.com, brc@zurich.ibm.com +# RFC3880 || J. Lennox, X. Wu, H. Schulzrinne || lennox@cs.columbia.edu, xiaotaow@cs.columbia.edu, schulzrinne@cs.columbia.edu +# RFC3881 || G. Marshall || glen.f.marshall@siemens.com +# RFC3882 || D. Turk || doughan.turk@bell.ca +# RFC3883 || S. Rao, A. Zinin, A. Roy || siraprao@hotmail.com, zinin@psg.com, akr@cisco.com +# RFC3884 || J. Touch, L. Eggert, Y. Wang || touch@isi.edu, lars.eggert@netlab.nec.de, yushunwa@isi.edu +# RFC3885 || E. Allman, T. Hansen || eric@Sendmail.COM, tony+msgtrk@maillennium.att.com +# RFC3886 || E. Allman || eric@Sendmail.COM +# RFC3887 || T. Hansen || tony+msgtrk@maillennium.att.com +# RFC3888 || T. Hansen || tony+msgtrk@maillennium.att.com +# RFC3889 || || +# RFC3890 || M. Westerlund || Magnus.Westerlund@ericsson.com +# RFC3891 || R. Mahy, B. Biggs, R. Dean || rohan@cisco.com, bbiggs@dumbterm.net, rfc@fdd.com +# RFC3892 || R. Sparks || RjS@xten.com +# RFC3893 || J. Peterson || jon.peterson@neustar.biz +# RFC3894 || J. Degener || jutta@sendmail.com +# RFC3895 || O. Nicklass, Ed. || orly_n@rad.com +# RFC3896 || O. Nicklass, Ed. || orly_n@rad.com +# RFC3897 || A. Barbir || abbieb@nortelnetworks.com +# RFC3898 || V. Kalusivalingam || vibhaska@cisco.com +# RFC3901 || A. Durand, J. Ihren || Alain.Durand@sun.com, johani@autonomica.se +# RFC3902 || M. Baker, M. Nottingham || distobj@acm.org, mnot@pobox.com +# RFC3903 || A. Niemi, Ed. || aki.niemi@nokia.com +# RFC3904 || C. Huitema, R. Austein, S. Satapati, R. van der Pol || huitema@microsoft.com, sra@isc.org, satapati@cisco.com, Ronald.vanderPol@nlnetlabs.nl +# RFC3905 || V. See, Ed. || vsee@microsoft.com +# RFC3906 || N. Shen, H. Smit || naiming@redback.com, hhwsmit@xs4all.nl +# RFC3909 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC3910 || V. Gurbani, Ed., A. Brusilovsky, I. Faynberg, J. Gato, H. Lu, M. Unmehopa || vkg@lucent.com, abrusilovsky@lucent.com, faynberg@lucent.com, jorge.gato@vodafone.com, huilanlu@lucent.com, unmehopa@lucent.com +# RFC3911 || R. Mahy, D. Petrie || rohan@airespace.com, dpetrie@pingtel.com +# RFC3912 || L. Daigle || leslie@verisignlabs.com +# RFC3913 || D. Thaler || dthaler@microsoft.com +# RFC3914 || A. Barbir, A. Rousskov || abbieb@nortelnetworks.com, rousskov@measurement-factory.com +# RFC3915 || S. Hollenbeck || shollenbeck@verisign.com +# RFC3916 || X. Xiao, Ed., D. McPherson, Ed., P. Pate, Ed. || xxiao@riverstonenet.com, danny@arbor.net, prayson.pate@overturenetworks.com +# RFC3917 || J. Quittek, T. Zseby, B. Claise, S. Zander || quittek@netlab.nec.de, zseby@fokus.fhg.de, bclaise@cisco.com, szander@swin.edu.au +# RFC3918 || D. Stopp, B. Hickman || debby@ixiacom.com, brooks.hickman@spirentcom.com +# RFC3919 || E. Stephan, J. Palet || emile.stephan@francetelecom.com, jordi.palet@consulintel.es +# RFC3920 || P. Saint-Andre, Ed. || ietf@stpeter.im +# RFC3921 || P. Saint-Andre, Ed. || ietf@stpeter.im +# RFC3922 || P. Saint-Andre || ietf@stpeter.im +# RFC3923 || P. Saint-Andre || ietf@stpeter.im +# RFC3924 || F. Baker, B. Foster, C. Sharp || fred@cisco.com, bfoster@cisco.com, chsharp@cisco.com +# RFC3925 || J. Littlefield || joshl@cisco.com +# RFC3926 || T. Paila, M. Luby, R. Lehtonen, V. Roca, R. Walsh || toni.paila@nokia.com, luby@digitalfountain.com, rami.lehtonen@teliasonera.com, vincent.roca@inrialpes.fr, rod.walsh@nokia.com +# RFC3927 || S. Cheshire, B. Aboba, E. Guttman || rfc@stuartcheshire.org, bernarda@microsoft.com, erik@spybeam.org +# RFC3928 || R. Megginson, Ed., M. Smith, O. Natkovich, J. Parham || rmegginson0224@aol.com, olgan@yahoo-inc.com, mcs@pearlcrescent.com, jeffparh@microsoft.com +# RFC3929 || T. Hardie || hardie@qualcomm.com +# RFC3930 || D. Eastlake 3rd || Donald.Eastlake@motorola.com +# RFC3931 || J. Lau, Ed., M. Townsley, Ed., I. Goyret, Ed. || jedlau@cisco.com, mark@townsley.net, igoyret@lucent.com +# RFC3932 || H. Alvestrand || harald@alvestrand.no +# RFC3933 || J. Klensin, S. Dawkins || john-ietf@jck.com, spencer@mcsr-labs.org +# RFC3934 || M. Wasserman || margaret@thingmagic.com +# RFC3935 || H. Alvestrand || harald@alvestrand.no +# RFC3936 || K. Kompella, J. Lang || kireeti@juniper.net, jplang@ieee.org +# RFC3937 || M. Steidl || mdirector@iptc.org +# RFC3938 || T. Hansen || tony+msgctxt@maillennium.att.com +# RFC3939 || G. Parsons, J. Maruszak || gparsons@nortelnetworks.com, jjmaruszak@sympatico.ca +# RFC3940 || B. Adamson, C. Bormann, M. Handley, J. Macker || adamson@itd.nrl.navy.mil, cabo@tzi.org, M.Handley@cs.ucl.ac.uk, macker@itd.nrl.navy.mil +# RFC3941 || B. Adamson, C. Bormann, M. Handley, J. Macker || adamson@itd.nrl.navy.mil, cabo@tzi.org, M.Handley@cs.ucl.ac.uk, macker@itd.nrl.navy.mil +# RFC3942 || B. Volz || volz@cisco.com +# RFC3943 || R. Friend || rfriend@hifn.com +# RFC3944 || T. Johnson, S. Okubo, S. Campos || Tyler_Johnson@unc.edu, sokubo@waseda.jp, simao.campos@itu.int +# RFC3945 || E. Mannie, Ed. || eric_mannie@hotmail.com +# RFC3946 || E. Mannie, D. Papadimitriou || eric_mannie@hotmail.com, dimitri.papadimitriou@alcatel.be +# RFC3947 || T. Kivinen, B. Swander, A. Huttunen, V. Volpe || kivinen@safenet-inc.com, Ari.Huttunen@F-Secure.com, briansw@microsoft.com, vvolpe@cisco.com +# RFC3948 || A. Huttunen, B. Swander, V. Volpe, L. DiBurro, M. Stenberg || Ari.Huttunen@F-Secure.com, briansw@microsoft.com, vvolpe@cisco.com, ldiburro@nortelnetworks.com, markus.stenberg@iki.fi +# RFC3949 || R. Buckley, D. Venable, L. McIntyre, G. Parsons, J. Rafferty || rbuckley@crt.xerox.com, dvenable@crt.xerox.com, lloyd10328@pacbell.net, gparsons@nortel.com, jraff@brooktrout.com +# RFC3950 || L. McIntyre, G. Parsons, J. Rafferty || lloyd10328@pacbell.net, gparsons@nortel.com, jraff@brooktrout.com +# RFC3951 || S. Andersen, A. Duric, H. Astrom, R. Hagen, W. Kleijn, J. Linden || sva@kom.auc.dk, alan.duric@telio.no, henrik.astrom@globalipsound.com, roar.hagen@globalipsound.com, bastiaan.kleijn@globalipsound.com, jan.linden@globalipsound.com +# RFC3952 || A. Duric, S. Andersen || alan.duric@telio.no, sva@kom.auc.dk +# RFC3953 || J. Peterson || jon.peterson@neustar.biz +# RFC3954 || B. Claise, Ed. || bclaise@cisco.com +# RFC3955 || S. Leinen || simon@switch.ch +# RFC3956 || P. Savola, B. Haberman || psavola@funet.fi, brian@innovationslab.net +# RFC3957 || C. Perkins, P. Calhoun || charles.perkins@nokia.com, pcalhoun@airespace.com +# RFC3958 || L. Daigle, A. Newton || leslie@thinkingcat.com, anewton@verisignlabs.com +# RFC3959 || G. Camarillo || Gonzalo.Camarillo@ericsson.com +# RFC3960 || G. Camarillo, H. Schulzrinne || Gonzalo.Camarillo@ericsson.com, schulzrinne@cs.columbia.edu +# RFC3961 || K. Raeburn || raeburn@mit.edu +# RFC3962 || K. Raeburn || raeburn@mit.edu +# RFC3963 || V. Devarapalli, R. Wakikawa, A. Petrescu, P. Thubert || vijay.devarapalli@nokia.com, ryuji@sfc.wide.ad.jp, Alexandru.Petrescu@motorola.com, pthubert@cisco.com +# RFC3964 || P. Savola, C. Patel || psavola@funet.fi, chirayu@chirayu.org +# RFC3965 || K. Toyoda, H. Ohno, J. Murai, D. Wing || toyoda.kiyoshi@jp.panasonic.com, hohno@ohnolab.org, jun@wide.ad.jp, dwing-ietf@fuggles.com +# RFC3966 || H. Schulzrinne || hgs@cs.columbia.edu +# RFC3967 || R. Bush, T. Narten || randy@psg.com, narten@us.ibm.com +# RFC3968 || G. Camarillo || Gonzalo.Camarillo@ericsson.com +# RFC3969 || G. Camarillo || Gonzalo.Camarillo@ericsson.com +# RFC3970 || K. Kompella || kireeti@juniper.net +# RFC3971 || J. Arkko, Ed., J. Kempf, B. Zill, P. Nikander || jari.arkko@ericsson.com, kempf@docomolabs-usa.com, bzill@microsoft.com, Pekka.Nikander@nomadiclab.com +# RFC3972 || T. Aura || tuomaura@microsoft.com +# RFC3973 || A. Adams, J. Nicholas, W. Siadak || ala@nexthop.com, jonathan.nicholas@itt.com, wfs@nexthop.com +# RFC3974 || M. Nakamura, J. Hagino || motonori@media.kyoto-u.ac.jp, itojun@iijlab.net +# RFC3975 || G. Huston, Ed., I. Leuca, Ed. || execd@iab.org, ileana.leuca@Cingular.com +# RFC3976 || V. K. Gurbani, F. Haerens, V. Rastogi || vkg@lucent.com, frans.haerens@alcatel.be, vidhi.rastogi@wipro.com +# RFC3977 || C. Feather || clive@davros.org +# RFC3978 || S. Bradner, Ed. || sob@harvard.edu +# RFC3979 || S. Bradner, Ed. || sob@harvard.edu +# RFC3980 || M. Krueger, M. Chadalapaka, R. Elliott || marjorie_krueger@hp.com, cbm@rose.hp.com, elliott@hp.com +# RFC3981 || A. Newton, M. Sanz || anewton@verisignlabs.com, sanz@denic.de +# RFC3982 || A. Newton, M. Sanz || anewton@verisignlabs.com, sanz@denic.de +# RFC3983 || A. Newton, M. Sanz || anewton@verisignlabs.com, sanz@denic.de +# RFC3984 || S. Wenger, M.M. Hannuksela, T. Stockhammer, M. Westerlund, D. Singer || stewe@stewe.org, miska.hannuksela@nokia.com, stockhammer@nomor.de, magnus.westerlund@ericsson.com, singer@apple.com +# RFC3985 || S. Bryant, Ed., P. Pate, Ed. || stbryant@cisco.com, prayson.pate@overturenetworks.com +# RFC3986 || T. Berners-Lee, R. Fielding, L. Masinter || timbl@w3.org, fielding@gbiv.com, LMM@acm.org +# RFC3987 || M. Duerst, M. Suignard || duerst@w3.org, michelsu@microsoft.com +# RFC3988 || B. Black, K. Kompella || ben@layer8.net, kireeti@juniper.net +# RFC3989 || M. Stiemerling, J. Quittek, T. Taylor || stiemerling@netlab.nec.de, quittek@netlab.nec.de, tom.taylor.stds@gmail.com +# RFC3990 || B. O'Hara, P. Calhoun, J. Kempf || bob@airespace.com, pcalhoun@airespace.com, kempf@docomolabs-usa.com +# RFC3991 || B. Foster, F. Andreasen || bfoster@cisco.com, fandreas@cisco.com +# RFC3992 || B. Foster, F. Andreasen || bfoster@cisco.com, fandreas@cisco.com +# RFC3993 || R. Johnson, T. Palaniappan, M. Stapp || raj@cisco.com, athenmoz@cisco.com, mjs@cisco.com +# RFC3994 || H. Schulzrinne || hgs@cs.columbia.edu +# RFC3995 || R. Herriot, T. Hastings || bob@herriot.com, tom.hastings@alum.mit.edu +# RFC3996 || R. Herriot, T. Hastings, H. Lewis || bob@herriot.com, tom.hastings@alum.mit.edu, harryl@us.ibm.com +# RFC3997 || T. Hastings, Ed., R. K. deBry, H. Lewis || tom.hastings@alum.mit.edu, debryro@uvsc.edu, harryl@us.ibm.com +# RFC3998 || C. Kugler, H. Lewis, T. Hastings, Ed. || kugler@us.ibm.com, harryl@us.ibm.com, tom.hastings@alum.mit.edu +# RFC4001 || M. Daniele, B. Haberman, S. Routhier, J. Schoenwaelder || michael.daniele@syamsoftware.com, brian@innovationslab.net, shawn.routhier@windriver.com, j.schoenwaelder@iu-bremen.de +# RFC4002 || R. Brandner, L. Conroy, R. Stastny || rudolf.brandner@siemens.com, lwc@roke.co.uk, richard.stastny@oefeg.at +# RFC4003 || L. Berger || lberger@movaz.com +# RFC4004 || P. Calhoun, T. Johansson, C. Perkins, T. Hiller, Ed., P. McCann || pcalhoun@cisco.com, tony.johansson@bytemobile.com, Charles.Perkins@nokia.com, tomhiller@lucent.com, mccap@lucent.com +# RFC4005 || P. Calhoun, G. Zorn, D. Spence, D. Mitton || pcalhoun@cisco.com, gwz@cisco.com, dspence@computer.org, dmitton@circularnetworks.com +# RFC4006 || H. Hakala, L. Mattila, J-P. Koskinen, M. Stura, J. Loughney || Harri.Hakala@ericsson.com, Leena.Mattila@ericsson.com, juha-pekka.koskinen@nokia.com, marco.stura@nokia.com, John.Loughney@nokia.com +# RFC4007 || S. Deering, B. Haberman, T. Jinmei, E. Nordmark, B. Zill || none, brian@innovationslab.net, jinmei@isl.rdc.toshiba.co.jp, Erik.Nordmark@sun.com, bzill@microsoft.com +# RFC4008 || R. Rohit, P. Srisuresh, R. Raghunarayan, N. Pai, C. Wang || rrohit74@hotmail.com, srisuresh@yahoo.com, raraghun@cisco.com, npai@cisco.com, cliffwang2000@yahoo.com +# RFC4009 || J. Park, S. Lee, J. Kim, J. Lee || khopri@kisa.or.kr, sjlee@kisa.or.kr, jykim@kisa.or.kr, jilee@kisa.or.kr +# RFC4010 || J. Park, S. Lee, J. Kim, J. Lee || khopri@kisa.or.kr, sjlee@kisa.or.kr, jykim@kisa.or.kr, jilee@kisa.or.kr +# RFC4011 || S. Waldbusser, J. Saperia, T. Hongal || waldbusser@nextbeacon.com, saperia@jdscons.com, hongal@riverstonenet.com +# RFC4012 || L. Blunk, J. Damas, F. Parent, A. Robachevsky || ljb@merit.edu, Joao_Damas@isc.org, Florent.Parent@hexago.com, andrei@ripe.net +# RFC4013 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC4014 || R. Droms, J. Schnizlein || rdroms@cisco.com, jschnizl@cisco.com +# RFC4015 || R. Ludwig, A. Gurtov || Reiner.Ludwig@ericsson.com, andrei.gurtov@cs.helsinki.fi +# RFC4016 || M. Parthasarathy || mohanp@sbcglobal.net +# RFC4017 || D. Stanley, J. Walker, B. Aboba || dstanley@agere.com, jesse.walker@intel.com, bernarda@microsoft.com +# RFC4018 || M. Bakke, J. Hufferd, K. Voruganti, M. Krueger, T. Sperry || mbakke@cisco.com, jlhufferd@comcast.net, kaladhar@us.ibm.com, marjorie_krueger@hp.com, todd_sperry@adaptec.com +# RFC4019 || G. Pelletier || ghyslain.pelletier@ericsson.com +# RFC4020 || K. Kompella, A. Zinin || kireeti@juniper.net, zinin@psg.com +# RFC4021 || G. Klyne, J. Palme || GK-IETF@ninebynine.org, jpalme@dsv.su.se +# RFC4022 || R. Raghunarayan, Ed. || raraghun@cisco.com +# RFC4023 || T. Worster, Y. Rekhter, E. Rosen, Ed. || tom.worster@motorola.com, yakov@juniper.net, erosen@cisco.com +# RFC4024 || G. Parsons, J. Maruszak || gparsons@nortel.com, jjmaruszak@sympatico.ca +# RFC4025 || M. Richardson || mcr@sandelman.ottawa.on.ca +# RFC4026 || L. Andersson, T. Madsen || loa@pi.se, tove.madsen@acreo.se +# RFC4027 || S. Josefsson || simon@josefsson.org +# RFC4028 || S. Donovan, J. Rosenberg || srd@cisco.com, jdrosen@cisco.com +# RFC4029 || M. Lind, V. Ksinant, S. Park, A. Baudot, P. Savola || mikael.lind@teliasonera.com, vladimir.ksinant@fr.thalesgroup.com, soohong.park@samsung.com, alain.baudot@francetelecom.com, psavola@funet.fi +# RFC4030 || M. Stapp, T. Lemon || mjs@cisco.com, Ted.Lemon@nominum.com +# RFC4031 || M. Carugi, Ed., D. McDysan, Ed. || marco.carugi@nortel.com, dave.mcdysan@mci.com +# RFC4032 || G. Camarillo, P. Kyzivat || Gonzalo.Camarillo@ericsson.com, pkyzivat@cisco.com +# RFC4033 || R. Arends, R. Austein, M. Larson, D. Massey, S. Rose || roy.arends@telin.nl, sra@isc.org, mlarson@verisign.com, massey@cs.colostate.edu, scott.rose@nist.gov +# RFC4034 || R. Arends, R. Austein, M. Larson, D. Massey, S. Rose || roy.arends@telin.nl, sra@isc.org, mlarson@verisign.com, massey@cs.colostate.edu, scott.rose@nist.gov +# RFC4035 || R. Arends, R. Austein, M. Larson, D. Massey, S. Rose || roy.arends@telin.nl, sra@isc.org, mlarson@verisign.com, massey@cs.colostate.edu, scott.rose@nist.gov +# RFC4036 || W. Sawyer || wsawyer@ieee.org +# RFC4037 || A. Rousskov || rousskov@measurement-factory.com +# RFC4038 || M-K. Shin, Ed., Y-G. Hong, J. Hagino, P. Savola, E. M. Castro || mshin@nist.gov, yghong@pec.etri.re.kr, itojun@iijlab.net, psavola@funet.fi, eva@gsyc.escet.urjc.es +# RFC4039 || S. Park, P. Kim, B. Volz || soohong.park@samsung.com, kimps@samsung.com, volz@cisco.com +# RFC4040 || R. Kreuter || ruediger.kreuter@siemens.com +# RFC4041 || A. Farrel || adrian@olddog.co.uk +# RFC4042 || M. Crispin || UTF9@Lingling.Panda.COM +# RFC4043 || D. Pinkas, T. Gindin || Denis.Pinkas@bull.net, tgindin@us.ibm.com +# RFC4044 || K. McCloghrie || kzm@cisco.com +# RFC4045 || G. Bourdon || gilles.bourdon@francetelecom.com +# RFC4046 || M. Baugher, R. Canetti, L. Dondeti, F. Lindholm || mbaugher@cisco.com, canetti@watson.ibm.com, ldondeti@qualcomm.com, fredrik.lindholm@ericsson.com +# RFC4047 || S. Allen, D. Wells || sla@ucolick.org, dwells@nrao.edu +# RFC4048 || B. Carpenter || brc@zurich.ibm.com +# RFC4049 || R. Housley || housley@vigilsec.com +# RFC4050 || S. Blake-Wilson, G. Karlinger, T. Kobayashi, Y. Wang || sblakewilson@bcisse.com, gregor.karlinger@cio.gv.at, kotetsu@isl.ntt.co.jp, yonwang@uncc.edu +# RFC4051 || D. Eastlake 3rd || Donald.Eastlake@motorola.com +# RFC4052 || L. Daigle, Ed., Internet Architecture Board || iab@iab.org, iab@iab.org +# RFC4053 || S. Trowbridge, S. Bradner, F. Baker || sjtrowbridge@lucent.com, sob@harvard.edu, fred@cisco.com +# RFC4054 || J. Strand, Ed., A. Chiu, Ed. || jls@research.att.com, chiu@research.att.com +# RFC4055 || J. Schaad, B. Kaliski, R. Housley || jimsch@exmsft.com, bkaliski@rsasecurity.com, housley@vigilsec.com +# RFC4056 || J. Schaad || jimsch@exmsft.com +# RFC4057 || J. Bound, Ed. || jim.bound@hp.com +# RFC4058 || A. Yegin, Ed., Y. Ohba, R. Penno, G. Tsirtsis, C. Wang || alper.yegin@samsung.com, yohba@tari.toshiba.com, rpenno@juniper.net, G.Tsirtsis@Flarion.com, cliffwangmail@yahoo.com +# RFC4059 || D. Linsenbardt, S. Pontius, A. Sturgeon || dlinsenbardt@spyrus.com, spontius@spyrus.com, asturgeon@spyrus.com +# RFC4060 || Q. Xie, D. Pearce || qxie1@email.mot.com, bdp003@motorola.com +# RFC4061 || V. Manral, R. White, A. Shaikh || vishwas@sinett.com, riw@cisco.com, ashaikh@research.att.com +# RFC4062 || V. Manral, R. White, A. Shaikh || vishwas@sinett.com, riw@cisco.com, ashaikh@research.att.com +# RFC4063 || V. Manral, R. White, A. Shaikh || vishwas@sinett.com, riw@cisco.com, ashaikh@research.att.com +# RFC4064 || A. Patel, K. Leung || alpesh@cisco.com, kleung@cisco.com +# RFC4065 || J. Kempf || kempf@docomolabs-usa.com +# RFC4066 || M. Liebsch, Ed., A. Singh, Ed., H. Chaskar, D. Funato, E. Shim || marco.liebsch@netlab.nec.de, asingh1@email.mot.com, hemant.chaskar@airtightnetworks.net, funato@mlab.yrp.nttdocomo.co.jp, eunsoo@research.panasonic.com +# RFC4067 || J. Loughney, Ed., M. Nakhjiri, C. Perkins, R. Koodli || john.loughney@nokia.com, madjid.nakhjiri@motorola.com, charles.perkins@.nokia.com, rajeev.koodli@nokia.com +# RFC4068 || R. Koodli, Ed. || Rajeev.Koodli@nokia.com +# RFC4069 || M. Dodge, B. Ray || mbdodge@ieee.org, rray@pesa.com +# RFC4070 || M. Dodge, B. Ray || mbdodge@ieee.org, rray@pesa.com +# RFC4071 || R. Austein, Ed., B. Wijnen, Ed. || sra@isc.org, bwijnen@lucent.com +# RFC4072 || P. Eronen, Ed., T. Hiller, G. Zorn || pe@iki.fi, tomhiller@lucent.com, gwz@cisco.com +# RFC4073 || R. Housley || housley@vigilsec.com +# RFC4074 || Y. Morishita, T. Jinmei || yasuhiro@jprs.co.jp, jinmei@isl.rdc.toshiba.co.jp +# RFC4075 || V. Kalusivalingam || vibhaska@cisco.com +# RFC4076 || T. Chown, S. Venaas, A. Vijayabhaskar || tjc@ecs.soton.ac.uk, venaas@uninett.no, vibhaska@cisco.com +# RFC4077 || A.B. Roach || adam@estacado.net +# RFC4078 || N. Earnshaw, S. Aoki, A. Ashley, W. Kameyama || nigel.earnshaw@rd.bbc.co.uk, shig@center.jfn.co.jp, aashley@ndsuk.com, wataru@waseda.jp +# RFC4079 || J. Peterson || jon.peterson@neustar.biz +# RFC4080 || R. Hancock, G. Karagiannis, J. Loughney, S. Van den Bosch || robert.hancock@roke.co.uk, g.karagiannis@ewi.utwente.nl, john.loughney@nokia.com, sven.van_den_bosch@alcatel.be +# RFC4081 || H. Tschofenig, D. Kroeselberg || Hannes.Tschofenig@siemens.com, Dirk.Kroeselberg@siemens.com +# RFC4082 || A. Perrig, D. Song, R. Canetti, J. D. Tygar, B. Briscoe || perrig@cmu.edu, dawnsong@cmu.edu, canetti@watson.ibm.com, doug.tygar@gmail.com, bob.briscoe@bt.com +# RFC4083 || M. Garcia-Martin || miguel.an.garcia@nokia.com +# RFC4084 || J. Klensin || john-ietf@jck.com +# RFC4085 || D. Plonka || plonka@doit.wisc.edu +# RFC4086 || D. Eastlake 3rd, J. Schiller, S. Crocker || Donald.Eastlake@motorola.com, jis@mit.edu, steve@stevecrocker.com +# RFC4087 || D. Thaler || dthaler@microsoft.com +# RFC4088 || D. Black, K. McCloghrie, J. Schoenwaelder || black_david@emc.com, kzm@cisco.com, j.schoenwaelder@iu-bremen.de +# RFC4089 || S. Hollenbeck, Ed., IAB and IESG || sah@428cobrajet.net, none, none +# RFC4090 || P. Pan, Ed., G. Swallow, Ed., A. Atlas, Ed. || ppan@hammerheadsystems.com, swallow@cisco.com, aatlas@avici.com +# RFC4091 || G. Camarillo, J. Rosenberg || Gonzalo.Camarillo@ericsson.com, jdrosen@cisco.com +# RFC4092 || G. Camarillo, J. Rosenberg || Gonzalo.Camarillo@ericsson.com, jdrosen@cisco.com +# RFC4093 || F. Adrangi, Ed., H. Levkowetz, Ed. || farid.adrangi@intel.com, henrik@levkowetz.com +# RFC4094 || J. Manner, X. Fu || jmanner@cs.helsinki.fi, fu@cs.uni-goettingen.de +# RFC4095 || C. Malamud || carl@media.org +# RFC4096 || C. Malamud || carl@media.org +# RFC4097 || M. Barnes, Ed. || mary.barnes@nortel.com +# RFC4098 || H. Berkowitz, E. Davies, Ed., S. Hares, P. Krishnaswamy, M. Lepp || hcb@gettcomm.com, elwynd@dial.pipex.com, skh@nexthop.com, padma.krishnaswamy@saic.com, mlepp@lepp.com +# RFC4101 || E. Rescorla, IAB || ekr@rtfm.com, iab@iab.org +# RFC4102 || P. Jones || paulej@packetizer.com +# RFC4103 || G. Hellstrom, P. Jones || gunnar.hellstrom@omnitor.se, paulej@packetizer.com +# RFC4104 || M. Pana, Ed., A. Reyes, A. Barba, D. Moron, M. Brunner || mpana@metasolv.com, mreyes@ac.upc.edu, telabm@mat.upc.es, dmor4477@hotmail.com, brunner@netlab.nec.de +# RFC4105 || J.-L. Le Roux, Ed., J.-P. Vasseur, Ed., J. Boyle, Ed. || jeanlouis.leroux@francetelecom.com, jpv@cisco.com, jboyle@pdnets.com +# RFC4106 || J. Viega, D. McGrew || viega@securesoftware.com, mcgrew@cisco.com +# RFC4107 || S. Bellovin, R. Housley || bellovin@acm.org, housley@vigilsec.com +# RFC4108 || R. Housley || housley@vigilsec.com +# RFC4109 || P. Hoffman || paul.hoffman@vpnc.org +# RFC4110 || R. Callon, M. Suzuki || rcallon@juniper.net, suzuki.muneyoshi@lab.ntt.co.jp +# RFC4111 || L. Fang, Ed. || luyuanfang@att.com +# RFC4112 || D. Eastlake 3rd || Donald.Eastlake@motorola.com +# RFC4113 || B. Fenner, J. Flick || fenner@research.att.com, john.flick@hp.com +# RFC4114 || S. Hollenbeck || shollenbeck@verisign.com +# RFC4115 || O. Aboul-Magd, S. Rabie || osama@nortel.com, rabie@nortel.com +# RFC4116 || J. Abley, K. Lindqvist, E. Davies, B. Black, V. Gill || jabley@isc.org, kurtis@kurtis.pp.se, elwynd@dial.pipex.com, ben@layer8.net, vgill@vijaygill.com +# RFC4117 || G. Camarillo, E. Burger, H. Schulzrinne, A. van Wijk || Gonzalo.Camarillo@ericsson.com, eburger@brooktrout.com, schulzrinne@cs.columbia.edu, a.vwijk@viataal.nl +# RFC4118 || L. Yang, P. Zerfos, E. Sadot || lily.l.yang@intel.com, pzerfos@cs.ucla.edu, esadot@avaya.com +# RFC4119 || J. Peterson || jon.peterson@neustar.biz +# RFC4120 || C. Neuman, T. Yu, S. Hartman, K. Raeburn || bcn@isi.edu, tlyu@mit.edu, hartmans-ietf@mit.edu, raeburn@mit.edu +# RFC4121 || L. Zhu, K. Jaganathan, S. Hartman || LZhu@microsoft.com, karthikj@microsoft.com, hartmans-ietf@mit.edu +# RFC4122 || P. Leach, M. Mealling, R. Salz || paulle@microsoft.com, michael@refactored-networks.com, rsalz@datapower.com +# RFC4123 || H. Schulzrinne, C. Agboh || hgs@cs.columbia.edu, charles.agboh@packetizer.com +# RFC4124 || F. Le Faucheur, Ed. || flefauch@cisco.com +# RFC4125 || F. Le Faucheur, W. Lai || flefauch@cisco.com, wlai@att.com +# RFC4126 || J. Ash || gash@att.com +# RFC4127 || F. Le Faucheur, Ed. || flefauch@cisco.com +# RFC4128 || W. Lai || wlai@att.com +# RFC4129 || R. Mukundan, K. Morneault, N. Mangalpally || ranjith.mukundan@wipro.com, kmorneau@cisco.com, narsim@nortelnetworks.com +# RFC4130 || D. Moberg, R. Drummond || dmoberg@cyclonecommerce.com, rvd2@drummondgroup.com +# RFC4131 || S. Green, K. Ozawa, E. Cardona, Ed., A. Katsnelson || rubbersoul3@yahoo.com, Kazuyoshi.Ozawa@toshiba.co.jp, katsnelson6@peoplepc.com, e.cardona@cablelabs.com +# RFC4132 || S. Moriai, A. Kato, M. Kanda || shiho@rd.scei.sony.co.jp, akato@po.ntts.co.jp, kanda.masayuki@lab.ntt.co.jp +# RFC4133 || A. Bierman, K. McCloghrie || andy@yumaworks.com, kzm@cisco.com +# RFC4134 || P. Hoffman, Ed. || phoffman@imc.org +# RFC4135 || JH. Choi, G. Daley || jinchoe@samsung.com, greg.daley@eng.monash.edu.au +# RFC4136 || P. Pillay-Esnault || ppe@cisco.com +# RFC4137 || J. Vollbrecht, P. Eronen, N. Petroni, Y. Ohba || jrv@mtghouse.com, pe@iki.fi, npetroni@cs.umd.edu, yohba@tari.toshiba.com +# RFC4138 || P. Sarolahti, M. Kojo || pasi.sarolahti@nokia.com, kojo@cs.helsinki.fi +# RFC4139 || D. Papadimitriou, J. Drake, J. Ash, A. Farrel, L. Ong || dimitri.papadimitriou@alcatel.be, John.E.Drake2@boeing.com, gash@att.com, adrian@olddog.co.uk, lyong@ciena.com +# RFC4140 || H. Soliman, C. Castelluccia, K. El Malki, L. Bellier || h.soliman@flarion.com, claude.castelluccia@inria.fr, karim@elmalki.homeip.net, ludovic.bellier@inria.fr +# RFC4141 || K. Toyoda, D. Crocker || toyoda.kiyoshi@jp.panasonic.com, dcrocker@bbiw.net +# RFC4142 || D. Crocker, G. Klyne || dcrocker@bbiw.net, GK-IETF@ninebynine.org +# RFC4143 || K. Toyoda, D. Crocker || toyoda.kiyoshi@jp.panasonic.com, dcrocker@bbiw.net +# RFC4144 || D. Eastlake 3rd || Donald.Eastlake@motorola.com +# RFC4145 || D. Yon, G. Camarillo || yon-comedia@rfdsoftware.com, Gonzalo.Camarillo@ericsson.com +# RFC4146 || R. Gellens || randy@qualcomm.com +# RFC4147 || G. Huston || gih@apnic.net +# RFC4148 || E. Stephan || emile.stephan@francetelecom.com +# RFC4149 || C. Kalbfleisch, R. Cole, D. Romascanu || ietf@kalbfleisch.us, robert.cole@jhuapl.edu, dromasca@gmail.com +# RFC4150 || R. Dietz, R. Cole || rdietz@hifn.com, robert.cole@jhuapl.edu +# RFC4151 || T. Kindberg, S. Hawke || timothy@hpl.hp.com, sandro@w3.org +# RFC4152 || K. Tesink, R. Fox || kaj@research.telcordia.com, rfox@telcordia.com +# RFC4153 || K. Fujimura, M. Terada, D. Eastlake 3rd || fujimura.ko@lab.ntt.co.jp, te@rex.yrp.nttdocomo.co.jp, Donald.Eastlake@motorola.com +# RFC4154 || M. Terada, K. Fujimura || te@rex.yrp.nttdocomo.co.jp, fujimura.ko@lab.ntt.co.jp +# RFC4155 || E. Hall || ehall@ntrg.com +# RFC4156 || P. Hoffman || paul.hoffman@vpnc.org +# RFC4157 || P. Hoffman || paul.hoffman@vpnc.org +# RFC4158 || M. Cooper, Y. Dzambasow, P. Hesse, S. Joseph, R. Nicholas || mcooper@orionsec.com, yuriy@anassoc.com, pmhesse@geminisecurity.com, susan.joseph@vdtg.com, richard.nicholas@it.baesystems.com +# RFC4159 || G. Huston || gih@apnic.net +# RFC4160 || K. Mimura, K. Yokoyama, T. Satoh, C. Kanaide, C. Allocchio || mimu@miyabi-labo.net, keiyoko@msn.com, zsatou@t-ns.co.jp, icemilk77@yahoo.co.jp, Claudio.Allocchio@garr.it +# RFC4161 || K. Mimura, K. Yokoyama, T. Satoh, K. Watanabe, C. Kanaide || mimu@miyabi-labo.net, keiyoko@msn.com, zsatou@t-ns.co.jp, knabe@ad.cyberhome.ne.jp, icemilk77@yahoo.co.jp +# RFC4162 || H.J. Lee, J.H. Yoon, J.I. Lee || jiinii@kisa.or.kr, jhyoon@kisa.or.kr, jilee@kisa.or.kr +# RFC4163 || L-E. Jonsson || lars-erik.jonsson@ericsson.com +# RFC4164 || G. Pelletier || ghyslain.pelletier@ericsson.com +# RFC4165 || T. George, B. Bidulock, R. Dantu, H. Schwarzbauer, K. Morneault || tgeorge_tx@verizon.net, bidulock@openss7.org, rdantu@unt.edu, HannsJuergen.Schwarzbauer@Siemens.com, kmorneau@cisco.com +# RFC4166 || L. Coene, J. Pastor-Balbas || lode.coene@siemens.com, J.Javier.Pastor@ericsson.com +# RFC4167 || A. Lindem || acee@cisco.com +# RFC4168 || J. Rosenberg, H. Schulzrinne, G. Camarillo || jdrosen@cisco.com, schulzrinne@cs.columbia.edu, Gonzalo.Camarillo@ericsson.com +# RFC4169 || V. Torvinen, J. Arkko, M. Naslund || vesa.torvinen@turkuamk.fi, jari.arkko@ericsson.com, mats.naslund@ericsson.com +# RFC4170 || B. Thompson, T. Koren, D. Wing || brucet@cisco.com, tmima@cisco.com, dwing-ietf@fuggles.com +# RFC4171 || J. Tseng, K. Gibbons, F. Travostino, C. Du Laney, J. Souza || joshtseng@yahoo.com, kevin.gibbons@mcdata.com, travos@nortel.com, cdl@rincon.com, joes@exmsft.com +# RFC4172 || C. Monia, R. Mullendore, F. Travostino, W. Jeong, M. Edwards || charles_monia@yahoo.com, Rod.Mullendore@MCDATA.com, travos@nortel.com, wayland@TroikaNetworks.com, mark_edwards@adaptec.com +# RFC4173 || P. Sarkar, D. Missimer, C. Sapuntzakis || psarkar@almaden.ibm.com, duncan.missimer@ieee.org, csapuntz@alum.mit.edu +# RFC4174 || C. Monia, J. Tseng, K. Gibbons || charles_monia@yahoo.com, joshtseng@yahoo.com, kevin.gibbons@mcdata.com +# RFC4175 || L. Gharai, C. Perkins || ladan@isi.edu, csp@csperkins.org +# RFC4176 || Y. El Mghazli, Ed., T. Nadeau, M. Boucadair, K. Chan, A. Gonguet || yacine.el_mghazli@alcatel.fr, tnadeau@cisco.com, mohamed.boucadair@francetelecom.com, khchan@nortel.com, arnaud.gonguet@alcatel.fr +# RFC4177 || G. Huston || gih@apnic.net +# RFC4178 || L. Zhu, P. Leach, K. Jaganathan, W. Ingersoll || lzhu@microsoft.com, paulle@microsoft.com, karthikj@microsoft.com, wyllys.ingersoll@sun.com +# RFC4179 || S. Kang || sukang@nca.or.kr +# RFC4180 || Y. Shafranovich || ietf@shaftek.org +# RFC4181 || C. Heard, Ed. || heard@pobox.com +# RFC4182 || E. Rosen || erosen@cisco.com +# RFC4183 || E. Warnicke || eaw@cisco.com +# RFC4184 || B. Link, T. Hager, J. Flaks || bdl@dolby.com, thh@dolby.com, jasonfl@microsoft.com +# RFC4185 || J. Klensin || john-ietf@jck.com +# RFC4186 || H. Haverinen, Ed., J. Salowey, Ed. || henry.haverinen@nokia.com, jsalowey@cisco.com +# RFC4187 || J. Arkko, H. Haverinen || jari.Arkko@ericsson.com, henry.haverinen@nokia.com +# RFC4188 || K. Norseth, Ed., E. Bell, Ed. || kenyon.c.norseth@L-3com.com, elbell@ntlworld.com +# RFC4189 || K. Ono, S. Tachimoto || ono.kumiko@lab.ntt.co.jp, kumiko@cs.columbia.edu, tachimoto.shinya@lab.ntt.co.jp +# RFC4190 || K. Carlberg, I. Brown, C. Beard || k.carlberg@cs.ucl.ac.uk, I.Brown@cs.ucl.ac.uk, BeardC@umkc.edu +# RFC4191 || R. Draves, D. Thaler || richdr@microsoft.com, dthaler@microsoft.com +# RFC4192 || F. Baker, E. Lear, R. Droms || fred@cisco.com, lear@cisco.com, rdroms@cisco.com +# RFC4193 || R. Hinden, B. Haberman || bob.hinden@gmail.com, brian@innovationslab.net +# RFC4194 || J. Strombergson, L. Walleij, P. Faltstrom || Joachim.Strombergson@InformAsic.com, triad@df.lth.se, paf@cisco.com +# RFC4195 || W. Kameyama || wataru@waseda.jp +# RFC4196 || H.J. Lee, J.H. Yoon, S.L. Lee, J.I. Lee || jiinii@kisa.or.kr, jhyoon@kisa.or.kr, sllee@kisa.or.kr, jilee@kisa.or.kr +# RFC4197 || M. Riegel, Ed. || maximilian.riegel@siemens.com +# RFC4198 || D. Tessman || dtessman@zelestra.com +# RFC4201 || K. Kompella, Y. Rekhter, L. Berger || kireeti@juniper.net, yakov@juniper.net, lberger@movaz.com +# RFC4202 || K. Kompella, Ed., Y. Rekhter, Ed. || kireeti@juniper.net, yakov@juniper.net +# RFC4203 || K. Kompella, Ed., Y. Rekhter, Ed. || kireeti@juniper.net, yakov@juniper.net +# RFC4204 || J. Lang, Ed. || jplang@ieee.org +# RFC4205 || K. Kompella, Ed., Y. Rekhter, Ed. || kireeti@juniper.net, yakov@juniper.net +# RFC4206 || K. Kompella, Y. Rekhter || kireeti@juniper.net, yakov@juniper.net +# RFC4207 || J. Lang, D. Papadimitriou || jplang@ieee.org, dimitri.papadimitriou@alcatel.be +# RFC4208 || G. Swallow, J. Drake, H. Ishimatsu, Y. Rekhter || swallow@cisco.com, John.E.Drake2@boeing.com, hirokazu.ishimatsu@g1m.jp, yakov@juniper.net +# RFC4209 || A. Fredette, Ed., J. Lang, Ed. || Afredette@HatterasNetworks.com, jplang@ieee.org +# RFC4210 || C. Adams, S. Farrell, T. Kause, T. Mononen || cadams@site.uottawa.ca, stephen.farrell@cs.tcd.ie, toka@ssh.com, tmononen@safenet-inc.com +# RFC4211 || J. Schaad || jimsch@exmsft.com +# RFC4212 || M. Blinov, C. Adams || mikblinov@online.ie, cadams@site.uottawa.ca +# RFC4213 || E. Nordmark, R. Gilligan || erik.nordmark@sun.com, bob.gilligan@acm.org +# RFC4214 || F. Templin, T. Gleeson, M. Talwar, D. Thaler || fltemplin@acm.org, tgleeson@cisco.com, mohitt@microsoft.com, dthaler@microsoft.com +# RFC4215 || J. Wiljakka, Ed. || juha.wiljakka@nokia.com +# RFC4216 || R. Zhang, Ed., J.-P. Vasseur, Ed. || raymond_zhang@infonet.com, jpv@cisco.com +# RFC4217 || P. Ford-Hutchinson || rfc4217@ford-hutchinson.com +# RFC4218 || E. Nordmark, T. Li || erik.nordmark@sun.com, Tony.Li@tony.li +# RFC4219 || E. Lear || lear@cisco.com +# RFC4220 || M. Dubuc, T. Nadeau, J. Lang || mdubuc@ncf.ca, tnadeau@cisco.com, jplang@ieee.org +# RFC4221 || T. Nadeau, C. Srinivasan, A. Farrel || tnadeau@cisco.com, cheenu@bloomberg.net, adrian@olddog.co.uk +# RFC4222 || G. Choudhury, Ed. || gchoudhury@att.com +# RFC4223 || P. Savola || psavola@funet.fi +# RFC4224 || G. Pelletier, L-E. Jonsson, K. Sandlund || ghyslain.pelletier@ericsson.com, lars-erik.jonsson@ericsson.com, kristofer.sandlund@ericsson.com +# RFC4225 || P. Nikander, J. Arkko, T. Aura, G. Montenegro, E. Nordmark || pekka.nikander@nomadiclab.com, jari.arkko@ericsson.com, Tuomaura@microsoft.com, gabriel_montenegro_2000@yahoo.com, erik.nordmark@sun.com +# RFC4226 || D. M'Raihi, M. Bellare, F. Hoornaert, D. Naccache, O. Ranen || davidietf@gmail.com, mihir@cs.ucsd.edu, frh@vasco.com, david.naccache@gemplus.com, Ohad.Ranen@ealaddin.com +# RFC4227 || E. O'Tuathail, M. Rose || eamon.otuathail@clipcode.com, mrose17@gmail.com +# RFC4228 || A. Rousskov || rousskov@measurement-factory.com +# RFC4229 || M. Nottingham, J. Mogul || mnot@pobox.com, JeffMogul@acm.org +# RFC4230 || H. Tschofenig, R. Graveman || Hannes.Tschofenig@siemens.com, rfg@acm.org +# RFC4231 || M. Nystrom || magnus@rsasecurity.com +# RFC4233 || K. Morneault, S. Rengasami, M. Kalla, G. Sidebottom || kmorneau@cisco.com, mkalla@telcordia.com, selvam@trideaworks.com, greg@signatustechnologies.com +# RFC4234 || D. Crocker, Ed., P. Overell || dcrocker@bbiw.net, paul@bayleaf.org.uk +# RFC4235 || J. Rosenberg, H. Schulzrinne, R. Mahy, Ed. || jdrosen@cisco.com, schulzrinne@cs.columbia.edu, rohan@ekabal.com +# RFC4236 || A. Rousskov, M. Stecher || rousskov@measurement-factory.com, martin.stecher@webwasher.com +# RFC4237 || G. Vaudreuil || GregV@ieee.org +# RFC4238 || G. Vaudreuil || GregV@ieee.org +# RFC4239 || S. McRae, G. Parsons || stuart.mcrae@uk.ibm.com, gparsons@nortel.com +# RFC4240 || E. Burger, Ed., J. Van Dyke, A. Spitzer || eburger@brooktrout.com, jvandyke@brooktrout.com, woof@brooktrout.com +# RFC4241 || Y. Shirasaki, S. Miyakawa, T. Yamasaki, A. Takenouchi || yasuhiro@nttv6.jp, miyakawa@nttv6.jp, t.yamasaki@ntt.com, takenouchi.ayako@lab.ntt.co.jp +# RFC4242 || S. Venaas, T. Chown, B. Volz || venaas@uninett.no, tjc@ecs.soton.ac.uk, volz@cisco.com +# RFC4243 || M. Stapp, R. Johnson, T. Palaniappan || mjs@cisco.com, raj@cisco.com, athenmoz@cisco.com +# RFC4244 || M. Barnes, Ed. || mary.barnes@nortel.com +# RFC4245 || O. Levin, R. Even || oritl@microsoft.com, roni.even@polycom.co.il +# RFC4246 || M. Dolan || md.1@newtbt.com +# RFC4247 || J. Ash, B. Goode, J. Hand, R. Zhang || gash@att.com, bgoode@att.com, jameshand@att.com, raymond.zhang@bt.infonet.com +# RFC4248 || P. Hoffman || paul.hoffman@vpnc.org +# RFC4249 || B. Lilly || blilly@erols.com +# RFC4250 || S. Lehtinen, C. Lonvick, Ed. || sjl@ssh.com, clonvick@cisco.com +# RFC4251 || T. Ylonen, C. Lonvick, Ed. || ylo@ssh.com, clonvick@cisco.com +# RFC4252 || T. Ylonen, C. Lonvick, Ed. || ylo@ssh.com, clonvick@cisco.com +# RFC4253 || T. Ylonen, C. Lonvick, Ed. || ylo@ssh.com, clonvick@cisco.com +# RFC4254 || T. Ylonen, C. Lonvick, Ed. || ylo@ssh.com, clonvick@cisco.com +# RFC4255 || J. Schlyter, W. Griffin || jakob@openssh.com, wgriffin@sparta.com +# RFC4256 || F. Cusack, M. Forssen || frank@savecore.net, maf@appgate.com +# RFC4257 || G. Bernstein, E. Mannie, V. Sharma, E. Gray || gregb@grotto-networking.com, eric.mannie@perceval.net, v.sharma@ieee.org, Eric.Gray@Marconi.com +# RFC4258 || D. Brungard, Ed. || dbrungard@att.com +# RFC4259 || M.-J. Montpetit, G. Fairhurst, H. Clausen, B. Collini-Nocker, H. Linder || mmontpetit@motorola.com, gorry@erg.abdn.ac.uk, h.d.clausen@ieee.org, bnocker@cosy.sbg.ac.at, hlinder@cosy.sbg.ac.at +# RFC4260 || P. McCann || mccap@lucent.com +# RFC4261 || J. Walker, A. Kulkarni, Ed. || jesse.walker@intel.com, amol.kulkarni@intel.com +# RFC4262 || S. Santesson || stefans@microsoft.com +# RFC4263 || B. Lilly || blilly@erols.com +# RFC4264 || T. Griffin, G. Huston || Timothy.Griffin@cl.cam.ac.uk, gih@apnic.net +# RFC4265 || B. Schliesser, T. Nadeau || bensons@savvis.net, tnadeau@cisco.com +# RFC4266 || P. Hoffman || paul.hoffman@vpnc.org +# RFC4267 || M. Froumentin || mf@w3.org +# RFC4268 || S. Chisholm, D. Perkins || schishol@nortel.com, dperkins@snmpinfo.com +# RFC4269 || H.J. Lee, S.J. Lee, J.H. Yoon, D.H. Cheon, J.I. Lee || jiinii@kisa.or.kr, sjlee@kisa.or.kr, jhyoon@kisa.or.kr, dhcheon@mmaa.or.kr, jilee@kisa.or.kr +# RFC4270 || P. Hoffman, B. Schneier || paul.hoffman@vpnc.org, schneier@counterpane.com +# RFC4271 || Y. Rekhter, Ed., T. Li, Ed., S. Hares, Ed. || yakov@juniper.net, tony.li@tony.li, skh@nexthop.com +# RFC4272 || S. Murphy || Sandy@tislabs.com +# RFC4273 || J. Haas, Ed., S. Hares, Ed. || jhaas@nexthop.com, skh@nexthop.com +# RFC4274 || D. Meyer, K. Patel || dmm@1-4-5.net, keyupate@cisco.com +# RFC4275 || S. Hares, D. Hares || skh@nexthop.com, dhares@hickoryhill-consulting.com +# RFC4276 || S. Hares, A. Retana || skh@nexthop.com, aretana@cisco.com +# RFC4277 || D. McPherson, K. Patel || danny@arbor.net, keyupate@cisco.com +# RFC4278 || S. Bellovin, A. Zinin || bellovin@acm.org, zinin@psg.com +# RFC4279 || P. Eronen, Ed., H. Tschofenig, Ed. || pe@iki.fi, Hannes.Tschofenig@siemens.com +# RFC4280 || K. Chowdhury, P. Yegani, L. Madour || kchowdhury@starentnetworks.com, pyegani@cisco.com, Lila.Madour@ericsson.com +# RFC4281 || R. Gellens, D. Singer, P. Frojdh || randy@qualcomm.com, singer@apple.com, Per.Frojdh@ericsson.com +# RFC4282 || B. Aboba, M. Beadles, J. Arkko, P. Eronen || bernarda@microsoft.com, mbeadles@endforce.com, jari.arkko@ericsson.com, pe@iki.fi +# RFC4283 || A. Patel, K. Leung, M. Khalil, H. Akhtar, K. Chowdhury || alpesh@cisco.com, kleung@cisco.com, mkhalil@nortel.com, haseebak@nortel.com, kchowdhury@starentnetworks.com +# RFC4284 || F. Adrangi, V. Lortz, F. Bari, P. Eronen || farid.adrangi@intel.com, victor.lortz@intel.com, farooq.bari@cingular.com, pe@iki.fi +# RFC4285 || A. Patel, K. Leung, M. Khalil, H. Akhtar, K. Chowdhury || alpesh@cisco.com, kleung@cisco.com, mkhalil@nortel.com, haseebak@nortel.com, kchowdhury@starentnetworks.com +# RFC4286 || B. Haberman, J. Martin || brian@innovationslab.net, jim@netzwert.ag +# RFC4287 || M. Nottingham, Ed., R. Sayre, Ed. || mnot@pobox.com, rfsayre@boswijck.com +# RFC4288 || N. Freed, J. Klensin || ned.freed@mrochek.com, klensin+ietf@jck.com +# RFC4289 || N. Freed, J. Klensin || ned.freed@mrochek.com, klensin+ietf@jck.com +# RFC4290 || J. Klensin || john-ietf@jck.com +# RFC4291 || R. Hinden, S. Deering || bob.hinden@gmail.com +# RFC4292 || B. Haberman || brian@innovationslab.net +# RFC4293 || S. Routhier, Ed. || sar@iwl.com +# RFC4294 || J. Loughney, Ed. || john.loughney@nokia.com +# RFC4295 || G. Keeni, K. Koide, K. Nagami, S. Gundavelli || glenn@cysols.com, koide@shiratori.riec.tohoku.ac.jp, nagami@inetcore.com, sgundave@cisco.com +# RFC4296 || S. Bailey, T. Talpey || steph@sandburst.com, thomas.talpey@netapp.com +# RFC4297 || A. Romanow, J. Mogul, T. Talpey, S. Bailey || allyn@cisco.com, JeffMogul@acm.org, thomas.talpey@netapp.com, steph@sandburst.com +# RFC4298 || J.-H. Chen, W. Lee, J. Thyssen || rchen@broadcom.com, wlee@broadcom.com, jthyssen@broadcom.com +# RFC4301 || S. Kent, K. Seo || kent@bbn.com, kseo@bbn.com +# RFC4302 || S. Kent || kent@bbn.com +# RFC4303 || S. Kent || kent@bbn.com +# RFC4304 || S. Kent || kent@bbn.com +# RFC4305 || D. Eastlake 3rd || Donald.Eastlake@Motorola.com +# RFC4306 || C. Kaufman, Ed. || charliek@microsoft.com +# RFC4307 || J. Schiller || jis@mit.edu +# RFC4308 || P. Hoffman || paul.hoffman@vpnc.org +# RFC4309 || R. Housley || housley@vigilsec.com +# RFC4310 || S. Hollenbeck || shollenbeck@verisign.com +# RFC4311 || R. Hinden, D. Thaler || bob.hinden@gmail.com, dthaler@microsoft.com +# RFC4312 || A. Kato, S. Moriai, M. Kanda || akato@po.ntts.co.jp, shiho@rd.scei.sony.co.jp, kanda@isl.ntt.co.jp +# RFC4313 || D. Oran || oran@cisco.com +# RFC4314 || A. Melnikov || alexey.melnikov@isode.com +# RFC4315 || M. Crispin || MRC@CAC.Washington.EDU +# RFC4316 || J. Reschke || julian.reschke@greenbytes.de +# RFC4317 || A. Johnston, R. Sparks || ajohnston@tello.com, rjsparks@estacado.net +# RFC4318 || D. Levi, D. Harrington || dlevi@nortel.com, ietfdbh@comcast.net +# RFC4319 || C. Sikes, B. Ray, R. Abbi || csikes@zhone.com, rray@pesa.com, Rajesh.Abbi@alcatel.com +# RFC4320 || R. Sparks || rjsparks@estacado.net +# RFC4321 || R. Sparks || rjsparks@estacado.net +# RFC4322 || M. Richardson, D.H. Redelmeier || mcr@sandelman.ottawa.on.ca, hugh@mimosa.com +# RFC4323 || M. Patrick, W. Murwin || michael.patrick@motorola.com, w.murwin@motorola.com +# RFC4324 || D. Royer, G. Babics, S. Mansour || Doug@IntelliCal.com, george.babics@oracle.com, smansour@ebay.com +# RFC4325 || S. Santesson, R. Housley || stefans@microsoft.com, housley@vigilsec.com +# RFC4326 || G. Fairhurst, B. Collini-Nocker || gorry@erg.abdn.ac.uk, bnocker@cosy.sbg.ac.at +# RFC4327 || M. Dubuc, T. Nadeau, J. Lang, E. McGinnis || dubuc.consulting@sympatico.ca, tnadeau@cisco.com, jplang@ieee.org, emcginnis@hammerheadsystems.com +# RFC4328 || D. Papadimitriou, Ed. || dimitri.papadimitriou@alcatel.be +# RFC4329 || B. Hoehrmann || bjoern@hoehrmann.de +# RFC4330 || D. Mills || mills@udel.edu +# RFC4331 || B. Korver, L. Dusseault || briank@networkresonance.com, lisa.dusseault@gmail.com +# RFC4332 || K. Leung, A. Patel, G. Tsirtsis, E. Klovning || kleung@cisco.com, alpesh@cisco.com, g.tsirtsis@flarion.com, espen@birdstep.com +# RFC4333 || G. Huston, Ed., B. Wijnen, Ed. || gih@apnic.net, bwijnen@lucent.com +# RFC4334 || R. Housley, T. Moore || housley@vigilsec.com, timmoore@microsoft.com +# RFC4335 || J. Galbraith, P. Remaker || galb-list@vandyke.com, remaker@cisco.com +# RFC4336 || S. Floyd, M. Handley, E. Kohler || floyd@icir.org, M.Handley@cs.ucl.ac.uk, kohler@cs.ucla.edu +# RFC4337 || Y Lim, D. Singer || young@netntv.co.kr, singer@apple.com +# RFC4338 || C. DeSanti, C. Carlson, R. Nixon || cds@cisco.com, craig.carlson@qlogic.com, bob.nixon@emulex.com +# RFC4339 || J. Jeong, Ed. || jjeong@cs.umn.edu +# RFC4340 || E. Kohler, M. Handley, S. Floyd || kohler@cs.ucla.edu, M.Handley@cs.ucl.ac.uk, floyd@icir.org +# RFC4341 || S. Floyd, E. Kohler || floyd@icir.org, kohler@cs.ucla.edu +# RFC4342 || S. Floyd, E. Kohler, J. Padhye || floyd@icir.org, kohler@cs.ucla.edu, padhye@microsoft.com +# RFC4343 || D. Eastlake 3rd || Donald.Eastlake@motorola.com +# RFC4344 || M. Bellare, T. Kohno, C. Namprempre || mihir@cs.ucsd.edu, tkohno@cs.ucsd.edu, meaw@alum.mit.edu +# RFC4345 || B. Harris || bjh21@bjh21.me.uk +# RFC4346 || T. Dierks, E. Rescorla || tim@dierks.org, ekr@rtfm.com +# RFC4347 || E. Rescorla, N. Modadugu || ekr@rtfm.com, nagendra@cs.stanford.edu +# RFC4348 || S. Ahmadi || sassan.ahmadi@ieee.org +# RFC4349 || C. Pignataro, M. Townsley || cpignata@cisco.com, mark@townsley.net +# RFC4350 || F. Hendrikx, C. Wallis || ferry.hendrikx@ssc.govt.nz, colin.wallis@ssc.govt.nz +# RFC4351 || G. Hellstrom, P. Jones || gunnar.hellstrom@omnitor.se, paulej@packetizer.com +# RFC4352 || J. Sjoberg, M. Westerlund, A. Lakaniemi, S. Wenger || Johan.Sjoberg@ericsson.com, Magnus.Westerlund@ericsson.com, ari.lakaniemi@nokia.com, Stephan.Wenger@nokia.com +# RFC4353 || J. Rosenberg || jdrosen@cisco.com +# RFC4354 || M. Garcia-Martin || miguel.an.garcia@nokia.com +# RFC4355 || R. Brandner, L. Conroy, R. Stastny || rudolf.brandner@siemens.com, lwc@roke.co.uk, Richard.stastny@oefeg.at +# RFC4356 || R. Gellens || randy@qualcomm.com +# RFC4357 || V. Popov, I. Kurepkin, S. Leontiev || vpopov@cryptopro.ru, kure@cryptopro.ru, lse@cryptopro.ru +# RFC4358 || D. Smith || dwight.smith@motorola.com +# RFC4359 || B. Weis || bew@cisco.com +# RFC4360 || S. Sangli, D. Tappan, Y. Rekhter || rsrihari@cisco.com, tappan@cisco.com, yakov@juniper.net +# RFC4361 || T. Lemon, B. Sommerfeld || mellon@nominum.com, sommerfeld@sun.com +# RFC4362 || L-E. Jonsson, G. Pelletier, K. Sandlund || lars-erik.jonsson@ericsson.com, ghyslain.pelletier@ericsson.com, kristofer.sandlund@ericsson.com +# RFC4363 || D. Levi, D. Harrington || dlevi@nortel.com, ietfdbh@comcast.net +# RFC4364 || E. Rosen, Y. Rekhter || erosen@cisco.com, yakov@juniper.net +# RFC4365 || E. Rosen || erosen@cisco.com +# RFC4366 || S. Blake-Wilson, M. Nystrom, D. Hopwood, J. Mikkelsen, T. Wright || sblakewilson@bcisse.com, magnus@rsasecurity.com, david.hopwood@blueyonder.co.uk, janm@transactionware.com, timothy.wright@vodafone.com +# RFC4367 || J. Rosenberg, Ed., IAB || jdrosen@cisco.com +# RFC4368 || T. Nadeau, S. Hegde || tnadeau@cisco.com, subrah@cisco.com +# RFC4369 || K. Gibbons, C. Monia, J. Tseng, F. Travostino || kevin.gibbons@mcdata.com, charles_monia@yahoo.com, joshtseng@yahoo.com, travos@nortel.com +# RFC4370 || R. Weltman || robw@worldspot.com +# RFC4371 || B. Carpenter, Ed., L. Lynch, Ed. || brc@zurich.ibm.com, llynch@darkwing.uoregon.edu +# RFC4372 || F. Adrangi, A. Lior, J. Korhonen, J. Loughney || farid.adrangi@intel.com, avi@bridgewatersystems.com, jouni.korhonen@teliasonera.com, john.loughney@nokia.com +# RFC4373 || R. Harrison, J. Sermersheim, Y. Dong || rharrison@novell.com, jimse@novell.com, yulindong@gmail.com +# RFC4374 || G. McCobb || mccobb@us.ibm.com +# RFC4375 || K. Carlberg || carlberg@g11.org.uk +# RFC4376 || P. Koskelainen, J. Ott, H. Schulzrinne, X. Wu || petri.koskelainen@nokia.com, jo@netlab.hut.fi, hgs@cs.columbia.edu, xiaotaow@cs.columbia.edu +# RFC4377 || T. Nadeau, M. Morrow, G. Swallow, D. Allan, S. Matsushima || tnadeau@cisco.com, mmorrow@cisco.com, swallow@cisco.com, dallan@nortel.com, satoru@ft.solteria.net +# RFC4378 || D. Allan, Ed., T. Nadeau, Ed. || dallan@nortel.com, tnadeau@cisco.com +# RFC4379 || K. Kompella, G. Swallow || kireeti@juniper.net, swallow@cisco.com +# RFC4380 || C. Huitema || huitema@microsoft.com +# RFC4381 || M. Behringer || mbehring@cisco.com +# RFC4382 || T. Nadeau, Ed., H. van der Linde, Ed. || tnadeau@cisco.com, havander@cisco.com +# RFC4383 || M. Baugher, E. Carrara || mbaugher@cisco.com, carrara@kth.se +# RFC4384 || D. Meyer || dmm@1-4-5.net +# RFC4385 || S. Bryant, G. Swallow, L. Martini, D. McPherson || stbryant@cisco.com, swallow@cisco.com, lmartini@cisco.com, danny@arbor.net +# RFC4386 || S. Boeyen, P. Hallam-Baker || sharon.boeyen@entrust.com, pbaker@VeriSign.com +# RFC4387 || P. Gutmann, Ed. || pgut001@cs.auckland.ac.nz +# RFC4388 || R. Woundy, K. Kinnear || richard_woundy@cable.comcast.com, kkinnear@cisco.com +# RFC4389 || D. Thaler, M. Talwar, C. Patel || dthaler@microsoft.com, mohitt@microsoft.com, chirayu@chirayu.org +# RFC4390 || V. Kashyap || vivk@us.ibm.com +# RFC4391 || J. Chu, V. Kashyap || jerry.chu@sun.com, vivk@us.ibm.com +# RFC4392 || V. Kashyap || vivk@us.ibm.com +# RFC4393 || H. Garudadri || hgarudadri@qualcomm.com +# RFC4394 || D. Fedyk, O. Aboul-Magd, D. Brungard, J. Lang, D. Papadimitriou || dwfedyk@nortel.com, osama@nortel.com, dbrungard@att.com, jplang@ieee.org, dimitri.papadimitriou@alcatel.be +# RFC4395 || T. Hansen, T. Hardie, L. Masinter || tony+urireg@maillennium.att.com, hardie@qualcomm.com, LMM@acm.org +# RFC4396 || J. Rey, Y. Matsui || jose.rey@eu.panasonic.com, matsui.yoshinori@jp.panasonic.com +# RFC4397 || I. Bryskin, A. Farrel || i_bryskin@yahoo.com, adrian@olddog.co.uk +# RFC4398 || S. Josefsson || simon@josefsson.org +# RFC4401 || N. Williams || Nicolas.Williams@sun.com +# RFC4402 || N. Williams || Nicolas.Williams@sun.com +# RFC4403 || B. Bergeson, K. Boogert, V. Nanjundaswamy || bruce.bergeson@novell.com, kent.boogert@novell.com, vijay.nanjundaswamy@oracle.com +# RFC4404 || R. Natarajan, A. Rijhsinghani || anil@charter.net, r.natarajan@f5.com +# RFC4405 || E. Allman, H. Katz || eric@sendmail.com, hkatz@microsoft.com +# RFC4406 || J. Lyon, M. Wong || jimlyon@microsoft.com, mengwong@dumbo.pobox.com +# RFC4407 || J. Lyon || jimlyon@microsoft.com +# RFC4408 || M. Wong, W. Schlitt || mengwong+spf@pobox.com, wayne@schlitt.net +# RFC4409 || R. Gellens, J. Klensin || g+ietf@qualcomm.com, john+ietf@jck.com +# RFC4410 || M. Pullen, F. Zhao, D. Cohen || mpullen@gmu.edu, fzhao@netlab.gmu.edu, danny.cohen@sun.com +# RFC4411 || J. Polk || jmpolk@cisco.com +# RFC4412 || H. Schulzrinne, J. Polk || hgs@cs.columbia.edu, jmpolk@cisco.com +# RFC4413 || M. West, S. McCann || mark.a.west@roke.co.uk, stephen.mccann@roke.co.uk +# RFC4414 || A. Newton || andy@hxr.us +# RFC4415 || R. Brandner, L. Conroy, R. Stastny || rudolf.brandner@siemens.com, lwc@roke.co.uk, Richard.stastny@oefeg.at +# RFC4416 || J. Wong, Ed. || j.k.wong@sympatico.ca +# RFC4417 || P. Resnick, Ed., P. Saint-Andre, Ed. || presnick@qti.qualcomm.com, ietf@stpeter.im +# RFC4418 || T. Krovetz, Ed. || tdk@acm.org +# RFC4419 || M. Friedl, N. Provos, W. Simpson || markus@openbsd.org, provos@citi.umich.edu, wsimpson@greendragon.com +# RFC4420 || A. Farrel, Ed., D. Papadimitriou, J.-P. Vasseur, A. Ayyangar || adrian@olddog.co.uk, dimitri.papadimitriou@alcatel.be, jpv@cisco.com, arthi@juniper.net +# RFC4421 || C. Perkins || csp@csperkins.org +# RFC4422 || A. Melnikov, Ed., K. Zeilenga, Ed. || Alexey.Melnikov@isode.com, Kurt@OpenLDAP.org +# RFC4423 || R. Moskowitz, P. Nikander || rgm@icsalabs.com, pekka.nikander@nomadiclab.com +# RFC4424 || S. Ahmadi || sassan.ahmadi@ieee.org +# RFC4425 || A. Klemets || Anders.Klemets@microsoft.com +# RFC4426 || J. Lang, Ed., B. Rajagopalan, Ed., D. Papadimitriou, Ed. || jplang@ieee.org, balar@microsoft.com, dimitri.papadimitriou@alcatel.be +# RFC4427 || E. Mannie, Ed., D. Papadimitriou, Ed. || eric.mannie@perceval.net, dimitri.papadimitriou@alcatel.be +# RFC4428 || D. Papadimitriou, Ed., E. Mannie, Ed. || dimitri.papadimitriou@alcatel.be, eric.mannie@perceval.net +# RFC4429 || N. Moore || sharkey@zoic.org +# RFC4430 || S. Sakane, K. Kamada, M. Thomas, J. Vilhuber || Shouichi.Sakane@jp.yokogawa.com, Ken-ichi.Kamada@jp.yokogawa.com, mat@cisco.com, vilhuber@cisco.com +# RFC4431 || M. Andrews, S. Weiler || Mark_Andrews@isc.org, weiler@tislabs.com +# RFC4432 || B. Harris || bjh21@bjh21.me.uk +# RFC4433 || M. Kulkarni, A. Patel, K. Leung || mkulkarn@cisco.com, alpesh@cisco.com, kleung@cisco.com +# RFC4434 || P. Hoffman || paul.hoffman@vpnc.org +# RFC4435 || Y. Nomura, R. Walsh, J-P. Luoma, H. Asaeda, H. Schulzrinne || nom@flab.fujitsu.co.jp, rod.walsh@nokia.com, juha-pekka.luoma@nokia.com, asaeda@wide.ad.jp, schulzrinne@cs.columbia.edu +# RFC4436 || B. Aboba, J. Carlson, S. Cheshire || bernarda@microsoft.com, james.d.carlson@sun.com, rfc@stuartcheshire.org +# RFC4437 || J. Whitehead, G. Clemm, J. Reschke, Ed. || ejw@cse.ucsc.edu, julian.reschke@greenbytes.degeoffrey.clemm@us.ibm.com, +# RFC4438 || C. DeSanti, V. Gaonkar, H.K. Vivek, K. McCloghrie, S. Gai || cds@cisco.com, vgaonkar@cisco.com, hvivek@cisco.com, kzm@cisco.com, none +# RFC4439 || C. DeSanti, V. Gaonkar, K. McCloghrie, S. Gai || cds@cisco.com, vgaonkar@cisco.com, kzm@cisco.com, none +# RFC4440 || S. Floyd, Ed., V. Paxson, Ed., A. Falk, Ed., IAB || floyd@acm.org, vern@icir.org, falk@isi.edu +# RFC4441 || B. Aboba, Ed. || bernarda@microsoft.com +# RFC4442 || S. Fries, H. Tschofenig || steffen.fries@siemens.com, Hannes.Tschofenig@siemens.com +# RFC4443 || A. Conta, S. Deering, M. Gupta, Ed. || aconta@txc.com, none, mukesh.gupta@tropos.com +# RFC4444 || J. Parker, Ed. || jeffp@middlebury.edu +# RFC4445 || J. Welch, J. Clark || Jim.Welch@ineoquest.com, jiclark@cisco.com +# RFC4446 || L. Martini || lmartini@cisco.com +# RFC4447 || L. Martini, Ed., E. Rosen, N. El-Aawar, T. Smith, G. Heron || lmartini@cisco.com, nna@level3.net, giles.heron@tellabs.com, erosen@cisco.com, tob@netapp.com +# RFC4448 || L. Martini, Ed., E. Rosen, N. El-Aawar, G. Heron || lmartini@cisco.com, nna@level3.net, giles.heron@tellabs.com, erosen@cisco.com +# RFC4449 || C. Perkins || charles.perkins@nokia.com +# RFC4450 || E. Lear, H. Alvestrand || lear@cisco.com, harald@alvestrand.no +# RFC4451 || D. McPherson, V. Gill || danny@arbor.net, VijayGill9@aol.com +# RFC4452 || H. Van de Sompel, T. Hammond, E. Neylon, S. Weibel || herbertv@lanl.gov, t.hammond@nature.com, eneylon@manifestsolutions.com, weibel@oclc.org +# RFC4453 || J. Rosenberg, G. Camarillo, Ed., D. Willis || jdrosen@cisco.com, Gonzalo.Camarillo@ericsson.com, dean.willis@softarmor.com +# RFC4454 || S. Singh, M. Townsley, C. Pignataro || sanjeevs@cisco.com, mark@townsley.net, cpignata@cisco.com +# RFC4455 || M. Hallak-Stamler, M. Bakke, Y. Lederman, M. Krueger, K. McCloghrie || michele@sanrad.com, mbakke@cisco.com, yaronled@bezeqint.net, marjorie_krueger@hp.com, kzm@cisco.com +# RFC4456 || T. Bates, E. Chen, R. Chandra || tbates@cisco.com, enkechen@cisco.com, rchandra@sonoasystems.com +# RFC4457 || G. Camarillo, G. Blanco || Gonzalo.Camarillo@ericsson.com, german.blanco@ericsson.com +# RFC4458 || C. Jennings, F. Audet, J. Elwell || fluffy@cisco.com, audet@nortel.com, john.elwell@siemens.com +# RFC4459 || P. Savola || psavola@funet.fi +# RFC4460 || R. Stewart, I. Arias-Rodriguez, K. Poon, A. Caro, M. Tuexen || randall@lakerest.net, ivan.arias-rodriguez@nokia.com, kacheong.poon@sun.com, acaro@bbn.com, tuexen@fh-muenster.de +# RFC4461 || S. Yasukawa, Ed. || yasukawa.seisho@lab.ntt.co.jp +# RFC4462 || J. Hutzelman, J. Salowey, J. Galbraith, V. Welch || jhutz+@cmu.edu, jsalowey@cisco.com, galb@vandyke.com, welch@mcs.anl.gov +# RFC4463 || S. Shanmugham, P. Monaco, B. Eberman || sarvi@cisco.com, peter.monaco@nuasis.com, brian.eberman@speechworks.com +# RFC4464 || A. Surtees, M. West || abigail.surtees@roke.co.uk, mark.a.west@roke.co.uk +# RFC4465 || A. Surtees, M. West || abigail.surtees@roke.co.uk, mark.a.west@roke.co.uk +# RFC4466 || A. Melnikov, C. Daboo || Alexey.Melnikov@isode.com, cyrus@daboo.name +# RFC4467 || M. Crispin || MRC@CAC.Washington.EDU +# RFC4468 || C. Newman || chris.newman@sun.com +# RFC4469 || P. Resnick || presnick@qti.qualcomm.com +# RFC4470 || S. Weiler, J. Ihren || weiler@tislabs.com, johani@autonomica.se +# RFC4471 || G. Sisson, B. Laurie || geoff@nominet.org.uk, ben@algroup.co.uk +# RFC4472 || A. Durand, J. Ihren, P. Savola || Alain_Durand@cable.comcast.com, johani@autonomica.se, psavola@funet.fi +# RFC4473 || Y. Nomura, R. Walsh, J-P. Luoma, J. Ott, H. Schulzrinne || nom@flab.fujitsu.co.jp, rod.walsh@nokia.com, juha-pekka.luoma@nokia.com, jo@netlab.tkk.fi, schulzrinne@cs.columbia.edu +# RFC4474 || J. Peterson, C. Jennings || jon.peterson@neustar.biz, fluffy@cisco.com +# RFC4475 || R. Sparks, Ed., A. Hawrylyshen, A. Johnston, J. Rosenberg, H. Schulzrinne || RjS@estacado.net, ahawrylyshen@ditechnetworks.com, alan@sipstation.com, jdrosen@cisco.com, hgs@cs.columbia.edu +# RFC4476 || C. Francis, D. Pinkas || Chris_S_Francis@Raytheon.com, Denis.Pinkas@bull.net +# RFC4477 || T. Chown, S. Venaas, C. Strauf || tjc@ecs.soton.ac.uk, venaas@uninett.no, strauf@rz.tu-clausthal.de +# RFC4478 || Y. Nir || ynir@checkpoint.com +# RFC4479 || J. Rosenberg || jdrosen@cisco.com +# RFC4480 || H. Schulzrinne, V. Gurbani, P. Kyzivat, J. Rosenberg || hgs+simple@cs.columbia.edu, vkg@lucent.com, pkyzivat@cisco.com, jdrosen@cisco.com +# RFC4481 || H. Schulzrinne || hgs+simple@cs.columbia.edu +# RFC4482 || H. Schulzrinne || hgs+simple@cs.columbia.edu +# RFC4483 || E. Burger, Ed. || eburger@cantata.com +# RFC4484 || J. Peterson, J. Polk, D. Sicker, H. Tschofenig || jon.peterson@neustar.biz, jmpolk@cisco.com, douglas.sicker@colorado.edu, Hannes.Tschofenig@siemens.com +# RFC4485 || J. Rosenberg, H. Schulzrinne || jdrosen@cisco.com, schulzrinne@cs.columbia.edu +# RFC4486 || E. Chen, V. Gillet || enkechen@cisco.com, vgi@opentransit.net +# RFC4487 || F. Le, S. Faccin, B. Patil, H. Tschofenig || franckle@cmu.edu, sfaccinstd@gmail.com, Basavaraj.Patil@nokia.com, Hannes.Tschofenig@siemens.com +# RFC4488 || O. Levin || oritl@microsoft.com +# RFC4489 || J-S. Park, M-K. Shin, H-J. Kim || pjs@etri.re.kr, myungki.shin@gmail.com, khj@etri.re.kr +# RFC4490 || S. Leontiev, Ed., G. Chudov, Ed. || lse@cryptopro.ru, chudov@cryptopro.ru +# RFC4491 || S. Leontiev, Ed., D. Shefanovski, Ed. || lse@cryptopro.ru, dbs@mts.ru +# RFC4492 || S. Blake-Wilson, N. Bolyard, V. Gupta, C. Hawk, B. Moeller || sblakewilson@safenet-inc.com, nelson@bolyard.com, vipul.gupta@sun.com, chris@corriente.net, bodo@openssl.org +# RFC4493 || JH. Song, R. Poovendran, J. Lee, T. Iwata || songlee@ee.washington.edu, radha@ee.washington.edu, icheol.lee@samsung.com, iwata@cse.nagoya-u.ac.jp +# RFC4494 || JH. Song, R. Poovendran, J. Lee || songlee@ee.washington.edu, radha@ee.washington.edu, jicheol.lee@samsung.com +# RFC4495 || J. Polk, S. Dhesikan || jmpolk@cisco.com, sdhesika@cisco.com +# RFC4496 || M. Stecher, A. Barbir || martin.stecher@webwasher.com, abbieb@nortel.com +# RFC4497 || J. Elwell, F. Derks, P. Mourot, O. Rousseau || john.elwell@siemens.com, frank.derks@nec-philips.com, Patrick.Mourot@alcatel.fr, Olivier.Rousseau@alcatel.fr +# RFC4498 || G. Keeni || glenn@cysols.com +# RFC4501 || S. Josefsson || simon@josefsson.org +# RFC4502 || S. Waldbusser || waldbusser@nextbeacon.com +# RFC4503 || M. Boesgaard, M. Vesterager, E. Zenner || mab@cryptico.com, mvp@cryptico.com, ez@cryptico.com +# RFC4504 || H. Sinnreich, Ed., S. Lass, C. Stredicke || henry@pulver.com, steven.lass@verizonbusiness.com, cs@snom.de +# RFC4505 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC4506 || M. Eisler, Ed. || email2mre-rfc4506@yahoo.com +# RFC4507 || J. Salowey, H. Zhou, P. Eronen, H. Tschofenig || jsalowey@cisco.com, hzhou@cisco.com, pe@iki.fi, Hannes.Tschofenig@siemens.com +# RFC4508 || O. Levin, A. Johnston || oritl@microsoft.com, ajohnston@ipstation.com +# RFC4509 || W. Hardaker || hardaker@tislabs.com +# RFC4510 || K. Zeilenga, Ed. || Kurt@OpenLDAP.org +# RFC4511 || J. Sermersheim, Ed. || jimse@novell.com +# RFC4512 || K. Zeilenga, Ed. || Kurt@OpenLDAP.org +# RFC4513 || R. Harrison, Ed. || roger_harrison@novell.com +# RFC4514 || K. Zeilenga, Ed. || Kurt@OpenLDAP.org +# RFC4515 || M. Smith, Ed., T. Howes || mcs@pearlcrescent.com, howes@opsware.com +# RFC4516 || M. Smith, Ed., T. Howes || mcs@pearlcrescent.com, howes@opsware.com +# RFC4517 || S. Legg, Ed. || steven.legg@eb2bcom.com +# RFC4518 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC4519 || A. Sciberras, Ed. || andrew.sciberras@eb2bcom.com +# RFC4520 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC4521 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC4522 || S. Legg || steven.legg@eb2bcom.com +# RFC4523 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC4524 || K. Zeilenga, Ed. || Kurt@OpenLDAP.org +# RFC4525 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC4526 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC4527 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC4528 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC4529 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC4530 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC4531 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC4532 || K. Zeilenga || Kurt@OpenLDAP.org +# RFC4533 || K. Zeilenga, J.H. Choi || Kurt@OpenLDAP.org, jongchoi@us.ibm.com +# RFC4534 || A Colegrove, H Harney || acc@sparta.com, hh@sparta.com +# RFC4535 || H. Harney, U. Meth, A. Colegrove, G. Gross || hh@sparta.com, umeth@sparta.com, acc@sparta.com, gmgross@identaware.com +# RFC4536 || P. Hoschka || ph@w3.org +# RFC4537 || L. Zhu, P. Leach, K. Jaganathan || lzhu@microsoft.com, paulle@microsoft.com, karthikj@microsoft.com +# RFC4538 || J. Rosenberg || jdrosen@cisco.com +# RFC4539 || T. Edwards || tedwards@pbs.org +# RFC4540 || M. Stiemerling, J. Quittek, C. Cadar || stiemerling@netlab.nec.de, quittek@netlab.nec.de, ccadar2@yahoo.com +# RFC4541 || M. Christensen, K. Kimball, F. Solensky || mjc@tt.dk, karen.kimball@hp.com, frank.solensky@calix.com +# RFC4542 || F. Baker, J. Polk || fred@cisco.com, jmpolk@cisco.com +# RFC4543 || D. McGrew, J. Viega || mcgrew@cisco.com, viega@list.org +# RFC4544 || M. Bakke, M. Krueger, T. McSweeney, J. Muchow || mbakke@cisco.com, marjorie_krueger@hp.com, tommcs@us.ibm.com, james.muchow@qlogic.com +# RFC4545 || M. Bakke, J. Muchow || mbakke@cisco.com, james.muchow@qlogic.com +# RFC4546 || D. Raftus, E. Cardona || david.raftus@ati.com, e.cardona@cablelabs.com +# RFC4547 || A. Ahmad, G. Nakanishi || azlina@cisco.com, gnakanishi@motorola.com +# RFC4548 || E. Gray, J. Rutemiller, G. Swallow || Eric.Gray@Marconi.com, John.Rutemiller@Marconi.com, swallow@cisco.com +# RFC4549 || A. Melnikov, Ed. || alexey.melnikov@isode.com +# RFC4550 || S. Maes, A. Melnikov || stephane.maes@oracle.com, Alexey.melnikov@isode.com +# RFC4551 || A. Melnikov, S. Hole || Alexey.Melnikov@isode.com, Steve.Hole@messagingdirect.com +# RFC4552 || M. Gupta, N. Melam || mukesh.gupta@tropos.com, nmelam@juniper.net +# RFC4553 || A. Vainshtein, Ed., YJ. Stein, Ed. || sasha@axerra.com, yaakov_s@rad.com +# RFC4554 || T. Chown || tjc@ecs.soton.ac.uk +# RFC4555 || P. Eronen || pe@iki.fi +# RFC4556 || L. Zhu, B. Tung || lzhu@microsoft.com, brian@aero.org +# RFC4557 || L. Zhu, K. Jaganathan, N. Williams || lzhu@microsoft.com, karthikj@microsoft.com, Nicolas.Williams@sun.com +# RFC4558 || Z. Ali, R. Rahman, D. Prairie, D. Papadimitriou || zali@cisco.com, rrahman@cisco.com, dprairie@cisco.com, dimitri.papadimitriou@alcatel.be +# RFC4559 || K. Jaganathan, L. Zhu, J. Brezak || karthikj@microsoft.com, lzhu@microsoft.com, jbrezak@microsoft.com +# RFC4560 || J. Quittek, Ed., K. White, Ed. || quittek@netlab.nec.de, wkenneth@us.ibm.com +# RFC4561 || J.-P. Vasseur, Ed., Z. Ali, S. Sivabalan || jpv@cisco.com, zali@cisco.com, msiva@cisco.com +# RFC4562 || T. Melsen, S. Blake || Torben.Melsen@ericsson.com, steven.blake@ericsson.com +# RFC4563 || E. Carrara, V. Lehtovirta, K. Norrman || carrara@kth.se, vesa.lehtovirta@ericsson.com, karl.norrman@ericsson.com +# RFC4564 || S. Govindan, Ed., H. Cheng, ZH. Yao, WH. Zhou, L. Yang || saravanan.govindan@sg.panasonic.com, hong.cheng@sg.panasonic.com, yaoth@huawei.com, zhouwenhui@chinamobile.com, lily.l.yang@intel.com +# RFC4565 || D. Loher, D. Nelson, O. Volinsky, B. Sarikaya || dplore@gmail.com, dnelson@enterasys.com, ovolinsky@colubris.com, sarikaya@ieee.org +# RFC4566 || M. Handley, V. Jacobson, C. Perkins || M.Handley@cs.ucl.ac.uk, van@packetdesign.com, csp@csperkins.org +# RFC4567 || J. Arkko, F. Lindholm, M. Naslund, K. Norrman, E. Carrara || jari.arkko@ericsson.com, fredrik.lindholm@ericsson.com, mats.naslund@ericsson.com, karl.norrman@ericsson.com, carrara@kth.se +# RFC4568 || F. Andreasen, M. Baugher, D. Wing || fandreas@cisco.com, mbaugher@cisco.com, dwing-ietf@fuggles.com +# RFC4569 || G. Camarillo || Gonzalo.Camarillo@ericsson.com +# RFC4570 || B. Quinn, R. Finlayson || rcq@boxnarrow.com, finlayson@live555.com +# RFC4571 || J. Lazzaro || lazzaro@cs.berkeley.edu +# RFC4572 || J. Lennox || lennox@cs.columbia.edu +# RFC4573 || R. Even, A. Lochbaum || roni.even@polycom.co.il, alochbaum@polycom.com +# RFC4574 || O. Levin, G. Camarillo || oritl@microsoft.com, Gonzalo.Camarillo@ericsson.com +# RFC4575 || J. Rosenberg, H. Schulzrinne, O. Levin, Ed. || jdrosen@cisco.com, schulzrinne@cs.columbia.edu, oritl@microsoft.com +# RFC4576 || E. Rosen, P. Psenak, P. Pillay-Esnault || erosen@cisco.com, ppsenak@cisco.com, ppe@cisco.com +# RFC4577 || E. Rosen, P. Psenak, P. Pillay-Esnault || erosen@cisco.com, ppsenak@cisco.com, ppe@cisco.com +# RFC4578 || M. Johnston, S. Venaas, Ed. || michael.johnston@intel.com, venaas@uninett.no +# RFC4579 || A. Johnston, O. Levin || alan@sipstation.com, oritl@microsoft.com +# RFC4580 || B. Volz || volz@cisco.com +# RFC4581 || M. Bagnulo, J. Arkko || marcelo@it.uc3m.es, jari.arkko@ericsson.com +# RFC4582 || G. Camarillo, J. Ott, K. Drage || Gonzalo.Camarillo@ericsson.com, jo@netlab.hut.fi, drage@lucent.com +# RFC4583 || G. Camarillo || Gonzalo.Camarillo@ericsson.com +# RFC4584 || S. Chakrabarti, E. Nordmark || samitac2@gmail.com, erik.nordmark@sun.com +# RFC4585 || J. Ott, S. Wenger, N. Sato, C. Burmeister, J. Rey || jo@acm.org, stewe@stewe.org, sato652@oki.com, carsten.burmeister@eu.panasonic.com, jose.rey@eu.panasonic.com +# RFC4586 || C. Burmeister, R. Hakenberg, A. Miyazaki, J. Ott, N. Sato, S. Fukunaga || carsten.burmeister@eu.panasonic.com, rolf.hakenberg@eu.panasonic.com, miyazaki.akihiro@jp.panasonic.com, jo@acm.org, sato652@oki.com, fukunaga444@oki.com +# RFC4587 || R. Even || roni.even@polycom.co.il +# RFC4588 || J. Rey, D. Leon, A. Miyazaki, V. Varsa, R. Hakenberg || jose.rey@eu.panasonic.com, davidleon123@yahoo.com, miyazaki.akihiro@jp.panasonic.com, viktor.varsa@nokia.com, rolf.hakenberg@eu.panasonic.com +# RFC4589 || H. Schulzrinne, H. Tschofenig || schulzrinne@cs.columbia.edu, Hannes.Tschofenig@siemens.com +# RFC4590 || B. Sterman, D. Sadolevsky, D. Schwartz, D. Williams, W. Beck || baruch@kayote.com, dscreat@dscreat.com, david@kayote.com, dwilli@cisco.com, beckw@t-systems.com +# RFC4591 || M. Townsley, G. Wilkie, S. Booth, S. Bryant, J. Lau || mark@townsley.net, gwilkie@cisco.com, jebooth@cisco.com, stbryant@cisco.com, jedlau@gmail.com +# RFC4592 || E. Lewis || ed.lewis@neustar.biz +# RFC4593 || A. Barbir, S. Murphy, Y. Yang || abbieb@nortel.com, sandy@sparta.com, yiya@cisco.com +# RFC4594 || J. Babiarz, K. Chan, F. Baker || babiarz@nortel.com, khchan@nortel.com, fred@cisco.com +# RFC4595 || F. Maino, D. Black || fmaino@cisco.com, black_david@emc.com +# RFC4596 || J. Rosenberg, P. Kyzivat || jdrosen@cisco.com, pkyzivat@cisco.com +# RFC4597 || R. Even, N. Ismail || roni.even@polycom.co.il, nismail@cisco.com +# RFC4598 || B. Link || bdl@dolby.com +# RFC4601 || B. Fenner, M. Handley, H. Holbrook, I. Kouvelas || fenner@research.att.com, M.Handley@cs.ucl.ac.uk, holbrook@arastra.com, kouvelas@cisco.com +# RFC4602 || T. Pusateri || pusateri@juniper.net +# RFC4603 || G. Zorn, G. Weber, R. Foltak || gwz@cisco.com, gdweber@cisco.com, rfoltak@cisco.com +# RFC4604 || H. Holbrook, B. Cain, B. Haberman || holbrook@cisco.com, bcain99@gmail.com, brian@innovationslab.net +# RFC4605 || B. Fenner, H. He, B. Haberman, H. Sandick || fenner@research.att.com, haixiang@nortelnetworks.com, brian@innovationslab.net, sandick@nc.rr.com +# RFC4606 || E. Mannie, D. Papadimitriou || eric.mannie@perceval.net, dimitri.papadimitriou@alcatel.be +# RFC4607 || H. Holbrook, B. Cain || holbrook@arastra.com, bcain99@gmail.com +# RFC4608 || D. Meyer, R. Rockell, G. Shepherd || dmm@1-4-5.net, rrockell@sprint.net, gjshep@gmail.com +# RFC4609 || P. Savola, R. Lehtonen, D. Meyer || psavola@funet.fi, rami.lehtonen@teliasonera.com, dmm@1-4-5.net +# RFC4610 || D. Farinacci, Y. Cai || dino@cisco.com, ycai@cisco.com +# RFC4611 || M. McBride, J. Meylor, D. Meyer || mcbride@cisco.com, jmeylor@cisco.com, dmm@1-4-5.net +# RFC4612 || P. Jones, H. Tamura || paulej@packetizer.com, tamura@cs.ricoh.co.jp +# RFC4613 || P. Frojdh, U. Lindgren, M. Westerlund || per.frojdh@ericsson.com, ulf.a.lindgren@ericsson.com, magnus.westerlund@ericsson.com +# RFC4614 || M. Duke, R. Braden, W. Eddy, E. Blanton || martin.duke@boeing.com, braden@isi.edu, weddy@grc.nasa.gov, eblanton@cs.purdue.edu +# RFC4615 || J. Song, R. Poovendran, J. Lee, T. Iwata || junhyuk.song@gmail.com, radha@ee.washington.edu, jicheol.lee@samsung.com, iwata@cse.nagoya-u.ac.jp +# RFC4616 || K. Zeilenga, Ed. || Kurt@OpenLDAP.org +# RFC4617 || J. Kornijenko || j.kornienko@abcsoftware.lv +# RFC4618 || L. Martini, E. Rosen, G. Heron, A. Malis || lmartini@cisco.com, erosen@cisco.com, giles.heron@tellabs.com, Andy.Malis@tellabs.com +# RFC4619 || L. Martini, Ed., C. Kawa, Ed., A. Malis, Ed. || lmartini@cisco.com, claude.kawa@oz.com, Andy.Malis@tellabs.com +# RFC4620 || M. Crawford, B. Haberman, Ed. || crawdad@fnal.gov, brian@innovationslab.net +# RFC4621 || T. Kivinen, H. Tschofenig || kivinen@safenet-inc.com, Hannes.Tschofenig@siemens.com +# RFC4622 || P. Saint-Andre || ietf@stpeter.im +# RFC4623 || A. Malis, M. Townsley || Andy.Malis@tellabs.com, mark@townsley.net +# RFC4624 || B. Fenner, D. Thaler || fenner@research.att.com, dthaler@microsoft.com +# RFC4625 || C. DeSanti, K. McCloghrie, S. Kode, S. Gai || cds@cisco.com, srinikode@yahoo.com, kzm@cisco.com, none +# RFC4626 || C. DeSanti, V. Gaonkar, K. McCloghrie, S. Gai || cds@cisco.com, vgaonkar@cisco.com, kzm@cisco.com, none +# RFC4627 || D. Crockford || douglas@crockford.com +# RFC4628 || R. Even || roni.even@polycom.co.il +# RFC4629 || J. Ott, C. Bormann, G. Sullivan, S. Wenger, R. Even, Ed. || jo@netlab.tkk.fi, cabo@tzi.org, garysull@microsoft.com, stewe@stewe.org, roni.even@polycom.co.il +# RFC4630 || R. Housley, S. Santesson || housley@vigilsec.com, stefans@microsoft.com +# RFC4631 || M. Dubuc, T. Nadeau, J. Lang, E. McGinnis, A. Farrel || dubuc.consulting@sympatico.ca, tnadeau@cisco.com, jplang@ieee.org, emcginnis@hammerheadsystems.com, adrian@olddog.co.uk +# RFC4632 || V. Fuller, T. Li || vaf@cisco.com, tli@tropos.com +# RFC4633 || S. Hartman || hartmans-ietf@mit.edu +# RFC4634 || D. Eastlake 3rd, T. Hansen || donald.eastlake@motorola.com, tony+shs@maillennium.att.com +# RFC4635 || D. Eastlake 3rd || Donald.Eastlake@motorola.com +# RFC4636 || C. Perkins || charles.perkins@nokia.com +# RFC4637 || || +# RFC4638 || P. Arberg, D. Kourkouzelis, M. Duckett, T. Anschutz, J. Moisand || parberg@redback.com, diamondk@redback.com, mike.duckett@bellsouth.com, tom.anschutz@bellsouth.com, jmoisand@juniper.net +# RFC4639 || R. Woundy, K. Marez || richard_woundy@cable.comcast.com, kevin.marez@motorola.com +# RFC4640 || A. Patel, Ed., G. Giaretta, Ed. || alpesh@cisco.com, gerardo.giaretta@telecomitalia.it +# RFC4641 || O. Kolkman, R. Gieben || olaf@nlnetlabs.nl, miek@miek.nl +# RFC4642 || K. Murchison, J. Vinocur, C. Newman || murch@andrew.cmu.edu, vinocur@cs.cornell.edu, Chris.Newman@sun.com +# RFC4643 || J. Vinocur, K. Murchison || vinocur@cs.cornell.edu, murch@andrew.cmu.edu +# RFC4644 || J. Vinocur, K. Murchison || vinocur@cs.cornell.edu, murch@andrew.cmu.edu +# RFC4645 || D. Ewell || dewell@adelphia.net +# RFC4646 || A. Phillips, M. Davis || addison@inter-locale.com, mark.davis@macchiato.com +# RFC4647 || A. Phillips, M. Davis || addison@inter-locale.com, mark.davis@macchiato.com +# RFC4648 || S. Josefsson || simon@josefsson.org +# RFC4649 || B. Volz || volz@cisco.com +# RFC4650 || M. Euchner || martin_euchner@hotmail.com +# RFC4651 || C. Vogt, J. Arkko || chvogt@tm.uka.de, jari.arkko@ericsson.com +# RFC4652 || D. Papadimitriou, Ed., L.Ong, J. Sadler, S. Shew, D. Ward || dimitri.papadimitriou@alcatel.be, lyong@ciena.com, jonathan.sadler@tellabs.com, sdshew@nortel.com, dward@cisco.com +# RFC4653 || S. Bhandarkar, A. L. N. Reddy, M. Allman, E. Blanton || sumitha@tamu.edu, reddy@ee.tamu.edu, mallman@icir.org, eblanton@cs.purdue.edu +# RFC4654 || J. Widmer, M. Handley || widmer@acm.org, m.handley@cs.ucl.ac.uk +# RFC4655 || A. Farrel, J.-P. Vasseur, J. Ash || adrian@olddog.co.uk, jpv@cisco.com, gash@att.com +# RFC4656 || S. Shalunov, B. Teitelbaum, A. Karp, J. Boote, M. Zekauskas || shalunov@internet2.edu, ben@internet2.edu, akarp@cs.wisc.edu, boote@internet2.edu, matt@internet2.edu +# RFC4657 || J. Ash, Ed., J.L. Le Roux, Ed. || gash@att.com, jeanlouis.leroux@orange-ft.com +# RFC4659 || J. De Clercq, D. Ooms, M. Carugi, F. Le Faucheur || jeremy.de_clercq@alcatel.be, dirk@onesparrow.com, marco.carugi@nortel.com, flefauch@cisco.com +# RFC4660 || H. Khartabil, E. Leppanen, M. Lonnfors, J. Costa-Requena || hisham.khartabil@telio.no, eva-maria.leppanen@nokia.com, mikko.lonnfors@nokia.com, jose.costa-requena@nokia.com +# RFC4661 || H. Khartabil, E. Leppanen, M. Lonnfors, J. Costa-Requena || hisham.khartabil@telio.no, eva-maria.leppanen@nokia.com, mikko.lonnfors@nokia.com, jose.costa-requena@nokia.com +# RFC4662 || A. B. Roach, B. Campbell, J. Rosenberg || adam@estacado.net, ben@estacado.net, jdrosen@cisco.com +# RFC4663 || D. Harrington || dbharrington@comcast.net +# RFC4664 || L. Andersson, Ed., E. Rosen, Ed. || loa@pi.se, erosen@cisco.com +# RFC4665 || W. Augustyn, Ed., Y. Serbest, Ed. || waldemar@wdmsys.com, yetik_serbest@labs.att.com +# RFC4666 || K. Morneault, Ed., J. Pastor-Balbas, Ed. || kmorneau@cisco.com, j.javier.pastor@ericsson.com +# RFC4667 || W. Luo || luo@cisco.com +# RFC4668 || D. Nelson || dnelson@enterasys.com +# RFC4669 || D. Nelson || dnelson@enterasys.com +# RFC4670 || D. Nelson || dnelson@enterasys.com +# RFC4671 || D. Nelson || dnelson@enterasys.com +# RFC4672 || S. De Cnodder, N. Jonnala, M. Chiba || stefaan.de_cnodder@alcatel.be, njonnala@cisco.com, mchiba@cisco.com +# RFC4673 || S. De Cnodder, N. Jonnala, M. Chiba || stefaan.de_cnodder@alcatel.be, njonnala@cisco.com, mchiba@cisco.com +# RFC4674 || J.L. Le Roux, Ed. || jeanlouis.leroux@francetelecom.com +# RFC4675 || P. Congdon, M. Sanchez, B. Aboba || paul.congdon@hp.com, mauricio.sanchez@hp.com, bernarda@microsoft.com +# RFC4676 || H. Schulzrinne || hgs+geopriv@cs.columbia.edu +# RFC4677 || P. Hoffman, S. Harris || paul.hoffman@vpnc.org, srh@umich.edu +# RFC4678 || A. Bivens || jbivens@us.ibm.com +# RFC4679 || V. Mammoliti, G. Zorn, P. Arberg, R. Rennison || vince@cisco.com, gwz@cisco.com, parberg@redback.com, robert.rennison@ecitele.com +# RFC4680 || S. Santesson || stefans@microsoft.com +# RFC4681 || S. Santesson, A. Medvinsky, J. Ball || stefans@microsoft.com, arimed@microsoft.com, joshball@microsoft.com +# RFC4682 || E. Nechamkin, J-F. Mule || enechamkin@broadcom.com, jf.mule@cablelabs.com +# RFC4683 || J. Park, J. Lee, H.. Lee, S. Park, T. Polk || khopri@kisa.or.kr, jilee@kisa.or.kr, hslee@kisa.or.kr, sjpark@bcqre.com, tim.polk@nist.gov +# RFC4684 || P. Marques, R. Bonica, L. Fang, L. Martini, R. Raszuk, K. Patel, J. Guichard || roque@juniper.net, rbonica@juniper.net, luyuanfang@att.com, lmartini@cisco.com, rraszuk@cisco.com, keyupate@cisco.com, jguichar@cisco.com +# RFC4685 || J. Snell || jasnell@gmail.com +# RFC4686 || J. Fenton || fenton@bluepopcorn.net +# RFC4687 || S. Yasukawa, A. Farrel, D. King, T. Nadeau || s.yasukawa@hco.ntt.co.jp, adrian@olddog.co.uk, daniel.king@aria-networks.com, tnadeau@cisco.com +# RFC4688 || S. Rushing || srushing@inmedius.com +# RFC4689 || S. Poretsky, J. Perser, S. Erramilli, S. Khurana || sporetsky@reefpoint.com, jerry@perser.org, shobha@research.telcordia.com, skhurana@motorola.com +# RFC4690 || J. Klensin, P. Faltstrom, C. Karp, IAB || john-ietf@jck.com, paf@cisco.com, ck@nic.museum, iab@iab.org +# RFC4691 || L. Andersson, Ed. || loa@pi.se +# RFC4692 || G. Huston || gih@apnic.net +# RFC4693 || H. Alvestrand || harald@alvestrand.no +# RFC4694 || J. Yu || james.yu@neustar.biz +# RFC4695 || J. Lazzaro, J. Wawrzynek || lazzaro@cs.berkeley.edu, johnw@cs.berkeley.edu +# RFC4696 || J. Lazzaro, J. Wawrzynek || lazzaro@cs.berkeley.edu, johnw@cs.berkeley.edu +# RFC4697 || M. Larson, P. Barber || mlarson@verisign.com, pbarber@verisign.com +# RFC4698 || E. Gunduz, A. Newton, S. Kerr || e.gunduz@computer.org, andy@hxr.us, shane@time-travellers.org +# RFC4701 || M. Stapp, T. Lemon, A. Gustafsson || mjs@cisco.com, mellon@nominum.com, gson@araneus.fi +# RFC4702 || M. Stapp, B. Volz, Y. Rekhter || mjs@cisco.com, volz@cisco.com, yakov@juniper.net +# RFC4703 || M. Stapp, B. Volz || mjs@cisco.com, volz@cisco.com +# RFC4704 || B. Volz || volz@cisco.com +# RFC4705 || R. Housley, A. Corry || housley@vigilsec.com, publications@gigabeam.com +# RFC4706 || M. Morgenstern, M. Dodge, S. Baillie, U. Bonollo || moti.Morgenstern@ecitele.com, mbdodge@ieee.org, scott.baillie@nec.com.au, umberto.bonollo@nec.com.au +# RFC4707 || P. Grau, V. Heinau, H. Schlichting, R. Schuettler || nas@fu-berlin.de, nas@fu-berlin.de, nas@fu-berlin.de, nas@fu-berlin.de +# RFC4708 || A. Miller || ak.miller@auckland.ac.nz +# RFC4709 || J. Reschke || julian.reschke@greenbytes.de +# RFC4710 || A. Siddiqui, D. Romascanu, E. Golovinsky || anwars@avaya.com, dromasca@gmail.com , gene@alertlogic.net +# RFC4711 || A. Siddiqui, D. Romascanu, E. Golovinsky || anwars@avaya.com, dromasca@gmail.com , gene@alertlogic.net +# RFC4712 || A. Siddiqui, D. Romascanu, E. Golovinsky, M. Rahman,Y. Kim || anwars@avaya.com, dromasca@gmail.com , gene@alertlogic.net, none, ybkim@broadcom.com +# RFC4713 || X. Lee, W. Mao, E. Chen, N. Hsu, J. Klensin || lee@cnnic.cn, mao@cnnic.cn, erin@twnic.net.tw, snw@twnic.net.tw, john+ietf@jck.com +# RFC4714 || A. Mankin, S. Hayes || mankin@psg.com, stephen.hayes@ericsson.com +# RFC4715 || M. Munakata, S. Schubert, T. Ohba || munakata.mayumi@lab.ntt.co.jp, shida@ntt-at.com, ohba.takumi@lab.ntt.co.jp +# RFC4716 || J. Galbraith, R. Thayer || galb@vandyke.com, rodney@canola-jones.com +# RFC4717 || L. Martini, J. Jayakumar, M. Bocci, N. El-Aawar, J. Brayley, G. Koleyni || lmartini@cisco.com, jjayakum@cisco.com, matthew.bocci@alcatel.co.uk, nna@level3.net, jeremy.brayley@ecitele.com, ghassem@nortelnetworks.com +# RFC4718 || P. Eronen, P. Hoffman || pe@iki.fi, paul.hoffman@vpnc.org +# RFC4719 || R. Aggarwal, Ed., M. Townsley, Ed., M. Dos Santos, Ed. || rahul@juniper.net, mark@townsley.net, mariados@cisco.com +# RFC4720 || A. Malis, D. Allan, N. Del Regno || Andy.Malis@tellabs.com, dallan@nortelnetworks.com, nick.delregno@mci.com +# RFC4721 || C. Perkins, P. Calhoun, J. Bharatia || charles.perkins@nokia.com, pcalhoun@cisco.com, jayshree@nortel.com +# RFC4722 || J. Van Dyke, E. Burger, Ed., A. Spitzer || jvandyke@cantata.com, eburger@cantata.com, woof@pingtel.com +# RFC4723 || T. Kosonen, T. White || timo.kosonen@nokia.com, twhite@midi.org +# RFC4724 || S. Sangli, E. Chen, R. Fernando, J. Scudder, Y. Rekhter || rsrihari@cisco.com, enkechen@cisco.com, rex@juniper.net, jgs@juniper.net, yakov@juniper.net +# RFC4725 || A. Mayrhofer, B. Hoeneisen || alexander.mayrhofer@enum.at, b.hoeneisen@ieee.org +# RFC4726 || A. Farrel, J.-P. Vasseur, A. Ayyangar || adrian@olddog.co.uk, jpv@cisco.com, arthi@nuovasystems.com +# RFC4727 || B. Fenner || fenner@research.att.com +# RFC4728 || D. Johnson, Y. Hu, D. Maltz || dbj@cs.rice.edu, yihchun@uiuc.edu, dmaltz@cs.cmu.edu +# RFC4729 || M. Abel || TPM@nfc-forum.org +# RFC4730 || E. Burger, M. Dolly || eburger@cantata.com, mdolly@att.com +# RFC4731 || A. Melnikov, D. Cridland || Alexey.Melnikov@isode.com, dave.cridland@inventuresystems.co.uk +# RFC4732 || M. Handley, Ed., E. Rescorla, Ed., IAB || M.Handley@cs.ucl.ac.uk, ekr@networkresonance.com, iab@ietf.org +# RFC4733 || H. Schulzrinne, T. Taylor || schulzrinne@cs.columbia.edu, tom.taylor.stds@gmail.com +# RFC4734 || H. Schulzrinne, T. Taylor || schulzrinne@cs.columbia.edu, tom.taylor.stds@gmail.com +# RFC4735 || T. Taylor || tom.taylor.stds@gmail.com +# RFC4736 || JP. Vasseur, Ed., Y. Ikejiri, R. Zhang || jpv@cisco.com, y.ikejiri@ntt.com, raymond_zhang@bt.infonet.com +# RFC4737 || A. Morton, L. Ciavattone, G. Ramachandran, S. Shalunov, J. Perser || acmorton@att.com, lencia@att.com, gomathi@att.com, shalunov@internet2.edu, jperser@veriwave.com +# RFC4738 || D. Ignjatic, L. Dondeti, F. Audet, P. Lin || dignjatic@polycom.com, dondeti@qualcomm.com, audet@nortel.com, linping@nortel.com +# RFC4739 || P. Eronen, J. Korhonen || pe@iki.fi, jouni.korhonen@teliasonera.com +# RFC4740 || M. Garcia-Martin, Ed., M. Belinchon, M. Pallares-Lopez, C. Canales-Valenzuela, K. Tammi || miguel.an.garcia@nokia.com, maria.carmen.belinchon@ericsson.com, miguel-angel.pallares@ericsson.com, carolina.canales@ericsson.com, kalle.tammi@nokia.com +# RFC4741 || R. Enns, Ed. || rpe@juniper.net +# RFC4742 || M. Wasserman, T. Goddard || margaret@thingmagic.com, ted.goddard@icesoft.com +# RFC4743 || T. Goddard || ted.goddard@icesoft.com +# RFC4744 || E. Lear, K. Crozier || lear@cisco.com, ken.crozier@gmail.com +# RFC4745 || H. Schulzrinne, H. Tschofenig, J. Morris, J. Cuellar, J. Polk, J. Rosenberg || schulzrinne@cs.columbia.edu, Hannes.Tschofenig@siemens.com, jmorris@cdt.org, Jorge.Cuellar@siemens.com, jmpolk@cisco.com, jdrosen@cisco.com +# RFC4746 || T. Clancy, W. Arbaugh || clancy@ltsnet.net, waa@cs.umd.edu +# RFC4747 || S. Kipp, G. Ramkumar, K. McCloghrie || scott.kipp@mcdata.com, gramkumar@stanfordalumni.org, kzm@cisco.com +# RFC4748 || S. Bradner, Ed. || sob@harvard.edu +# RFC4749 || A. Sollaud || aurelien.sollaud@orange-ft.com +# RFC4750 || D. Joyal, Ed., P. Galecki, Ed., S. Giacalone, Ed., R. Coltun, F. Baker || djoyal@nortel.com, pgalecki@airvana.com, spencer.giacalone@gmail.com, fred@cisco.com +# RFC4752 || A. Melnikov, Ed. || Alexey.Melnikov@isode.com +# RFC4753 || D. Fu, J. Solinas || defu@orion.ncsc.mil, jasolin@orion.ncsc.mil +# RFC4754 || D. Fu, J. Solinas || defu@orion.ncsc.mil, jasolin@orion.ncsc.mil +# RFC4755 || V. Kashyap || vivk@us.ibm.com +# RFC4756 || A. Li || adamli@hyervision.com +# RFC4757 || K. Jaganathan, L. Zhu, J. Brezak || karthikj@microsoft.com, lzhu@microsoft.com, jbrezak@microsoft.com +# RFC4758 || M. Nystroem || magnus@rsasecurity.com +# RFC4759 || R. Stastny, R. Shockey, L. Conroy || Richard.stastny@oefeg.at, richard.shockey@neustar.biz, lconroy@insensate.co.uk +# RFC4760 || T. Bates, R. Chandra, D. Katz, Y. Rekhter || tbates@cisco.com, rchandra@sonoasystems.com, dkatz@juniper.com, yakov@juniper.com +# RFC4761 || K. Kompella, Ed., Y. Rekhter, Ed. || kireeti@juniper.net, yakov@juniper.net +# RFC4762 || M. Lasserre, Ed., V. Kompella, Ed. || mlasserre@alcatel-lucent.com, vach.kompella@alcatel-lucent.com +# RFC4763 || M. Vanderveen, H. Soliman || mvandervn@yahoo.com, solimanhs@gmail.com +# RFC4764 || F. Bersani, H. Tschofenig || bersani_florent@yahoo.fr, Hannes.Tschofenig@siemens.com +# RFC4765 || H. Debar, D. Curry, B. Feinstein || herve.debar@orange-ftgroup.com, david_a_curry@glic.com, feinstein@acm.org +# RFC4766 || M. Wood, M. Erlinger || mark1@iss.net, mike@cs.hmc.edu +# RFC4767 || B. Feinstein, G. Matthews || bfeinstein@acm.org, gmatthew@nas.nasa.gov +# RFC4768 || S. Hartman || hartmans-ietf@mit.edu +# RFC4769 || J. Livingood, R. Shockey || jason_livingood@cable.comcast.com, richard.shockey@neustar.biz +# RFC4770 || C. Jennings, J. Reschke, Ed. || fluffy@cisco.com, julian.reschke@greenbytes.de +# RFC4771 || V. Lehtovirta, M. Naslund, K. Norrman || vesa.lehtovirta@ericsson.com, mats.naslund@ericsson.com, karl.norrman@ericsson.com +# RFC4772 || S. Kelly || scott@hyperthought.com +# RFC4773 || G. Huston || gih@apnic.net +# RFC4774 || S. Floyd || floyd@icir.org +# RFC4775 || S. Bradner, B. Carpenter, Ed., T. Narten || sob@harvard.edu, brc@zurich.ibm.com, narten@us.ibm.com +# RFC4776 || H. Schulzrinne || hgs+geopriv@cs.columbia.edu +# RFC4777 || T. Murphy Jr., P. Rieth, J. Stevens || murphyte@us.ibm.com, rieth@us.ibm.com, jssteven@us.ibm.com +# RFC4778 || M. Kaeo || merike@doubleshotsecurity.com +# RFC4779 || S. Asadullah, A. Ahmed, C. Popoviciu, P. Savola, J. Palet || sasad@cisco.com, adahmed@cisco.com, cpopovic@cisco.com, psavola@funet.fi, jordi.palet@consulintel.es +# RFC4780 || K. Lingle, J-F. Mule, J. Maeng, D. Walker || klingle@cisco.com, jf.mule@cablelabs.com, jmaeng@austin.rr.com, drwalker@rogers.com +# RFC4781 || Y. Rekhter, R. Aggarwal || yakov@juniper.net, rahul@juniper.net +# RFC4782 || S. Floyd, M. Allman, A. Jain, P. Sarolahti || floyd@icir.org, mallman@icir.org, a.jain@f5.com, pasi.sarolahti@iki.fi +# RFC4783 || L. Berger, Ed. || lberger@labn.net +# RFC4784 || C. Carroll, F. Quick || Christopher.Carroll@ropesgray.com, fquick@qualcomm.com +# RFC4785 || U. Blumenthal, P. Goel || urimobile@optonline.net, Purushottam.Goel@intel.com +# RFC4786 || J. Abley, K. Lindqvist || jabley@ca.afilias.info, kurtis@kurtis.pp.se +# RFC4787 || F. Audet, Ed., C. Jennings || audet@nortel.com, fluffy@cisco.com +# RFC4788 || Q. Xie, R. Kapoor || Qiaobing.Xie@Motorola.com, rkapoor@qualcomm.com +# RFC4789 || J. Schoenwaelder, T. Jeffree || j.schoenwaelder@iu-bremen.de, tony@jeffree.co.uk +# RFC4790 || C. Newman, M. Duerst, A. Gulbrandsen || chris.newman@sun.com, duerst@it.aoyama.ac.jp, arnt@oryx.com +# RFC4791 || C. Daboo, B. Desruisseaux, L. Dusseault || cyrus@daboo.name, bernard.desruisseaux@oracle.com, lisa.dusseault@gmail.com +# RFC4792 || S. Legg || steven.legg@eb2bcom.com +# RFC4793 || M. Nystroem || magnus@rsasecurity.com +# RFC4794 || B. Fenner || fenner@research.att.com +# RFC4795 || B. Aboba, D. Thaler, L. Esibov || bernarda@microsoft.com, dthaler@microsoft.com, levone@microsoft.com +# RFC4796 || J. Hautakorpi, G. Camarillo || Jani.Hautakorpi@ericsson.com, Gonzalo.Camarillo@ericsson.com +# RFC4797 || Y. Rekhter, R. Bonica, E. Rosen || yakov@juniper.net, rbonica@juniper.net, erosen@cisco.com +# RFC4798 || J. De Clercq, D. Ooms, S. Prevost, F. Le Faucheur || jeremy.de_clercq@alcatel-lucent.be, dirk@onesparrow.com, stuart.prevost@bt.com, flefauch@cisco.com +# RFC4801 || T. Nadeau, Ed., A. Farrel, Ed. || tnadeau@cisco.com, adrian@olddog.co.uk +# RFC4802 || T. Nadeau, Ed., A. Farrel, Ed. || tnadeau@cisco.com, adrian@olddog.co.uk +# RFC4803 || T. Nadeau, Ed., A. Farrel, Ed. || tnadeau@cisco.com, adrian@olddog.co.uk +# RFC4804 || F. Le Faucheur, Ed. || flefauch@cisco.com +# RFC4805 || O. Nicklass, Ed. || orly_n@rad.com +# RFC4806 || M. Myers, H. Tschofenig || mmyers@fastq.com, Hannes.Tschofenig@siemens.com +# RFC4807 || M. Baer, R. Charlet, W. Hardaker, R. Story, C. Wang || baerm@tislabs.com, rcharlet@alumni.calpoly.edu, hardaker@tislabs.com, rstory@ipsp.revelstone.com, cliffwangmail@yahoo.com +# RFC4808 || S. Bellovin || bellovin@acm.org +# RFC4809 || C. Bonatti, Ed., S. Turner, Ed., G. Lebovitz, Ed. || Bonattic@ieca.com, Turners@ieca.com, gregory.ietf@gmail.com +# RFC4810 || C. Wallace, U. Pordesch, R. Brandner || cwallace@cygnacom.com, ulrich.pordesch@zv.fraunhofer.de, ralf.brandner@intercomponentware.com +# RFC4811 || L. Nguyen, A. Roy, A. Zinin || lhnguyen@cisco.com, akr@cisco.com, alex.zinin@alcatel-lucent.com +# RFC4812 || L. Nguyen, A. Roy, A. Zinin || lhnguyen@cisco.com, akr@cisco.com, alex.zinin@alcatel-lucent.com +# RFC4813 || B. Friedman, L. Nguyen, A. Roy, D. Yeung, A. Zinin || friedman@cisco.com, lhnguyen@cisco.com, akr@cisco.com, myeung@cisco.com, alex.zinin@alcatel-lucent.com +# RFC4814 || D. Newman, T. Player || dnewman@networktest.com, timmons.player@spirent.com +# RFC4815 || L-E. Jonsson, K. Sandlund, G. Pelletier, P. Kremer || lars-erik.jonsson@ericsson.com, kristofer.sandlund@ericsson.com, ghyslain.pelletier@ericsson.com, peter.kremer@ericsson.com +# RFC4816 || A. Malis, L. Martini, J. Brayley, T. Walsh || andrew.g.malis@verizon.com, lmartini@cisco.com, jeremy.brayley@ecitele.com, twalsh@juniper.net +# RFC4817 || M. Townsley, C. Pignataro, S. Wainner, T. Seely, J. Young || mark@townsley.net, cpignata@cisco.com, swainner@cisco.com, tseely@sprint.net, young@jsyoung.net +# RFC4818 || J. Salowey, R. Droms || jsalowey@cisco.com, rdroms@cisco.com +# RFC4819 || J. Galbraith, J. Van Dyke, J. Bright || galb@vandyke.com, jpv@vandyke.com, jon@siliconcircus.com +# RFC4820 || M. Tuexen, R. Stewart, P. Lei || tuexen@fh-muenster.de, randall@lakerest.net, peterlei@cisco.com +# RFC4821 || M. Mathis, J. Heffner || mathis@psc.edu, jheffner@psc.edu +# RFC4822 || R. Atkinson, M. Fanto || rja@extremenetworks.com, mattjf@umd.edu +# RFC4823 || T. Harding, R. Scott || tharding@us.axway.com, rscott@us.axway.com +# RFC4824 || J. Hofmueller, Ed., A. Bachmann, Ed., IO. zmoelnig, Ed. || ip-sfs@mur.at, ip-sfs@mur.at, ip-sfs@mur.at +# RFC4825 || J. Rosenberg || jdrosen@cisco.com +# RFC4826 || J. Rosenberg || jdrosen@cisco.com +# RFC4827 || M. Isomaki, E. Leppanen || markus.isomaki@nokia.com, eva-maria.leppanen@nokia.com +# RFC4828 || S. Floyd, E. Kohler || floyd@icir.org, kohler@cs.ucla.edu +# RFC4829 || J. de Oliveira, Ed., JP. Vasseur, Ed., L. Chen, C. Scoglio || jau@ece.drexel.edu, jpv@cisco.com, leonardo.c.chen@verizon.com, caterina@eece.ksu.edu +# RFC4830 || J. Kempf, Ed. || kempf@docomolabs-usa.com +# RFC4831 || J. Kempf, Ed. || kempf@docomolabs-usa.com +# RFC4832 || C. Vogt, J. Kempf || chvogt@tm.uka.de, kempf@docomolabs-usa.com +# RFC4833 || E. Lear, P. Eggert || lear@cisco.com, eggert@cs.ucla.edu +# RFC4834 || T. Morin, Ed. || thomas.morin@orange-ftgroup.com +# RFC4835 || V. Manral || vishwas@ipinfusion.com +# RFC4836 || E. Beili || edward.beili@actelis.com +# RFC4837 || L. Khermosh || lior_khermosh@pmc-sierra.com +# RFC4838 || V. Cerf, S. Burleigh, A. Hooke, L. Torgerson, R. Durst, K. Scott, K. Fall, H. Weiss || vint@google.com, Scott.Burleigh@jpl.nasa.gov, Adrian.Hooke@jpl.nasa.gov, ltorgerson@jpl.nasa.gov, durst@mitre.org, kscott@mitre.org, kfall@intel.com, howard.weiss@sparta.com +# RFC4839 || G. Conboy, J. Rivlin, J. Ferraiolo || gc@ebooktechnologies.com, john@ebooktechnologies.com, jferrai@us.ibm.com +# RFC4840 || B. Aboba, Ed., E. Davies, D. Thaler || bernarda@microsoft.com, elwynd@dial.pipex.com, dthaler@microsoft.com +# RFC4841 || C. Heard, Ed. || heard@pobox.com +# RFC4842 || A. Malis, P. Pate, R. Cohen, Ed., D. Zelig || andrew.g.malis@verizon.com, prayson.pate@overturenetworks.com, ronc@resolutenetworks.com, davidz@corrigent.com +# RFC4843 || P. Nikander, J. Laganier, F. Dupont || pekka.nikander@nomadiclab.com, julien.ietf@laposte.net, Francis.Dupont@fdupont.fr +# RFC4844 || L. Daigle, Ed., Internet Architecture Board || leslie@thinkingcat.com, iab@iab.org +# RFC4845 || L. Daigle, Ed., Internet Architecture Board || leslie@thinkingcat.com, iab@iab.org +# RFC4846 || J. Klensin, Ed., D. Thaler, Ed. || john-ietf@jck.com, dthaler@microsoft.com +# RFC4847 || T. Takeda, Ed. || takeda.tomonori@lab.ntt.co.jp +# RFC4848 || L. Daigle || leslie@thinkingcat.com +# RFC4849 || P. Congdon, M. Sanchez, B. Aboba || paul.congdon@hp.com, mauricio.sanchez@hp.com, bernarda@microsoft.com +# RFC4850 || D. Wysochanski || wysochanski@pobox.com +# RFC4851 || N. Cam-Winget, D. McGrew, J. Salowey, H. Zhou || ncamwing@cisco.com, mcgrew@cisco.com, jsalowey@cisco.com, hzhou@cisco.com +# RFC4852 || J. Bound, Y. Pouffary, S. Klynsma, T. Chown, D. Green || jim.bound@hp.com, Yanick.pouffary@hp.com, tjc@ecs.soton.ac.uk, green@commandinformation.com, sklynsma@mitre.org +# RFC4853 || R. Housley || housley@vigilsec.com +# RFC4854 || P. Saint-Andre || ietf@stpeter.im +# RFC4855 || S. Casner || casner@acm.org +# RFC4856 || S. Casner || casner@acm.org +# RFC4857 || E. Fogelstroem, A. Jonsson, C. Perkins || eva.fogelstrom@ericsson.com, annika.jonsson@ericsson.com, charles.perkins@nsn.com +# RFC4858 || H. Levkowetz, D. Meyer, L. Eggert, A. Mankin || henrik@levkowetz.com, dmm@1-4-5.net, lars.eggert@nokia.com, mankin@psg.com +# RFC4859 || A. Farrel || adrian@olddog.co.uk +# RFC4860 || F. Le Faucheur, B. Davie, P. Bose, C. Christou, M. Davenport || flefauch@cisco.com, bds@cisco.com, pratik.bose@lmco.com, christou_chris@bah.com, davenport_michael@bah.com +# RFC4861 || T. Narten, E. Nordmark, W. Simpson, H. Soliman || narten@us.ibm.com, erik.nordmark@sun.com, william.allen.simpson@gmail.com, hesham@elevatemobile.com +# RFC4862 || S. Thomson, T. Narten, T. Jinmei || sethomso@cisco.com, narten@us.ibm.com, jinmei@isl.rdc.toshiba.co.jp +# RFC4863 || L. Martini, G. Swallow || lmartini@cisco.com, swallow@cisco.com +# RFC4864 || G. Van de Velde, T. Hain, R. Droms, B. Carpenter, E. Klein || gunter@cisco.com, alh-ietf@tndh.net, rdroms@cisco.com, brc@zurich.ibm.com, ericlklein.ipv6@gmail.com +# RFC4865 || G. White, G. Vaudreuil || g.a.white@comcast.net, GregV@ieee.org +# RFC4866 || J. Arkko, C. Vogt, W. Haddad || jari.arkko@ericsson.com, chvogt@tm.uka.de, wassim.haddad@ericsson.com +# RFC4867 || J. Sjoberg, M. Westerlund, A. Lakaniemi, Q. Xie || Johan.Sjoberg@ericsson.com, Magnus.Westerlund@ericsson.com, ari.lakaniemi@nokia.com, Qiaobing.Xie@motorola.com +# RFC4868 || S. Kelly, S. Frankel || scott@hyperthought.com, sheila.frankel@nist.gov +# RFC4869 || L. Law, J. Solinas || lelaw@orion.ncsc.mil, jasolin@orion.ncsc.mil +# RFC4870 || M. Delany || markd+domainkeys@yahoo-inc.com +# RFC4871 || E. Allman, J. Callas, M. Delany, M. Libbey, J. Fenton, M. Thomas || eric+dkim@sendmail.org, jon@pgp.com, markd+dkim@yahoo-inc.com, mlibbeymail-mailsig@yahoo.com, fenton@bluepopcorn.net, mat@cisco.com +# RFC4872 || J.P. Lang, Ed., Y. Rekhter, Ed., D. Papadimitriou, Ed. || jplang@ieee.org, yakov@juniper.net, dimitri.papadimitriou@alcatel-lucent.be +# RFC4873 || L. Berger, I. Bryskin, D. Papadimitriou, A. Farrel || lberger@labn.net, IBryskin@advaoptical.com, dimitri.papadimitriou@alcatel-lucent.be, adrian@olddog.co.uk +# RFC4874 || CY. Lee, A. Farrel, S. De Cnodder || c.yin.lee@gmail.com, adrian@olddog.co.uk, stefaan.de_cnodder@alcatel-lucent.be +# RFC4875 || R. Aggarwal, Ed., D. Papadimitriou, Ed., S. Yasukawa, Ed. || rahul@juniper.net, yasukawa.seisho@lab.ntt.co.jp, Dimitri.Papadimitriou@alcatel-lucent.be +# RFC4876 || B. Neal-Joslin, Ed., L. Howard, M. Ansari || bob_joslin@hp.com, lukeh@padl.com, morteza@infoblox.com +# RFC4877 || V. Devarapalli, F. Dupont || vijay.devarapalli@azairenet.com, Francis.Dupont@fdupont.fr +# RFC4878 || M. Squire || msquire@hatterasnetworks.com +# RFC4879 || T. Narten || narten@us.ibm.com +# RFC4880 || J. Callas, L. Donnerhacke, H. Finney, D. Shaw, R. Thayer || jon@callas.org, lutz@iks-jena.de, hal@finney.org, dshaw@jabberwocky.com, rodney@canola-jones.com +# RFC4881 || K. El Malki, Ed. || karim@athonet.com +# RFC4882 || R. Koodli || rajeev.koodli@nokia.com +# RFC4883 || G. Feher, K. Nemeth, A. Korn, I. Cselenyi || Gabor.Feher@tmit.bme.hu, Krisztian.Nemeth@tmit.bme.hu, Andras.Korn@tmit.bme.hu, Istvan.Cselenyi@teliasonera.com +# RFC4884 || R. Bonica, D. Gan, D. Tappan, C. Pignataro || rbonica@juniper.net, derhwagan@yahoo.com, Dan.Tappan@gmail.com, cpignata@cisco.com +# RFC4885 || T. Ernst, H-Y. Lach || thierry.ernst@inria.fr, hong-yon.lach@motorola.com +# RFC4886 || T. Ernst || thierry.ernst@inria.fr +# RFC4887 || P. Thubert, R. Wakikawa, V. Devarapalli || pthubert@cisco.com, ryuji@sfc.wide.ad.jp, vijay.devarapalli@azairenet.com +# RFC4888 || C. Ng, P. Thubert, M. Watari, F. Zhao || chanwah.ng@sg.panasonic.com, pthubert@cisco.com, watari@kddilabs.jp, fanzhao@ucdavis.edu +# RFC4889 || C. Ng, F. Zhao, M. Watari, P. Thubert || chanwah.ng@sg.panasonic.com, fanzhao@ucdavis.edu, watari@kddilabs.jp, pthubert@cisco.com +# RFC4890 || E. Davies, J. Mohacsi || elwynd@dial.pipex.com, mohacsi@niif.hu +# RFC4891 || R. Graveman, M. Parthasarathy, P. Savola, H. Tschofenig || rfg@acm.org, mohanp@sbcglobal.net, psavola@funet.fi, Hannes.Tschofenig@nsn.com +# RFC4892 || S. Woolf, D. Conrad || woolf@isc.org, david.conrad@icann.org +# RFC4893 || Q. Vohra, E. Chen || quaizar.vohra@gmail.com, enkechen@cisco.com +# RFC4894 || P. Hoffman || paul.hoffman@vpnc.org +# RFC4895 || M. Tuexen, R. Stewart, P. Lei, E. Rescorla || tuexen@fh-muenster.de, randall@lakerest.net, peterlei@cisco.com, ekr@rtfm.com +# RFC4896 || A. Surtees, M. West, A.B. Roach || abigail.surtees@roke.co.uk, mark.a.west@roke.co.uk, adam@estacado.net +# RFC4897 || J. Klensin, S. Hartman || john-ietf@jck.com, hartmans-ietf@mit.edu +# RFC4898 || M. Mathis, J. Heffner, R. Raghunarayan || mathis@psc.edu, jheffner@psc.edu, raraghun@cisco.com +# RFC4901 || J. Ash, Ed., J. Hand, Ed., A. Malis, Ed. || gash5107@yahoo.com, jameshand@att.com, andrew.g.malis@verizon.com +# RFC4902 || M. Stecher || martin.stecher@webwasher.com +# RFC4903 || D. Thaler || dthaler@microsoft.com +# RFC4904 || V. Gurbani, C. Jennings || vkg@alcatel-lucent.com, fluffy@cisco.com +# RFC4905 || L. Martini, Ed., E. Rosen, Ed., N. El-Aawar, Ed. || lmartini@cisco.com, erosen@cisco.com, nna@level3.net +# RFC4906 || L. Martini, Ed., E. Rosen, Ed., N. El-Aawar, Ed. || lmartini@cisco.com, erosen@cisco.com, nna@level3.net +# RFC4907 || B. Aboba, Ed. || bernarda@microsoft.com +# RFC4908 || K. Nagami, S. Uda, N. Ogashiwa, H. Esaki, R. Wakikawa, H. Ohnishi || nagami@inetcore.com, zin@jaist.ac.jp, ogashiwa@wide.ad.jp, hiroshi@wide.ad.jp, ryuji@sfc.wide.ad.jp, ohnishi.hiroyuki@lab.ntt.co.jp +# RFC4909 || L. Dondeti, Ed., D. Castleford, F. Hartung || ldondeti@qualcomm.com, david.castleford@orange-ftgroup.com, frank.hartung@ericsson.com +# RFC4910 || S. Legg, D. Prager || steven.legg@eb2bcom.com, dap@austhink.com +# RFC4911 || S. Legg || steven.legg@eb2bcom.com +# RFC4912 || S. Legg || steven.legg@eb2bcom.com +# RFC4913 || S. Legg || steven.legg@eb2bcom.com +# RFC4914 || S. Legg || steven.legg@eb2bcom.com +# RFC4915 || P. Psenak, S. Mirtorabi, A. Roy, L. Nguyen, P. Pillay-Esnault || ppsenak@cisco.com, sina@force10networks.com, akr@cisco.com, lhnguyen@cisco.com, ppe@cisco.com +# RFC4916 || J. Elwell || john.elwell@siemens.com +# RFC4917 || V. Sastry, K. Leung, A. Patel || venkat.s@samsung.com, kleung@cisco.com, alpesh@cisco.com +# RFC4918 || L. Dusseault, Ed. || lisa.dusseault@gmail.com +# RFC4919 || N. Kushalnagar, G. Montenegro, C. Schumacher || nandakishore.kushalnagar@intel.com, gabriel.montenegro@microsoft.com, schumacher@danfoss.com +# RFC4920 || A. Farrel, Ed., A. Satyanarayana, A. Iwata, N. Fujita, G. Ash || adrian@olddog.co.uk, asatyana@cisco.com, a-iwata@ah.jp.nec.com, n-fujita@bk.jp.nec.com, gash5107@yahoo.com +# RFC4923 || F. Baker, P. Bose || fred@cisco.com, pratik.bose@lmco.com +# RFC4924 || B. Aboba, Ed., E. Davies || bernarda@microsoft.com, elwynd@dial.pipex.com +# RFC4925 || X. Li, Ed., S. Dawkins, Ed., D. Ward, Ed., A. Durand, Ed. || xing@cernet.edu.cn, spencer@mcsr-labs.org, dward@cisco.com, alain_durand@cable.comcast.com +# RFC4926 || T.Kalin, M.Molina || tomaz.kalin@dante.org.uk, maurizio.molina@dante.org.uk +# RFC4927 || J.-L. Le Roux, Ed. || jeanlouis.leroux@orange-ftgroup.com +# RFC4928 || G. Swallow, S. Bryant, L. Andersson || stbryant@cisco.com, swallow@cisco.com, loa@pi.se +# RFC4929 || L. Andersson, Ed., A. Farrel, Ed. || loa@pi.se, adrian@olddog.co.uk +# RFC4930 || S. Hollenbeck || shollenbeck@verisign.com +# RFC4931 || S. Hollenbeck || shollenbeck@verisign.com +# RFC4932 || S. Hollenbeck || shollenbeck@verisign.com +# RFC4933 || S. Hollenbeck || shollenbeck@verisign.com +# RFC4934 || S. Hollenbeck || shollenbeck@verisign.com +# RFC4935 || C. DeSanti, H.K. Vivek, K. McCloghrie, S. Gai || cds@cisco.com, hvivek@cisco.com, kzm@cisco.com, sgai@nuovasystems.com +# RFC4936 || C. DeSanti, H.K. Vivek, K. McCloghrie, S. Gai || cds@cisco.com, hvivek@cisco.com, kzm@cisco.com, sgai@nuovasystems.com +# RFC4937 || P. Arberg, V. Mammoliti || parberg@redback.com, vince@cisco.com +# RFC4938 || B. Berry, H. Holgate || bberry@cisco.com, hholgate@cisco.com +# RFC4939 || K. Gibbons, G. Ramkumar, S. Kipp || kgibbons@yahoo.com, gramkumar@stanfordalumni.org, skipp@brocade.com +# RFC4940 || K. Kompella, B. Fenner || kireeti@juniper.net, fenner@research.att.com +# RFC4941 || T. Narten, R. Draves, S. Krishnan || narten@us.ibm.com, richdr@microsoft.com, suresh.krishnan@ericsson.com +# RFC4942 || E. Davies, S. Krishnan, P. Savola || elwynd@dial.pipex.com, suresh.krishnan@ericsson.com, psavola@funet.fi +# RFC4943 || S. Roy, A. Durand, J. Paugh || sebastien.roy@sun.com, alain_durand@cable.comcast.com, jim.paugh@nominum.com +# RFC4944 || G. Montenegro, N. Kushalnagar, J. Hui, D. Culler || gabriel.montenegro@microsoft.com, nandakishore.kushalnagar@intel.com, jhui@archrock.com, dculler@archrock.com +# RFC4945 || B. Korver || briank@networkresonance.com +# RFC4946 || J. Snell || jasnell@gmail.com +# RFC4947 || G. Fairhurst, M. Montpetit || gorry@erg.abdn.ac.uk, mmontpetit@motorola.com +# RFC4948 || L. Andersson, E. Davies, L. Zhang || loa@pi.se, elwynd@dial.pipex.com, lixia@cs.ucla.edu +# RFC4949 || R. Shirey || rwshirey4949@verizon.net +# RFC4950 || R. Bonica, D. Gan, D. Tappan, C. Pignataro || rbonica@juniper.net, derhwagan@yahoo.com, dan.tappan@gmail.com, cpignata@cisco.com +# RFC4951 || V. Jain, Ed. || vipinietf@yahoo.com +# RFC4952 || J. Klensin, Y. Ko || john-ietf@jck.com, yw@mrko.pe.kr +# RFC4953 || J. Touch || touch@isi.edu +# RFC4954 || R. Siemborski, Ed., A. Melnikov, Ed. || robsiemb@google.com, Alexey.Melnikov@isode.com +# RFC4955 || D. Blacka || davidb@verisign.com +# RFC4956 || R. Arends, M. Kosters, D. Blacka || roy@nominet.org.uk, markk@verisign.com, davidb@verisign.com +# RFC4957 || S. Krishnan, Ed., N. Montavont, E. Njedjou, S. Veerepalli, A. Yegin, Ed. || suresh.krishnan@ericsson.com, nicolas.montavont@enst-bretagne.fr, eric.njedjou@orange-ftgroup.com, sivav@qualcomm.com, a.yegin@partner.samsung.com +# RFC4958 || K. Carlberg || carlberg@g11.org.uk +# RFC4959 || R. Siemborski, A. Gulbrandsen || robsiemb@google.com, arnt@oryx.com +# RFC4960 || R. Stewart, Ed. || randall@lakerest.net +# RFC4961 || D. Wing || dwing-ietf@fuggles.com +# RFC4962 || R. Housley, B. Aboba || housley@vigilsec.com, bernarda@microsoft.com +# RFC4963 || J. Heffner, M. Mathis, B. Chandler || jheffner@psc.edu, mathis@psc.edu, bchandle@gmail.com +# RFC4964 || A. Allen, Ed., J. Holm, T. Hallin || aallen@rim.com, Jan.Holm@ericsson.com, thallin@motorola.com +# RFC4965 || J-F. Mule, W. Townsley || jf.mule@cablelabs.com, mark@townsley.net +# RFC4966 || C. Aoun, E. Davies || ietf@energizeurnet.com, elwynd@dial.pipex.com +# RFC4967 || B. Rosen || br@brianrosen.net +# RFC4968 || S. Madanapalli, Ed. || smadanapalli@gmail.com +# RFC4969 || A. Mayrhofer || alexander.mayrhofer@enum.at +# RFC4970 || A. Lindem, Ed., N. Shen, JP. Vasseur, R. Aggarwal, S. Shaffer || acee@redback.com, naiming@cisco.com, jpv@cisco.com, rahul@juniper.net, sshaffer@bridgeport-networks.com +# RFC4971 || JP. Vasseur, Ed., N. Shen, Ed., R. Aggarwal, Ed. || jpv@cisco.com, naiming@cisco.com, rahul@juniper.net +# RFC4972 || JP. Vasseur, Ed., JL. Leroux, Ed., S. Yasukawa, S. Previdi, P. Psenak, P. Mabbey || jpv@cisco.com, jeanlouis.leroux@orange-ftgroup.com, s.yasukawa@hco.ntt.co.jp, sprevidi@cisco.com, ppsenak@cisco.com, Paul_Mabey@cable.comcast.com +# RFC4973 || P. Srisuresh, P. Joseph || srisuresh@yahoo.com, paul_95014@yahoo.com +# RFC4974 || D. Papadimitriou, A. Farrel || dimitri.papadimitriou@alcatel-lucent.be, adrian@olddog.co.uk +# RFC4975 || B. Campbell, Ed., R. Mahy, Ed., C. Jennings, Ed. || ben@estacado.net, rohan@ekabal.com, fluffy@cisco.com +# RFC4976 || C. Jennings, R. Mahy, A. B. Roach || fluffy@cisco.com, rohan@ekabal.com, adam@estacado.net +# RFC4977 || G. Tsirtsis, H. Soliman || tsirtsis@qualcomm.com, hesham@elevatemobile.com +# RFC4978 || A. Gulbrandsen || arnt@oryx.com +# RFC4979 || A. Mayrhofer || alexander.mayrhofer@enum.at +# RFC4980 || C. Ng, T. Ernst, E. Paik, M. Bagnulo || chanwah.ng@sg.panasonic.com, thierry.ernst@inria.fr, euna@kt.co.kr, marcelo@it.uc3m.es +# RFC4981 || J. Risson, T. Moors || jr@tuffit.com, t.moors@unsw.edu.au +# RFC4982 || M. Bagnulo, J. Arkko || marcelo@it.uc3m.es, jari.arkko@ericsson.com +# RFC4983 || C. DeSanti, H.K. Vivek, K. McCloghrie, S. Gai || cds@cisco.com, hvivek@cisco.com, kzm@cisco.com, sgai@nuovasystems.com +# RFC4984 || D. Meyer, Ed., L. Zhang, Ed., K. Fall, Ed. || dmm@1-4-5.net, lixia@cs.ucla.edu, kfall@intel.com +# RFC4985 || S. Santesson || stefans@microsoft.com +# RFC4986 || H. Eland, R. Mundy, S. Crocker, S. Krishnaswamy || heland@afilias.info, mundy@sparta.com, steve@shinkuro.com, suresh@sparta.com +# RFC4987 || W. Eddy || weddy@grc.nasa.gov +# RFC4988 || R. Koodli, C. Perkins || rajeev.koodli@nokia.com, charles.perkins@nokia.com +# RFC4990 || K. Shiomoto, R. Papneja, R. Rabbat || shiomoto.kohei@lab.ntt.co.jp, rabbat@alum.mit.edu, rpapneja@isocore.com +# RFC4991 || A. Newton || andy@hxr.us +# RFC4992 || A. Newton || andy@hxr.us +# RFC4993 || A. Newton || andy@hxr.us +# RFC4994 || S. Zeng, B. Volz, K. Kinnear, J. Brzozowski || szeng@cisco.com, volz@cisco.com, kkinnear@cisco.com, john_brzozowski@cable.comcast.com +# RFC4995 || L-E. Jonsson, G. Pelletier, K. Sandlund || lars-erik@lejonsson.com, ghyslain.pelletier@ericsson.com, kristofer.sandlund@ericsson.com +# RFC4996 || G. Pelletier, K. Sandlund, L-E. Jonsson, M. West || ghyslain.pelletier@ericsson.com, kristofer.sandlund@ericsson.com, lars-erik@lejonsson.com, mark.a.west@roke.co.uk +# RFC4997 || R. Finking, G. Pelletier || robert.finking@roke.co.uk, ghyslain.pelletier@ericsson.com +# RFC4998 || T. Gondrom, R. Brandner, U. Pordesch || tobias.gondrom@opentext.com, ralf.brandner@intercomponentware.com, ulrich.pordesch@zv.fraunhofer.de +# RFC5000 || RFC Editor || rfc-editor@rfc-editor.org +# RFC5001 || R. Austein || sra@isc.org +# RFC5002 || G. Camarillo, G. Blanco || Gonzalo.Camarillo@ericsson.com, German.Blanco@ericsson.com +# RFC5003 || C. Metz, L. Martini, F. Balus, J. Sugimoto || chmetz@cisco.com, lmartini@cisco.com, florin.balus@alcatel-lucent.com, sugimoto@nortel.com +# RFC5004 || E. Chen, S. Sangli || enkechen@cisco.com, rsrihari@cisco.com +# RFC5005 || M. Nottingham || mnot@pobox.com +# RFC5006 || J. Jeong, Ed., S. Park, L. Beloeil, S. Madanapalli || jjeong@cs.umn.edu, soohong.park@samsung.com, luc.beloeil@orange-ftgroup.com, smadanapalli@gmail.com +# RFC5007 || J. Brzozowski, K. Kinnear, B. Volz, S. Zeng || john_brzozowski@cable.comcast.com, kkinnear@cisco.com, volz@cisco.com, szeng@cisco.com +# RFC5008 || R. Housley, J. Solinas || housley@vigilsec.com, jasolin@orion.ncsc.mil +# RFC5009 || R. Ejza || ejzak@alcatel-lucent.com +# RFC5010 || K. Kinnear, M. Normoyle, M. Stapp || kkinnear@cisco.com, mnormoyle@cisco.com, mjs@cisco.com +# RFC5011 || M. StJohns || mstjohns@comcast.net +# RFC5012 || H. Schulzrinne, R. Marshall, Ed. || hgs+ecrit@cs.columbia.edu, rmarshall@telecomsys.com +# RFC5013 || J. Kunze, T. Baker || jak@ucop.edu, tbaker@tbaker.de +# RFC5014 || E. Nordmark, S. Chakrabarti, J. Laganier || Erik.Nordmark@Sun.com, samitac2@gmail.com, julien.IETF@laposte.net +# RFC5015 || M. Handley, I. Kouvelas, T. Speakman, L. Vicisano || M.Handley@cs.ucl.ac.uk, kouvelas@cisco.com, speakman@cisco.com, lorenzo@digitalfountain.com +# RFC5016 || M. Thomas || mat@cisco.com +# RFC5017 || D. McWalter, Ed. || dmcw@dataconnection.com +# RFC5018 || G. Camarillo || Gonzalo.Camarillo@ericsson.com +# RFC5019 || A. Deacon, R. Hurst || alex@verisign.com, rmh@microsoft.com +# RFC5020 || K. Zeilenga || Kurt.Zeilenga@Isode.COM +# RFC5021 || S. Josefsson || simon@josefsson.org +# RFC5022 || J. Van Dyke, E. Burger, Ed., A. Spitzer || jvandyke@cantata.com, eburger@standardstrack.com, woof@pingtel.com +# RFC5023 || J. Gregorio, Ed., B. de hOra, Ed. || joe@bitworking.org, bill@dehora.net +# RFC5024 || I. Friend || ieuan.friend@dip.co.uk +# RFC5025 || J. Rosenberg || jdrosen@cisco.com +# RFC5026 || G. Giaretta, Ed., J. Kempf, V. Devarapalli, Ed. || gerardog@qualcomm.com, kempf@docomolabs-usa.com, vijay.devarapalli@azairenet.com +# RFC5027 || F. Andreasen, D. Wing || fandreas@cisco.com, dwing-ietf@fuggles.com +# RFC5028 || R. Mahy || rohan@ekabal.com +# RFC5029 || JP. Vasseur, S. Previdi || jpv@cisco.com, sprevidi@cisco.com +# RFC5030 || M. Nakhjiri, Ed., K. Chowdhury, A. Lior, K. Leung || madjid.nakhjiri@motorola.com, kchowdhury@starentnetworks.com, avi@bridgewatersystems.com, kleung@cisco.com +# RFC5031 || H. Schulzrinne || hgs+ecrit@cs.columbia.edu +# RFC5032 || E. Burger, Ed. || eric.burger@bea.com +# RFC5033 || S. Floyd, M. Allman || floyd@icir.org, mallman@icir.org +# RFC5034 || R. Siemborski, A. Menon-Sen || robsiemb@google.com, ams@oryx.com +# RFC5035 || J. Schaad || jimsch@exmsft.com +# RFC5036 || L. Andersson, Ed., I. Minei, Ed., B. Thomas, Ed. || loa@pi.se, ina@juniper.net, rhthomas@cisco.com +# RFC5037 || L. Andersson, Ed., I. Minei, Ed., B. Thomas, Ed. || loa@pi.se, ina@juniper.net, rhthomas@cisco.com +# RFC5038 || B. Thomas, L. Andersson || loa@pi.se, rhthomas@cisco.com +# RFC5039 || J. Rosenberg, C. Jennings || jdrosen@cisco.com, fluffy@cisco.com +# RFC5040 || R. Recio, B. Metzler, P. Culley, J. Hilland, D. Garcia || recio@us.ibm.com, bmt@zurich.ibm.com, paul.culley@hp.com, jeff.hilland@hp.com, Dave.Garcia@StanfordAlumni.org +# RFC5041 || H. Shah, J. Pinkerton, R. Recio, P. Culley || hemal@broadcom.com, jpink@microsoft.com, recio@us.ibm.com, paul.culley@hp.com +# RFC5042 || J. Pinkerton, E. Deleganes || jpink@windows.microsoft.com, deleganes@yahoo.com +# RFC5043 || C. Bestler, Ed., R. Stewart, Ed. || caitlin.bestler@neterion.com, randall@lakerest.net +# RFC5044 || P. Culley, U. Elzur, R. Recio, S. Bailey, J. Carrier || paul.culley@hp.com, uri@broadcom.com, recio@us.ibm.com, steph@sandburst.com, carrier@cray.com +# RFC5045 || C. Bestler, Ed., L. Coene || caitlin.bestler@neterion.com, lode.coene@nsn.com +# RFC5046 || M. Ko, M. Chadalapaka, J. Hufferd, U. Elzur, H. Shah, P. Thaler || mako@us.ibm.com, cbm@rose.hp.com, jhufferd@brocade.com, Uri@Broadcom.com, hemal@broadcom.com, pthaler@broadcom.com +# RFC5047 || M. Chadalapaka, J. Hufferd, J. Satran, H. Shah || cbm@rose.hp.com, jhufferd@brocade.com, Julian_Satran@il.ibm.com, hemal@broadcom.com +# RFC5048 || M. Chadalapaka, Ed. || cbm@rose.hp.com +# RFC5049 || C. Bormann, Z. Liu, R. Price, G. Camarillo, Ed. || cabo@tzi.org, zhigang.c.liu@nokia.com, richard.price@eads.com, Gonzalo.Camarillo@ericsson.com +# RFC5050 || K. Scott, S. Burleigh || kscott@mitre.org, Scott.Burleigh@jpl.nasa.gov +# RFC5051 || M. Crispin || MRC@CAC.Washington.EDU +# RFC5052 || M. Watson, M. Luby, L. Vicisano || mark@digitalfountain.com, luby@digitalfountain.com, lorenzo@digitalfountain.com +# RFC5053 || M. Luby, A. Shokrollahi, M. Watson, T. Stockhammer || luby@digitalfountain.com, amin.shokrollahi@epfl.ch, mark@digitalfountain.com, stockhammer@nomor.de +# RFC5054 || D. Taylor, T. Wu, N. Mavrogiannopoulos, T. Perrin || dtaylor@gnutls.org, thomwu@cisco.com, nmav@gnutls.org, trevp@trevp.net +# RFC5055 || T. Freeman, R. Housley, A. Malpani, D. Cooper, W. Polk || trevorf@microsoft.com, housley@vigilsec.com, ambarish@yahoo.com, david.cooper@nist.gov, wpolk@nist.gov +# RFC5056 || N. Williams || Nicolas.Williams@sun.com +# RFC5057 || R. Sparks || RjS@estacado.net +# RFC5058 || R. Boivie, N. Feldman, Y. Imai, W. Livens, D. Ooms || rhboivie@us.ibm.com, nkfeldman@yahoo.com, ug@xcast.jp, wim@livens.net, dirk@onesparrow.com +# RFC5059 || N. Bhaskar, A. Gall, J. Lingard, S. Venaas || nidhi@arastra.com, alexander.gall@switch.ch, jchl@arastra.com, venaas@uninett.no +# RFC5060 || R. Sivaramu, J. Lingard, D. McWalter, B. Joshi, A. Kessler || raghava@cisco.com, jchl@arastra.com, dmcw@dataconnection.com, bharat_joshi@infosys.com, kessler@cisco.com +# RFC5061 || R. Stewart, Q. Xie, M. Tuexen, S. Maruyama, M. Kozuka || randall@lakerest.net, Qiaobing.Xie@motorola.com, tuexen@fh-muenster.de, mail@marushin.gr.jp, ma-kun@kozuka.jp +# RFC5062 || R. Stewart, M. Tuexen, G. Camarillo || randall@lakerest.net, tuexen@fh-muenster.de, Gonzalo.Camarillo@ericsson.com +# RFC5063 || A. Satyanarayana, Ed., R. Rahman, Ed. || asatyana@cisco.com, rrahman@cisco.com +# RFC5064 || M. Duerst || duerst@it.aoyama.ac.jp +# RFC5065 || P. Traina, D. McPherson, J. Scudder || bgp-confederations@st04.pst.org, danny@arbor.net, jgs@juniper.net +# RFC5066 || E. Beili || edward.beili@actelis.com +# RFC5067 || S. Lind, P. Pfautz || sdlind@att.com, ppfautz@att.com +# RFC5068 || C. Hutzler, D. Crocker, P. Resnick, E. Allman, T. Finch || cdhutzler@aol.com, dcrocker@bbiw.net, presnick@qti.qualcomm.com, eric+ietf-smtp@sendmail.org, dot@dotat.at +# RFC5069 || T. Taylor, Ed., H. Tschofenig, H. Schulzrinne, M. Shanmugam || tom.taylor.stds@gmail.com, Hannes.Tschofenig@nsn.com, hgs+ecrit@cs.columbia.edu, murugaraj.shanmugam@detecon.com +# RFC5070 || R. Danyliw, J. Meijer, Y. Demchenko || rdd@cert.org, jan@flyingcloggies.nl, demch@chello.nl +# RFC5071 || D. Hankins || David_Hankins@isc.org +# RFC5072 || S. Varada, Ed., D. Haskins, E. Allen || varada@txc.com +# RFC5073 || J.P. Vasseur, Ed., J.L. Le Roux, Ed. || jpv@cisco.com, jeanlouis.leroux@orange-ftgroup.com +# RFC5074 || S. Weiler || weiler@tislabs.com +# RFC5075 || B. Haberman, Ed., R. Hinden || brian@innovationslab.net, bob.hinden@gmail.com +# RFC5076 || B. Hoeneisen || hoeneisen@switch.ch +# RFC5077 || J. Salowey, H. Zhou, P. Eronen, H. Tschofenig || jsalowey@cisco.com, hzhou@cisco.com, pe@iki.fi, Hannes.Tschofenig@nsn.com +# RFC5078 || S. Dawkins || spencer@mcsr-labs.org +# RFC5079 || J. Rosenberg || jdrosen@cisco.com +# RFC5080 || D. Nelson, A. DeKok || dnelson@elbrysnetworks.com, aland@freeradius.org +# RFC5081 || N. Mavrogiannopoulos || nmav@gnutls.org +# RFC5082 || V. Gill, J. Heasley, D. Meyer, P. Savola, Ed., C. Pignataro || vijay@umbc.edu, heas@shrubbery.net, dmm@1-4-5.net, psavola@funet.fi, cpignata@cisco.com +# RFC5083 || R. Housley || housley@vigilsec.com +# RFC5084 || R. Housley || housley@vigilsec.com +# RFC5085 || T. Nadeau, Ed., C. Pignataro, Ed. || tnadeau@lucidvision.com, cpignata@cisco.com +# RFC5086 || A. Vainshtein, Ed., I. Sasson, E. Metz, T. Frost, P. Pate || sasha@axerra.com, israel@axerra.com, e.t.metz@telecom.tno.nl, tfrost@symmetricom.com, prayson.pate@overturenetworks.com +# RFC5087 || Y(J). Stein, R. Shashoua, R. Insler, M. Anavi || yaakov_s@rad.com, ronen_s@rad.com, ron_i@rad.com, motty@radusa.com +# RFC5088 || JL. Le Roux, Ed., JP. Vasseur, Ed., Y. Ikejiri, R. Zhang || jeanlouis.leroux@orange-ftgroup.com, jpv@cisco.com, y.ikejiri@ntt.com, raymond.zhang@bt.com +# RFC5089 || JL. Le Roux, Ed., JP. Vasseur, Ed., Y. Ikejiri, R. Zhang || jeanlouis.leroux@orange-ftgroup.com, jpv@cisco.com, y.ikejiri@ntt.com, raymond.zhang@bt.com +# RFC5090 || B. Sterman, D. Sadolevsky, D. Schwartz, D. Williams, W. Beck || baruch@kayote.com, dscreat@dscreat.com, david@kayote.com, dwilli@cisco.com, beckw@t-systems.com +# RFC5091 || X. Boyen, L. Martin || xavier@voltage.com, martin@voltage.com +# RFC5092 || A. Melnikov, Ed., C. Newman || Alexey.Melnikov@isode.com, chris.newman@sun.com +# RFC5093 || G. Hunt || geoff.hunt@bt.com +# RFC5094 || V. Devarapalli, A. Patel, K. Leung || vijay.devarapalli@azairenet.com, alpesh@cisco.com, kleung@cisco.com +# RFC5095 || J. Abley, P. Savola, G. Neville-Neil || jabley@ca.afilias.info, psavola@funet.fi, gnn@neville-neil.com +# RFC5096 || V. Devarapalli || vijay.devarapalli@azairenet.com +# RFC5097 || G. Renker, G. Fairhurst || gerrit@erg.abdn.ac.uk, gorry@erg.abdn.ac.uk +# RFC5098 || G. Beacham, S. Kumar, S. Channabasappa || gordon.beacham@motorola.com, satish.kumar@ti.com, Sumanth@cablelabs.com +# RFC5101 || B. Claise, Ed. || bclaise@cisco.com +# RFC5102 || J. Quittek, S. Bryant, B. Claise, P. Aitken, J. Meyer || quittek@netlab.nec.de, stbryant@cisco.com, bclaise@cisco.com, paitken@cisco.com, jemeyer@paypal.com +# RFC5103 || B. Trammell, E. Boschi || bht@cert.org, elisa.boschi@hitachi-eu.com +# RFC5104 || S. Wenger, U. Chandra, M. Westerlund, B. Burman || stewe@stewe.org, Umesh.1.Chandra@nokia.com, magnus.westerlund@ericsson.com, bo.burman@ericsson.com +# RFC5105 || O. Lendl || otmar.lendl@enum.at +# RFC5106 || H. Tschofenig, D. Kroeselberg, A. Pashalidis, Y. Ohba, F. Bersani || Hannes.Tschofenig@nsn.com, Dirk.Kroeselberg@nsn.com, pashalidis@nw.neclab.eu, yohba@tari.toshiba.com, florent.ftrd@gmail.com +# RFC5107 || R. Johnson, J. Kumarasamy, K. Kinnear, M. Stapp || raj@cisco.com, jayk@cisco.com, kkinnear@cisco.com, mjs@cisco.com +# RFC5109 || A. Li, Ed. || adamli@hyervision.com +# RFC5110 || P. Savola || psavola@funet.fi +# RFC5111 || B. Aboba, L. Dondeti || bernarda@microsoft.com, ldondeti@qualcomm.com +# RFC5112 || M. Garcia-Martin || miguel.garcia@nsn.com +# RFC5113 || J. Arkko, B. Aboba, J. Korhonen, Ed., F. Bari || jari.arkko@ericsson.com, bernarda@microsoft.com, jouni.korhonen@teliasonera.com, farooq.bari@att.com +# RFC5114 || M. Lepinski, S. Kent || mlepinski@bbn.com, kent@bbn.com +# RFC5115 || K. Carlberg, P. O'Hanlon || carlberg@g11.org.uk, p.ohanlon@cs.ucl.ac.uk +# RFC5116 || D. McGrew || mcgrew@cisco.com +# RFC5117 || M. Westerlund, S. Wenger || magnus.westerlund@ericsson.com, stewe@stewe.org +# RFC5118 || V. Gurbani, C. Boulton, R. Sparks || vkg@alcatel-lucent.com, cboulton@ubiquitysoftware.com, RjS@estacado.net +# RFC5119 || T. Edwards || thomas.edwards@fox.com +# RFC5120 || T. Przygienda, N. Shen, N. Sheth || prz@net4u.ch, naiming@cisco.com, nsheth@juniper.net +# RFC5121 || B. Patil, F. Xia, B. Sarikaya, JH. Choi, S. Madanapalli || basavaraj.patil@nsn.com, xiayangsong@huawei.com, sarikaya@ieee.org, jinchoe@samsung.com, smadanapalli@gmail.com +# RFC5122 || P. Saint-Andre || ietf@stpeter.im +# RFC5123 || R. White, B. Akyol || riw@cisco.com, bora@cisco.com +# RFC5124 || J. Ott, E. Carrara || jo@comnet.tkk.fi, carrara@kth.se +# RFC5125 || T. Taylor || tom.taylor.stds@gmail.com +# RFC5126 || D. Pinkas, N. Pope, J. Ross || Denis.Pinkas@bull.net, nick.pope@thales-esecurity.com, ross@secstan.com +# RFC5127 || K. Chan, J. Babiarz, F. Baker || khchan@nortel.com, babiarz@nortel.com, fred@cisco.com +# RFC5128 || P. Srisuresh, B. Ford, D. Kegel || srisuresh@yahoo.com, baford@mit.edu, dank06@kegel.com +# RFC5129 || B. Davie, B. Briscoe, J. Tay || bsd@cisco.com, bob.briscoe@bt.com, june.tay@bt.com +# RFC5130 || S. Previdi, M. Shand, Ed., C. Martin || sprevidi@cisco.com, mshand@cisco.com, chris@ipath.net +# RFC5131 || D. McWalter, Ed. || dmcw@dataconnection.com +# RFC5132 || D. McWalter, D. Thaler, A. Kessler || dmcw@dataconnection.com, dthaler@windows.microsoft.com, kessler@cisco.com +# RFC5133 || M. Tuexen, K. Morneault || tuexen@fh-muenster.de, kmorneau@cisco.com +# RFC5134 || M. Mealling || michael@refactored-networks.com +# RFC5135 || D. Wing, T. Eckert || dwing-ietf@fuggles.com, eckert@cisco.com +# RFC5136 || P. Chimento, J. Ishac || Philip.Chimento@jhuapl.edu, jishac@nasa.gov +# RFC5137 || J. Klensin || john-ietf@jck.com +# RFC5138 || S. Cox || Simon.Cox@csiro.au +# RFC5139 || M. Thomson, J. Winterbottom || martin.thomson@andrew.com, james.winterbottom@andrew.com +# RFC5140 || M. Bangalore, R. Kumar, J. Rosenberg, H. Salama, D.N. Shah || manjax@cisco.com, rajneesh@cisco.com, jdrosen@cisco.com, hsalama@citexsoftware.com, dhaval@moowee.tv +# RFC5141 || J. Goodwin, H. Apel || goodwin@iso.org, apel@iso.org +# RFC5142 || B. Haley, V. Devarapalli, H. Deng, J. Kempf || brian.haley@hp.com, vijay.devarapalli@azairenet.com, kempf@docomolabs-usa.com, denghui@chinamobile.com +# RFC5143 || A. Malis, J. Brayley, J. Shirron, L. Martini, S. Vogelsang || andrew.g.malis@verizon.com, jeremy.brayley@ecitele.com, john.shirron@ecitele.com, lmartini@cisco.com, steve.vogelsang@alcatel-lucent.com +# RFC5144 || A. Newton, M. Sanz || andy@arin.net, sanz@denic.de +# RFC5145 || K. Shiomoto, Ed. || shiomoto.kohei@lab.ntt.co.jp +# RFC5146 || K. Kumaki, Ed. || ke-kumaki@kddi.com +# RFC5147 || E. Wilde, M. Duerst || dret@berkeley.edu, duerst@it.aoyama.ac.jp +# RFC5148 || T. Clausen, C. Dearlove, B. Adamson || T.Clausen@computer.org, chris.dearlove@baesystems.com, adamson@itd.nrl.navy.mil +# RFC5149 || J. Korhonen, U. Nilsson, V. Devarapalli || jouni.korhonen@teliasonera.com, ulf.s.nilsson@teliasonera.com, vijay.devarapalli@azairenet.com +# RFC5150 || A. Ayyangar, K. Kompella, JP. Vasseur, A. Farrel || arthi@juniper.net, kireeti@juniper.net, jpv@cisco.com, adrian@olddog.co.uk +# RFC5151 || A. Farrel, Ed., A. Ayyangar, JP. Vasseur || adrian@olddog.co.uk, arthi@juniper.net, jpv@cisco.com +# RFC5152 || JP. Vasseur, Ed., A. Ayyangar, Ed., R. Zhang || jpv@cisco.com, arthi@juniper.net, raymond.zhang@bt.com +# RFC5153 || E. Boschi, L. Mark, J. Quittek, M. Stiemerling, P. Aitken || elisa.boschi@hitachi-eu.com, lutz.mark@fokus.fraunhofer.de, quittek@nw.neclab.eu, stiemerling@nw.neclab.eu, paitken@cisco.com +# RFC5154 || J. Jee, Ed., S. Madanapalli, J. Mandin || jhjee@etri.re.kr, smadanapalli@gmail.com, j_mandin@yahoo.com +# RFC5155 || B. Laurie, G. Sisson, R. Arends, D. Blacka || ben@links.org, geoff-s@panix.com, roy@nominet.org.uk, davidb@verisign.com +# RFC5156 || M. Blanchet || Marc.Blanchet@viagenie.ca +# RFC5157 || T. Chown || tjc@ecs.soton.ac.uk +# RFC5158 || G. Huston || gih@apnic.net +# RFC5159 || L. Dondeti, Ed., A. Jerichow || ldondeti@qualcomm.com, anja.jerichow@nsn.com +# RFC5160 || P. Levis, M. Boucadair || pierre.levis@orange-ftgroup.com, mohamed.boucadair@orange-ftgroup.com +# RFC5161 || A. Gulbrandsen, Ed., A. Melnikov, Ed. || arnt@oryx.com, Alexey.Melnikov@isode.com +# RFC5162 || A. Melnikov, D. Cridland, C. Wilson || Alexey.Melnikov@isode.com, dave.cridland@isode.com, corby@computer.org +# RFC5163 || G. Fairhurst, B. Collini-Nocker || gorry@erg.abdn.ac.uk, bnocker@cosy.sbg.ac.at +# RFC5164 || T. Melia, Ed. || tmelia@cisco.com +# RFC5165 || C. Reed || creed@opengeospatial.org +# RFC5166 || S. Floyd, Ed. || floyd@icir.org +# RFC5167 || M. Dolly, R. Even || mdolly@att.com, roni.even@polycom.co.il +# RFC5168 || O. Levin, R. Even, P. Hagendorf || oritl@microsoft.com, roni.even@polycom.co.il, pierre@radvision.com +# RFC5169 || T. Clancy, M. Nakhjiri, V. Narayanan, L. Dondeti || clancy@LTSnet.net, madjid.nakhjiri@motorola.com, vidyan@qualcomm.com, ldondeti@qualcomm.com +# RFC5170 || V. Roca, C. Neumann, D. Furodet || vincent.roca@inria.fr, christoph.neumann@thomson.net, david.furodet@st.com +# RFC5171 || M. Foschiano || foschia@cisco.com +# RFC5172 || S. Varada, Ed. || varada@ieee.org +# RFC5173 || J. Degener, P. Guenther || jutta@pobox.com, guenther@sendmail.com +# RFC5174 || J-P. Evain || evain@ebu.ch +# RFC5175 || B. Haberman, Ed., R. Hinden || brian@innovationslab.net, bob.hinden@gmail.com +# RFC5176 || M. Chiba, G. Dommety, M. Eklund, D. Mitton, B. Aboba || mchiba@cisco.com, gdommety@cisco.com, meklund@cisco.com, david@mitton.com, bernarda@microsoft.com +# RFC5177 || K. Leung, G. Dommety, V. Narayanan, A. Petrescu || kleung@cisco.com, gdommety@cisco.com, vidyan@qualcomm.com, alexandru.petrescu@motorola.com +# RFC5178 || N. Williams, A. Melnikov || Nicolas.Williams@sun.com, Alexey.Melnikov@isode.com +# RFC5179 || N. Williams || Nicolas.Williams@sun.com +# RFC5180 || C. Popoviciu, A. Hamza, G. Van de Velde, D. Dugatkin || cpopovic@cisco.com, ahamza@cisco.com, gunter@cisco.com, diego@fastsoft.com +# RFC5181 || M-K. Shin, Ed., Y-H. Han, S-E. Kim, D. Premec || myungki.shin@gmail.com, yhhan@kut.ac.kr, sekim@kt.co.kr, domagoj.premec@siemens.com +# RFC5182 || A. Melnikov || Alexey.Melnikov@isode.com +# RFC5183 || N. Freed || ned.freed@mrochek.com +# RFC5184 || F. Teraoka, K. Gogo, K. Mitsuya, R. Shibui, K. Mitani || tera@ics.keio.ac.jp, gogo@tera.ics.keio.ac.jp, mitsuya@sfc.wide.ad.jp, shibrie@tera.ics.keio.ac.jp, koki@tera.ics.keio.ac.jp, rajeev_koodli@yahoo.com +# RFC5185 || S. Mirtorabi, P. Psenak, A. Lindem, Ed., A. Oswal || sina@nuovasystems.com, ppsenak@cisco.com, acee@redback.com, aoswal@redback.com +# RFC5186 || B. Haberman, J. Martin || brian@innovationslab.net, jim@wovensystems.com +# RFC5187 || P. Pillay-Esnault, A. Lindem || ppe@cisco.com, acee@redback.com +# RFC5188 || H. Desineni, Q. Xie || hd@qualcomm.com, Qiaobing.Xie@Gmail.com +# RFC5189 || M. Stiemerling, J. Quittek, T. Taylor || stiemerling@nw.neclab.eu, quittek@nw.neclab.eu, tom.taylor.stds@gmail.com +# RFC5190 || J. Quittek, M. Stiemerling, P. Srisuresh || quittek@nw.neclab.eu, stiemerling@nw.neclab.eu, srisuresh@yahoo.com +# RFC5191 || D. Forsberg, Y. Ohba, Ed., B. Patil, H. Tschofenig, A. Yegin || dan.forsberg@nokia.com, yohba@tari.toshiba.com, basavaraj.patil@nsn.com, hannes.tschofenig@nsn.com, a.yegin@partner.samsung.com +# RFC5192 || L. Morand, A. Yegin, S. Kumar, S. Madanapalli || lionel.morand@orange-ftgroup.com, a.yegin@partner.samsung.com, surajk@techmahindra.com, syam@samsung.com +# RFC5193 || P. Jayaraman, R. Lopez, Y. Ohba, Ed., M. Parthasarathy, A. Yegin || prakash_jayaraman@net.com, rafa@um.es, yohba@tari.toshiba.com, mohanp@sbcglobal.net, a.yegin@partner.samsung.com +# RFC5194 || A. van Wijk, Ed., G. Gybels, Ed. || guido.gybels@rnid.org.uk, arnoud@realtimetext.org +# RFC5195 || H. Ould-Brahim, D. Fedyk, Y. Rekhter || hbrahim@nortel.com, yakov@juniper.net, dwfedyk@nortel.com +# RFC5196 || M. Lonnfors, K. Kiss || mikko.lonnfors@nokia.com, krisztian.kiss@nokia.com +# RFC5197 || S. Fries, D. Ignjatic || steffen.fries@siemens.com, dignjatic@polycom.com +# RFC5198 || J. Klensin, M. Padlipsky || john-ietf@jck.com, the.map@alum.mit.edu +# RFC5201 || R. Moskowitz, P. Nikander, P. Jokela, Ed., T. Henderson || rgm@icsalabs.com, pekka.nikander@nomadiclab.com, petri.jokela@nomadiclab.com, thomas.r.henderson@boeing.com +# RFC5202 || P. Jokela, R. Moskowitz, P. Nikander || petri.jokela@nomadiclab.com, rgm@icsalabs.com, pekka.nikander@nomadiclab.com +# RFC5203 || J. Laganier, T. Koponen, L. Eggert || julien.ietf@laposte.net, teemu.koponen@iki.fi, lars.eggert@nokia.com +# RFC5204 || J. Laganier, L. Eggert || julien.ietf@laposte.net, lars.eggert@nokia.com +# RFC5205 || P. Nikander, J. Laganier || pekka.nikander@nomadiclab.com, julien.ietf@laposte.net +# RFC5206 || P. Nikander, T. Henderson, Ed., C. Vogt, J. Arkko || pekka.nikander@nomadiclab.com, thomas.r.henderson@boeing.com, christian.vogt@ericsson.com, jari.arkko@ericsson.com +# RFC5207 || M. Stiemerling, J. Quittek, L. Eggert || stiemerling@netlab.nec.de, quittek@nw.neclab.eu, lars.eggert@nokia.com +# RFC5208 || B. Kaliski || kaliski_burt@emc.com +# RFC5209 || P. Sangster, H. Khosravi, M. Mani, K. Narayan, J. Tardo || Paul_Sangster@symantec.com, hormuzd.m.khosravi@intel.com, mmani@avaya.com, kaushik@cisco.com, joseph.tardo@nevisnetworks.com +# RFC5210 || J. Wu, J. Bi, X. Li, G. Ren, K. Xu, M. Williams || jianping@cernet.edu.cn, junbi@cernet.edu.cn, xing@cernet.edu.cn, rg03@mails.tsinghua.edu.cn, xuke@csnet1.cs.tsinghua.edu.cn, miw@juniper.net +# RFC5211 || J. Curran || jcurran@istaff.org +# RFC5212 || K. Shiomoto, D. Papadimitriou, JL. Le Roux, M. Vigoureux, D. Brungard || shiomoto.kohei@lab.ntt.co.jp, dimitri.papadimitriou@alcatel-lucent.be, jeanlouis.leroux@orange-ftgroup.com, martin.vigoureux@alcatel-lucent.fr, dbrungard@att.com +# RFC5213 || S. Gundavelli, Ed., K. Leung, V. Devarapalli, K. Chowdhury, B. Patil || sgundave@cisco.com, kleung@cisco.com, vijay@wichorus.com, kchowdhury@starentnetworks.com, basavaraj.patil@nokia.com +# RFC5214 || F. Templin, T. Gleeson, D. Thaler || fred.l.templin@boeing.com, tgleeson@cisco.com, dthaler@microsoft.com +# RFC5215 || L. Barbato || lu_zero@gentoo.org +# RFC5216 || D. Simon, B. Aboba, R. Hurst || dansimon@microsoft.com, bernarda@microsoft.com, rmh@microsoft.com +# RFC5217 || M. Shimaoka, Ed., N. Hastings, R. Nielsen || m-shimaoka@secom.co.jp, nelson.hastings@nist.gov, nielsen_rebecca@bah.com +# RFC5218 || D. Thaler, B. Aboba || dthaler@microsoft.com, bernarda@microsoft.com +# RFC5219 || R. Finlayson || finlayson@live555.com +# RFC5220 || A. Matsumoto, T. Fujisaki, R. Hiromi, K. Kanayama || arifumi@nttv6.net, fujisaki@nttv6.net, hiromi@inetcore.com, kanayama_kenichi@intec-si.co.jp +# RFC5221 || A. Matsumoto, T. Fujisaki, R. Hiromi, K. Kanayama || arifumi@nttv6.net, fujisaki@nttv6.net, hiromi@inetcore.com, kanayama_kenichi@intec-si.co.jp +# RFC5222 || T. Hardie, A. Newton, H. Schulzrinne, H. Tschofenig || hardie@qualcomm.com, andy@hxr.us, hgs+ecrit@cs.columbia.edu, Hannes.Tschofenig@nsn.com +# RFC5223 || H. Schulzrinne, J. Polk, H. Tschofenig || hgs+ecrit@cs.columbia.edu, jmpolk@cisco.com, Hannes.Tschofenig@nsn.com +# RFC5224 || M. Brenner || mrbrenner@alcatel-lucent.com +# RFC5225 || G. Pelletier, K. Sandlund || ghyslain.pelletier@ericsson.com, kristofer.sandlund@ericsson.com +# RFC5226 || T. Narten, H. Alvestrand || narten@us.ibm.com, Harald@Alvestrand.no +# RFC5227 || S. Cheshire || rfc@stuartcheshire.org +# RFC5228 || P. Guenther, Ed., T. Showalter, Ed. || guenther@sendmail.com, tjs@psaux.com +# RFC5229 || K. Homme || kjetilho@ifi.uio.no +# RFC5230 || T. Showalter, N. Freed, Ed. || tjs@psaux.com, ned.freed@mrochek.com +# RFC5231 || W. Segmuller, B. Leiba || werewolf@us.ibm.com, leiba@watson.ibm.com +# RFC5232 || A. Melnikov || alexey.melnikov@isode.com +# RFC5233 || K. Murchison || murch@andrew.cmu.edu +# RFC5234 || D. Crocker, Ed., P. Overell || dcrocker@bbiw.net, paul@bayleaf.org.uk +# RFC5235 || C. Daboo || cyrus@daboo.name +# RFC5236 || A. Jayasumana, N. Piratla, T. Banka, A. Bare, R. Whitner || Anura.Jayasumana@colostate.edu, Nischal.Piratla@telekom.de, Tarun.Banka@colostate.edu, abhijit_bare@agilent.com, rick_whitner@agilent.com +# RFC5237 || J. Arkko, S. Bradner || jari.arkko@piuha.net, sob@harvard.edu +# RFC5238 || T. Phelan || tphelan@sonusnet.com +# RFC5239 || M. Barnes, C. Boulton, O. Levin || mary.barnes@nortel.com, cboulton@avaya.com, oritl@microsoft.com +# RFC5240 || B. Joshi, R. Bijlani || bharat_joshi@infosys.com, rainab@gmail.com +# RFC5241 || A. Falk, S. Bradner || falk@bbn.com, sob@harvard.edu +# RFC5242 || J. Klensin, H. Alvestrand || john+ietf@jck.com, harald@alvestrand.no +# RFC5243 || R. Ogier || rich.ogier@earthlink.net +# RFC5244 || H. Schulzrinne, T. Taylor || schulzrinne@cs.columbia.edu, tom.taylor.stds@gmail.com +# RFC5245 || J. Rosenberg || jdrosen@jdrosen.net +# RFC5246 || T. Dierks, E. Rescorla || tim@dierks.org, ekr@rtfm.com +# RFC5247 || B. Aboba, D. Simon, P. Eronen || bernarda@microsoft.com, dansimon@microsoft.com, pe@iki.fi +# RFC5248 || T. Hansen, J. Klensin || tony+mailesc@maillennium.att.com, john+ietf@jck.com +# RFC5249 || D. Harrington, Ed. || dharrington@huawei.com +# RFC5250 || L. Berger, I. Bryskin, A. Zinin, R. Coltun || lberger@labn.net, ibryskin@advaoptical.com, alex.zinin@alcatel-lucent.com, none +# RFC5251 || D. Fedyk, Ed., Y. Rekhter, Ed., D. Papadimitriou, R. Rabbat, L. Berger || dwfedyk@nortel.com, yakov@juniper.net, Dimitri.Papadimitriou@alcatel-lucent.be, rabbat@alum.mit.edu, lberger@labn.net +# RFC5252 || I. Bryskin, L. Berger || ibryskin@advaoptical.com, lberger@labn.net +# RFC5253 || T. Takeda, Ed. || takeda.tomonori@lab.ntt.co.jp +# RFC5254 || N. Bitar, Ed., M. Bocci, Ed., L. Martini, Ed. || nabil.bitar@verizon.com, matthew.bocci@alcatel-lucent.co.uk, lmartini@cisco.com +# RFC5255 || C. Newman, A. Gulbrandsen, A. Melnikov || chris.newman@sun.com, arnt@oryx.com, Alexey.Melnikov@isode.com +# RFC5256 || M. Crispin, K. Murchison || IMAP+SORT+THREAD@Lingling.Panda.COM, murch@andrew.cmu.edu +# RFC5257 || C. Daboo, R. Gellens || cyrus@daboo.name, randy@qualcomm.com +# RFC5258 || B. Leiba, A. Melnikov || leiba@watson.ibm.com, Alexey.Melnikov@isode.com +# RFC5259 || A. Melnikov, Ed., P. Coates, Ed. || Alexey.Melnikov@isode.com, peter.coates@Sun.COM +# RFC5260 || N. Freed || ned.freed@mrochek.com +# RFC5261 || J. Urpalainen || jari.urpalainen@nokia.com +# RFC5262 || M. Lonnfors, E. Leppanen, H. Khartabil, J. Urpalainen || mikko.lonnfors@nokia.com, eva.leppanen@saunalahti.fi, hisham.khartabil@gmail.com, jari.urpalainen@nokia.com +# RFC5263 || M. Lonnfors, J. Costa-Requena, E. Leppanen, H. Khartabil || mikko.lonnfors@nokia.com, jose.costa-requena@nokia.com, eva.leppanen@saunalahti.fi, hisham.khartabil@gmail.com +# RFC5264 || A. Niemi, M. Lonnfors, E. Leppanen || aki.niemi@nokia.com, mikko.lonnfors@nokia.com, eva.leppanen@saunalaht.fi +# RFC5265 || S. Vaarala, E. Klovning || sami.vaarala@iki.fi, espen@birdstep.com +# RFC5266 || V. Devarapalli, P. Eronen || vijay@wichorus.com, pe@iki.fi +# RFC5267 || D. Cridland, C. King || dave.cridland@isode.com, cking@mumbo.ca +# RFC5268 || R. Koodli, Ed. || rkoodli@starentnetworks.com[ +# RFC5269 || J. Kempf, R. Koodli || kempf@docomolabs-usa.com, rkoodli@starentnetworks.com +# RFC5270 || H. Jang, J. Jee, Y. Han, S. Park, J. Cha || heejin.jang@gmail.com, jhjee@etri.re.kr, yhhan@kut.ac.kr, soohong.park@samsung.com, jscha@etri.re.kr +# RFC5271 || H. Yokota, G. Dommety || yokota@kddilabs.jp, gdommety@cisco.com +# RFC5272 || J. Schaad, M. Myers || jimsch@nwlink.com, mmyers@fastq.com +# RFC5273 || J. Schaad, M. Myers || jimsch@nwlink.com, mmyers@fastq.com +# RFC5274 || J. Schaad, M. Myers || jimsch@nwlink.com, mmyers@fastq.com +# RFC5275 || S. Turner || turners@ieca.com +# RFC5276 || C. Wallace || cwallace@cygnacom.com +# RFC5277 || S. Chisholm, H. Trevino || schishol@nortel.com, htrevino@cisco.com +# RFC5278 || J. Livingood, D. Troshynski || jason_livingood@cable.comcast.com, dtroshynski@acmepacket.com +# RFC5279 || A. Monrad, S. Loreto || atle.monrad@ericsson.com, Salvatore.Loreto@ericsson.com +# RFC5280 || D. Cooper, S. Santesson, S. Farrell, S. Boeyen, R. Housley, W. Polk || david.cooper@nist.gov, stefans@microsoft.com, stephen.farrell@cs.tcd.ie, sharon.boeyen@entrust.com, housley@vigilsec.com, wpolk@nist.gov +# RFC5281 || P. Funk, S. Blake-Wilson || PaulFunk@alum.mit.edu, sblakewilson@nl.safenet-inc.com +# RFC5282 || D. Black, D. McGrew || black_david@emc.com, mcgrew@cisco.com +# RFC5283 || B. Decraene, JL. Le Roux, I. Minei || bruno.decraene@orange-ftgroup.com, jeanlouis.leroux@orange-ftgroup.com, ina@juniper.net +# RFC5284 || G. Swallow, A. Farrel || swallow@cisco.com, adrian@olddog.co.uk +# RFC5285 || D. Singer, H. Desineni || singer@apple.com, hd@qualcomm.com +# RFC5286 || A. Atlas, Ed., A. Zinin, Ed. || alia.atlas@bt.com, alex.zinin@alcatel-lucent.com +# RFC5287 || A. Vainshtein, Y(J). Stein || Alexander.Vainshtein@ecitele.com, yaakov_s@rad.com +# RFC5288 || J. Salowey, A. Choudhury, D. McGrew || jsalowey@cisco.com, abhijitc@cisco.com, mcgrew@cisco.com +# RFC5289 || E. Rescorla || ekr@rtfm.com +# RFC5290 || S. Floyd, M. Allman || floyd@icir.org, mallman@icir.org +# RFC5291 || E. Chen, Y. Rekhter || enkechen@cisco.com, yakov@juniper.net +# RFC5292 || E. Chen, S. Sangli || enkechen@cisco.com, rsrihari@cisco.com +# RFC5293 || J. Degener, P. Guenther || jutta@pobox.com, guenther@sendmail.com +# RFC5294 || P. Savola, J. Lingard || psavola@funet.fi, jchl@arastra.com +# RFC5295 || J. Salowey, L. Dondeti, V. Narayanan, M. Nakhjiri || jsalowey@cisco.com, ldondeti@qualcomm.com, vidyan@qualcomm.com, madjid.nakhjiri@motorola.com +# RFC5296 || V. Narayanan, L. Dondeti || vidyan@qualcomm.com, ldondeti@qualcomm.com +# RFC5297 || D. Harkins || dharkins@arubanetworks.com +# RFC5298 || T. Takeda, Ed., A. Farrel, Ed., Y. Ikejiri, JP. Vasseur || takeda.tomonori@lab.ntt.co.jp, y.ikejiri@ntt.com, adrian@olddog.co.uk, jpv@cisco.com +# RFC5301 || D. McPherson, N. Shen || danny@arbor.net, naiming@cisco.com +# RFC5302 || T. Li, H. Smit, T. Przygienda || tony.li@tony.li, hhw.smit@xs4all.nl, prz@net4u.ch +# RFC5303 || D. Katz, R. Saluja, D. Eastlake 3rd || dkatz@juniper.net, rajesh.saluja@tenetindia.com, d3e3e3@gmail.com +# RFC5304 || T. Li, R. Atkinson || tony.li@tony.li, rja@extremenetworks.com +# RFC5305 || T. Li, H. Smit || tony.li@tony.li, hhwsmit@xs4all.nl +# RFC5306 || M. Shand, L. Ginsberg || mshand@cisco.com, ginsberg@cisco.com +# RFC5307 || K. Kompella, Ed., Y. Rekhter, Ed. || kireeti@juniper.net, yakov@juniper.net +# RFC5308 || C. Hopps || chopps@cisco.com +# RFC5309 || N. Shen, Ed., A. Zinin, Ed. || naiming@cisco.com, alex.zinin@alcatel-lucent.com +# RFC5310 || M. Bhatia, V. Manral, T. Li, R. Atkinson, R. White, M. Fanto || manav@alcatel-lucent.com, vishwas@ipinfusion.com, tony.li@tony.li, rja@extremenetworks.com, riw@cisco.com, mfanto@aegisdatasecurity.com +# RFC5311 || D. McPherson, Ed., L. Ginsberg, S. Previdi, M. Shand || danny@arbor.net, ginsberg@cisco.com, sprevidi@cisco.com, mshand@cisco.com +# RFC5316 || M. Chen, R. Zhang, X. Duan || mach@huawei.com, zhangrenhai@huawei.com, duanxiaodong@chinamobile.com +# RFC5317 || S. Bryant, Ed., L. Andersson, Ed. || stbryant@cisco.com, loa@pi.nu +# RFC5318 || J. Hautakorpi, G. Camarillo || Jani.Hautakorpi@ericsson.com, Gonzalo.Camarillo@ericsson.com +# RFC5320 || F. Templin, Ed. || fltemplin@acm.org +# RFC5321 || J. Klensin || john+smtp@jck.com +# RFC5322 || P. Resnick, Ed. || presnick@qti.qualcomm.com +# RFC5323 || J. Reschke, Ed., S. Reddy, J. Davis, A. Babich || julian.reschke@greenbytes.de, Surendra.Reddy@mitrix.com, jrd3@alum.mit.edu, ababich@us.ibm.com +# RFC5324 || C. DeSanti, F. Maino, K. McCloghrie || cds@cisco.com, fmaino@cisco.com, kzm@cisco.com +# RFC5325 || S. Burleigh, M. Ramadas, S. Farrell || Scott.Burleigh@jpl.nasa.gov, mramadas@gmail.com, stephen.farrell@cs.tcd.ie +# RFC5326 || M. Ramadas, S. Burleigh, S. Farrell || mramadas@gmail.com, Scott.Burleigh@jpl.nasa.gov, stephen.farrell@cs.tcd.ie +# RFC5327 || S. Farrell, M. Ramadas, S. Burleigh || stephen.farrell@cs.tcd.ie, mramadas@gmail.com, Scott.Burleigh@jpl.nasa.gov +# RFC5328 || A. Adolf, P. MacAvock || alexander.adolf@micronas.com, macavock@dvb.org +# RFC5329 || K. Ishiguro, V. Manral, A. Davey, A. Lindem, Ed. || kunihiro@ipinfusion.com, vishwas@ipinfusion.com, Alan.Davey@dataconnection.com, acee@redback.com +# RFC5330 || JP. Vasseur, Ed., M. Meyer, K. Kumaki, A. Bonda || jpv@cisco.com, matthew.meyer@bt.com, ke-kumaki@kddi.com, alberto.tempiabonda@telecomitalia.it +# RFC5331 || R. Aggarwal, Y. Rekhter, E. Rosen || rahul@juniper.net, yakov@juniper.net, erosen@cisco.com +# RFC5332 || T. Eckert, E. Rosen, Ed., R. Aggarwal, Y. Rekhter || eckert@cisco.com, erosen@cisco.com, rahul@juniper.net, yakov@juniper.net +# RFC5333 || R. Mahy, B. Hoeneisen || rohan@ekabal.com, bernie@ietf.hoeneisen.ch +# RFC5334 || I. Goncalves, S. Pfeiffer, C. Montgomery || justivo@gmail.com, silvia@annodex.net, monty@xiph.org +# RFC5335 || A. Yang, Ed. || abelyang@twnic.net.tw +# RFC5336 || J. Yao, Ed., W. Mao, Ed. || yaojk@cnnic.cn, maowei_ietf@cnnic.cn +# RFC5337 || C. Newman, A. Melnikov, Ed. || chris.newman@sun.com, Alexey.Melnikov@isode.com +# RFC5338 || T. Henderson, P. Nikander, M. Komu || thomas.r.henderson@boeing.com, pekka.nikander@nomadiclab.com, miika@iki.fi +# RFC5339 || JL. Le Roux, Ed., D. Papadimitriou, Ed. || jeanlouis.leroux@orange-ftgroup.com, dimitri.papadimitriou@alcatel-lucent.be +# RFC5340 || R. Coltun, D. Ferguson, J. Moy, A. Lindem || none, dennis@juniper.net, jmoy@sycamorenet.com, acee@redback.com +# RFC5341 || C. Jennings, V. Gurbani || fluffy@cisco.com, vkg@alcatel-lucent.com +# RFC5342 || D. Eastlake 3rd || d3e3e3@gmail.com +# RFC5343 || J. Schoenwaelder || j.schoenwaelder@jacobs-university.de +# RFC5344 || A. Houri, E. Aoki, S. Parameswar || avshalom@il.ibm.com, aoki@aol.net, Sriram.Parameswar@microsoft.com +# RFC5345 || J. Schoenwaelder || j.schoenwaelder@jacobs-university.de +# RFC5346 || J. Lim, W. Kim, C. Park, L. Conroy || jhlim@nida.or.kr, wkim@nida.or.kr, ckp@nida.or.kr, lconroy@insensate.co.uk +# RFC5347 || F. Andreasen, D. Hancock || fandreas@cisco.com, d.hancock@cablelabs.com +# RFC5348 || S. Floyd, M. Handley, J. Padhye, J. Widmer || floyd@icir.org, M.Handley@cs.ucl.ac.uk, padhye@microsoft.com, widmer@acm.org +# RFC5349 || L. Zhu, K. Jaganathan, K. Lauter || lzhu@microsoft.com, karthikj@microsoft.com, klauter@microsoft.com +# RFC5350 || J. Manner, A. McDonald || jukka.manner@tkk.fi, andrew.mcdonald@roke.co.uk +# RFC5351 || P. Lei, L. Ong, M. Tuexen, T. Dreibholz || peterlei@cisco.com, Lyong@Ciena.com, tuexen@fh-muenster.de, dreibh@iem.uni-due.de +# RFC5352 || R. Stewart, Q. Xie, M. Stillman, M. Tuexen || randall@lakerest.net, Qiaobing.Xie@gmail.org, maureen.stillman@nokia.com, tuexen@fh-muenster.de +# RFC5353 || Q. Xie, R. Stewart, M. Stillman, M. Tuexen, A. Silverton || Qiaobing.Xie@gmail.org, randall@lakerest.net, maureen.stillman@nokia.com, tuexen@fh-muenster.de, ajs.ietf@gmail.com +# RFC5354 || R. Stewart, Q. Xie, M. Stillman, M. Tuexen || randall@lakerest.net, Qiaobing.Xie@gmail.org, maureen.stillman@nokia.com, tuexen@fh-muenster.de +# RFC5355 || M. Stillman, Ed., R. Gopal, E. Guttman, S. Sengodan, M. Holdrege || maureen.stillman@nokia.com, ram.gopal@nsn.com, Erik.Guttman@sun.com, Senthil.sengodan@nsn.com, Holdrege@gmail.com +# RFC5356 || T. Dreibholz, M. Tuexen || dreibh@iem.uni-due.de, tuexen@fh-muenster.de +# RFC5357 || K. Hedayat, R. Krzanowski, A. Morton, K. Yum, J. Babiarz || khedayat@brixnet.com, roman.krzanowski@verizon.com, acmorton@att.com, kyum@juniper.net, babiarz@nortel.com +# RFC5358 || J. Damas, F. Neves || Joao_Damas@isc.org, fneves@registro.br +# RFC5359 || A. Johnston, Ed., R. Sparks, C. Cunningham, S. Donovan, K. Summers || alan@sipstation.com, RjS@nostrum.com, chrcunni@cisco.com, srd@cisco.com, ksummers@sonusnet.com +# RFC5360 || J. Rosenberg, G. Camarillo, Ed., D. Willis || jdrosen@cisco.com, Gonzalo.Camarillo@ericsson.com, dean.willis@softarmor.com +# RFC5361 || G. Camarillo || Gonzalo.Camarillo@ericsson.com +# RFC5362 || G. Camarillo || Gonzalo.Camarillo@ericsson.com +# RFC5363 || G. Camarillo, A.B. Roach || Gonzalo.Camarillo@ericsson.com, Adam.Roach@tekelec.com +# RFC5364 || M. Garcia-Martin, G. Camarillo || miguel.a.garcia@ericsson.com, Gonzalo.Camarillo@ericsson.com +# RFC5365 || M. Garcia-Martin, G. Camarillo || miguel.a.garcia@ericsson.com, Gonzalo.Camarillo@ericsson.com +# RFC5366 || G. Camarillo, A. Johnston || Gonzalo.Camarillo@ericsson.com, alan@sipstation.com +# RFC5367 || G. Camarillo, A.B. Roach, O. Levin || Gonzalo.Camarillo@ericsson.com, Adam.Roach@tekelec.com, oritl@microsoft.com +# RFC5368 || G. Camarillo, A. Niemi, M. Isomaki, M. Garcia-Martin, H. Khartabil || Gonzalo.Camarillo@ericsson.com, Aki.Niemi@nokia.com, markus.isomaki@nokia.com, miguel.a.garcia@ericsson.com, hisham.khartabil@gmail.com +# RFC5369 || G. Camarillo || Gonzalo.Camarillo@ericsson.com +# RFC5370 || G. Camarillo || Gonzalo.Camarillo@ericsson.com +# RFC5371 || S. Futemma, E. Itakura, A. Leung || satosi-f@sm.sony.co.jp, itakura@sm.sony.co.jp, andrew@ualberta.net +# RFC5372 || A. Leung, S. Futemma, E. Itakura || andrew@ualberta.net, satosi-f@sm.sony.co.jp, itakura@sm.sony.co.jp +# RFC5373 || D. Willis, Ed., A. Allen || dean.willis@softarmor.com, aallen@rim.com +# RFC5374 || B. Weis, G. Gross, D. Ignjatic || bew@cisco.com, gmgross@securemulticast.net, dignjatic@polycom.com +# RFC5375 || G. Van de Velde, C. Popoviciu, T. Chown, O. Bonness, C. Hahn || gunter@cisco.com, cpopovic@cisco.com, tjc@ecs.soton.ac.uk, Olaf.Bonness@t-systems.com, HahnC@t-systems.com +# RFC5376 || N. Bitar, R. Zhang, K. Kumaki || nabil.n.bitar@verizon.com, ke-kumaki@kddi.com, Raymond.zhang@bt.com +# RFC5377 || J. Halpern, Ed. || jmh@joelhalpern.com +# RFC5378 || S. Bradner, Ed., J. Contreras, Ed. || sob@harvard.edu, jorge.contreras@wilmerhale.com +# RFC5379 || M. Munakata, S. Schubert, T. Ohba || munakata.mayumi@lab.ntt.co.jp, shida@ntt-at.com, ohba.takumi@lab.ntt.co.jp +# RFC5380 || H. Soliman, C. Castelluccia, K. ElMalki, L. Bellier || hesham@elevatemobile.com, claude.castelluccia@inria.fr, karim@athonet.com, ludovic.bellier@inria.fr +# RFC5381 || T. Iijima, Y. Atarashi, H. Kimura, M. Kitani, H. Okita || tomoyuki.iijima@alaxala.com, atarashi@alaxala.net, h-kimura@alaxala.net, makoto.kitani@alaxala.com, hideki.okita.pf@hitachi.com +# RFC5382 || S. Guha, Ed., K. Biswas, B. Ford, S. Sivakumar, P. Srisuresh || saikat@cs.cornell.edu, kbiswas@cisco.com, baford@mpi-sws.org, ssenthil@cisco.com, srisuresh@yahoo.com +# RFC5383 || R. Gellens || randy@qualcomm.com +# RFC5384 || A. Boers, I. Wijnands, E. Rosen || aboers@cisco.com, ice@cisco.com, erosen@cisco.com +# RFC5385 || J. Touch || touch@isi.edu +# RFC5386 || N. Williams, M. Richardson || Nicolas.Williams@sun.com, mcr@sandelman.ottawa.on.ca +# RFC5387 || J. Touch, D. Black, Y. Wang || touch@isi.edu, black_david@emc.com, yu-shun.wang@microsoft.com +# RFC5388 || S. Niccolini, S. Tartarelli, J. Quittek, T. Dietz, M. Swany || saverio.niccolini@nw.neclab.eu, sandra.tartarelli@nw.neclab.eu, quittek@nw.neclab.eu, thomas.dietz@nw.neclab.eu, swany@UDel.Edu +# RFC5389 || J. Rosenberg, R. Mahy, P. Matthews, D. Wing || jdrosen@cisco.com, rohan@ekabal.com, philip_matthews@magma.ca, dwing-ietf@fuggles.com +# RFC5390 || J. Rosenberg || jdrosen@cisco.com +# RFC5391 || A. Sollaud || aurelien.sollaud@orange-ftgroup.com +# RFC5392 || M. Chen, R. Zhang, X. Duan || mach@huawei.com, zhangrenhai@huawei.com, duanxiaodong@chinamobile.com +# RFC5393 || R. Sparks, Ed., S. Lawrence, A. Hawrylyshen, B. Campen || RjS@nostrum.com, scott.lawrence@nortel.com, alan.ietf@polyphase.ca, bcampen@estacado.net +# RFC5394 || I. Bryskin, D. Papadimitriou, L. Berger, J. Ash || ibryskin@advaoptical.com, dimitri.papadimitriou@alcatel.be, lberger@labn.net, gash5107@yahoo.com +# RFC5395 || D. Eastlake 3rd || d3e3e3@gmail.com +# RFC5396 || G. Huston, G. Michaelson || gih@apnic.net, ggm@apnic.net +# RFC5397 || W. Sanchez, C. Daboo || wsanchez@wsanchez.net, cyrus@daboo.name +# RFC5398 || G. Huston || gih@apnic.net +# RFC5401 || B. Adamson, C. Bormann, M. Handley, J. Macker || adamson@itd.nrl.navy.mil, cabo@tzi.org, M.Handley@cs.ucl.ac.uk, macker@itd.nrl.navy.mil +# RFC5402 || T. Harding, Ed. || tharding@us.axway.com +# RFC5403 || M. Eisler || mike@eisler.com +# RFC5404 || M. Westerlund, I. Johansson || magnus.westerlund@ericsson.com, ingemar.s.johansson@ericsson.com +# RFC5405 || L. Eggert, G. Fairhurst || lars.eggert@nokia.com, gorry@erg.abdn.ac.uk +# RFC5406 || S. Bellovin || bellovin@acm.org +# RFC5407 || M. Hasebe, J. Koshiko, Y. Suzuki, T. Yoshikawa, P. Kyzivat || hasebe.miki@east.ntt.co.jp, j.koshiko@east.ntt.co.jp, suzuki.yasushi@lab.ntt.co.jp, tomoyuki.yoshikawa@east.ntt.co.jp, pkyzivat@cisco.com +# RFC5408 || G. Appenzeller, L. Martin, M. Schertler || appenz@cs.stanford.edu, martin@voltage.com, mschertler@us.axway.com +# RFC5409 || L. Martin, M. Schertler || martin@voltage.com, mschertler@us.axway.com +# RFC5410 || A. Jerichow, Ed., L. Piron || anja.jerichow@nsn.com, laurent.piron@nagravision.com +# RFC5411 || J. Rosenberg || jdrosen@cisco.com +# RFC5412 || P. Calhoun, R. Suri, N. Cam-Winget, M. Williams, S. Hares, B. O'Hara, S. Kelly || pcalhoun@cisco.com, rsuri@cisco.com, ncamwing@cisco.com, gwhiz@gwhiz.com, shares@ndzh.com, bob.ohara@computer.org, scott@hyperthought.com +# RFC5413 || P. Narasimhan, D. Harkins, S. Ponnuswamy || partha@arubanetworks.com, dharkins@arubanetworks.com, subbu@arubanetworks.com +# RFC5414 || S. Iino, S. Govindan, M. Sugiura, H. Cheng || iino.satoshi@jp.panasonic.com, saravanan.govindan@sg.panasonic.com, sugiura.mikihito@jp.panasonic.com, hong.cheng@sg.panasonic.com +# RFC5415 || P. Calhoun, Ed., M. Montemurro, Ed., D. Stanley, Ed. || pcalhoun@cisco.com, mmontemurro@rim.com, dstanley@arubanetworks.com +# RFC5416 || P. Calhoun, Ed., M. Montemurro, Ed., D. Stanley, Ed. || pcalhoun@cisco.com, mmontemurro@rim.com, dstanley@arubanetworks.com +# RFC5417 || P. Calhoun || pcalhoun@cisco.com +# RFC5418 || S. Kelly, T. Clancy || scott@hyperthought.com, clancy@LTSnet.net +# RFC5419 || B. Patil, G. Dommety || basavaraj.patil@nokia.com, gdommety@cisco.com +# RFC5420 || A. Farrel, Ed., D. Papadimitriou, JP. Vasseur, A. Ayyangarps || adrian@olddog.co.uk, dimitri.papadimitriou@alcatel.be, jpv@cisco.com, arthi@juniper.net +# RFC5421 || N. Cam-Winget, H. Zhou || ncamwing@cisco.com, hzhou@cisco.com +# RFC5422 || N. Cam-Winget, D. McGrew, J. Salowey, H. Zhou || ncamwing@cisco.com, mcgrew@cisco.com, jsalowey@cisco.com, hzhou@cisco.com +# RFC5423 || R. Gellens, C. Newman || rg+ietf@qualcomm.com, chris.newman@sun.com +# RFC5424 || R. Gerhards || rgerhards@adiscon.com +# RFC5425 || F. Miao, Ed., Y. Ma, Ed., J. Salowey, Ed. || miaofy@huawei.com, myz@huawei.com, jsalowey@cisco.com +# RFC5426 || A. Okmianski || aokmians@cisco.com +# RFC5427 || G. Keeni || glenn@cysols.com +# RFC5428 || S. Channabasappa, W. De Ketelaere, E. Nechamkin || Sumanth@cablelabs.com, deketelaere@tComLabs.com, enechamkin@broadcom.com +# RFC5429 || A. Stone, Ed. || aaron@serendipity.palo-alto.ca.us +# RFC5430 || M. Salter, E. Rescorla, R. Housley || msalter@restarea.ncsc.mil, ekr@rtfm.com, housley@vigilsec.com +# RFC5431 || D. Sun || dongsun@alcatel-lucent.com +# RFC5432 || J. Polk, S. Dhesikan, G. Camarillo || jmpolk@cisco.com, sdhesika@cisco.com, Gonzalo.Camarillo@ericsson.com +# RFC5433 || T. Clancy, H. Tschofenig || clancy@ltsnet.net, Hannes.Tschofenig@gmx.net +# RFC5434 || T. Narten || narten@us.ibm.com +# RFC5435 || A. Melnikov, Ed., B. Leiba, Ed., W. Segmuller, T. Martin || Alexey.Melnikov@isode.com, leiba@watson.ibm.com, werewolf@us.ibm.com, timmartin@alumni.cmu.edu +# RFC5436 || B. Leiba, M. Haardt || leiba@watson.ibm.com, michael.haardt@freenet.ag +# RFC5437 || P. Saint-Andre, A. Melnikov || ietf@stpeter.im, Alexey.Melnikov@isode.com +# RFC5438 || E. Burger, H. Khartabil || eburger@standardstrack.com, hisham.khartabil@gmail.com +# RFC5439 || S. Yasukawa, A. Farrel, O. Komolafe || s.yasukawa@hco.ntt.co.jp, adrian@olddog.co.uk, femi@cisco.com +# RFC5440 || JP. Vasseur, Ed., JL. Le Roux, Ed. || jpv@cisco.com, jeanlouis.leroux@orange-ftgroup.com +# RFC5441 || JP. Vasseur, Ed., R. Zhang, N. Bitar, JL. Le Roux || jpv@cisco.com, raymond.zhang@bt.com, nabil.n.bitar@verizon.com, jeanlouis.leroux@orange-ftgroup.com +# RFC5442 || E. Burger, G. Parsons || eburger@standardstrack.com, gparsons@nortel.com +# RFC5443 || M. Jork, A. Atlas, L. Fang || Markus.Jork@genband.com, alia.atlas@bt.com, lufang@cisco.com +# RFC5444 || T. Clausen, C. Dearlove, J. Dean, C. Adjih || T.Clausen@computer.org, chris.dearlove@baesystems.com, jdean@itd.nrl.navy.mil, Cedric.Adjih@inria.fr +# RFC5445 || M. Watson || mark@digitalfountain.com +# RFC5446 || J. Korhonen, U. Nilsson || jouni.nospam@gmail.com, ulf.s.nilsson@teliasonera.com +# RFC5447 || J. Korhonen, Ed., J. Bournelle, H. Tschofenig, C. Perkins, K. Chowdhury || jouni.nospam@gmail.com, julien.bournelle@orange-ftgroup.com, Hannes.Tschofenig@nsn.com, charliep@wichorus.com, kchowdhury@starentnetworks.com +# RFC5448 || J. Arkko, V. Lehtovirta, P. Eronen || jari.arkko@piuha.net, vesa.lehtovirta@ericsson.com, pe@iki.fi +# RFC5449 || E. Baccelli, P. Jacquet, D. Nguyen, T. Clausen || Emmanuel.Baccelli@inria.fr, Philippe.Jacquet@inria.fr, dang.nguyen@crc.ca, T.Clausen@computer.org +# RFC5450 || D. Singer, H. Desineni || singer@apple.com, hd@qualcomm.com +# RFC5451 || M. Kucherawy || msk+ietf@sendmail.com +# RFC5452 || A. Hubert, R. van Mook || bert.hubert@netherlabs.nl, remco@eu.equinix.com +# RFC5453 || S. Krishnan || suresh.krishnan@ericsson.com +# RFC5454 || G. Tsirtsis, V. Park, H. Soliman || tsirtsis@googlemail.com, vpark@qualcomm.com, hesham@elevatemobile.com +# RFC5455 || S. Sivabalan, Ed., J. Parker, S. Boutros, K. Kumaki || msiva@cisco.com, jdparker@cisco.com, sboutros@cisco.com, ke-kumaki@kddi.com +# RFC5456 || M. Spencer, B. Capouch, E. Guy, Ed., F. Miller, K. Shumard || markster@digium.com, brianc@saintjoe.edu, edguy@emcsw.com, mail@frankwmiller.net, kshumard@gmail.com +# RFC5457 || E. Guy, Ed. || edguy@emcsw.com +# RFC5458 || H. Cruickshank, P. Pillai, M. Noisternig, S. Iyengar || h.cruickshank@surrey.ac.uk, p.pillai@bradford.ac.uk, mnoist@cosy.sbg.ac.at, sunil.iyengar@logica.com +# RFC5459 || A. Sollaud || aurelien.sollaud@orange-ftgroup.com +# RFC5460 || M. Stapp || mjs@cisco.com +# RFC5461 || F. Gont || fernando@gont.com.ar +# RFC5462 || L. Andersson, R. Asati || loa@pi.nu, rajiva@cisco.com +# RFC5463 || N. Freed || ned.freed@mrochek.com +# RFC5464 || C. Daboo || cyrus@daboo.name +# RFC5465 || A. Gulbrandsen, C. King, A. Melnikov || arnt@oryx.com, Curtis.King@isode.com, Alexey.Melnikov@isode.com +# RFC5466 || A. Melnikov, C. King || Alexey.Melnikov@isode.com, Curtis.King@isode.com +# RFC5467 || L. Berger, A. Takacs, D. Caviglia, D. Fedyk, J. Meuric || lberger@labn.net, attila.takacs@ericsson.com, diego.caviglia@ericsson.com, dwfedyk@nortel.com, julien.meuric@orange-ftgroup.com +# RFC5468 || S. Dasgupta, J. de Oliveira, JP. Vasseur || sukrit@ece.drexel.edu, jau@ece.drexel.edu, jpv@cisco.com +# RFC5469 || P. Eronen, Ed. || pe@iki.fi +# RFC5470 || G. Sadasivan, N. Brownlee, B. Claise, J. Quittek || gsadasiv@rohati.com, n.brownlee@auckland.ac.nz, bclaise@cisco.com, quittek@nw.neclab.eu +# RFC5471 || C. Schmoll, P. Aitken, B. Claise || carsten.schmoll@fokus.fraunhofer.de, paitken@cisco.com, bclaise@cisco.com +# RFC5472 || T. Zseby, E. Boschi, N. Brownlee, B. Claise || tanja.zseby@fokus.fraunhofer.de, elisa.boschi@hitachi-eu.com, nevil@caida.org, bclaise@cisco.com +# RFC5473 || E. Boschi, L. Mark, B. Claise || elisa.boschi@hitachi-eu.com, lutz.mark@ifam.fraunhofer.de, bclaise@cisco.com +# RFC5474 || N. Duffield, Ed., D. Chiou, B. Claise, A. Greenberg, M. Grossglauser, J. Rexford || duffield@research.att.com, Derek@ece.utexas.edu, bclaise@cisco.com, albert@microsoft.com, matthias.grossglauser@epfl.ch, jrex@cs.princeton.edu +# RFC5475 || T. Zseby, M. Molina, N. Duffield, S. Niccolini, F. Raspall || tanja.zseby@fokus.fraunhofer.de, maurizio.molina@dante.org.uk, duffield@research.att.com, saverio.niccolini@netlab.nec.de, fredi@entel.upc.es +# RFC5476 || B. Claise, Ed., A. Johnson, J. Quittek || bclaise@cisco.com, andrjohn@cisco.com, quittek@nw.neclab.eu +# RFC5477 || T. Dietz, B. Claise, P. Aitken, F. Dressler, G. Carle || Thomas.Dietz@nw.neclab.eu, bclaise@cisco.com, paitken@cisco.com, dressler@informatik.uni-erlangen.de, carle@informatik.uni-tuebingen.de +# RFC5478 || J. Polk || jmpolk@cisco.com +# RFC5479 || D. Wing, Ed., S. Fries, H. Tschofenig, F. Audet || dwing-ietf@fuggles.com, steffen.fries@siemens.com, Hannes.Tschofenig@nsn.com, audet@nortel.com +# RFC5480 || S. Turner, D. Brown, K. Yiu, R. Housley, T. Polk || turners@ieca.com, kelviny@microsoft.com, dbrown@certicom.com, housley@vigilsec.com, wpolk@nist.gov +# RFC5481 || A. Morton, B. Claise || acmorton@att.com, bclaise@cisco.com +# RFC5482 || L. Eggert, F. Gont || lars.eggert@nokia.com, fernando@gont.com.ar +# RFC5483 || L. Conroy, K. Fujiwara || lconroy@insensate.co.uk, fujiwara@jprs.co.jp +# RFC5484 || D. Singer || singer@apple.com +# RFC5485 || R. Housley || housley@vigilsec.com +# RFC5486 || D. Malas, Ed., D. Meyer, Ed. || d.malas@cablelabs.com, dmm@1-4-5.net +# RFC5487 || M. Badra || badra@isima.fr +# RFC5488 || S. Gundavelli, G. Keeni, K. Koide, K. Nagami || sgundave@cisco.com, glenn@cysols.com, ka-koide@kddi.com, nagami@inetcore.com +# RFC5489 || M. Badra, I. Hajjeh || badra@isima.fr, ibrahim.hajjeh@ineovation.fr +# RFC5490 || A. Melnikov || Alexey.Melnikov@isode.com +# RFC5491 || J. Winterbottom, M. Thomson, H. Tschofenig || james.winterbottom@andrew.com, martin.thomson@andrew.com, Hannes.Tschofenig@gmx.net +# RFC5492 || J. Scudder, R. Chandra || jgs@juniper.net, rchandra@sonoasystems.com +# RFC5493 || D. Caviglia, D. Bramanti, D. Li, D. McDysan || diego.caviglia@ericsson.com, dino.bramanti@ericsson.com, danli@huawei.com, dave.mcdysan@verizon.com +# RFC5494 || J. Arkko, C. Pignataro || jari.arkko@piuha.net, cpignata@cisco.com +# RFC5495 || D. Li, J. Gao, A. Satyanarayana, S. Bardalai || danli@huawei.com, gjhhit@huawei.com, asatyana@cisco.com, snigdho.bardalai@us.fujitsu.com +# RFC5496 || IJ. Wijnands, A. Boers, E. Rosen || ice@cisco.com, aboers@cisco.com, erosen@cisco.com +# RFC5497 || T. Clausen, C. Dearlove || T.Clausen@computer.org, chris.dearlove@baesystems.com +# RFC5498 || I. Chakeres || ian.chakeres@gmail.com +# RFC5501 || Y. Kamite, Ed., Y. Wada, Y. Serbest, T. Morin, L. Fang || y.kamite@ntt.com, wada.yuichiro@lab.ntt.co.jp, yetik_serbest@labs.att.com, thomas.morin@francetelecom.com, lufang@cisco.com +# RFC5502 || J. van Elburg || HansErik.van.Elburg@ericsson.com +# RFC5503 || F. Andreasen, B. McKibben, B. Marshall || fandreas@cisco.com, B.McKibben@cablelabs.com, wtm@research.att.com +# RFC5504 || K. Fujiwara, Ed., Y. Yoneya, Ed. || fujiwara@jprs.co.jp, yone@jprs.co.jp +# RFC5505 || B. Aboba, D. Thaler, L. Andersson, S. Cheshire || bernarda@microsoft.com, dthaler@microsoft.com, loa.andersson@ericsson.com, cheshire@apple.com +# RFC5506 || I. Johansson, M. Westerlund || ingemar.s.johansson@ericsson.com, magnus.westerlund@ericsson.com +# RFC5507 || IAB, P. Faltstrom, Ed., R. Austein, Ed., P. Koch, Ed. || iab@iab.org, paf@cisco.com, sra@isc.org, pk@denic.de +# RFC5508 || P. Srisuresh, B. Ford, S. Sivakumar, S. Guha || srisuresh@yahoo.com, baford@mpi-sws.org, ssenthil@cisco.com, saikat@cs.cornell.edu +# RFC5509 || S. Loreto || Salvatore.Loreto@ericsson.com +# RFC5510 || J. Lacan, V. Roca, J. Peltotalo, S. Peltotalo || jerome.lacan@isae.fr, vincent.roca@inria.fr, jani.peltotalo@tut.fi, sami.peltotalo@tut.fi +# RFC5511 || A. Farrel || adrian@olddog.co.uk +# RFC5512 || P. Mohapatra, E. Rosen || pmohapat@cisco.com, erosen@cisco.com +# RFC5513 || A. Farrel || adrian@olddog.co.uk +# RFC5514 || E. Vyncke || evyncke@cisco.com +# RFC5515 || V. Mammoliti, C. Pignataro, P. Arberg, J. Gibbons, P. Howard || vince@cisco.com, cpignata@cisco.com, parberg@redback.com, jgibbons@juniper.net, howsoft@mindspring.com +# RFC5516 || M. Jones, L. Morand || mark.jones@bridgewatersystems.com, lionel.morand@orange-ftgroup.com +# RFC5517 || S. HomChaudhuri, M. Foschiano || sanjibhc@gmail.com, foschia@cisco.com +# RFC5518 || P. Hoffman, J. Levine, A. Hathcock || paul.hoffman@domain-assurance.org, john.levine@domain-assurance.org, arvel.hathcock@altn.com +# RFC5519 || J. Chesterfield, B. Haberman, Ed. || julian.chesterfield@cl.cam.ac.uk, brian@innovationslab.net +# RFC5520 || R. Bradford, Ed., JP. Vasseur, A. Farrel || rbradfor@cisco.com, jpv@cisco.com, adrian@olddog.co.uk +# RFC5521 || E. Oki, T. Takeda, A. Farrel || oki@ice.uec.ac.jp, takeda.tomonori@lab.ntt.co.jp, adrian@olddog.co.uk +# RFC5522 || W. Eddy, W. Ivancic, T. Davis || weddy@grc.nasa.gov, William.D.Ivancic@grc.nasa.gov, Terry.L.Davis@boeing.com +# RFC5523 || L. Berger || lberger@labn.net +# RFC5524 || D. Cridland || dave.cridland@isode.com +# RFC5525 || T. Dreibholz, J. Mulik || dreibh@iem.uni-due.de, jaiwant@mulik.com +# RFC5526 || J. Livingood, P. Pfautz, R. Stastny || jason_livingood@cable.comcast.com, ppfautz@att.com, richard.stastny@gmail.com +# RFC5527 || M. Haberler, O. Lendl, R. Stastny || ietf@mah.priv.at, otmar.lendl@enum.at, richardstastny@gmail.com +# RFC5528 || A. Kato, M. Kanda, S. Kanno || akato@po.ntts.co.jp, kanda.masayuki@lab.ntt.co.jp, kanno-s@po.ntts.co.jp +# RFC5529 || A. Kato, M. Kanda, S. Kanno || akato@po.ntts.co.jp, kanda.masayuki@lab.ntt.co.jp, kanno-s@po.ntts.co.jp +# RFC5530 || A. Gulbrandsen || arnt@oryx.com +# RFC5531 || R. Thurlow || robert.thurlow@sun.com +# RFC5532 || T. Talpey, C. Juszczak || tmtalpey@gmail.com, chetnh@earthlink.net +# RFC5533 || E. Nordmark, M. Bagnulo || erik.nordmark@sun.com, marcelo@it.uc3m.es +# RFC5534 || J. Arkko, I. van Beijnum || jari.arkko@ericsson.com, iljitsch@muada.com +# RFC5535 || M. Bagnulo || marcelo@it.uc3m.es +# RFC5536 || K. Murchison, Ed., C. Lindsey, D. Kohn || murch@andrew.cmu.edu, chl@clerew.man.ac.uk, dan@dankohn.com +# RFC5537 || R. Allbery, Ed., C. Lindsey || rra@stanford.edu, chl@clerew.man.ac.uk +# RFC5538 || F. Ellermann || hmdmhdfmhdjmzdtjmzdtzktdkztdjz@gmail.com +# RFC5539 || M. Badra || badra@isima.fr +# RFC5540 || RFC Editor || rfc-editor@rfc-editor.org +# RFC5541 || JL. Le Roux, JP. Vasseur, Y. Lee || jeanlouis.leroux@orange-ftgroup.com, jpv@cisco.com, ylee@huawei.com +# RFC5542 || T. Nadeau, Ed., D. Zelig, Ed., O. Nicklass, Ed. || tom.nadeau@bt.com, davidz@oversi.com, orlyn@radvision.com +# RFC5543 || H. Ould-Brahim, D. Fedyk, Y. Rekhter || hbrahim@nortel.com, donald.fedyk@alcatel-lucent.com, yakov@juniper.com +# RFC5544 || A. Santoni || adriano.santoni@actalis.it +# RFC5545 || B. Desruisseaux, Ed. || bernard.desruisseaux@oracle.com +# RFC5546 || C. Daboo, Ed. || cyrus@daboo.name +# RFC5547 || M. Garcia-Martin, M. Isomaki, G. Camarillo, S. Loreto, P. Kyzivat || miguel.a.garcia@ericsson.com, markus.isomaki@nokia.com, Gonzalo.Camarillo@ericsson.com, Salvatore.Loreto@ericsson.com, pkyzivat@cisco.com +# RFC5548 || M. Dohler, Ed., T. Watteyne, Ed., T. Winter, Ed., D. Barthel, Ed. || mischa.dohler@cttc.es, watteyne@eecs.berkeley.edu, wintert@acm.org, Dominique.Barthel@orange-ftgroup.com +# RFC5549 || F. Le Faucheur, E. Rosen || flefauch@cisco.com, erosen@cisco.com +# RFC5550 || D. Cridland, Ed., A. Melnikov, Ed., S. Maes, Ed. || dave.cridland@isode.com, Alexey.Melnikov@isode.com, stephane.maes@oracle.com +# RFC5551 || R. Gellens, Ed. || rg+ietf@qualcomm.com +# RFC5552 || D. Burke, M. Scott || daveburke@google.com, Mark.Scott@genesyslab.com +# RFC5553 || A. Farrel, Ed., R. Bradford, JP. Vasseur || adrian@olddog.co.uk, rbradfor@cisco.com, jpv@cisco.com +# RFC5554 || N. Williams || Nicolas.Williams@sun.com +# RFC5555 || H. Soliman, Ed. || hesham@elevatemobile.com +# RFC5556 || J. Touch, R. Perlman || touch@isi.edu, Radia.Perlman@sun.com +# RFC5557 || Y. Lee, JL. Le Roux, D. King, E. Oki || ylee@huawei.com, jeanlouis.leroux@orange-ftgroup.com, daniel@olddog.co.uk, oki@ice.uec.ac.jp +# RFC5558 || F. Templin, Ed. || fltemplin@acm.org +# RFC5559 || P. Eardley, Ed. || philip.eardley@bt.com +# RFC5560 || H. Uijterwaal || henk@ripe.net +# RFC5561 || B. Thomas, K. Raza, S. Aggarwal, R. Aggarwal, JL. Le Roux || bobthomas@alum.mit.edu, skraza@cisco.com, shivani@juniper.net, rahul@juniper.net, jeanlouis.leroux@orange-ftgroup.com +# RFC5562 || A. Kuzmanovic, A. Mondal, S. Floyd, K. Ramakrishnan || akuzma@northwestern.edu, a-mondal@northwestern.edu, floyd@icir.org, kkrama@research.att.com +# RFC5563 || K. Leung, G. Dommety, P. Yegani, K. Chowdhury || kleung@cisco.com, gdommety@cisco.com, pyegani@cisco.com, kchowdhury@starentnetworks.com +# RFC5564 || A. El-Sherbiny, M. Farah, I. Oueichek, A. Al-Zoman || El-sherbiny@un.org, farah14@un.org, oueichek@scs-net.org, azoman@citc.gov.sa +# RFC5565 || J. Wu, Y. Cui, C. Metz, E. Rosen || jianping@cernet.edu.cn, yong@csnet1.cs.tsinghua.edu.cn, chmetz@cisco.com, erosen@cisco.com +# RFC5566 || L. Berger, R. White, E. Rosen || lberger@labn.net, riw@cisco.com, erosen@cisco.com +# RFC5567 || T. Melanchuk, Ed. || tim.melanchuk@gmail.com +# RFC5568 || R. Koodli, Ed. || rkoodli@starentnetworks.com +# RFC5569 || R. Despres || remi.despres@free.fr +# RFC5570 || M. StJohns, R. Atkinson, G. Thomas || mstjohns@comcast.net, rja@extremenetworks.com, none +# RFC5571 || B. Storer, C. Pignataro, Ed., M. Dos Santos, B. Stevant, Ed., L. Toutain, J. Tremblay || bstorer@cisco.com, cpignata@cisco.com, mariados@cisco.com, bruno.stevant@telecom-bretagne.eu, laurent.toutain@telecom-bretagne.eu, jf@jftremblay.com +# RFC5572 || M. Blanchet, F. Parent || Marc.Blanchet@viagenie.ca, Florent.Parent@beon.ca +# RFC5573 || M. Thomson || martin.thomson@andrew.com +# RFC5574 || G. Herlein, J. Valin, A. Heggestad, A. Moizard || gherlein@herlein.com, jean-marc.valin@usherbrooke.ca, aeh@db.org, jack@atosc.org +# RFC5575 || P. Marques, N. Sheth, R. Raszuk, B. Greene, J. Mauch, D. McPherson || roque@cisco.com, nsheth@juniper.net, raszuk@cisco.com, bgreene@juniper.net, jmauch@us.ntt.net, danny@arbor.net +# RFC5576 || J. Lennox, J. Ott, T. Schierl || jonathan@vidyo.com, jo@acm.org, ts@thomas-schierl.de +# RFC5577 || P. Luthi, R. Even || patrick.luthi@tandberg.no, ron.even.tlv@gmail.com +# RFC5578 || B. Berry, Ed., S. Ratliff, E. Paradise, T. Kaiser, M. Adams || bberry@cisco.com, sratliff@cisco.com, pdice@cisco.com, timothy.kaiser@harris.com, Michael.D.Adams@L-3com.com +# RFC5579 || F. Templin, Ed. || fltemplin@acm.org +# RFC5580 || H. Tschofenig, Ed., F. Adrangi, M. Jones, A. Lior, B. Aboba || Hannes.Tschofenig@gmx.net, farid.adrangi@intel.com, mark.jones@bridgewatersystems.com, avi@bridgewatersystems.com, bernarda@microsoft.com +# RFC5581 || D. Shaw || dshaw@jabberwocky.com +# RFC5582 || H. Schulzrinne || hgs+ecrit@cs.columbia.edu +# RFC5583 || T. Schierl, S. Wenger || ts@thomas-schierl.de, stewe@stewe.org +# RFC5584 || M. Hatanaka, J. Matsumoto || actech@jp.sony.com, actech@jp.sony.com +# RFC5585 || T. Hansen, D. Crocker, P. Hallam-Baker || tony+dkimov@maillennium.att.com, dcrocker@bbiw.net, phillip@hallambaker.com +# RFC5586 || M. Bocci, Ed., M. Vigoureux, Ed., S. Bryant, Ed. || matthew.bocci@alcatel-lucent.com, martin.vigoureux@alcatel-lucent.com, stbryant@cisco.com +# RFC5587 || N. Williams || Nicolas.Williams@sun.com +# RFC5588 || N. Williams || Nicolas.Williams@sun.com +# RFC5589 || R. Sparks, A. Johnston, Ed., D. Petrie || RjS@nostrum.com, alan@sipstation.com, dan.ietf@SIPez.com +# RFC5590 || D. Harrington, J. Schoenwaelder || ietfdbh@comcast.net, j.schoenwaelder@jacobs-university.de +# RFC5591 || D. Harrington, W. Hardaker || ietfdbh@comcast.net, ietf@hardakers.net +# RFC5592 || D. Harrington, J. Salowey, W. Hardaker || ietfdbh@comcast.net, jsalowey@cisco.com, ietf@hardakers.net +# RFC5593 || N. Cook || neil.cook@noware.co.uk +# RFC5594 || J. Peterson, A. Cooper || jon.peterson@neustar.biz, acooper@cdt.org +# RFC5595 || G. Fairhurst || gorry@erg.abdn.ac.uk +# RFC5596 || G. Fairhurst || gorry@erg.abdn.ac.uk +# RFC5597 || R. Denis-Courmont || rem@videolan.org +# RFC5598 || D. Crocker || dcrocker@bbiw.net +# RFC5601 || T. Nadeau, Ed., D. Zelig, Ed. || thomas.nadeau@bt.com, davidz@oversi.com +# RFC5602 || D. Zelig, Ed., T. Nadeau, Ed. || davidz@oversi.com, tom.nadeau@bt.com +# RFC5603 || D. Zelig, Ed., T. Nadeau, Ed. || davidz@oversi.com, tom.nadeau@bt.com +# RFC5604 || O. Nicklass || orlyn@radvision.com +# RFC5605 || O. Nicklass, T. Nadeau || orlyn@radvision.com, tom.nadeau@bt.com +# RFC5606 || J. Peterson, T. Hardie, J. Morris || jon.peterson@neustar.biz, hardie@qualcomm.com, jmorris@cdt.org +# RFC5607 || D. Nelson, G. Weber || dnelson@elbrysnetworks.com, gdweber@gmail.com +# RFC5608 || K. Narayan, D. Nelson || kaushik_narayan@yahoo.com, dnelson@elbrysnetworks.com +# RFC5609 || V. Fajardo, Ed., Y. Ohba, R. Marin-Lopez || vfajardo@research.telcordia.com, yoshihiro.ohba@toshiba.co.jp, rafa@um.es +# RFC5610 || E. Boschi, B. Trammell, L. Mark, T. Zseby || elisa.boschi@hitachi-eu.com, brian.trammell@hitachi-eu.com, lutz.mark@ifam.fraunhofer.de, tanja.zseby@fokus.fraunhofer.de +# RFC5611 || A. Vainshtein, S. Galtzur || Alexander.Vainshtein@ecitele.com, sharon.galtzur@rebellion.co.uk +# RFC5612 || P. Eronen, D. Harrington || pe@iki.fi, dharrington@huawei.com +# RFC5613 || A. Zinin, A. Roy, L. Nguyen, B. Friedman, D. Yeung || alex.zinin@alcatel-lucent.com, akr@cisco.com, lhnguyen@cisco.com, barryf@google.com, myeung@cisco.com +# RFC5614 || R. Ogier, P. Spagnolo || rich.ogier@earthlink.net, phillipspagnolo@gmail.com +# RFC5615 || C. Groves, Y. Lin || Christian.Groves@nteczone.com, linyangbo@huawei.com +# RFC5616 || N. Cook || neil.cook@noware.co.uk +# RFC5617 || E. Allman, J. Fenton, M. Delany, J. Levine || eric+dkim@sendmail.org, fenton@bluepopcorn.net, markd+dkim@yahoo-inc.com, standards@taugh.com +# RFC5618 || A. Morton, K. Hedayat || acmorton@att.com, kaynam.hedayat@exfo.com +# RFC5619 || S. Yamamoto, C. Williams, H. Yokota, F. Parent || shu@nict.go.jp, carlw@mcsr-labs.org, yokota@kddilabs.jp, Florent.Parent@beon.ca +# RFC5620 || O. Kolkman, Ed., IAB || olaf@nlnetlabs.nl, iab@iab.org +# RFC5621 || G. Camarillo || Gonzalo.Camarillo@ericsson.com +# RFC5622 || S. Floyd, E. Kohler || floyd@icir.org, kohler@cs.ucla.edu +# RFC5623 || E. Oki, T. Takeda, JL. Le Roux, A. Farrel || oki@ice.uec.ac.jp, takeda.tomonori@lab.ntt.co.jp, jeanlouis.leroux@orange-ftgroup.com, adrian@olddog.co.uk +# RFC5624 || J. Korhonen, Ed., H. Tschofenig, E. Davies || jouni.korhonen@nsn.com, Hannes.Tschofenig@gmx.net, elwynd@dial.pipex.com +# RFC5625 || R. Bellis || ray.bellis@nominet.org.uk +# RFC5626 || C. Jennings, Ed., R. Mahy, Ed., F. Audet, Ed. || fluffy@cisco.com, rohan@ekabal.com, francois.audet@skypelabs.com +# RFC5627 || J. Rosenberg || jdrosen@cisco.com +# RFC5628 || P. Kyzivat || pkyzivat@cisco.com +# RFC5629 || J. Rosenberg || jdrosen@cisco.com +# RFC5630 || F. Audet || francois.audet@skypelabs.com +# RFC5631 || R. Shacham, H. Schulzrinne, S. Thakolsri, W. Kellerer || shacham@cs.columbia.edu, hgs@cs.columbia.edu, thakolsri@docomolab-euro.com, kellerer@docomolab-euro.com +# RFC5632 || C. Griffiths, J. Livingood, L. Popkin, R. Woundy, Y. Yang || chris_griffiths@cable.comcast.com, jason_livingood@cable.comcast.com, laird@pando.com, richard_woundy@cable.comcast.com, yry@cs.yale.edu +# RFC5633 || S. Dawkins, Ed. || spencer@wonderhamster.org +# RFC5634 || G. Fairhurst, A. Sathiaseelan || gorry@erg.abdn.ac.uk, arjuna@erg.abdn.ac.uk +# RFC5635 || W. Kumari, D. McPherson || warren@kumari.net, danny@arbor.net +# RFC5636 || S. Park, H. Park, Y. Won, J. Lee, S. Kent || shpark@kisa.or.kr, hrpark@kisa.or.kr, yjwon@kisa.or.kr, jilee@kisa.or.kr, kent@bbn.com +# RFC5637 || G. Giaretta, I. Guardini, E. Demaria, J. Bournelle, R. Lopez || gerardo@qualcomm.com, ivano.guardini@telecomitalia.it, elena.demaria@telecomitalia.it, julien.bournelle@gmail.com, rafa@dif.um.es +# RFC5638 || H. Sinnreich, Ed., A. Johnston, E. Shim, K. Singh || henrys@adobe.com, alan@sipstation.com, eunsooshim@gmail.com, kns10@cs.columbia.edu +# RFC5639 || M. Lochter, J. Merkle || manfred.lochter@bsi.bund.de, johannes.merkle@secunet.com +# RFC5640 || C. Filsfils, P. Mohapatra, C. Pignataro || cfilsfil@cisco.com, pmohapat@cisco.com, cpignata@cisco.com +# RFC5641 || N. McGill, C. Pignataro || nmcgill@cisco.com, cpignata@cisco.com +# RFC5642 || S. Venkata, S. Harwani, C. Pignataro, D. McPherson || svenkata@google.com, sharwani@cisco.com, cpignata@cisco.com, danny@arbor.net +# RFC5643 || D. Joyal, Ed., V. Manral, Ed. || djoyal@nortel.com, vishwas@ipinfusion.com +# RFC5644 || E. Stephan, L. Liang, A. Morton || emile.stephan@orange-ftgroup.com, L.Liang@surrey.ac.uk, acmorton@att.com +# RFC5645 || D. Ewell, Ed. || doug@ewellic.org +# RFC5646 || A. Phillips, Ed., M. Davis, Ed. || addison@inter-locale.com, markdavis@google.com +# RFC5647 || K. Igoe, J. Solinas || kmigoe@nsa.gov, jasolin@orion.ncsc.mil +# RFC5648 || R. Wakikawa, Ed., V. Devarapalli, G. Tsirtsis, T. Ernst, K. Nagami || ryuji.wakikawa@gmail.com, vijay@wichorus.com, Tsirtsis@gmail.com, thierry.ernst@inria.fr, nagami@inetcore.com +# RFC5649 || R. Housley, M. Dworkin || housley@vigilsec.com, dworkin@nist.gov +# RFC5650 || M. Morgenstern, S. Baillie, U. Bonollo || moti.Morgenstern@ecitele.com, scott.baillie@nec.com.au, umberto.bonollo@nec.com.au +# RFC5651 || M. Luby, M. Watson, L. Vicisano || luby@qti.qualcomm.com, watson@qualcomm.com, vicisano@qualcomm.com +# RFC5652 || R. Housley || housley@vigilsec.com +# RFC5653 || M. Upadhyay, S. Malkani || m.d.upadhyay+ietf@gmail.com, Seema.Malkani@gmail.com +# RFC5654 || B. Niven-Jenkins, Ed., D. Brungard, Ed., M. Betts, Ed., N. Sprecher, S. Ueno || benjamin.niven-jenkins@bt.com, dbrungard@att.com, malcolm.betts@huawei.com, nurit.sprecher@nsn.com, satoshi.ueno@ntt.com +# RFC5655 || B. Trammell, E. Boschi, L. Mark, T. Zseby, A. Wagner || brian.trammell@hitachi-eu.com, elisa.boschi@hitachi-eu.com, lutz.mark@ifam.fraunhofer.de, tanja.zseby@fokus.fraunhofer.de, arno@wagner.name +# RFC5656 || D. Stebila, J. Green || douglas@stebila.ca, jonathan.green@queensu.ca +# RFC5657 || L. Dusseault, R. Sparks || lisa.dusseault@gmail.com, RjS@nostrum.com +# RFC5658 || T. Froment, C. Lebel, B. Bonnaerens || thomas.froment@tech-invite.com, Christophe.Lebel@alcatel-lucent.fr, ben.bonnaerens@alcatel-lucent.be +# RFC5659 || M. Bocci, S. Bryant || matthew.bocci@alcatel-lucent.com, stbryant@cisco.com +# RFC5660 || N. Williams || Nicolas.Williams@sun.com +# RFC5661 || S. Shepler, Ed., M. Eisler, Ed., D. Noveck, Ed. || shepler@storspeed.com, mike@eisler.com, dnoveck@netapp.com +# RFC5662 || S. Shepler, Ed., M. Eisler, Ed., D. Noveck, Ed. || shepler@storspeed.com, mike@eisler.com, dnoveck@netapp.com +# RFC5663 || D. Black, S. Fridella, J. Glasgow || black_david@emc.com, stevef@nasuni.com, jglasgow@aya.yale.edu +# RFC5664 || B. Halevy, B. Welch, J. Zelenka || bhalevy@panasas.com, welch@panasas.com, jimz@panasas.com +# RFC5665 || M. Eisler || mike@eisler.com +# RFC5666 || T. Talpey, B. Callaghan || tmtalpey@gmail.com, brentc@apple.com +# RFC5667 || T. Talpey, B. Callaghan || tmtalpey@gmail.com, brentc@apple.com +# RFC5668 || Y. Rekhter, S. Sangli, D. Tappan || yakov@juniper.net, rsrihari@cisco.com, Dan.Tappan@Gmail.com +# RFC5669 || S. Yoon, J. Kim, H. Park, H. Jeong, Y. Won || seokung@kisa.or.kr, seopo@kisa.or.kr, hrpark@kisa.or.kr, hcjung@kisa.or.kr, yjwon@kisa.or.kr +# RFC5670 || P. Eardley, Ed. || philip.eardley@bt.com +# RFC5671 || S. Yasukawa, A. Farrel, Ed. || yasukawa.seisho@lab.ntt.co.jp, adrian@olddog.co.uk +# RFC5672 || D. Crocker, Ed. || dcrocker@bbiw.net +# RFC5673 || K. Pister, Ed., P. Thubert, Ed., S. Dwars, T. Phinney || kpister@dustnetworks.com, pthubert@cisco.com, sicco.dwars@shell.com, tom.phinney@cox.net +# RFC5674 || S. Chisholm, R. Gerhards || schishol@nortel.com, rgerhards@adiscon.com +# RFC5675 || V. Marinov, J. Schoenwaelder || v.marinov@jacobs-university.de, j.schoenwaelder@jacobs-university.de +# RFC5676 || J. Schoenwaelder, A. Clemm, A. Karmakar || j.schoenwaelder@jacobs-university.de, alex@cisco.com, akarmaka@cisco.com +# RFC5677 || T. Melia, Ed., G. Bajko, S. Das, N. Golmie, JC. Zuniga || telemaco.melia@alcatel-lucent.com, Gabor.Bajko@nokia.com, subir@research.telcordia.com, nada.golmie@nist.gov, j.c.zuniga@ieee.org +# RFC5678 || G. Bajko, S. Das || gabor.bajko@nokia.com, subir@research.telcordia.com +# RFC5679 || G. Bajko || gabor.bajko@nokia.com +# RFC5680 || S. Dawkins, Ed. || spencer@wonderhamster.org +# RFC5681 || M. Allman, V. Paxson, E. Blanton || mallman@icir.org, vern@icir.org, eblanton@cs.purdue.edu +# RFC5682 || P. Sarolahti, M. Kojo, K. Yamamoto, M. Hata || pasi.sarolahti@iki.fi, kojo@cs.helsinki.fi, yamamotokaz@nttdocomo.co.jp, hatama@s1.nttdocomo.co.jp +# RFC5683 || A. Brusilovsky, I. Faynberg, Z. Zeltsan, S. Patel || Alec.Brusilovsky@alcatel-lucent.com, igor.faynberg@alcatel-lucent.com, zeltsan@alcatel-lucent.com, sarvar@google.com +# RFC5684 || P. Srisuresh, B. Ford || srisuresh@yahoo.com, bryan.ford@yale.edu +# RFC5685 || V. Devarapalli, K. Weniger || vijay@wichorus.com, kilian.weniger@googlemail.com +# RFC5686 || Y. Hiwasaki, H. Ohmuro || hiwasaki.yusuke@lab.ntt.co.jp, ohmuro.hitoshi@lab.ntt.co.jp +# RFC5687 || H. Tschofenig, H. Schulzrinne || Hannes.Tschofenig@gmx.net, hgs+ecrit@cs.columbia.edu +# RFC5688 || J. Rosenberg || jdrosen@jdrosen.net +# RFC5689 || C. Daboo || cyrus@daboo.name +# RFC5690 || S. Floyd, A. Arcia, D. Ros, J. Iyengar || floyd@icir.org, ae.arcia@telecom-bretagne.eu, David.Ros@telecom-bretagne.eu, jiyengar@fandm.edu +# RFC5691 || F. de Bont, S. Doehla, M. Schmidt, R. Sperschneider || frans.de.bont@philips.com, stefan.doehla@iis.fraunhofer.de, malte.schmidt@dolby.com, ralph.sperschneider@iis.fraunhofer.de +# RFC5692 || H. Jeon, S. Jeong, M. Riegel || hongseok.jeon@gmail.com, sjjeong@etri.re.kr, maximilian.riegel@nsn.com +# RFC5693 || J. Seedorf, E. Burger || jan.seedorf@nw.neclab.eu, eburger@standardstrack.com +# RFC5694 || G. Camarillo, Ed., IAB || Gonzalo.Camarillo@ericsson.com, iab@iab.org +# RFC5695 || A. Akhter, R. Asati, C. Pignataro || aakhter@cisco.com, rajiva@cisco.com, cpignata@cisco.com +# RFC5696 || T. Moncaster, B. Briscoe, M. Menth || toby.moncaster@bt.com, bob.briscoe@bt.com, menth@informatik.uni-wuerzburg.de +# RFC5697 || S. Farrell || stephen.farrell@cs.tcd.ie +# RFC5698 || T. Kunz, S. Okunick, U. Pordesch || thomas.kunz@sit.fraunhofer.de, susanne.okunick@pawisda.de, ulrich.pordesch@zv.fraunhofer.de +# RFC5701 || Y. Rekhter || yakov@juniper.net +# RFC5702 || J. Jansen || jelte@NLnetLabs.nl +# RFC5703 || T. Hansen, C. Daboo || tony+sieveloop@maillennium.att.com, cyrus@daboo.name +# RFC5704 || S. Bryant, Ed., M. Morrow, Ed., IAB || stbryant@cisco.com, mmorrow@cisco.com, iab@iab.org +# RFC5705 || E. Rescorla || ekr@rtfm.com +# RFC5706 || D. Harrington || ietfdbh@comcast.net +# RFC5707 || A. Saleem, Y. Xin, G. Sharratt || adnan.saleem@RadiSys.com, yong.xin@RadiSys.com, garland.sharratt@gmail.com +# RFC5708 || A. Keromytis || angelos@cs.columbia.edu +# RFC5709 || M. Bhatia, V. Manral, M. Fanto, R. White, M. Barnes, T. Li, R. Atkinson || manav@alcatel-lucent.com, vishwas@ipinfusion.com, mfanto@aegisdatasecurity.com, riw@cisco.com, mjbarnes@cisco.com, tony.li@tony.li, rja@extremenetworks.com +# RFC5710 || L. Berger, D. Papadimitriou, JP. Vasseur || lberger@labn.net, Dimitri.Papadimitriou@alcatel-lucent.be, jpv@cisco.com +# RFC5711 || JP. Vasseur, Ed., G. Swallow, I. Minei || jpv@cisco.com, swallow@cisco.com, ina@juniper.net +# RFC5712 || M. Meyer, Ed., JP. Vasseur, Ed. || matthew.meyer@bt.com, jpv@cisco.com +# RFC5713 || H. Moustafa, H. Tschofenig, S. De Cnodder || hassnaa.moustafa@orange-ftgroup.com, Hannes.Tschofenig@gmx.net, stefaan.de_cnodder@alcatel-lucent.com +# RFC5714 || M. Shand, S. Bryant || mshand@cisco.com, stbryant@cisco.com +# RFC5715 || M. Shand, S. Bryant || mshand@cisco.com, stbryant@cisco.com +# RFC5716 || J. Lentini, C. Everhart, D. Ellard, R. Tewari, M. Naik || jlentini@netapp.com, everhart@netapp.com, dellard@bbn.com, tewarir@us.ibm.com, manoj@almaden.ibm.com +# RFC5717 || B. Lengyel, M. Bjorklund || balazs.lengyel@ericsson.com, mbj@tail-f.com +# RFC5718 || D. Beller, A. Farrel || dieter.beller@alcatel-lucent.com, adrian@olddog.co.uk +# RFC5719 || D. Romascanu, H. Tschofenig || dromasca@gmail.com , Hannes.Tschofenig@gmx.net +# RFC5720 || F. Templin || fltemplin@acm.org +# RFC5721 || R. Gellens, C. Newman || rg+ietf@qualcomm.com, chris.newman@sun.com +# RFC5722 || S. Krishnan || suresh.krishnan@ericsson.com +# RFC5723 || Y. Sheffer, H. Tschofenig || yaronf@checkpoint.com, Hannes.Tschofenig@gmx.net +# RFC5724 || E. Wilde, A. Vaha-Sipila || dret@berkeley.edu, antti.vaha-sipila@nokia.com +# RFC5725 || A. Begen, D. Hsu, M. Lague || abegen@cisco.com, dohsu@cisco.com, mlague@cisco.com +# RFC5726 || Y. Qiu, F. Zhao, Ed., R. Koodli || qiuying@i2r.a-star.edu.sg, fanzhao@google.com, rkoodli@cisco.com +# RFC5727 || J. Peterson, C. Jennings, R. Sparks || jon.peterson@neustar.biz, fluffy@cisco.com, rjsparks@nostrum.com +# RFC5728 || S. Combes, P. Amundsen, M. Lambert, H-P. Lexow || stephane.combes@esa.int, pca@verisat.no, micheline.lambert@advantechamt.com, hlexow@stmi.com +# RFC5729 || J. Korhonen, Ed., M. Jones, L. Morand, T. Tsou || jouni.nospam@gmail.com, Mark.Jones@bridgewatersystems.com, Lionel.morand@orange-ftgroup.com, tena@huawei.com +# RFC5730 || S. Hollenbeck || shollenbeck@verisign.com +# RFC5731 || S. Hollenbeck || shollenbeck@verisign.com +# RFC5732 || S. Hollenbeck || shollenbeck@verisign.com +# RFC5733 || S. Hollenbeck || shollenbeck@verisign.com +# RFC5734 || S. Hollenbeck || shollenbeck@verisign.com +# RFC5735 || M. Cotton, L. Vegoda || michelle.cotton@icann.org, leo.vegoda@icann.org +# RFC5736 || G. Huston, M. Cotton, L. Vegoda || gih@apnic.net, michelle.cotton@icann.org, leo.vegoda@icann.org +# RFC5737 || J. Arkko, M. Cotton, L. Vegoda || jari.arkko@piuha.net, michelle.cotton@icann.org, leo.vegoda@icann.org +# RFC5738 || P. Resnick, C. Newman || presnick@qti.qualcomm.com, chris.newman@sun.com +# RFC5739 || P. Eronen, J. Laganier, C. Madson || pe@iki.fi, julienl@qualcomm.com, cmadson@cisco.com +# RFC5740 || B. Adamson, C. Bormann, M. Handley, J. Macker || adamson@itd.nrl.navy.mil, cabo@tzi.org, M.Handley@cs.ucl.ac.uk, macker@itd.nrl.navy.mil +# RFC5741 || L. Daigle, Ed., O. Kolkman, Ed., IAB || leslie@thinkingcat.com, olaf@nlnetlabs.nl, iab@iab.org +# RFC5742 || H. Alvestrand, R. Housley || harald@alvestrand.no, housley@vigilsec.com +# RFC5743 || A. Falk || falk@bbn.com +# RFC5744 || R. Braden, J. Halpern || braden@isi.edu, jhalpern@redback.com +# RFC5745 || A. Malis, Ed., IAB || andrew.g.malis@verizon.com, iab@iab.org +# RFC5746 || E. Rescorla, M. Ray, S. Dispensa, N. Oskov || ekr@rtfm.com, marsh@extendedsubset.com, dispensa@phonefactor.com, nasko.oskov@microsoft.com +# RFC5747 || J. Wu, Y. Cui, X. Li, M. Xu, C. Metz || jianping@cernet.edu.cn, cy@csnet1.cs.tsinghua.edu.cn, xing@cernet.edu.cn, xmw@csnet1.cs.tsinghua.edu.cn, chmetz@cisco.com +# RFC5748 || S. Yoon, J. Jeong, H. Kim, H. Jeong, Y. Won || seokung@kisa.or.kr, jijeong@kisa.or.kr, rinyfeel@kisa.or.kr, hcjung@kisa.or.kr, yjwon@kisa.or.kr +# RFC5749 || K. Hoeper, Ed., M. Nakhjiri, Y. Ohba, Ed. || khoeper@motorola.com, madjid.nakhjiri@motorola.com, yoshihiro.ohba@toshiba.co.jp +# RFC5750 || B. Ramsdell, S. Turner || blaker@gmail.com, turners@ieca.com +# RFC5751 || B. Ramsdell, S. Turner || blaker@gmail.com, turners@ieca.com +# RFC5752 || S. Turner, J. Schaad || turners@ieca.com, jimsch@exmsft.com +# RFC5753 || S. Turner, D. Brown || turners@ieca.com, dbrown@certicom.com +# RFC5754 || S. Turner || turners@ieca.com +# RFC5755 || S. Farrell, R. Housley, S. Turner || stephen.farrell@cs.tcd.ie, housley@vigilsec.com, turners@ieca.com +# RFC5756 || S. Turner, D. Brown, K. Yiu, R. Housley, T. Polk || turners@ieca.com, dbrown@certicom.com, kelviny@microsoft.com, housley@vigilsec.com, wpolk@nist.gov +# RFC5757 || T. Schmidt, M. Waehlisch, G. Fairhurst || schmidt@informatik.haw-hamburg.de, mw@link-lab.net, gorry@erg.abdn.ac.uk +# RFC5758 || Q. Dang, S. Santesson, K. Moriarty, D. Brown, T. Polk || quynh.dang@nist.gov, sts@aaa-sec.com, Moriarty_Kathleen@emc.com, dbrown@certicom.com, tim.polk@nist.gov +# RFC5759 || J. Solinas, L. Zieglar || jasolin@orion.ncsc.mil, llziegl@tycho.ncsc.mil +# RFC5760 || J. Ott, J. Chesterfield, E. Schooler || jo@acm.org, julianchesterfield@cantab.net, eve_schooler@acm.org +# RFC5761 || C. Perkins, M. Westerlund || csp@csperkins.org, magnus.westerlund@ericsson.com +# RFC5762 || C. Perkins || csp@csperkins.org +# RFC5763 || J. Fischl, H. Tschofenig, E. Rescorla || jason.fischl@skype.net, Hannes.Tschofenig@gmx.net, ekr@rtfm.com +# RFC5764 || D. McGrew, E. Rescorla || mcgrew@cisco.com, ekr@rtfm.com +# RFC5765 || H. Schulzrinne, E. Marocco, E. Ivov || hgs@cs.columbia.edu, enrico.marocco@telecomitalia.it, emcho@sip-communicator.org +# RFC5766 || R. Mahy, P. Matthews, J. Rosenberg || rohan@ekabal.com, philip_matthews@magma.ca, jdrosen@jdrosen.net +# RFC5767 || M. Munakata, S. Schubert, T. Ohba || munakata.mayumi@lab.ntt.co.jp, shida@ntt-at.com, ohba.takumi@lab.ntt.co.jp +# RFC5768 || J. Rosenberg || jdrosen@jdrosen.net +# RFC5769 || R. Denis-Courmont || remi.denis-courmont@nokia.com +# RFC5770 || M. Komu, T. Henderson, H. Tschofenig, J. Melen, A. Keranen, Ed. || miika@iki.fi, thomas.r.henderson@boeing.com, Hannes.Tschofenig@gmx.net, jan.melen@ericsson.com, ari.keranen@ericsson.com +# RFC5771 || M. Cotton, L. Vegoda, D. Meyer || michelle.cotton@icann.org, leo.vegoda@icann.org, dmm@1-4-5.net +# RFC5772 || A. Doria, E. Davies, F. Kastenholz || avri@ltu.se, elwynd@dial.pipex.com, frank@bbn.com +# RFC5773 || E. Davies, A. Doria || elwynd@dial.pipex.com, avri@acm.org +# RFC5774 || K. Wolf, A. Mayrhofer || karlheinz.wolf@nic.at, alexander.mayrhofer@nic.at +# RFC5775 || M. Luby, M. Watson, L. Vicisano || luby@qti.qualcomm.com, watson@qualcomm.com, vicisano@qualcomm.com +# RFC5776 || V. Roca, A. Francillon, S. Faurite || vincent.roca@inria.fr, aurelien.francillon@inria.fr, faurite@lcpc.fr +# RFC5777 || J. Korhonen, H. Tschofenig, M. Arumaithurai, M. Jones, Ed., A. Lior || jouni.korhonen@nsn.com, Hannes.Tschofenig@gmx.net, mayutan.arumaithurai@gmail.com, mark.jones@bridgewatersystems.com, avi@bridgewatersystems.com +# RFC5778 || J. Korhonen, Ed., H. Tschofenig, J. Bournelle, G. Giaretta, M. Nakhjiri || jouni.nospam@gmail.com, Hannes.Tschofenig@gmx.net, julien.bournelle@orange-ftgroup.com, gerardo.giaretta@gmail.com, madjid.nakhjiri@motorola.com +# RFC5779 || J. Korhonen, Ed., J. Bournelle, K. Chowdhury, A. Muhanna, U. Meyer || jouni.nospam@gmail.com, julien.bournelle@orange-ftgroup.com, kchowdhury@cisco.com, Ahmad.muhanna@ericsson.com, meyer@umic.rwth-aachen.de +# RFC5780 || D. MacDonald, B. Lowekamp || derek.macdonald@gmail.com, bbl@lowekamp.net +# RFC5781 || S. Weiler, D. Ward, R. Housley || weiler@tislabs.com, dward@juniper.net, housley@vigilsec.com +# RFC5782 || J. Levine || standards@taugh.com +# RFC5783 || M. Welzl, W. Eddy || michawe@ifi.uio.no, wes@mti-systems.com +# RFC5784 || N. Freed, S. Vedam || ned.freed@mrochek.com, Srinivas.Sv@Sun.COM +# RFC5785 || M. Nottingham, E. Hammer-Lahav || mnot@mnot.net, eran@hueniverse.com +# RFC5786 || R. Aggarwal, K. Kompella || rahul@juniper.net, kireeti@juniper.net +# RFC5787 || D. Papadimitriou || dimitri.papadimitriou@alcatel-lucent.be +# RFC5788 || A. Melnikov, D. Cridland || Alexey.Melnikov@isode.com, dave.cridland@isode.com +# RFC5789 || L. Dusseault, J. Snell || lisa.dusseault@gmail.com, jasnell@gmail.com +# RFC5790 || H. Liu, W. Cao, H. Asaeda || Liuhui47967@huawei.com, caowayne@huawei.com, asaeda@wide.ad.jp +# RFC5791 || J. Reschke, J. Kunze || julian.reschke@greenbytes.de, jak@ucop.edu +# RFC5792 || P. Sangster, K. Narayan || Paul_Sangster@symantec.com, kaushik@cisco.com +# RFC5793 || R. Sahita, S. Hanna, R. Hurst, K. Narayan || Ravi.Sahita@intel.com, shanna@juniper.net, Ryan.Hurst@microsoft.com, kaushik@cisco.com +# RFC5794 || J. Lee, J. Lee, J. Kim, D. Kwon, C. Kim || jklee@ensec.re.kr, jlee05@ensec.re.kr, jaeheon@ensec.re.kr, ds_kwon@ensec.re.kr, jbr@ensec.re.kr +# RFC5795 || K. Sandlund, G. Pelletier, L-E. Jonsson || kristofer.sandlund@ericsson.com, ghyslain.pelletier@ericsson.com, lars-erik@lejonsson.com +# RFC5796 || W. Atwood, S. Islam, M. Siami || bill@cse.concordia.ca, Salekul.Islam@emt.inrs.ca, mzrsm@yahoo.ca +# RFC5797 || J. Klensin, A. Hoenes || john+ietf@jck.com, ah@TR-Sys.de +# RFC5798 || S. Nadas, Ed. || stephen.nadas@ericsson.com +# RFC5801 || S. Josefsson, N. Williams || simon@josefsson.org, Nicolas.Williams@oracle.com +# RFC5802 || C. Newman, A. Menon-Sen, A. Melnikov, N. Williams || chris.newman@oracle.com, ams@toroid.org, Alexey.Melnikov@isode.com, Nicolas.Williams@oracle.com +# RFC5803 || A. Melnikov || alexey.melnikov@isode.com +# RFC5804 || A. Melnikov, Ed., T. Martin || Alexey.Melnikov@isode.com, timmartin@alumni.cmu.edu +# RFC5805 || K. Zeilenga || Kurt.Zeilenga@Isode.COM +# RFC5806 || S. Levy, M. Mohali, Ed. || stlevy@cisco.com, marianne.mohali@orange-ftgroup.com +# RFC5807 || Y. Ohba, A. Yegin || yoshihiro.ohba@toshiba.co.jp, alper.yegin@yegin.org +# RFC5808 || R. Marshall, Ed. || rmarshall@telecomsys.com +# RFC5810 || A. Doria, Ed., J. Hadi Salim, Ed., R. Haas, Ed., H. Khosravi, Ed., W. Wang, Ed., L. Dong, R. Gopal, J. Halpern || avri@ltu.se, hadi@mojatatu.com, rha@zurich.ibm.com, hormuzd.m.khosravi@intel.com, wmwang@mail.zjgsu.edu.cn, donglg@zjgsu.edu.cn, ram.gopal@nsn.com, jmh@joelhalpern.com +# RFC5811 || J. Hadi Salim, K. Ogawa || hadi@mojatatu.com, ogawa.kentaro@lab.ntt.co.jp +# RFC5812 || J. Halpern, J. Hadi Salim || jmh@joelhalpern.com, hadi@mojatatu.com +# RFC5813 || R. Haas || rha@zurich.ibm.com +# RFC5814 || W. Sun, Ed., G. Zhang, Ed. || sunwq@mit.edu, zhangguoying@mail.ritt.com.cn +# RFC5815 || T. Dietz, Ed., A. Kobayashi, B. Claise, G. Muenz || Thomas.Dietz@nw.neclab.eu, akoba@nttv6.net, bclaise@cisco.com, muenz@net.in.tum.de +# RFC5816 || S. Santesson, N. Pope || sts@aaa-sec.com, nick.pope@thales-esecurity.com +# RFC5817 || Z. Ali, JP. Vasseur, A. Zamfir, J. Newton || zali@cisco.com, jpv@cisco.com, ancaz@cisco.com, jonathan.newton@cw.com +# RFC5818 || D. Li, H. Xu, S. Bardalai, J. Meuric, D. Caviglia || danli@huawei.com, xuhuiying@huawei.com, snigdho.bardalai@us.fujitsu.com, julien.meuric@orange-ftgroup.com, diego.caviglia@ericsson.com +# RFC5819 || A. Melnikov, T. Sirainen || Alexey.Melnikov@isode.com, tss@iki.fi +# RFC5820 || A. Roy, Ed., M. Chandra, Ed. || akr@cisco.com, mw.chandra@gmail.com +# RFC5824 || K. Kumaki, Ed., R. Zhang, Y. Kamite || ke-kumaki@kddi.com, raymond.zhang@bt.com, y.kamite@ntt.com +# RFC5825 || K. Fujiwara, B. Leiba || fujiwara@jprs.co.jp, barryleiba@computer.org +# RFC5826 || A. Brandt, J. Buron, G. Porcu || abr@sdesigns.dk, jbu@sdesigns.dk, gporcu@gmail.com +# RFC5827 || M. Allman, K. Avrachenkov, U. Ayesta, J. Blanton, P. Hurtig || mallman@icir.org, k.avrachenkov@sophia.inria.fr, urtzi@laas.fr, jblanton@irg.cs.ohiou.edu, per.hurtig@kau.se +# RFC5828 || D. Fedyk, L. Berger, L. Andersson || donald.fedyk@alcatel-lucent.com, lberger@labn.net, loa.andersson@ericsson.com +# RFC5829 || A. Brown, G. Clemm, J. Reschke, Ed. || albertcbrown@us.ibm.com, geoffrey.clemm@us.ibm.com, julian.reschke@greenbytes.de +# RFC5830 || V. Dolmatov, Ed. || dol@cryptocom.ru +# RFC5831 || V. Dolmatov, Ed. || dol@cryptocom.ru +# RFC5832 || V. Dolmatov, Ed. || dol@cryptocom.ru +# RFC5833 || Y. Shi, Ed., D. Perkins, Ed., C. Elliott, Ed., Y. Zhang, Ed. || rishyang@gmail.com, dperkins@dsperkins.com, chelliot@pobox.com, yzhang@fortinet.com +# RFC5834 || Y. Shi, Ed., D. Perkins, Ed., C. Elliott, Ed., Y. Zhang, Ed. || rishyang@gmail.com, dperkins@dsperkins.com, chelliot@pobox.com, yzhang@fortinet.com +# RFC5835 || A. Morton, Ed., S. Van den Berghe, Ed. || acmorton@att.com, steven.van_den_berghe@alcatel-lucent.com +# RFC5836 || Y. Ohba, Ed., Q. Wu, Ed., G. Zorn, Ed. || oshihiro.ohba@toshiba.co.jp, sunseawq@huawei.com, gwz@net-zen.net +# RFC5837 || A. Atlas, Ed., R. Bonica, Ed., C. Pignataro, Ed., N. Shen, JR. Rivers || alia.atlas@bt.com, rbonica@juniper.net, cpignata@cisco.com, naiming@cisco.com, jrrivers@yahoo.com +# RFC5838 || A. Lindem, Ed., S. Mirtorabi, A. Roy, M. Barnes, R. Aggarwal || acee.lindem@ericsson.com, smirtora@cisco.com, akr@cisco.com, mjbarnes@cisco.com, rahul@juniper.net +# RFC5839 || A. Niemi, D. Willis, Ed. || aki.niemi@nokia.com, dean.willis@softarmor.com +# RFC5840 || K. Grewal, G. Montenegro, M. Bhatia || ken.grewal@intel.com, gabriel.montenegro@microsoft.com, manav.bhatia@alcatel-lucent.com +# RFC5841 || R. Hay, W. Turkal || rhay@google.com, turkal@google.com +# RFC5842 || G. Clemm, J. Crawford, J. Reschke, Ed., J. Whitehead || geoffrey.clemm@us.ibm.com, ccjason@us.ibm.com, julian.reschke@greenbytes.de, ejw@cse.ucsc.edu +# RFC5843 || A. Bryan || anthonybryan@gmail.com +# RFC5844 || R. Wakikawa, S. Gundavelli || ryuji@us.toyota-itc.com, sgundave@cisco.com +# RFC5845 || A. Muhanna, M. Khalil, S. Gundavelli, K. Leung || ahmad.muhanna@ericsson.com, Mohamed.khalil@ericsson.com, sgundave@cisco.com, kleung@cisco.com +# RFC5846 || A. Muhanna, M. Khalil, S. Gundavelli, K. Chowdhury, P. Yegani || ahmad.muhanna@ericsson.com, mohamed.khalil@ericsson.com, sgundave@cisco.com, kchowdhu@cisco.com, pyegani@juniper.net +# RFC5847 || V. Devarapalli, Ed., R. Koodli, Ed., H. Lim, N. Kant, S. Krishnan, J. Laganier || vijay@wichorus.com, rkoodli@cisco.com, hlim@stoke.com, nishi@stoke.com, suresh.krishnan@ericsson.com, julienl@qualcomm.com +# RFC5848 || J. Kelsey, J. Callas, A. Clemm || john.kelsey@nist.gov, jon@callas.org, alex@cisco.com +# RFC5849 || E. Hammer-Lahav, Ed. || eran@hueniverse.com +# RFC5850 || R. Mahy, R. Sparks, J. Rosenberg, D. Petrie, A. Johnston, Ed. || rohan@ekabal.com, rjsparks@nostrum.com, jdrosen@jdrosen.net, dpetrie@sipez.com, alan@sipstation.com +# RFC5851 || S. Ooghe, N. Voigt, M. Platnic, T. Haag, S. Wadhwa || sven.ooghe@alcatel-lucent.com, norbert.voigt@nsn.com, mplatnic@gmail.com, haagt@telekom.de, swadhwa@juniper.net +# RFC5852 || D. Caviglia, D. Ceccarelli, D. Bramanti, D. Li, S. Bardalai || diego.caviglia@ericsson.com, daniele.ceccarelli@ericsson.com, none, danli@huawei.com, sbardalai@gmail.com +# RFC5853 || J. Hautakorpi, Ed., G. Camarillo, R. Penfield, A. Hawrylyshen, M. Bhatia || Jani.Hautakorpi@ericsson.com, Gonzalo.Camarillo@ericsson.com, bpenfield@acmepacket.com, alan.ietf@polyphase.ca, mbhatia@3clogic.com +# RFC5854 || A. Bryan, T. Tsujikawa, N. McNab, P. Poeml || anthonybryan@gmail.com, tatsuhiro.t@gmail.com, neil@nabber.org, peter@poeml.de +# RFC5855 || J. Abley, T. Manderson || joe.abley@icann.org, terry.manderson@icann.org +# RFC5856 || E. Ertekin, R. Jasani, C. Christou, C. Bormann || ertekin_emre@bah.com, ro@breakcheck.com, christou_chris@bah.com, cabo@tzi.org +# RFC5857 || E. Ertekin, C. Christou, R. Jasani, T. Kivinen, C. Bormann || ertekin_emre@bah.com, christou_chris@bah.com, ro@breakcheck.com, kivinen@iki.fi, cabo@tzi.org +# RFC5858 || E. Ertekin, C. Christou, C. Bormann || ertekin_emre@bah.com, christou_chris@bah.com, cabo@tzi.org +# RFC5859 || R. Johnson || raj@cisco.com +# RFC5860 || M. Vigoureux, Ed., D. Ward, Ed., M. Betts, Ed. || martin.vigoureux@alcatel-lucent.com, dward@juniper.net, malcolm.betts@rogers.com +# RFC5861 || M. Nottingham || mnot@yahoo-inc.com +# RFC5862 || S. Yasukawa, A. Farrel || yasukawa.seisho@lab.ntt.co.jp, adrian@olddog.co.uk +# RFC5863 || T. Hansen, E. Siegel, P. Hallam-Baker, D. Crocker || tony+dkimov@maillennium.att.com, dkim@esiegel.net, phillip@hallambaker.com, dcrocker@bbiw.net +# RFC5864 || R. Allbery || rra@stanford.edu +# RFC5865 || F. Baker, J. Polk, M. Dolly || fred@cisco.com, jmpolk@cisco.com, mdolly@att.com +# RFC5866 || D. Sun, Ed., P. McCann, H. Tschofenig, T. Tsou, A. Doria, G. Zorn, Ed. || d.sun@alcatel-lucent.com, pete.mccann@motorola.com, Hannes.Tschofenig@gmx.net, tena@huawei.com, avri@ltu.se, gwz@net-zen.net +# RFC5867 || J. Martocci, Ed., P. De Mil, N. Riou, W. Vermeylen || jerald.p.martocci@jci.com, pieter.demil@intec.ugent.be, nicolas.riou@fr.schneider-electric.com, wouter@vooruit.be +# RFC5868 || S. Sakane, K. Kamada, S. Zrelli, M. Ishiyama || Shouichi.Sakane@jp.yokogawa.com, Ken-ichi.Kamada@jp.yokogawa.com, Saber.Zrelli@jp.yokogawa.com, masahiro@isl.rdc.toshiba.co.jp +# RFC5869 || H. Krawczyk, P. Eronen || hugokraw@us.ibm.com, pe@iki.fi +# RFC5870 || A. Mayrhofer, C. Spanring || alexander.mayrhofer@ipcom.at, christian@spanring.eu +# RFC5871 || J. Arkko, S. Bradner || jari.arkko@piuha.net, sob@harvard.edu +# RFC5872 || J. Arkko, A. Yegin || jari.arkko@piuha.net, alper.yegin@yegin.org +# RFC5873 || Y. Ohba, A. Yegin || yoshihiro.ohba@toshiba.co.jp, alper.yegin@yegin.org +# RFC5874 || J. Rosenberg, J. Urpalainen || jdrosen.net, jari.urpalainen@nokia.com +# RFC5875 || J. Urpalainen, D. Willis, Ed. || jari.urpalainen@nokia.com, dean.willis@softarmor.com +# RFC5876 || J. Elwell || john.elwell@siemens-enterprise.com +# RFC5877 || R. Housley || housley@vigilsec.com +# RFC5878 || M. Brown, R. Housley || mark@redphonesecurity.com, housley@vigilsec.com +# RFC5879 || T. Kivinen, D. McDonald || kivinen@iki.fi, danmcd@opensolaris.org +# RFC5880 || D. Katz, D. Ward || dkatz@juniper.net, dward@juniper.net +# RFC5881 || D. Katz, D. Ward || dkatz@juniper.net, dward@juniper.net +# RFC5882 || D. Katz, D. Ward || dkatz@juniper.net, dward@juniper.net +# RFC5883 || D. Katz, D. Ward || dkatz@juniper.net, dward@juniper.net +# RFC5884 || R. Aggarwal, K. Kompella, T. Nadeau, G. Swallow || rahul@juniper.net, kireeti@juniper.net, tom.nadeau@bt.com, swallow@cisco.com +# RFC5885 || T. Nadeau, Ed., C. Pignataro, Ed. || tom.nadeau@bt.com, cpignata@cisco.com +# RFC5886 || JP. Vasseur, Ed., JL. Le Roux, Y. Ikejiri || jpv@cisco.com, jeanlouis.leroux@orange-ftgroup.com, y.ikejiri@ntt.com +# RFC5887 || B. Carpenter, R. Atkinson, H. Flinck || brian.e.carpenter@gmail.com, rja@extremenetworks.com, hannu.flinck@nsn.com +# RFC5888 || G. Camarillo, H. Schulzrinne || Gonzalo.Camarillo@ericsson.com, schulzrinne@cs.columbia.edu +# RFC5889 || E. Baccelli, Ed., M. Townsley, Ed. || Emmanuel.Baccelli@inria.fr, mark@townsley.net +# RFC5890 || J. Klensin || john+ietf@jck.com +# RFC5891 || J. Klensin || john+ietf@jck.com +# RFC5892 || P. Faltstrom, Ed. || paf@cisco.com +# RFC5893 || H. Alvestrand, Ed., C. Karp || harald@alvestrand.no, ck@nic.museum +# RFC5894 || J. Klensin || john+ietf@jck.com +# RFC5895 || P. Resnick, P. Hoffman || presnick@qti.qualcomm.com, paul.hoffman@vpnc.org +# RFC5896 || L. Hornquist Astrand, S. Hartman || lha@apple.com, hartmans-ietf@mit.edu +# RFC5897 || J. Rosenberg || jdrosen@jdrosen.net +# RFC5898 || F. Andreasen, G. Camarillo, D. Oran, D. Wing || fandreas@cisco.com, Gonzalo.Camarillo@ericsson.com, oran@cisco.com, dwing-ietf@fuggles.com +# RFC5901 || P. Cain, D. Jevans || pcain@coopercain.com, dave.jevans@antiphishing.org +# RFC5902 || D. Thaler, L. Zhang, G. Lebovitz || dthaler@microsoft.com, lixia@cs.ucla.edu, gregory.ietf@gmail.com, iab@iab.org +# RFC5903 || D. Fu, J. Solinas || defu@orion.ncsc.mil, jasolin@orion.ncsc.mil +# RFC5904 || G. Zorn || gwz@net-zen.net +# RFC5905 || D. Mills, J. Martin, Ed., J. Burbank, W. Kasch || mills@udel.edu, jrmii@isc.org, jack.burbank@jhuapl.edu, william.kasch@jhuapl.edu +# RFC5906 || B. Haberman, Ed., D. Mills || brian@innovationslab.net, mills@udel.edu +# RFC5907 || H. Gerstung, C. Elliott, B. Haberman, Ed. || heiko.gerstung@meinberg.de, chelliot@pobox.com, brian@innovationslab.net +# RFC5908 || R. Gayraud, B. Lourdelet || richard.gayraud@free.fr, blourdel@cisco.com +# RFC5909 || J-M. Combes, S. Krishnan, G. Daley || jeanmichel.combes@orange-ftgroup.com, Suresh.Krishnan@ericsson.com, hoskuld@hotmail.com +# RFC5910 || J. Gould, S. Hollenbeck || jgould@verisign.com, shollenbeck@verisign.com +# RFC5911 || P. Hoffman, J. Schaad || paul.hoffman@vpnc.org, jimsch@exmsft.com +# RFC5912 || P. Hoffman, J. Schaad || paul.hoffman@vpnc.org, jimsch@exmsft.com +# RFC5913 || S. Turner, S. Chokhani || turners@ieca.com, SChokhani@cygnacom.com +# RFC5914 || R. Housley, S. Ashmore, C. Wallace || housley@vigilsec.com, srashmo@radium.ncsc.mil, cwallace@cygnacom.com +# RFC5915 || S. Turner, D. Brown || turners@ieca.com, dbrown@certicom.com +# RFC5916 || S. Turner || turners@ieca.com +# RFC5917 || S. Turner || turners@ieca.com +# RFC5918 || R. Asati, I. Minei, B. Thomas || rajiva@cisco.com, ina@juniper.net, bobthomas@alum.mit.edu +# RFC5919 || R. Asati, P. Mohapatra, E. Chen, B. Thomas || rajiva@cisco.com, pmohapat@cisco.com, chenying220@huawei.com, bobthomas@alum.mit.edu +# RFC5920 || L. Fang, Ed. || lufang@cisco.com +# RFC5921 || M. Bocci, Ed., S. Bryant, Ed., D. Frost, Ed., L. Levrau, L. Berger || matthew.bocci@alcatel-lucent.com, stbryant@cisco.com, danfrost@cisco.com, lieven.levrau@alcatel-lucent.com, lberger@labn.net +# RFC5922 || V. Gurbani, S. Lawrence, A. Jeffrey || vkg@alcatel-lucent.com, scott-ietf@skrb.org, ajeffrey@alcatel-lucent.com +# RFC5923 || V. Gurbani, Ed., R. Mahy, B. Tate || vkg@alcatel-lucent.com, rohan@ekabal.com, brett@broadsoft.com +# RFC5924 || S. Lawrence, V. Gurbani || scott-ietf@skrb.org, vkg@bell-labs.com +# RFC5925 || J. Touch, A. Mankin, R. Bonica || touch@isi.edu, mankin@psg.com, rbonica@juniper.net +# RFC5926 || G. Lebovitz, E. Rescorla || gregory.ietf@gmail.com, ekr@rtfm.com +# RFC5927 || F. Gont || fernando@gont.com.ar +# RFC5928 || M. Petit-Huguenin || petithug@acm.org +# RFC5929 || J. Altman, N. Williams, L. Zhu || jaltman@secure-endpoints.com, Nicolas.Williams@oracle.com, larry.zhu@microsoft.com +# RFC5930 || S. Shen, Y. Mao, NSS. Murthy || shenshuo@cnnic.cn, yumao9@gmail.com, ssmurthy.nittala@freescale.com +# RFC5931 || D. Harkins, G. Zorn || dharkins@arubanetworks.com, gwz@net-zen.net +# RFC5932 || A. Kato, M. Kanda, S. Kanno || kato.akihiro@po.ntts.co.jp, kanda.masayuki@lab.ntt.co.jp, kanno.satoru@po.ntts.co.jp +# RFC5933 || V. Dolmatov, Ed., A. Chuprina, I. Ustinov || dol@cryptocom.ru, ran@cryptocom.ru, igus@cryptocom.ru +# RFC5934 || R. Housley, S. Ashmore, C. Wallace || housley@vigilsec.com, srashmo@radium.ncsc.mil, cwallace@cygnacom.com +# RFC5935 || M. Ellison, B. Natale || ietf@ellisonsoftware.com, rnatale@mitre.org +# RFC5936 || E. Lewis, A. Hoenes, Ed. || ed.lewis@neustar.biz, ah@TR-Sys.de +# RFC5937 || S. Ashmore, C. Wallace || srashmo@radium.ncsc.mil, cwallace@cygnacom.com +# RFC5938 || A. Morton, M. Chiba || acmorton@att.com, mchiba@cisco.com +# RFC5939 || F. Andreasen || fandreas@cisco.com +# RFC5940 || S. Turner, R. Housley || turners@ieca.com, housley@vigilsec.com +# RFC5941 || D. M'Raihi, S. Boeyen, M. Grandcolas, S. Bajaj || davidietf@gmail.com, sharon.boeyen@entrust.com, michael.grandcolas@hotmail.com, sbajaj@verisign.com +# RFC5942 || H. Singh, W. Beebee, E. Nordmark || shemant@cisco.com, wbeebee@cisco.com, erik.nordmark@oracle.com +# RFC5943 || B. Haberman, Ed. || brian@innovationslab.net +# RFC5944 || C. Perkins, Ed. || charliep@computer.org +# RFC5945 || F. Le Faucheur, J. Manner, D. Wing, A. Guillou || flefauch@cisco.com, jukka.manner@tkk.fi, dwing-ietf@fuggles.com, allan.guillou@sfr.com +# RFC5946 || F. Le Faucheur, J. Manner, A. Narayanan, A. Guillou, H. Malik || flefauch@cisco.com, jukka.manner@tkk.fi, ashokn@cisco.com, allan.guillou@sfr.com, Hemant.Malik@airtel.in +# RFC5947 || J. Elwell, H. Kaplan || john.elwell@siemens-enterprise.com, hkaplan@acmepacket.com +# RFC5948 || S. Madanapalli, S. Park, S. Chakrabarti, G. Montenegro || smadanapalli@gmail.com, soohong.park@samsung.com, samitac@ipinfusion.com, gabriel.montenegro@microsoft.com +# RFC5949 || H. Yokota, K. Chowdhury, R. Koodli, B. Patil, F. Xia || yokota@kddilabs.jp, kchowdhu@cisco.com, rkoodli@cisco.com, basavaraj.patil@nokia.com, xiayangsong@huawei.com +# RFC5950 || S. Mansfield, Ed., E. Gray, Ed., K. Lam, Ed. || scott.mansfield@ericsson.com, eric.gray@ericsson.com, Kam.Lam@alcatel-lucent.com +# RFC5951 || K. Lam, S. Mansfield, E. Gray || Kam.Lam@Alcatel-Lucent.com, Scott.Mansfield@Ericsson.com, Kam.Lam@Alcatel-Lucent.com +# RFC5952 || S. Kawamura, M. Kawashima || kawamucho@mesh.ad.jp, kawashimam@vx.jp.nec.com +# RFC5953 || W. Hardaker || ietf@hardakers.net +# RFC5954 || V. Gurbani, Ed., B. Carpenter, Ed., B. Tate, Ed. || vkg@bell-labs.com, brian.e.carpenter@gmail.com, brett@broadsoft.com +# RFC5955 || A. Santoni || adriano.santoni@actalis.it +# RFC5956 || A. Begen || abegen@cisco.com +# RFC5957 || D. Karp || dkarp@zimbra.com +# RFC5958 || S. Turner || turners@ieca.com +# RFC5959 || S. Turner || turners@ieca.com +# RFC5960 || D. Frost, Ed., S. Bryant, Ed., M. Bocci, Ed. || danfrost@cisco.com, stbryant@cisco.com, matthew.bocci@alcatel-lucent.com +# RFC5961 || A. Ramaiah, R. Stewart, M. Dalal || ananth@cisco.com, randall@lakerest.net, mdalal@cisco.com +# RFC5962 || H. Schulzrinne, V. Singh, H. Tschofenig, M. Thomson || hgs@cs.columbia.edu, vs2140@cs.columbia.edu, Hannes.Tschofenig@gmx.net, martin.thomson@andrew.com +# RFC5963 || R. Gagliano || rogaglia@cisco.com +# RFC5964 || J. Winterbottom, M. Thomson || james.winterbottom@andrew.com, martin.thomson@andrew.com +# RFC5965 || Y. Shafranovich, J. Levine, M. Kucherawy || ietf@shaftek.org, standards@taugh.com, msk@cloudmark.com +# RFC5966 || R. Bellis || ray.bellis@nominet.org.uk +# RFC5967 || S. Turner || turners@ieca.com +# RFC5968 || J. Ott, C. Perkins || jo@netlab.tkk.fi, csp@csperkins.org +# RFC5969 || W. Townsley, O. Troan || mark@townsley.net, ot@cisco.com +# RFC5970 || T. Huth, J. Freimann, V. Zimmer, D. Thaler || thuth@de.ibm.com, jfrei@de.ibm.com, vincent.zimmer@intel.com, dthaler@microsoft.com +# RFC5971 || H. Schulzrinne, R. Hancock || hgs+nsis@cs.columbia.edu, robert.hancock@roke.co.uk +# RFC5972 || T. Tsenov, H. Tschofenig, X. Fu, Ed., C. Aoun, E. Davies || tseno.tsenov@mytum.de, Hannes.Tschofenig@nsn.com, fu@cs.uni-goettingen.de, cedaoun@yahoo.fr, elwynd@dial.pipex.com +# RFC5973 || M. Stiemerling, H. Tschofenig, C. Aoun, E. Davies || Martin.Stiemerling@neclab.eu, Hannes.Tschofenig@nsn.com, cedaoun@yahoo.fr, elwynd@dial.pipex.com +# RFC5974 || J. Manner, G. Karagiannis, A. McDonald || jukka.manner@tkk.fi, karagian@cs.utwente.nl, andrew.mcdonald@roke.co.uk +# RFC5975 || G. Ash, Ed., A. Bader, Ed., C. Kappler, Ed., D. Oran, Ed. || gash5107@yahoo.com, Attila.Bader@ericsson.com, cornelia.kappler@cktecc.de, oran@cisco.com +# RFC5976 || G. Ash, A. Morton, M. Dolly, P. Tarapore, C. Dvorak, Y. El Mghazli || gash5107@yahoo.com, acmorton@att.com, mdolly@att.com, tarapore@att.com, cdvorak@att.com, yacine.el_mghazli@alcatel.fr +# RFC5977 || A. Bader, L. Westberg, G. Karagiannis, C. Kappler, T. Phelan || Attila.Bader@ericsson.com, Lars.Westberg@ericsson.com, g.karagiannis@ewi.utwente.nl, cornelia.kappler@cktecc.de, tphelan@sonusnet.com +# RFC5978 || J. Manner, R. Bless, J. Loughney, E. Davies, Ed. || jukka.manner@tkk.fi, bless@kit.edu, john.loughney@nokia.com, elwynd@folly.org.uk +# RFC5979 || C. Shen, H. Schulzrinne, S. Lee, J. Bang || charles@cs.columbia.edu, hgs@cs.columbia.edu, sung1.lee@samsung.com, jh0278.bang@samsung.com +# RFC5980 || T. Sanda, Ed., X. Fu, S. Jeong, J. Manner, H. Tschofenig || sanda.takako@jp.panasonic.com, fu@cs.uni-goettingen.de, shjeong@hufs.ac.kr, jukka.manner@tkk.fi, Hannes.Tschofenig@nsn.com +# RFC5981 || J. Manner, M. Stiemerling, H. Tschofenig, R. Bless, Ed. || jukka.manner@tkk.fi, martin.stiemerling@neclab.eu, Hannes.Tschofenig@gmx.net, roland.bless@kit.edu +# RFC5982 || A. Kobayashi, Ed., B. Claise, Ed. || akoba@nttv6.net, bclaise@cisco.com +# RFC5983 || R. Gellens || rg+ietf@qualcomm.com +# RFC5984 || K-M. Moller || kalle@tankesaft.se +# RFC5985 || M. Barnes, Ed. || mary.ietf.barnes@gmail.com +# RFC5986 || M. Thomson, J. Winterbottom || martin.thomson@andrew.com, james.winterbottom@andrew.com +# RFC5987 || J. Reschke || julian.reschke@greenbytes.de +# RFC5988 || M. Nottingham || mnot@mnot.net +# RFC5989 || A.B. Roach || adam@nostrum.com +# RFC5990 || J. Randall, B. Kaliski, J. Brainard, S. Turner || jdrandall@comcast.net, kaliski_burt@emc.com, jbrainard@rsa.com, turners@ieca.com +# RFC5991 || D. Thaler, S. Krishnan, J. Hoagland || dthaler@microsoft.com, suresh.krishnan@ericsson.com, Jim_Hoagland@symantec.com +# RFC5992 || S. Sharikov, D. Miloshevic, J. Klensin || s.shar@regtime.net, dmiloshevic@afilias.info, john-ietf@jck.com +# RFC5993 || X. Duan, S. Wang, M. Westerlund, K. Hellwig, I. Johansson || duanxiaodong@chinamobile.com, wangshuaiyu@chinamobile.com, magnus.westerlund@ericsson.com, karl.hellwig@ericsson.com, ingemar.s.johansson@ericsson.com +# RFC5994 || S. Bryant, Ed., M. Morrow, G. Swallow, R. Cherukuri, T. Nadeau, N. Harrison, B. Niven-Jenkins || stbryant@cisco.com, mmorrow@cisco.com, swallow@cisco.com, cherukuri@juniper.net, thomas.nadeau@huawei.com, neil.2.harrison@bt.com, ben@niven-jenkins.co.uk +# RFC5995 || J. Reschke || julian.reschke@greenbytes.de +# RFC5996 || C. Kaufman, P. Hoffman, Y. Nir, P. Eronen || charliek@microsoft.com, paul.hoffman@vpnc.org, ynir@checkpoint.com, pe@iki.fi +# RFC5997 || A. DeKok || aland@freeradius.org +# RFC5998 || P. Eronen, H. Tschofenig, Y. Sheffer || pe@iki.fi, Hannes.Tschofenig@gmx.net, yaronf.ietf@gmail.com +# RFC6001 || D. Papadimitriou, M. Vigoureux, K. Shiomoto, D. Brungard, JL. Le Roux || dimitri.papadimitriou@alcatel-lucent.com, martin.vigoureux@alcatel-lucent.fr, shiomoto.kohei@lab.ntt.co.jp, dbrungard@att.com, jean-louis.leroux@rd.francetelecom.com +# RFC6002 || L. Berger, D. Fedyk || lberger@labn.net, donald.fedyk@alcatel-lucent.com +# RFC6003 || D. Papadimitriou || dimitri.papadimitriou@alcatel-lucent.be +# RFC6004 || L. Berger, D. Fedyk || lberger@labn.net, donald.fedyk@alcatel-lucent.com +# RFC6005 || L. Berger, D. Fedyk || lberger@labn.net, donald.fedyk@alcatel-lucent.com +# RFC6006 || Q. Zhao, Ed., D. King, Ed., F. Verhaeghe, T. Takeda, Z. Ali, J. Meuric || qzhao@huawei.com, daniel@olddog.co.uk, fabien.verhaeghe@gmail.com, takeda.tomonori@lab.ntt.co.jp, zali@cisco.com, julien.meuric@orange-ftgroup.com +# RFC6007 || I. Nishioka, D. King || i-nishioka@cb.jp.nec.com, daniel@olddog.co.uk +# RFC6008 || M. Kucherawy || msk@cloudmark.com +# RFC6009 || N. Freed || ned.freed@mrochek.com +# RFC6010 || R. Housley, S. Ashmore, C. Wallace || housley@vigilsec.com, srashmo@radium.ncsc.mil, cwallace@cygnacom.com +# RFC6011 || S. Lawrence, Ed., J. Elwell || scott-ietf@skrb.org, john.elwell@siemens-enterprise.com +# RFC6012 || J. Salowey, T. Petch, R. Gerhards, H. Feng || jsalowey@cisco.com, tomSecurity@network-engineer.co.uk, rgerhards@adiscon.com, fhyfeng@gmail.com +# RFC6013 || W. Simpson || William.Allen.Simpson@Gmail.com +# RFC6014 || P. Hoffman || paul.hoffman@vpnc.org +# RFC6015 || A. Begen || abegen@cisco.com +# RFC6016 || B. Davie, F. Le Faucheur, A. Narayanan || bsd@cisco.com, flefauch@cisco.com, ashokn@cisco.com +# RFC6017 || K. Meadors, Ed. || kyle@drummondgroup.com +# RFC6018 || F. Baker, W. Harrop, G. Armitage || fred@cisco.com, wazz@bud.cc.swin.edu.au, garmitage@swin.edu.au +# RFC6019 || R. Housley || housley@vigilsec.com +# RFC6020 || M. Bjorklund, Ed. || mbj@tail-f.com +# RFC6021 || J. Schoenwaelder, Ed. || j.schoenwaelder@jacobs-university.de +# RFC6022 || M. Scott, M. Bjorklund || mark.scott@ericsson.com, mbj@tail-f.com +# RFC6023 || Y. Nir, H. Tschofenig, H. Deng, R. Singh || ynir@checkpoint.com, Hannes.Tschofenig@gmx.net, denghui02@gmail.com, rsj@cisco.com +# RFC6024 || R. Reddy, C. Wallace || r.reddy@radium.ncsc.mil, cwallace@cygnacom.com +# RFC6025 || C. Wallace, C. Gardiner || cwallace@cygnacom.com, gardiner@bbn.com +# RFC6026 || R. Sparks, T. Zourzouvillys || RjS@nostrum.com, theo@crazygreek.co.uk +# RFC6027 || Y. Nir || ynir@checkpoint.com +# RFC6028 || G. Camarillo, A. Keranen || Gonzalo.Camarillo@ericsson.com, Ari.Keranen@ericsson.com +# RFC6029 || I. Rimac, V. Hilt, M. Tomsu, V. Gurbani, E. Marocco || rimac@bell-labs.com, volkerh@bell-labs.com, marco.tomsu@alcatel-lucent.com, vkg@bell-labs.com, enrico.marocco@telecomitalia.it +# RFC6030 || P. Hoyer, M. Pei, S. Machani || phoyer@actividentity.com, mpei@verisign.com, smachani@diversinet.com +# RFC6031 || S. Turner, R. Housley || turners@ieca.com, housley@vigilsec.com +# RFC6032 || S. Turner, R. Housley || turners@ieca.com, housley@vigilsec.com +# RFC6033 || S. Turner || turners@ieca.com +# RFC6034 || D. Thaler || dthaler@microsoft.com +# RFC6035 || A. Pendleton, A. Clark, A. Johnston, H. Sinnreich || aspen@telchemy.com, alan.d.clark@telchemy.com, alan.b.johnston@gmail.com, henry.sinnreich@gmail.com +# RFC6036 || B. Carpenter, S. Jiang || brian.e.carpenter@gmail.com, shengjiang@huawei.com +# RFC6037 || E. Rosen, Ed., Y. Cai, Ed., IJ. Wijnands || erosen@cisco.com, ycai@cisco.com, ice@cisco.com +# RFC6038 || A. Morton, L. Ciavattone || acmorton@att.com, lencia@att.com +# RFC6039 || V. Manral, M. Bhatia, J. Jaeggli, R. White || vishwas@ipinfusion.com, manav.bhatia@alcatel-lucent.com, joel.jaeggli@nokia.com, riw@cisco.com +# RFC6040 || B. Briscoe || bob.briscoe@bt.com +# RFC6041 || A. Crouch, H. Khosravi, A. Doria, Ed., X. Wang, K. Ogawa || alan.crouch@intel.com, hormuzd.m.khosravi@intel.com, avri@acm.org, carly.wang@huawei.com, ogawa.kentaro@lab.ntt.co.jp +# RFC6042 || A. Keromytis || angelos@cs.columbia.edu +# RFC6043 || J. Mattsson, T. Tian || john.mattsson@ericsson.com, tian.tian1@zte.com.cn +# RFC6044 || M. Mohali || marianne.mohali@orange-ftgroup.com +# RFC6045 || K. Moriarty || Moriarty_Kathleen@EMC.com +# RFC6046 || K. Moriarty, B. Trammell || Moriarty_Kathleen@EMC.com, trammell@tik.ee.ethz.ch +# RFC6047 || A. Melnikov, Ed. || Alexey.Melnikov@isode.com +# RFC6048 || J. Elie || julien@trigofacile.com +# RFC6049 || A. Morton, E. Stephan || acmorton@att.com, emile.stephan@orange-ftgroup.com +# RFC6050 || K. Drage || drage@alcatel-lucent.com +# RFC6051 || C. Perkins, T. Schierl || csp@csperkins.org, ts@thomas-schierl.de +# RFC6052 || C. Bao, C. Huitema, M. Bagnulo, M. Boucadair, X. Li || congxiao@cernet.edu.cn, huitema@microsoft.com, marcelo@it.uc3m.es, mohamed.boucadair@orange-ftgroup.com, xing@cernet.edu.cn +# RFC6053 || E. Haleplidis, K. Ogawa, W. Wang, J. Hadi Salim || ehalep@ece.upatras.gr, ogawa.kentaro@lab.ntt.co.jp, wmwang@mail.zjgsu.edu.cn, hadi@mojatatu.com +# RFC6054 || D. McGrew, B. Weis || mcgrew@cisco.com, bew@cisco.com +# RFC6055 || D. Thaler, J. Klensin, S. Cheshire || dthaler@microsoft.com, john+ietf@jck.com, cheshire@apple.com +# RFC6056 || M. Larsen, F. Gont || michael.larsen@tieto.com, fernando@gont.com.ar +# RFC6057 || C. Bastian, T. Klieber, J. Livingood, J. Mills, R. Woundy || chris_bastian@cable.comcast.com, tom_klieber@cable.comcast.com, jason_livingood@cable.comcast.com, jim_mills@cable.comcast.com, richard_woundy@cable.comcast.com +# RFC6058 || M. Liebsch, Ed., A. Muhanna, O. Blume || marco.liebsch@neclab.eu, ahmad.muhanna@ericsson.com, oliver.blume@alcatel-lucent.de +# RFC6059 || S. Krishnan, G. Daley || suresh.krishnan@ericsson.com, hoskuld@hotmail.com +# RFC6060 || D. Fedyk, H. Shah, N. Bitar, A. Takacs || donald.fedyk@alcatel-lucent.com, hshah@ciena.com, nabil.n.bitar@verizon.com, attila.takacs@ericsson.com +# RFC6061 || B. Rosen || br@brianrosen.net +# RFC6062 || S. Perreault, Ed., J. Rosenberg || simon.perreault@viagenie.ca, jdrosen@jdrosen.net +# RFC6063 || A. Doherty, M. Pei, S. Machani, M. Nystrom || andrea.doherty@rsa.com, mpei@verisign.com, smachani@diversinet.com, mnystrom@microsoft.com +# RFC6064 || M. Westerlund, P. Frojdh || magnus.westerlund@ericsson.com, per.frojdh@ericsson.com +# RFC6065 || K. Narayan, D. Nelson, R. Presuhn, Ed. || kaushik_narayan@yahoo.com, d.b.nelson@comcast.net, randy_presuhn@mindspring.com +# RFC6066 || D. Eastlake 3rd || d3e3e3@gmail.com +# RFC6067 || M. Davis, A. Phillips, Y. Umaoka || mark@macchiato.com, addison@lab126.com, yoshito_umaoka@us.ibm.com +# RFC6068 || M. Duerst, L. Masinter, J. Zawinski || duerst@it.aoyama.ac.jp, LMM@acm.org, jwz@jwz.org +# RFC6069 || A. Zimmermann, A. Hannemann || zimmermann@cs.rwth-aachen.de, hannemann@nets.rwth-aachen.de +# RFC6070 || S. Josefsson || simon@josefsson.org +# RFC6071 || S. Frankel, S. Krishnan || sheila.frankel@nist.gov, suresh.krishnan@ericsson.com +# RFC6072 || C. Jennings, J. Fischl, Ed. || fluffy@cisco.com, jason.fischl@skype.net +# RFC6073 || L. Martini, C. Metz, T. Nadeau, M. Bocci, M. Aissaoui || lmartini@cisco.com, chmetz@cisco.com, tnadeau@lucidvision.com, matthew.bocci@alcatel-lucent.co.uk, mustapha.aissaoui@alcatel-lucent.com +# RFC6074 || E. Rosen, B. Davie, V. Radoaca, W. Luo || erosen@cisco.com, bsd@cisco.com, vasile.radoaca@alcatel-lucent.com, luo@weiluo.net +# RFC6075 || D. Cridland || dave.cridland@isode.com +# RFC6076 || D. Malas, A. Morton || d.malas@cablelabs.com, acmorton@att.com +# RFC6077 || D. Papadimitriou, Ed., M. Welzl, M. Scharf, B. Briscoe || dimitri.papadimitriou@alcatel-lucent.be, michawe@ifi.uio.no, michael.scharf@googlemail.com, bob.briscoe@bt.com +# RFC6078 || G. Camarillo, J. Melen || Gonzalo.Camarillo@ericsson.com, Jan.Melen@ericsson.com +# RFC6079 || G. Camarillo, P. Nikander, J. Hautakorpi, A. Keranen, A. Johnston || Gonzalo.Camarillo@ericsson.com, Pekka.Nikander@ericsson.com, Jani.Hautakorpi@ericsson.com, Ari.Keranen@ericsson.com, alan.b.johnston@gmail.com +# RFC6080 || D. Petrie, S. Channabasappa, Ed. || dan.ietf@SIPez.com, sumanth@cablelabs.com +# RFC6081 || D. Thaler || dthaler@microsoft.com +# RFC6082 || K. Whistler, G. Adams, M. Duerst, R. Presuhn, Ed., J. Klensin || kenw@sybase.com, glenn@skynav.com, duerst@it.aoyama.ac.jp, randy_presuhn@mindspring.com, john+ietf@jck.com +# RFC6083 || M. Tuexen, R. Seggelmann, E. Rescorla || tuexen@fh-muenster.de, seggelmann@fh-muenster.de, ekr@networkresonance.com +# RFC6084 || X. Fu, C. Dickmann, J. Crowcroft || fu@cs.uni-goettingen.de, mail@christian-dickmann.de, jon.crowcroft@cl.cam.ac.uk +# RFC6085 || S. Gundavelli, M. Townsley, O. Troan, W. Dec || sgundave@cisco.com, townsley@cisco.com, ot@cisco.com, wdec@cisco.com +# RFC6086 || C. Holmberg, E. Burger, H. Kaplan || christer.holmberg@ericsson.com, eburger@standardstrack.com, hkaplan@acmepacket.com +# RFC6087 || A. Bierman || andy@yumaworks.com +# RFC6088 || G. Tsirtsis, G. Giarreta, H. Soliman, N. Montavont || tsirtsis@qualcomm.com, gerardog@qualcomm.com, hesham@elevatemobile.com, nicolas.montavont@telecom-bretagne.eu +# RFC6089 || G. Tsirtsis, H. Soliman, N. Montavont, G. Giaretta, K. Kuladinithi || tsirtsis@qualcomm.com, hesham@elevatemobile.com, nicolas.montavont@telecom-bretagne.eu, gerardog@qualcomm.com, koo@comnets.uni-bremen.de +# RFC6090 || D. McGrew, K. Igoe, M. Salter || mcgrew@cisco.com, kmigoe@nsa.gov, msalter@restarea.ncsc.mil +# RFC6091 || N. Mavrogiannopoulos, D. Gillmor || nikos.mavrogiannopoulos@esat.kuleuven.be, dkg@fifthhorseman.net +# RFC6092 || J. Woodyatt, Ed. || jhw@apple.com +# RFC6093 || F. Gont, A. Yourtchenko || fernando@gont.com.ar, ayourtch@cisco.com +# RFC6094 || M. Bhatia, V. Manral || manav.bhatia@alcatel-lucent.com, vishwas@ipinfusion.com +# RFC6095 || B. Linowski, M. Ersue, S. Kuryla || bernd.linowski.ext@nsn.com, mehmet.ersue@nsn.com, s.kuryla@gmail.com +# RFC6096 || M. Tuexen, R. Stewart || tuexen@fh-muenster.de, randall@lakerest.net +# RFC6097 || J. Korhonen, V. Devarapalli || jouni.nospam@gmail.com, dvijay@gmail.com +# RFC6098 || H. Deng, H. Levkowetz, V. Devarapalli, S. Gundavelli, B. Haley || denghui02@gmail.com, henrik@levkowetz.com, dvijay@gmail.com, sgundave@cisco.com, brian.haley@hp.com +# RFC6101 || A. Freier, P. Karlton, P. Kocher || nikos.mavrogiannopoulos@esat.kuleuven.be +# RFC6104 || T. Chown, S. Venaas || tjc@ecs.soton.ac.uk, stig@cisco.com +# RFC6105 || E. Levy-Abegnoli, G. Van de Velde, C. Popoviciu, J. Mohacsi || elevyabe@cisco.com, gunter@cisco.com, chip@technodyne.com, mohacsi@niif.hu +# RFC6106 || J. Jeong, S. Park, L. Beloeil, S. Madanapalli || pjeong@brocade.com, soohong.park@samsung.com, luc.beloeil@orange-ftgroup.com, smadanapalli@gmail.com +# RFC6107 || K. Shiomoto, Ed., A. Farrel, Ed. || shiomoto.kohei@lab.ntt.co.jp, adrian@olddog.co.uk +# RFC6108 || C. Chung, A. Kasyanov, J. Livingood, N. Mody, B. Van Lieu || chae_chung@cable.comcast.com, alexander_kasyanov@cable.comcast.com, jason_livingood@cable.comcast.com, nirmal_mody@cable.comcast.com, brian@vanlieu.net +# RFC6109 || C. Petrucci, F. Gennai, A. Shahin, A. Vinciarelli || petrucci@digitpa.gov.it, francesco.gennai@isti.cnr.it, alba.shahin@isti.cnr.it, alessandro.vinciarelli@gmail.com +# RFC6110 || L. Lhotka, Ed. || ladislav@lhotka.name +# RFC6111 || L. Zhu || lzhu@microsoft.com +# RFC6112 || L. Zhu, P. Leach, S. Hartman || larry.zhu@microsoft.com, paulle@microsoft.com, hartmans-ietf@mit.edu +# RFC6113 || S. Hartman, L. Zhu || hartmans-ietf@mit.edu, larry.zhu@microsoft.com +# RFC6114 || M. Katagi, S. Moriai || Masanobu.Katagi@jp.sony.com, clefia-q@jp.sony.com +# RFC6115 || T. Li, Ed. || tony.li@tony.li +# RFC6116 || S. Bradner, L. Conroy, K. Fujiwara || sob@harvard.edu, lconroy@insensate.co.uk, fujiwara@jprs.co.jp +# RFC6117 || B. Hoeneisen, A. Mayrhofer, J. Livingood || bernie@ietf.hoeneisen.ch, alexander.mayrhofer@enum.at, jason_livingood@cable.comcast.com +# RFC6118 || B. Hoeneisen, A. Mayrhofer || bernie@ietf.hoeneisen.ch, alexander.mayrhofer@enum.at +# RFC6119 || J. Harrison, J. Berger, M. Bartlett || jon.harrison@metaswitch.com, jon.berger@metaswitch.com, mike.bartlett@metaswitch.com +# RFC6120 || P. Saint-Andre || ietf@stpeter.im +# RFC6121 || P. Saint-Andre || ietf@stpeter.im +# RFC6122 || P. Saint-Andre || ietf@stpeter.im +# RFC6123 || A. Farrel || adrian@olddog.co.uk +# RFC6124 || Y. Sheffer, G. Zorn, H. Tschofenig, S. Fluhrer || yaronf.ietf@gmail.com, gwz@net-zen.net, Hannes.Tschofenig@gmx.net, sfluhrer@cisco.com +# RFC6125 || P. Saint-Andre, J. Hodges || ietf@stpeter.im, Jeff.Hodges@PayPal.com +# RFC6126 || J. Chroboczek || jch@pps.jussieu.fr +# RFC6127 || J. Arkko, M. Townsley || jari.arkko@piuha.net, townsley@cisco.com +# RFC6128 || A. Begen || abegen@cisco.com +# RFC6129 || L. Romary, S. Lundberg || laurent.romary@inria.fr, slu@kb.dk +# RFC6130 || T. Clausen, C. Dearlove, J. Dean || T.Clausen@computer.org, chris.dearlove@baesystems.com, jdean@itd.nrl.navy.mil +# RFC6131 || R. George, B. Leiba || robinsgv@gmail.com, barryleiba@computer.org +# RFC6132 || R. George, B. Leiba || robinsgv@gmail.com, barryleiba@computer.org +# RFC6133 || R. George, B. Leiba, A. Melnikov || robinsgv@gmail.com, barryleiba@computer.org, Alexey.Melnikov@isode.com +# RFC6134 || A. Melnikov, B. Leiba || Alexey.Melnikov@isode.com, barryleiba@computer.org +# RFC6135 || C. Holmberg, S. Blau || christer.holmberg@ericsson.com, staffan.blau@ericsson.com +# RFC6136 || A. Sajassi, Ed., D. Mohan, Ed. || sajassi@cisco.com, mohand@nortel.com +# RFC6137 || D. Zisiadis, Ed., S. Kopsidas, Ed., M. Tsavli, Ed., G. Cessieux, Ed. || dzisiadis@iti.gr, spyros@uth.gr, sttsavli@uth.gr, Guillaume.Cessieux@cc.in2p3.fr +# RFC6138 || S. Kini, Ed., W. Lu, Ed. || sriganesh.kini@ericsson.com, wenhu.lu@ericsson.com +# RFC6139 || S. Russert, Ed., E. Fleischman, Ed., F. Templin, Ed. || russerts@hotmail.com, eric.fleischman@boeing.com, fltemplin@acm.org +# RFC6140 || A.B. Roach || adam@nostrum.com +# RFC6141 || G. Camarillo, Ed., C. Holmberg, Y. Gao || Gonzalo.Camarillo@ericsson.com, Christer.Holmberg@ericsson.com, gao.yang2@zte.com.cn +# RFC6142 || A. Moise, J. Brodkin || avy@fdos.ca, jonathan.brodkin@fdos.ca +# RFC6143 || T. Richardson, J. Levine || standards@realvnc.com, standards@taugh.com +# RFC6144 || F. Baker, X. Li, C. Bao, K. Yin || fred@cisco.com, xing@cernet.edu.cn, congxiao@cernet.edu.cn, kyin@cisco.com +# RFC6145 || X. Li, C. Bao, F. Baker || xing@cernet.edu.cn, congxiao@cernet.edu.cn, fred@cisco.com +# RFC6146 || M. Bagnulo, P. Matthews, I. van Beijnum || marcelo@it.uc3m.es, philip_matthews@magma.ca, iljitsch@muada.com +# RFC6147 || M. Bagnulo, A. Sullivan, P. Matthews, I. van Beijnum || marcelo@it.uc3m.es, ajs@shinkuro.com, philip_matthews@magma.ca, iljitsch@muada.com +# RFC6148 || P. Kurapati, R. Desetti, B. Joshi || kurapati@juniper.net, ramakrishnadtv@infosys.com, bharat_joshi@infosys.com +# RFC6149 || S. Turner, L. Chen || turners@ieca.com, lily.chen@nist.gov +# RFC6150 || S. Turner, L. Chen || turners@ieca.com, lily.chen@nist.gov +# RFC6151 || S. Turner, L. Chen || turners@ieca.com, lily.chen@nist.gov +# RFC6152 || J. Klensin, N. Freed, M. Rose, D. Crocker, Ed. || john+ietf@jck.com, ned.freed@mrochek.com, mrose17@gmail.com, dcrocker@bbiw.net +# RFC6153 || S. Das, G. Bajko || subir@research.Telcordia.com, gabor.bajko@nokia.com +# RFC6154 || B. Leiba, J. Nicolson || barryleiba@computer.org, nicolson@google.com +# RFC6155 || J. Winterbottom, M. Thomson, H. Tschofenig, R. Barnes || james.winterbottom@andrew.com, martin.thomson@andrew.com, Hannes.Tschofenig@gmx.net, rbarnes@bbn.com +# RFC6156 || G. Camarillo, O. Novo, S. Perreault, Ed. || Gonzalo.Camarillo@ericsson.com, Oscar.Novo@ericsson.com, simon.perreault@viagenie.ca +# RFC6157 || G. Camarillo, K. El Malki, V. Gurbani || Gonzalo.Camarillo@ericsson.com, karim@athonet.com, vkg@bell-labs.com +# RFC6158 || A. DeKok, Ed., G. Weber || aland@freeradius.org, gdweber@gmail.com +# RFC6159 || T. Tsou, G. Zorn, T. Taylor, Ed. || tena@huawei.com, gwz@net-zen.net, tom.taylor.stds@gmail.com +# RFC6160 || S. Turner || turners@ieca.com +# RFC6161 || S. Turner || turners@ieca.com +# RFC6162 || S. Turner || turners@ieca.com +# RFC6163 || Y. Lee, Ed., G. Bernstein, Ed., W. Imajuku || ylee@huawei.com, gregb@grotto-networking.com, imajuku.wataru@lab.ntt.co.jp +# RFC6164 || M. Kohno, B. Nitzan, R. Bush, Y. Matsuzaki, L. Colitti, T. Narten || mkohno@juniper.net, nitzan@juniper.net, randy@psg.com, maz@iij.ad.jp, lorenzo@google.com, narten@us.ibm.com +# RFC6165 || A. Banerjee, D. Ward || ayabaner@cisco.com, dward@juniper.net +# RFC6166 || S. Venaas || stig@cisco.com +# RFC6167 || M. Phillips, P. Adams, D. Rokicki, E. Johnson || m8philli@uk.ibm.com, phil_adams@us.ibm.com, derek.rokicki@softwareag.com, eric@tibco.com +# RFC6168 || W. Hardaker || ietf@hardakers.net +# RFC6169 || S. Krishnan, D. Thaler, J. Hoagland || suresh.krishnan@ericsson.com, dthaler@microsoft.com, Jim_Hoagland@symantec.com +# RFC6170 || S. Santesson, R. Housley, S. Bajaj, L. Rosenthol || sts@aaa-sec.com, housley@vigilsec.com, siddharthietf@gmail.com, leonardr@adobe.com +# RFC6171 || K. Zeilenga || Kurt.Zeilenga@Isode.COM +# RFC6172 || D. Black, D. Peterson || david.black@emc.com, david.peterson@brocade.com +# RFC6173 || P. Venkatesen, Ed. || prakashvn@hcl.com +# RFC6174 || E. Juskevicius || edj.etc@gmail.com +# RFC6175 || E. Juskevicius || edj.etc@gmail.com +# RFC6176 || S. Turner, T. Polk || turners@ieca.com, tim.polk@nist.gov +# RFC6177 || T. Narten, G. Huston, L. Roberts || narten@us.ibm.com, gih@apnic.net, lea.roberts@stanford.edu +# RFC6178 || D. Smith, J. Mullooly, W. Jaeger, T. Scholl || djsmith@cisco.com, jmullool@cisco.com, wjaeger@att.com, tscholl@nlayer.net +# RFC6179 || F. Templin, Ed. || fltemplin@acm.org +# RFC6180 || J. Arkko, F. Baker || jari.arkko@piuha.net, fred@cisco.com +# RFC6181 || M. Bagnulo || marcelo@it.uc3m.es +# RFC6182 || A. Ford, C. Raiciu, M. Handley, S. Barre, J. Iyengar || alan.ford@roke.co.uk, c.raiciu@cs.ucl.ac.uk, m.handley@cs.ucl.ac.uk, sebastien.barre@uclouvain.be, jiyengar@fandm.edu +# RFC6183 || A. Kobayashi, B. Claise, G. Muenz, K. Ishibashi || akoba@orange.plala.or.jp, bclaise@cisco.com, muenz@net.in.tum.de, ishibashi.keisuke@lab.ntt.co.jp +# RFC6184 || Y.-K. Wang, R. Even, T. Kristensen, R. Jesup || yekuiwang@huawei.com, even.roni@huawei.com, tom.kristensen@tandberg.com, rjesup@wgate.com +# RFC6185 || T. Kristensen, P. Luthi || tom.kristensen@tandberg.com, patrick.luthi@tandberg.com +# RFC6186 || C. Daboo || cyrus@daboo.name +# RFC6187 || K. Igoe, D. Stebila || kmigoe@nsa.gov, douglas@stebila.ca +# RFC6188 || D. McGrew || mcgrew@cisco.com +# RFC6189 || P. Zimmermann, A. Johnston, Ed., J. Callas || prz@mit.edu, alan.b.johnston@gmail.com, jon@callas.org +# RFC6190 || S. Wenger, Y.-K. Wang, T. Schierl, A. Eleftheriadis || stewe@stewe.org, yekui.wang@huawei.com, ts@thomas-schierl.de, alex@vidyo.com +# RFC6191 || F. Gont || fernando@gont.com.ar +# RFC6192 || D. Dugal, C. Pignataro, R. Dunn || dave@juniper.net, cpignata@cisco.com, rodunn@cisco.com +# RFC6193 || M. Saito, D. Wing, M. Toyama || ma.saito@nttv6.jp, dwing-ietf@fuggles.com, toyama.masashi@lab.ntt.co.jp +# RFC6194 || T. Polk, L. Chen, S. Turner, P. Hoffman || tim.polk@nist.gov, lily.chen@nist.gov, turners@ieca.com, paul.hoffman@vpnc.org +# RFC6195 || D. Eastlake 3rd || d3e3e3@gmail.com +# RFC6196 || A. Melnikov || Alexey.Melnikov@isode.com +# RFC6197 || K. Wolf || karlheinz.wolf@nic.at +# RFC6198 || B. Decraene, P. Francois, C. Pelsser, Z. Ahmad, A.J. Elizondo Armengol, T. Takeda || bruno.decraene@orange-ftgroup.com, francois@info.ucl.ac.be, cristel@iij.ad.jp, zubair.ahmad@orange-ftgroup.com, ajea@tid.es, takeda.tomonori@lab.ntt.co.jp +# RFC6201 || R. Asati, C. Pignataro, F. Calabria, C. Olvera || rajiva@cisco.com, cpignata@cisco.com, fcalabri@cisco.com, cesar.olvera@consulintel.es +# RFC6202 || S. Loreto, P. Saint-Andre, S. Salsano, G. Wilkins || salvatore.loreto@ericsson.com, ietf@stpeter.im, stefano.salsano@uniroma2.it, gregw@webtide.com +# RFC6203 || T. Sirainen || tss@iki.fi +# RFC6204 || H. Singh, W. Beebee, C. Donley, B. Stark, O. Troan, Ed. || shemant@cisco.com, wbeebee@cisco.com, c.donley@cablelabs.com, barbara.stark@att.com, ot@cisco.com +# RFC6205 || T. Otani, Ed., D. Li, Ed. || tm-otani@kddi.com, danli@huawei.com +# RFC6206 || P. Levis, T. Clausen, J. Hui, O. Gnawali, J. Ko || pal@cs.stanford.edu, T.Clausen@computer.org, jhui@archrock.com, gnawali@cs.stanford.edu, jgko@cs.jhu.edu +# RFC6207 || R. Denenberg, Ed. || rden@loc.gov +# RFC6208 || K. Sankar, Ed., A. Jones || ksankar@cisco.com, arnold.jones@snia.org +# RFC6209 || W. Kim, J. Lee, J. Park, D. Kwon || whkim5@ensec.re.kr, jklee@ensec.re.kr, jhpark@ensec.re.kr, ds_kwon@ensec.re.kr +# RFC6210 || J. Schaad || ietf@augustcellars.com +# RFC6211 || J. Schaad || ietf@augustcellars.com +# RFC6212 || M. Kucherawy || msk@cloudmark.com +# RFC6213 || C. Hopps, L. Ginsberg || chopps@cisco.com, ginsberg@cisco.com +# RFC6214 || B. Carpenter, R. Hinden || brian.e.carpenter@gmail.com, bob.hinden@gmail.com +# RFC6215 || M. Bocci, L. Levrau, D. Frost || matthew.bocci@alcatel-lucent.com, lieven.levrau@alcatel-lucent.com, danfrost@cisco.com +# RFC6216 || C. Jennings, K. Ono, R. Sparks, B. Hibbard, Ed. || fluffy@cisco.com, kumiko@cs.columbia.edu, Robert.Sparks@tekelec.com, Brian.Hibbard@tekelec.com +# RFC6217 || T. Ritter || tom@ritter.vg +# RFC6218 || G. Zorn, T. Zhang, J. Walker, J. Salowey || gwz@net-zen.net, tzhang@advistatech.com, jesse.walker@intel.com, jsalowey@cisco.com +# RFC6219 || X. Li, C. Bao, M. Chen, H. Zhang, J. Wu || xing@cernet.edu.cn, congxiao@cernet.edu.cn, fibrib@gmail.com, neilzh@gmail.com, jianping@cernet.edu.cn +# RFC6220 || D. McPherson, Ed., O. Kolkman, Ed., J. Klensin, Ed., G. Huston, Ed., Internet Architecture Board || dmcpherson@verisign.com, olaf@NLnetLabs.nl, john+ietf@jck.com, gih@apnic.net +# RFC6221 || D. Miles, Ed., S. Ooghe, W. Dec, S. Krishnan, A. Kavanagh || david.miles@alcatel-lucent.com, sven.ooghe@alcatel-lucent.com, wdec@cisco.com, suresh.krishnan@ericsson.com, alan.kavanagh@ericsson.com +# RFC6222 || A. Begen, C. Perkins, D. Wing || abegen@cisco.com, csp@csperkins.org, dwing-ietf@fuggles.com +# RFC6223 || C. Holmberg || christer.holmberg@ericsson.com +# RFC6224 || T. Schmidt, M. Waehlisch, S. Krishnan || schmidt@informatik.haw-hamburg.de, mw@link-lab.net, suresh.krishnan@ericsson.com +# RFC6225 || J. Polk, M. Linsner, M. Thomson, B. Aboba, Ed. || jmpolk@cisco.com, marc.linsner@cisco.com, martin.thomson@andrew.com, bernard_aboba@hotmail.com +# RFC6226 || B. Joshi, A. Kessler, D. McWalter || bharat_joshi@infosys.com, kessler@cisco.com, david@mcwalter.eu +# RFC6227 || T. Li, Ed. || tli@cisco.com +# RFC6228 || C. Holmberg || christer.holmberg@ericsson.com +# RFC6229 || J. Strombergson, S. Josefsson || joachim@secworks.se, simon@josefsson.org +# RFC6230 || C. Boulton, T. Melanchuk, S. McGlashan || chris@ns-technologies.com, timm@rainwillow.com, smcg.stds01@mcglashan.org +# RFC6231 || S. McGlashan, T. Melanchuk, C. Boulton || smcg.stds01@mcglashan.org, timm@rainwillow.com, chris@ns-technologies.com +# RFC6232 || F. Wei, Y. Qin, Z. Li, T. Li, J. Dong || weifang@chinamobile.com, qinyue@chinamobile.com, lizhenqiang@chinamobile.com, tony.li@tony.li, dongjie_dj@huawei.com +# RFC6233 || T. Li, L. Ginsberg || tony.li@tony.li, ginsberg@cisco.com +# RFC6234 || D. Eastlake 3rd, T. Hansen || d3e3e3@gmail.com, tony+shs@maillennium.att.com +# RFC6235 || E. Boschi, B. Trammell || boschie@tik.ee.ethz.ch, trammell@tik.ee.ethz.ch +# RFC6236 || I. Johansson, K. Jung || ingemar.s.johansson@ericsson.com, kyunghun.jung@samsung.com +# RFC6237 || B. Leiba, A. Melnikov || barryleiba@computer.org, Alexey.Melnikov@isode.com +# RFC6238 || D. M'Raihi, S. Machani, M. Pei, J. Rydell || davidietf@gmail.com, smachani@diversinet.com, Mingliang_Pei@symantec.com, johanietf@gmail.com +# RFC6239 || K. Igoe || kmigoe@nsa.gov +# RFC6240 || D. Zelig, Ed., R. Cohen, Ed., T. Nadeau, Ed. || david_zelig@pmc-sierra.com, ronc@resolutenetworks.com, Thomas.Nadeau@ca.com +# RFC6241 || R. Enns, Ed., M. Bjorklund, Ed., J. Schoenwaelder, Ed., A. Bierman, Ed. || rob.enns@gmail.com, mbj@tail-f.com, j.schoenwaelder@jacobs-university.de, andy@yumaworks.com +# RFC6242 || M. Wasserman || mrw@painless-security.com +# RFC6243 || A. Bierman, B. Lengyel || andy@yumaworks.com, balazs.lengyel@ericsson.com +# RFC6244 || P. Shafer || phil@juniper.net +# RFC6245 || P. Yegani, K. Leung, A. Lior, K. Chowdhury, J. Navali || pyegani@juniper.net, kleung@cisco.com, avi@bridgewatersystems.com, kchowdhu@cisco.com, jnavali@cisco.com +# RFC6246 || A. Sajassi, Ed., F. Brockners, D. Mohan, Ed., Y. Serbest || sajassi@cisco.com, fbrockne@cisco.com, dinmohan@hotmail.com, yetik_serbest@labs.att.com +# RFC6247 || L. Eggert || lars.eggert@nokia.com +# RFC6248 || A. Morton || acmorton@att.com +# RFC6249 || A. Bryan, N. McNab, T. Tsujikawa, P. Poeml, H. Nordstrom || anthonybryan@gmail.com, neil@nabber.org, tatsuhiro.t@gmail.com, peter@poeml.de, henrik@henriknordstrom.net +# RFC6250 || D. Thaler || dthaler@microsoft.com +# RFC6251 || S. Josefsson || simon@josefsson.org +# RFC6252 || A. Dutta, Ed., V. Fajardo, Y. Ohba, K. Taniuchi, H. Schulzrinne || ashutosh.dutta@ieee.org, vf0213@gmail.com, yoshihiro.ohba@toshiba.co.jp, kenichi.taniuchi@toshiba.co.jp, hgs@cs.columbia.edu +# RFC6253 || T. Heer, S. Varjonen || heer@cs.rwth-aachen.de, samu.varjonen@hiit.fi +# RFC6254 || M. McFadden || mark.mcfadden@icann.org +# RFC6255 || M. Blanchet || Marc.Blanchet@viagenie.ca +# RFC6256 || W. Eddy, E. Davies || wes@mti-systems.com, elwynd@folly.org.uk +# RFC6257 || S. Symington, S. Farrell, H. Weiss, P. Lovell || susan@mitre.org, stephen.farrell@cs.tcd.ie, howard.weiss@sparta.com, dtnbsp@gmail.com +# RFC6258 || S. Symington || susan@mitre.org +# RFC6259 || S. Symington || susan@mitre.org +# RFC6260 || S. Burleigh || Scott.C.Burleigh@jpl.nasa.gov +# RFC6261 || A. Keranen || ari.keranen@ericsson.com +# RFC6262 || S. Ikonin || ikonin@spiritdsp.com +# RFC6263 || X. Marjou, A. Sollaud || xavier.marjou@orange-ftgroup.com, aurelien.sollaud@orange-ftgroup.com +# RFC6264 || S. Jiang, D. Guo, B. Carpenter || jiangsheng@huawei.com, guoseu@huawei.com, brian.e.carpenter@gmail.com +# RFC6265 || A. Barth || abarth@eecs.berkeley.edu +# RFC6266 || J. Reschke || julian.reschke@greenbytes.de +# RFC6267 || V. Cakulev, G. Sundaram || violeta.cakulev@alcatel-lucent.com, ganesh.sundaram@alcatel-lucent.com +# RFC6268 || J. Schaad, S. Turner || ietf@augustcellars.com, turners@ieca.com +# RFC6269 || M. Ford, Ed., M. Boucadair, A. Durand, P. Levis, P. Roberts || ford@isoc.org, mohamed.boucadair@orange-ftgroup.com, adurand@juniper.net, pierre.levis@orange-ftgroup.com, roberts@isoc.org +# RFC6270 || M. Yevstifeyev || evnikita2@gmail.com +# RFC6271 || J-F. Mule || jf.mule@cablelabs.com +# RFC6272 || F. Baker, D. Meyer || fred@cisco.com, dmm@cisco.com +# RFC6273 || A. Kukec, S. Krishnan, S. Jiang || ana.kukec@fer.hr, suresh.krishnan@ericsson.com, jiangsheng@huawei.com +# RFC6274 || F. Gont || fernando@gont.com.ar +# RFC6275 || C. Perkins, Ed., D. Johnson, J. Arkko || charliep@computer.org, dbj@cs.rice.edu, jari.arkko@ericsson.com +# RFC6276 || R. Droms, P. Thubert, F. Dupont, W. Haddad, C. Bernardos || rdroms@cisco.com, pthubert@cisco.com, fdupont@isc.org, Wassim.Haddad@ericsson.com, cjbc@it.uc3m.es +# RFC6277 || S. Santesson, P. Hallam-Baker || sts@aaa-sec.com, hallam@gmail.com +# RFC6278 || J. Herzog, R. Khazan || jherzog@ll.mit.edu, rkh@ll.mit.edu +# RFC6279 || M. Liebsch, Ed., S. Jeong, Q. Wu || liebsch@neclab.eu, sjjeong@etri.re.kr, sunseawq@huawei.com +# RFC6280 || R. Barnes, M. Lepinski, A. Cooper, J. Morris, H. Tschofenig, H. Schulzrinne || rbarnes@bbn.com, mlepinski@bbn.com, acooper@cdt.org, jmorris@cdt.org, Hannes.Tschofenig@gmx.net, hgs@cs.columbia.edu +# RFC6281 || S. Cheshire, Z. Zhu, R. Wakikawa, L. Zhang || cheshire@apple.com, zhenkai@ucla.edu, ryuji@jp.toyota-itc.com, lixia@cs.ucla.edu +# RFC6282 || J. Hui, Ed., P. Thubert || jhui@archrock.com, pthubert@cisco.com +# RFC6283 || A. Jerman Blazic, S. Saljic, T. Gondrom || aljosa@setcce.si, svetlana.saljic@setcce.si, tobias.gondrom@gondrom.org +# RFC6284 || A. Begen, D. Wing, T. Van Caenegem || abegen@cisco.com, dwing-ietf@fuggles.com, Tom.Van_Caenegem@alcatel-lucent.com +# RFC6285 || B. Ver Steeg, A. Begen, T. Van Caenegem, Z. Vax || billvs@cisco.com, abegen@cisco.com, Tom.Van_Caenegem@alcatel-lucent.be, zeevvax@microsoft.com +# RFC6286 || E. Chen, J. Yuan || enkechen@cisco.com, jenny@cisco.com +# RFC6287 || D. M'Raihi, J. Rydell, S. Bajaj, S. Machani, D. Naccache || davidietf@gmail.com, johanietf@gmail.com, siddharthietf@gmail.com, smachani@diversinet.com, david.naccache@ens.fr +# RFC6288 || C. Reed || creed@opengeospatial.org +# RFC6289 || E. Cardona, S. Channabasappa, J-F. Mule || e.cardona@cablelabs.com, sumanth@cablelabs.com, jf.mule@cablelabs.com +# RFC6290 || Y. Nir, Ed., D. Wierbowski, F. Detienne, P. Sethi || ynir@checkpoint.com, wierbows@us.ibm.com, fd@cisco.com, psethi@cisco.com +# RFC6291 || L. Andersson, H. van Helvoort, R. Bonica, D. Romascanu, S. Mansfield || loa.andersson@ericsson.com, huub.van.helvoort@huawei.com, rbonica@juniper.net, dromasca@gmail.com , scott.mansfield@ericsson.com +# RFC6292 || P. Hoffman || paul.hoffman@vpnc.org +# RFC6293 || P. Hoffman || paul.hoffman@vpnc.org +# RFC6294 || Q. Hu, B. Carpenter || qhu009@aucklanduni.ac.nz, brian.e.carpenter@gmail.com +# RFC6295 || J. Lazzaro, J. Wawrzynek || lazzaro@cs.berkeley.edu, johnw@cs.berkeley.edu +# RFC6296 || M. Wasserman, F. Baker || mrw@painless-security.com, fred@cisco.com +# RFC6297 || M. Welzl, D. Ros || michawe@ifi.uio.no, david.ros@telecom-bretagne.eu +# RFC6298 || V. Paxson, M. Allman, J. Chu, M. Sargent || vern@icir.org, mallman@icir.org, hkchu@google.com, mts71@case.edu +# RFC6301 || Z. Zhu, R. Wakikawa, L. Zhang || zhenkai@cs.ucla.edu, ryuji.wakikawa@gmail.com, lixia@cs.ucla.edu +# RFC6302 || A. Durand, I. Gashinsky, D. Lee, S. Sheppard || adurand@juniper.net, igor@yahoo-inc.com, donn@fb.com, Scott.Sheppard@att.com +# RFC6303 || M. Andrews || marka@isc.org +# RFC6304 || J. Abley, W. Maton || joe.abley@icann.org, wmaton@ryouko.imsb.nrc.ca +# RFC6305 || J. Abley, W. Maton || joe.abley@icann.org, wmaton@ryouko.imsb.nrc.ca +# RFC6306 || P. Frejborg || pfrejborg@gmail.com +# RFC6307 || D. Black, Ed., L. Dunbar, Ed., M. Roth, R. Solomon || david.black@emc.com, ldunbar@huawei.com, MRoth@infinera.com, ronens@corrigent.com +# RFC6308 || P. Savola || psavola@funet.fi +# RFC6309 || J. Arkko, A. Keranen, J. Mattsson || jari.arkko@piuha.net, ari.keranen@ericsson.com, john.mattsson@ericsson.com +# RFC6310 || M. Aissaoui, P. Busschbach, L. Martini, M. Morrow, T. Nadeau, Y(J). Stein || mustapha.aissaoui@alcatel-lucent.com, busschbach@alcatel-lucent.com, lmartini@cisco.com, mmorrow@cisco.com, Thomas.Nadeau@ca.com, yaakov_s@rad.com +# RFC6311 || R. Singh, Ed., G. Kalyani, Y. Nir, Y. Sheffer, D. Zhang || rsj@cisco.com, kagarigi@cisco.com, ynir@checkpoint.com, yaronf.ietf@gmail.com, zhangdacheng@huawei.com +# RFC6312 || R. Koodli || rkoodli@cisco.com +# RFC6313 || B. Claise, G. Dhandapani, P. Aitken, S. Yates || bclaise@cisco.com, gowri@cisco.com, paitken@cisco.com, syates@cisco.com +# RFC6314 || C. Boulton, J. Rosenberg, G. Camarillo, F. Audet || chris@ns-technologies.com, jdrosen@jdrosen.net, Gonzalo.Camarillo@ericsson.com, francois.audet@skype.net +# RFC6315 || E. Guy, K. Darilion || edguy@CleverSpoke.com, klaus.darilion@nic.at +# RFC6316 || M. Komu, M. Bagnulo, K. Slavov, S. Sugimoto, Ed. || miika@iki.fi, marcelo@it.uc3m.es, kristian.slavov@ericsson.com, shinta@sfc.wide.ad.jp +# RFC6317 || M. Komu, T. Henderson || miika@iki.fi, thomas.r.henderson@boeing.com +# RFC6318 || R. Housley, J. Solinas || housley@vigilsec.com, jasolin@orion.ncsc.mil +# RFC6319 || M. Azinger, L. Vegoda || marla.azinger@ftr.com, leo.vegoda@icann.org +# RFC6320 || S. Wadhwa, J. Moisand, T. Haag, N. Voigt, T. Taylor, Ed. || sanjay.wadhwa@alcatel-lucent.com, jmoisand@juniper.net, haagt@telekom.de, norbert.voigt@nsn.com, tom.taylor.stds@gmail.com +# RFC6321 || C. Daboo, M. Douglass, S. Lees || cyrus@daboo.name, douglm@rpi.edu, steven.lees@microsoft.com +# RFC6322 || P. Hoffman || paul.hoffman@vpnc.org +# RFC6323 || G. Renker, G. Fairhurst || gerrit@erg.abdn.ac.uk, gorry@erg.abdn.ac.uk +# RFC6324 || G. Nakibly, F. Templin || gnakibly@yahoo.com, fltemplin@acm.org +# RFC6325 || R. Perlman, D. Eastlake 3rd, D. Dutt, S. Gai, A. Ghanwani || Radia@alum.mit.edu, d3e3e3@gmail.com, ddutt@cisco.com, silvano@ip6.com, anoop@alumni.duke.edu +# RFC6326 || D. Eastlake, A. Banerjee, D. Dutt, R. Perlman, A. Ghanwani || d3e3e3@gmail.com, ayabaner@cisco.com, ddutt@cisco.com, Radia@alum.mit.edu, anoop@alumni.duke.edu +# RFC6327 || D. Eastlake 3rd, R. Perlman, A. Ghanwani, D. Dutt, V. Manral || d3e3e3@gmail.com, Radia@alum.mit.edu, anoop@alumni.duke.edu, ddutt@cisco.com, vishwas.manral@hp.com +# RFC6328 || D. Eastlake 3rd || d3e3e3@gmail.com +# RFC6329 || D. Fedyk, Ed., P. Ashwood-Smith, Ed., D. Allan, A. Bragg, P. Unbehagen || Donald.Fedyk@alcatel-lucent.com, Peter.AshwoodSmith@huawei.com, david.i.allan@ericsson.com, nbragg@ciena.com, unbehagen@avaya.com +# RFC6330 || M. Luby, A. Shokrollahi, M. Watson, T. Stockhammer, L. Minder || luby@qti.qualcomm.com, amin.shokrollahi@epfl.ch, watsonm@netflix.com, stockhammer@nomor.de, lminder@qualcomm.com +# RFC6331 || A. Melnikov || Alexey.Melnikov@isode.com +# RFC6332 || A. Begen, E. Friedrich || abegen@cisco.com, efriedri@cisco.com +# RFC6333 || A. Durand, R. Droms, J. Woodyatt, Y. Lee || adurand@juniper.net, rdroms@cisco.com, jhw@apple.com, yiu_lee@cable.comcast.com +# RFC6334 || D. Hankins, T. Mrugalski || dhankins@google.com, tomasz.mrugalski@eti.pg.gda.pl +# RFC6335 || M. Cotton, L. Eggert, J. Touch, M. Westerlund, S. Cheshire || michelle.cotton@icann.org, lars.eggert@nokia.com, touch@isi.edu, magnus.westerlund@ericsson.com, cheshire@apple.com +# RFC6336 || M. Westerlund, C. Perkins || magnus.westerlund@ericsson.com, csp@csperkins.org +# RFC6337 || S. Okumura, T. Sawada, P. Kyzivat || shinji.okumura@softfront.jp, tu-sawada@kddi.com, pkyzivat@alum.mit.edu +# RFC6338 || V. Giralt, R. McDuff || victoriano@uma.es, r.mcduff@uq.edu.au +# RFC6339 || S. Josefsson, L. Hornquist Astrand || simon@josefsson.org, lha@apple.com +# RFC6340 || R. Presuhn || randy_presuhn@mindspring.com +# RFC6341 || K. Rehor, Ed., L. Portman, Ed., A. Hutton, R. Jain || krehor@cisco.com, leon.portman@nice.com, andrew.hutton@siemens-enterprise.com, rajnish.jain@ipc.com +# RFC6342 || R. Koodli || rkoodli@cisco.com +# RFC6343 || B. Carpenter || brian.e.carpenter@gmail.com +# RFC6344 || G. Bernstein, Ed., D. Caviglia, R. Rabbat, H. van Helvoort || gregb@grotto-networking.com, diego.caviglia@ericsson.com, rabbat@alum.mit.edu, hhelvoort@huawei.com +# RFC6345 || P. Duffy, S. Chakrabarti, R. Cragie, Y. Ohba, Ed., A. Yegin || paduffy@cisco.com, samita.chakrabarti@ericsson.com, robert.cragie@gridmerge.com, yoshihiro.ohba@toshiba.co.jp, a.yegin@partner.samsung.com +# RFC6346 || R. Bush, Ed. || randy@psg.com +# RFC6347 || E. Rescorla, N. Modadugu || ekr@rtfm.com, nagendra@cs.stanford.edu +# RFC6348 || JL. Le Roux, Ed., T. Morin, Ed. || jeanlouis.leroux@orange-ftgroup.com, thomas.morin@orange-ftgroup.com +# RFC6349 || B. Constantine, G. Forget, R. Geib, R. Schrage || barry.constantine@jdsu.com, gilles.forget@sympatico.ca, Ruediger.Geib@telekom.de, reinhard@schrageconsult.com +# RFC6350 || S. Perreault || simon.perreault@viagenie.ca +# RFC6351 || S. Perreault || simon.perreault@viagenie.ca +# RFC6352 || C. Daboo || cyrus@daboo.name +# RFC6353 || W. Hardaker || ietf@hardakers.net +# RFC6354 || Q. Xie || Qiaobing.Xie@gmail.com +# RFC6355 || T. Narten, J. Johnson || narten@us.ibm.com, jarrod.b.johnson@gmail.com +# RFC6356 || C. Raiciu, M. Handley, D. Wischik || costin.raiciu@cs.pub.ro, m.handley@cs.ucl.ac.uk, d.wischik@cs.ucl.ac.uk +# RFC6357 || V. Hilt, E. Noel, C. Shen, A. Abdelal || volker.hilt@alcatel-lucent.com, eric.noel@att.com, charles@cs.columbia.edu, aabdelal@sonusnet.com +# RFC6358 || P. Hoffman || paul.hoffman@vpnc.org +# RFC6359 || S. Ginoza, M. Cotton, A. Morris || sginoza@amsl.com, michelle.cotton@icann.org, amorris@amsl.com +# RFC6360 || R. Housley || housley@vigilsec.com +# RFC6361 || J. Carlson, D. Eastlake 3rd || carlsonj@workingcode.com, d3e3e3@gmail.com +# RFC6362 || K. Meadors, Ed. || kyle@drummondgroup.com +# RFC6363 || M. Watson, A. Begen, V. Roca || watsonm@netflix.com, abegen@cisco.com, vincent.roca@inria.fr +# RFC6364 || A. Begen || abegen@cisco.com +# RFC6365 || P. Hoffman, J. Klensin || paul.hoffman@vpnc.org, john+ietf@jck.com +# RFC6366 || J. Valin, K. Vos || jmvalin@jmvalin.ca, koen.vos@skype.net +# RFC6367 || S. Kanno, M. Kanda || kanno.satoru@po.ntts.co.jp, kanda.masayuki@lab.ntt.co.jp +# RFC6368 || P. Marques, R. Raszuk, K. Patel, K. Kumaki, T. Yamagata || pedro.r.marques@gmail.com, robert@raszuk.net, keyupate@cisco.com, ke-kumaki@kddi.com, to-yamagata@kddi.com +# RFC6369 || E. Haleplidis, O. Koufopavlou, S. Denazis || ehalep@ece.upatras.gr, odysseas@ece.upatras.gr, sdena@upatras.gr +# RFC6370 || M. Bocci, G. Swallow, E. Gray || matthew.bocci@alcatel-lucent.com, swallow@cisco.com, eric.gray@ericsson.com +# RFC6371 || I. Busi, Ed., D. Allan, Ed. || Italo.Busi@alcatel-lucent.com, david.i.allan@ericsson.com +# RFC6372 || N. Sprecher, Ed., A. Farrel, Ed. || nurit.sprecher@nsn.com, adrian@olddog.co.uk +# RFC6373 || L. Andersson, Ed., L. Berger, Ed., L. Fang, Ed., N. Bitar, Ed., E. Gray, Ed. || loa.andersson@ericsson.com, lberger@labn.net, lufang@cisco.com, nabil.n.bitar@verizon.com, Eric.Gray@Ericsson.com +# RFC6374 || D. Frost, S. Bryant || danfrost@cisco.com, stbryant@cisco.com +# RFC6375 || D. Frost, Ed., S. Bryant, Ed. || danfrost@cisco.com, stbryant@cisco.com +# RFC6376 || D. Crocker, Ed., T. Hansen, Ed., M. Kucherawy, Ed. || dcrocker@bbiw.net, tony+dkimov@maillennium.att.com, msk@cloudmark.com +# RFC6377 || M. Kucherawy || msk@cloudmark.com +# RFC6378 || Y. Weingarten, Ed., S. Bryant, E. Osborne, N. Sprecher, A. Fulignoli, Ed. || yaacov.weingarten@nsn.com, stbryant@cisco.com, eosborne@cisco.com, nurit.sprecher@nsn.com, annamaria.fulignoli@ericsson.com +# RFC6379 || L. Law, J. Solinas || lelaw@orion.ncsc.mil, jasolin@orion.ncsc.mil +# RFC6380 || K. Burgin, M. Peck || kwburgi@tycho.ncsc.mil, mpeck@mitre.org +# RFC6381 || R. Gellens, D. Singer, P. Frojdh || rg+ietf@qualcomm.com, singer@apple.com, Per.Frojdh@ericsson.com +# RFC6382 || D. McPherson, R. Donnelly, F. Scalzo || dmcpherson@verisign.com, rdonnelly@verisign.com, fscalzo@verisign.com +# RFC6383 || K. Shiomoto, A. Farrel || shiomoto.kohei@lab.ntt.co.jp, adrian@olddog.co.uk +# RFC6384 || I. van Beijnum || iljitsch@muada.com +# RFC6385 || M. Barnes, A. Doria, H. Alvestrand, B. Carpenter || mary.ietf.barnes@gmail.com, avri@acm.org, harald@alvestrand.no, brian.e.carpenter@gmail.com +# RFC6386 || J. Bankoski, J. Koleszar, L. Quillio, J. Salonen, P. Wilkins, Y. Xu || jimbankoski@google.com, jkoleszar@google.com, louquillio@google.com, jsalonen@google.com, paulwilkins@google.com, yaowu@google.com +# RFC6387 || A. Takacs, L. Berger, D. Caviglia, D. Fedyk, J. Meuric || attila.takacs@ericsson.com, lberger@labn.net, diego.caviglia@ericsson.com, donald.fedyk@alcatel-lucent.com, julien.meuric@orange.com +# RFC6388 || IJ. Wijnands, Ed., I. Minei, Ed., K. Kompella, B. Thomas || ice@cisco.com, ina@juniper.net, kireeti@juniper.net, bobthomas@alum.mit.edu +# RFC6389 || R. Aggarwal, JL. Le Roux || raggarwa_1@yahoo.com, jeanlouis.leroux@orange-ftgroup.com +# RFC6390 || A. Clark, B. Claise || alan.d.clark@telchemy.com, bclaise@cisco.com +# RFC6391 || S. Bryant, Ed., C. Filsfils, U. Drafz, V. Kompella, J. Regan, S. Amante || stbryant@cisco.com, cfilsfil@cisco.com, Ulrich.Drafz@telekom.de, vach.kompella@alcatel-lucent.com, joe.regan@alcatel-lucent.com, shane@level3.net +# RFC6392 || R. Alimi, Ed., A. Rahman, Ed., Y. Yang, Ed. || ralimi@google.com, Akbar.Rahman@InterDigital.com, yry@cs.yale.edu +# RFC6393 || M. Yevstifeyev || evnikita2@gmail.com +# RFC6394 || R. Barnes || rbarnes@bbn.com +# RFC6395 || S. Gulrajani, S. Venaas || sameerg@cisco.com, stig@cisco.com +# RFC6396 || L. Blunk, M. Karir, C. Labovitz || ljb@merit.edu, mkarir@merit.edu, labovit@deepfield.net +# RFC6397 || T. Manderson || terry.manderson@icann.org +# RFC6398 || F. Le Faucheur, Ed. || flefauch@cisco.com +# RFC6401 || F. Le Faucheur, J. Polk, K. Carlberg || flefauch@cisco.com, jmpolk@cisco.com, carlberg@g11.org.uk +# RFC6402 || J. Schaad || jimsch@augustcellars.com +# RFC6403 || L. Zieglar, S. Turner, M. Peck || llziegl@tycho.ncsc.mil, turners@ieca.com, mpeck@alumni.virginia.edu +# RFC6404 || J. Seedorf, S. Niccolini, E. Chen, H. Scholz || jan.seedorf@nw.neclab.eu, saverio.niccolini@.neclab.eu, eric.chen@lab.ntt.co.jp, hendrik.scholz@voipfuture.com +# RFC6405 || A. Uzelac, Ed., Y. Lee, Ed. || adam.uzelac@globalcrossing.com, yiu_lee@cable.comcast.com +# RFC6406 || D. Malas, Ed., J. Livingood, Ed. || d.malas@cablelabs.com, Jason_Livingood@cable.comcast.com +# RFC6407 || B. Weis, S. Rowles, T. Hardjono || bew@cisco.com, sheela@cisco.com, hardjono@mit.edu +# RFC6408 || M. Jones, J. Korhonen, L. Morand || mark@azu.ca, jouni.nospam@gmail.com, lionel.morand@orange-ftgroup.com +# RFC6409 || R. Gellens, J. Klensin || rg+ietf@qualcomm.com, john-ietf@jck.com +# RFC6410 || R. Housley, D. Crocker, E. Burger || housley@vigilsec.com, dcrocker@bbiw.net, eburger@standardstrack.com +# RFC6411 || M. Behringer, F. Le Faucheur, B. Weis || mbehring@cisco.com, flefauch@cisco.com, bew@cisco.com +# RFC6412 || S. Poretsky, B. Imhoff, K. Michielsen || sporetsky@allot.com, bimhoff@planetspork.com, kmichiel@cisco.com +# RFC6413 || S. Poretsky, B. Imhoff, K. Michielsen || sporetsky@allot.com, bimhoff@planetspork.com, kmichiel@cisco.com +# RFC6414 || S. Poretsky, R. Papneja, J. Karthik, S. Vapiwala || sporetsky@allot.com, rajiv.papneja@huawei.com, jkarthik@cisco.com, svapiwal@cisco.com +# RFC6415 || E. Hammer-Lahav, Ed., B. Cook || eran@hueniverse.com, romeda@gmail.com +# RFC6416 || M. Schmidt, F. de Bont, S. Doehla, J. Kim || malte.schmidt@dolby.com, frans.de.bont@philips.com, stefan.doehla@iis.fraunhofer.de, kjh1905m@naver.com +# RFC6417 || P. Eardley, L. Eggert, M. Bagnulo, R. Winter || philip.eardley@bt.com, lars.eggert@nokia.com, marcelo@it.uc3m.es, rolf.winter@neclab.eu +# RFC6418 || M. Blanchet, P. Seite || Marc.Blanchet@viagenie.ca, pierrick.seite@orange.com +# RFC6419 || M. Wasserman, P. Seite || mrw@painless-security.com, pierrick.seite@orange-ftgroup.com +# RFC6420 || Y. Cai, H. Ou || ycai@cisco.com, hou@cisco.com +# RFC6421 || D. Nelson, Ed. || d.b.nelson@comcast.net +# RFC6422 || T. Lemon, Q. Wu || mellon@nominum.com, sunseawq@huawei.com +# RFC6423 || H. Li, L. Martini, J. He, F. Huang || lihan@chinamobile.com, lmartini@cisco.com, hejia@huawei.com, feng.f.huang@alcatel-sbell.com.cn +# RFC6424 || N. Bahadur, K. Kompella, G. Swallow || nitinb@juniper.net, kireeti@juniper.net, swallow@cisco.com +# RFC6425 || S. Saxena, Ed., G. Swallow, Z. Ali, A. Farrel, S. Yasukawa, T. Nadeau || ssaxena@cisco.com, swallow@cisco.com, zali@cisco.com, adrian@olddog.co.uk, yasukawa.seisho@lab.ntt.co.jp, thomas.nadeau@ca.com +# RFC6426 || E. Gray, N. Bahadur, S. Boutros, R. Aggarwal || eric.gray@ericsson.com, nitinb@juniper.net, sboutros@cisco.com, raggarwa_1@yahoo.com +# RFC6427 || G. Swallow, Ed., A. Fulignoli, Ed., M. Vigoureux, Ed., S. Boutros, D. Ward || swallow@cisco.com, annamaria.fulignoli@ericsson.com, martin.vigoureux@alcatel-lucent.com, sboutros@cisco.com, dward@juniper.net +# RFC6428 || D. Allan, Ed., G. Swallow, Ed., J. Drake, Ed. || david.i.allan@ericsson.com, swallow@cisco.com, jdrake@juniper.net +# RFC6429 || M. Bashyam, M. Jethanandani, A. Ramaiah || mbashyam@ocarinanetworks.com, mjethanandani@gmail.com, ananth@cisco.com +# RFC6430 || K. Li, B. Leiba || likepeng@huawei.com, barryleiba@computer.org +# RFC6431 || M. Boucadair, P. Levis, G. Bajko, T. Savolainen, T. Tsou || mohamed.boucadair@orange.com, pierre.levis@orange.com, gabor.bajko@nokia.com, teemu.savolainen@nokia.com, tina.tsou.zouting@huawei.com +# RFC6432 || R. Jesske, L. Liess || r.jesske@telekom.de, L.Liess@telekom.de +# RFC6433 || P. Hoffman || paul.hoffman@vpnc.org +# RFC6434 || E. Jankiewicz, J. Loughney, T. Narten || edward.jankiewicz@sri.com, john.loughney@nokia.com, narten@us.ibm.com +# RFC6435 || S. Boutros, Ed., S. Sivabalan, Ed., R. Aggarwal, Ed., M. Vigoureux, Ed., X. Dai, Ed. || sboutros@cisco.com, msiva@cisco.com, raggarwa_1@yahoo.com, martin.vigoureux@alcatel-lucent.com, dai.xuehui@zte.com.cn +# RFC6436 || S. Amante, B. Carpenter, S. Jiang || shane@level3.net, brian.e.carpenter@gmail.com, shengjiang@huawei.com +# RFC6437 || S. Amante, B. Carpenter, S. Jiang, J. Rajahalme || shane@level3.net, brian.e.carpenter@gmail.com, jiangsheng@huawei.com, jarno.rajahalme@nsn.com +# RFC6438 || B. Carpenter, S. Amante || brian.e.carpenter@gmail.com, shane@level3.net +# RFC6439 || R. Perlman, D. Eastlake, Y. Li, A. Banerjee, F. Hu || Radia@alum.mit.edu, d3e3e3@gmail.com, liyizhou@huawei.com, ayabaner@cisco.com, hu.fangwei@zte.com.cn +# RFC6440 || G. Zorn, Q. Wu, Y. Wang || gwz@net-zen.net, sunseawq@huawei.com, w52006@huawei.com +# RFC6441 || L. Vegoda || leo.vegoda@icann.org +# RFC6442 || J. Polk, B. Rosen, J. Peterson || jmpolk@cisco.com, br@brianrosen.net, jon.peterson@neustar.biz +# RFC6443 || B. Rosen, H. Schulzrinne, J. Polk, A. Newton || br@brianrosen.net, hgs@cs.columbia.edu, jmpolk@cisco.com, andy@hxr.us +# RFC6444 || H. Schulzrinne, L. Liess, H. Tschofenig, B. Stark, A. Kuett || hgs+ecrit@cs.columbia.edu, L.Liess@telekom.de, Hannes.Tschofenig@gmx.net, barbara.stark@att.com, andres.kytt@skype.net +# RFC6445 || T. Nadeau, Ed., A. Koushik, Ed., R. Cetin, Ed. || thomas.nadeau@ca.com, kkoushik@cisco.com, riza.cetin@alcatel.be +# RFC6446 || A. Niemi, K. Kiss, S. Loreto || aki.niemi@nokia.com, krisztian.kiss@nokia.com, salvatore.loreto@ericsson.com +# RFC6447 || R. Mahy, B. Rosen, H. Tschofenig || rohan@ekabal.com, br@brianrosen.net, Hannes.Tschofenig@gmx.net +# RFC6448 || R. Yount || rjy@cmu.edu +# RFC6449 || J. Falk, Ed. || ietf@cybernothing.org +# RFC6450 || S. Venaas || stig@cisco.com +# RFC6451 || A. Forte, H. Schulzrinne || forte@att.com, hgs@cs.columbia.edu +# RFC6452 || P. Faltstrom, Ed., P. Hoffman, Ed. || paf@cisco.com, paul.hoffman@vpnc.org +# RFC6453 || F. Dijkstra, R. Hughes-Jones || Freek.Dijkstra@sara.nl, Richard.Hughes-Jones@dante.net +# RFC6454 || A. Barth || ietf@adambarth.com +# RFC6455 || I. Fette, A. Melnikov || ifette+ietf@google.com, Alexey.Melnikov@isode.com +# RFC6456 || H. Li, R. Zheng, A. Farrel || hongyu.lihongyu@huawei.com, robin@huawei.com, adrian@olddog.co.uk +# RFC6457 || T. Takeda, Ed., A. Farrel || takeda.tomonori@lab.ntt.co.jp, adrian@olddog.co.uk +# RFC6458 || R. Stewart, M. Tuexen, K. Poon, P. Lei, V. Yasevich || randall@lakerest.net, tuexen@fh-muenster.de, ka-cheong.poon@oracle.com, peterlei@cisco.com, vladislav.yasevich@hp.com +# RFC6459 || J. Korhonen, Ed., J. Soininen, B. Patil, T. Savolainen, G. Bajko, K. Iisakkila || jouni.nospam@gmail.com, jonne.soininen@renesasmobile.com, basavaraj.patil@nokia.com, teemu.savolainen@nokia.com, gabor.bajko@nokia.com, kaisu.iisakkila@renesasmobile.com +# RFC6460 || M. Salter, R. Housley || misalte@nsa.gov, housley@vigilsec.com +# RFC6461 || S. Channabasappa, Ed. || sumanth@cablelabs.com +# RFC6462 || A. Cooper || acooper@cdt.org +# RFC6463 || J. Korhonen, Ed., S. Gundavelli, H. Yokota, X. Cui || jouni.nospam@gmail.com, sri.gundavelli@cisco.com, yokota@kddilabs.jp, Xiangsong.Cui@huawei.com +# RFC6464 || J. Lennox, Ed., E. Ivov, E. Marocco || jonathan@vidyo.com, emcho@jitsi.org, enrico.marocco@telecomitalia.it +# RFC6465 || E. Ivov, Ed., E. Marocco, Ed., J. Lennox || emcho@jitsi.org, enrico.marocco@telecomitalia.it, jonathan@vidyo.com +# RFC6466 || G. Salgueiro || gsalguei@cisco.com +# RFC6467 || T. Kivinen || kivinen@iki.fi +# RFC6468 || A. Melnikov, B. Leiba, K. Li || Alexey.Melnikov@isode.com, barryleiba@computer.org, likepeng@huawei.com +# RFC6469 || K. Kobayashi, K. Mishima, S. Casner, C. Bormann || ikob@riken.jp, three@sfc.wide.ad.jp, casner@acm.org, cabo@tzi.org +# RFC6470 || A. Bierman || andy@yumaworks.com +# RFC6471 || C. Lewis, M. Sergeant || clewisbcp@cauce.org, matt@sergeant.org +# RFC6472 || W. Kumari, K. Sriram || warren@kumari.net, ksriram@nist.gov +# RFC6473 || P. Saint-Andre || ietf@stpeter.im +# RFC6474 || K. Li, B. Leiba || likepeng@huawei.com, barryleiba@computer.org +# RFC6475 || G. Keeni, K. Koide, S. Gundavelli, R. Wakikawa || glenn@cysols.com, ka-koide@kddi.com, sgundave@cisco.com, ryuji@us.toyota-itc.com +# RFC6476 || P. Gutmann || pgut001@cs.auckland.ac.nz +# RFC6477 || A. Melnikov, G. Lunt || Alexey.Melnikov@isode.com, graeme.lunt@smhs.co.uk +# RFC6478 || L. Martini, G. Swallow, G. Heron, M. Bocci || lmartini@cisco.com, swallow@cisco.com, giheron@cisco.com, matthew.bocci@alcatel-lucent.com +# RFC6479 || X. Zhang, T. Tsou || xiangyang.zhang@huawei.com, tena@huawei.com +# RFC6480 || M. Lepinski, S. Kent || mlepinski@bbn.com, kent@bbn.com +# RFC6481 || G. Huston, R. Loomans, G. Michaelson || gih@apnic.net, robertl@apnic.net, ggm@apnic.net +# RFC6482 || M. Lepinski, S. Kent, D. Kong || mlepinski@bbn.com, skent@bbn.com, dkong@bbn.com +# RFC6483 || G. Huston, G. Michaelson || gih@apnic.net, ggm@apnic.net +# RFC6484 || S. Kent, D. Kong, K. Seo, R. Watro || skent@bbn.com, dkong@bbn.com, kseo@bbn.com, rwatro@bbn.com +# RFC6485 || G. Huston || gih@apnic.net +# RFC6486 || R. Austein, G. Huston, S. Kent, M. Lepinski || sra@isc.org, gih@apnic.net, kent@bbn.com, mlepinski@bbn.com +# RFC6487 || G. Huston, G. Michaelson, R. Loomans || gih@apnic.net, ggm@apnic.net, robertl@apnic.net +# RFC6488 || M. Lepinski, A. Chi, S. Kent || mlepinski@bbn.com, achi@bbn.com, kent@bbn.com +# RFC6489 || G. Huston, G. Michaelson, S. Kent || gih@apnic.net, ggm@apnic.net, kent@bbn.com +# RFC6490 || G. Huston, S. Weiler, G. Michaelson, S. Kent || gih@apnic.net, weiler@sparta.com, ggm@apnic.net, kent@bbn.com +# RFC6491 || T. Manderson, L. Vegoda, S. Kent || terry.manderson@icann.org, leo.vegoda@icann.org, kent@bbn.com +# RFC6492 || G. Huston, R. Loomans, B. Ellacott, R. Austein || gih@apnic.net, robertl@apnic.net, bje@apnic.net, sra@hactrn.net +# RFC6493 || R. Bush || randy@psg.com +# RFC6494 || R. Gagliano, S. Krishnan, A. Kukec || rogaglia@cisco.com, suresh.krishnan@ericsson.com, ana.kukec@enterprisearchitects.com +# RFC6495 || R. Gagliano, S. Krishnan, A. Kukec || rogaglia@cisco.com, suresh.krishnan@ericsson.com, ana.kukec@enterprisearchitects.com +# RFC6496 || S. Krishnan, J. Laganier, M. Bonola, A. Garcia-Martinez || suresh.krishnan@ericsson.com, julien.ietf@gmail.com, marco.bonola@gmail.com, alberto@it.uc3m.es +# RFC6497 || M. Davis, A. Phillips, Y. Umaoka, C. Falk || mark@macchiato.com, addison@lab126.com, yoshito_umaoka@us.ibm.com, court@infiauto.com +# RFC6498 || J. Stone, R. Kumar, F. Andreasen || joestone@cisco.com, rkumar@cisco.com, fandreas@cisco.com +# RFC6501 || O. Novo, G. Camarillo, D. Morgan, J. Urpalainen || Oscar.Novo@ericsson.com, Gonzalo.Camarillo@ericsson.com, Dave.Morgan@fmr.com, jari.urpalainen@nokia.com +# RFC6502 || G. Camarillo, S. Srinivasan, R. Even, J. Urpalainen || Gonzalo.Camarillo@ericsson.com, srivatsa.srinivasan@gmail.com, ron.even.tlv@gmail.com, jari.urpalainen@nokia.com +# RFC6503 || M. Barnes, C. Boulton, S. Romano, H. Schulzrinne || mary.ietf.barnes@gmail.com, chris@ns-technologies.com, spromano@unina.it, hgs+xcon@cs.columbia.edu +# RFC6504 || M. Barnes, L. Miniero, R. Presta, S P. Romano || mary.ietf.barnes@gmail.com, lorenzo@meetecho.com, roberta.presta@unina.it, spromano@unina.it +# RFC6505 || S. McGlashan, T. Melanchuk, C. Boulton || smcg.stds01@mcglashan.org, timm@rainwillow.com, chris@ns-technologies.com +# RFC6506 || M. Bhatia, V. Manral, A. Lindem || manav.bhatia@alcatel-lucent.com, vishwas.manral@hp.com, acee.lindem@ericsson.com +# RFC6507 || M. Groves || Michael.Groves@cesg.gsi.gov.uk +# RFC6508 || M. Groves || Michael.Groves@cesg.gsi.gov.uk +# RFC6509 || M. Groves || Michael.Groves@cesg.gsi.gov.uk +# RFC6510 || L. Berger, G. Swallow || lberger@labn.net, swallow@cisco.com +# RFC6511 || Z. Ali, G. Swallow, R. Aggarwal || zali@cisco.com, swallow@cisco.com, raggarwa_1@yahoo.com +# RFC6512 || IJ. Wijnands, E. Rosen, M. Napierala, N. Leymann || ice@cisco.com, erosen@cisco.com, mnapierala@att.com, n.leymann@telekom.de +# RFC6513 || E. Rosen, Ed., R. Aggarwal, Ed. || erosen@cisco.com, raggarwa_1@yahoo.com +# RFC6514 || R. Aggarwal, E. Rosen, T. Morin, Y. Rekhter || raggarwa_1@yahoo.com, erosen@cisco.com, thomas.morin@orange-ftgroup.com, yakov@juniper.net +# RFC6515 || R. Aggarwal, E. Rosen || raggarwa_1@yahoo.com, erosen@cisco.com +# RFC6516 || Y. Cai, E. Rosen, Ed., I. Wijnands || ycai@cisco.com, erosen@cisco.com, ice@cisco.com +# RFC6517 || T. Morin, Ed., B. Niven-Jenkins, Ed., Y. Kamite, R. Zhang, N. Leymann, N. Bitar || thomas.morin@orange.com, ben@niven-jenkins.co.uk, y.kamite@ntt.com, raymond.zhang@alcatel-lucent.com, n.leymann@telekom.de, nabil.n.bitar@verizon.com +# RFC6518 || G. Lebovitz, M. Bhatia || gregory.ietf@gmail.com, manav.bhatia@alcatel-lucent.com +# RFC6519 || R. Maglione, A. Durand || roberta.maglione@telecomitalia.it, adurand@juniper.net +# RFC6520 || R. Seggelmann, M. Tuexen, M. Williams || seggelmann@fh-muenster.de, tuexen@fh-muenster.de, michael.glenn.williams@gmail.com +# RFC6521 || A. Makela, J. Korhonen || antti.t.makela@iki.fi, jouni.nospam@gmail.com +# RFC6522 || M. Kucherawy, Ed. || msk@cloudmark.com +# RFC6525 || R. Stewart, M. Tuexen, P. Lei || randall@lakerest.net, tuexen@fh-muenster.de, peterlei@cisco.com +# RFC6526 || B. Claise, P. Aitken, A. Johnson, G. Muenz || bclaise@cisco.com, paitken@cisco.com, andrjohn@cisco.com, muenz@net.in.tum.de +# RFC6527 || K. Tata || tata_kalyan@yahoo.com +# RFC6528 || F. Gont, S. Bellovin || fgont@si6networks.com, bellovin@acm.org +# RFC6529 || A. McKenzie, S. Crocker || amckenzie3@yahoo.com, steve@stevecrocker.com +# RFC6530 || J. Klensin, Y. Ko || john-ietf@jck.com, yangwooko@gmail.com +# RFC6531 || J. Yao, W. Mao || yaojk@cnnic.cn, maowei_ietf@cnnic.cn +# RFC6532 || A. Yang, S. Steele, N. Freed || abelyang@twnic.net.tw, Shawn.Steele@microsoft.com, ned+ietf@mrochek.com +# RFC6533 || T. Hansen, Ed., C. Newman, A. Melnikov || tony+eaidsn@maillennium.att.com, chris.newman@oracle.com, Alexey.Melnikov@isode.com +# RFC6534 || N. Duffield, A. Morton, J. Sommers || duffield@research.att.com, acmorton@att.com, jsommers@colgate.edu +# RFC6535 || B. Huang, H. Deng, T. Savolainen || bill.huang@chinamobile.com, denghui@chinamobile.com, teemu.savolainen@nokia.com +# RFC6536 || A. Bierman, M. Bjorklund || andy@yumaworks.com, mbj@tail-f.com +# RFC6537 || J. Ahrenholz || jeffrey.m.ahrenholz@boeing.com +# RFC6538 || T. Henderson, A. Gurtov || thomas.r.henderson@boeing.com, gurtov@ee.oulu.fi +# RFC6539 || V. Cakulev, G. Sundaram, I. Broustis || violeta.cakulev@alcatel-lucent.com, ganesh.sundaram@alcatel-lucent.com, ioannis.broustis@alcatel-lucent.com +# RFC6540 || W. George, C. Donley, C. Liljenstolpe, L. Howard || wesley.george@twcable.com, C.Donley@cablelabs.com, cdl@asgaard.org, lee.howard@twcable.com +# RFC6541 || M. Kucherawy || msk@cloudmark.com +# RFC6542 || S. Emery || shawn.emery@oracle.com +# RFC6543 || S. Gundavelli || sgundave@cisco.com +# RFC6544 || J. Rosenberg, A. Keranen, B. B. Lowekamp, A. B. Roach || jdrosen@jdrosen.net, ari.keranen@ericsson.com, bbl@lowekamp.net, adam@nostrum.com +# RFC6545 || K. Moriarty || Kathleen.Moriarty@emc.com +# RFC6546 || B. Trammell || trammell@tik.ee.ethz.ch +# RFC6547 || W. George || wesley.george@twcable.com +# RFC6548 || N. Brownlee, Ed., IAB || n.brownlee@auckland.ac.nz, iab@iab.org +# RFC6549 || A. Lindem, A. Roy, S. Mirtorabi || acee.lindem@ericsson.com, akr@cisco.com, sina@cisco.com +# RFC6550 || T. Winter, Ed., P. Thubert, Ed., A. Brandt, J. Hui, R. Kelsey, P. Levis, K. Pister, R. Struik, JP. Vasseur, R. Alexander || wintert@acm.org, pthubert@cisco.com, abr@sdesigns.dk, jhui@archrock.com, kelsey@ember.com, pal@cs.stanford.edu, kpister@dustnetworks.com, rstruik.ext@gmail.com, jpv@cisco.com, roger.alexander@cooperindustries.com +# RFC6551 || JP. Vasseur, Ed., M. Kim, Ed., K. Pister, N. Dejean, D. Barthel || jpv@cisco.com, mjkim@kt.com, kpister@dustnetworks.com, nicolas.dejean@coronis.com, dominique.barthel@orange-ftgroup.com +# RFC6552 || P. Thubert, Ed. || pthubert@cisco.com +# RFC6553 || J. Hui, JP. Vasseur || jonhui@cisco.com, jpv@cisco.com +# RFC6554 || J. Hui, JP. Vasseur, D. Culler, V. Manral || jonhui@cisco.com, jpv@cisco.com, culler@cs.berkeley.edu, vishwas.manral@hp.com +# RFC6555 || D. Wing, A. Yourtchenko || dwing-ietf@fuggles.com, ayourtch@cisco.com +# RFC6556 || F. Baker || fred@cisco.com +# RFC6557 || E. Lear, P. Eggert || lear@cisco.com, eggert@cs.ucla.edu +# RFC6558 || A. Melnikov, B. Leiba, K. Li || Alexey.Melnikov@isode.com, barryleiba@computer.org, likepeng@huawei.com +# RFC6559 || D. Farinacci, IJ. Wijnands, S. Venaas, M. Napierala || dino@cisco.com, ice@cisco.com, stig@cisco.com, mnapierala@att.com +# RFC6560 || G. Richards || gareth.richards@rsa.com +# RFC6561 || J. Livingood, N. Mody, M. O'Reirdan || jason_livingood@cable.comcast.com, nirmal_mody@cable.comcast.com, michael_oreirdan@cable.comcast.com +# RFC6562 || C. Perkins, JM. Valin || csp@csperkins.org, jmvalin@jmvalin.ca +# RFC6563 || S. Jiang, D. Conrad, B. Carpenter || jiangsheng@huawei.com, drc@cloudflare.com, brian.e.carpenter@gmail.com +# RFC6564 || S. Krishnan, J. Woodyatt, E. Kline, J. Hoagland, M. Bhatia || suresh.krishnan@ericsson.com, jhw@apple.com, ek@google.com, Jim_Hoagland@symantec.com, manav.bhatia@alcatel-lucent.com +# RFC6565 || P. Pillay-Esnault, P. Moyer, J. Doyle, E. Ertekin, M. Lundberg || ppe@cisco.com, pete@pollere.net, jdoyle@doyleassociates.net, ertekin_emre@bah.com, lundberg_michael@bah.com +# RFC6566 || Y. Lee, Ed., G. Bernstein, Ed., D. Li, G. Martinelli || leeyoung@huawei.com, gregb@grotto-networking.com, danli@huawei.com, giomarti@cisco.com +# RFC6567 || A. Johnston, L. Liess || alan.b.johnston@gmail.com, laura.liess.dt@gmail.com +# RFC6568 || E. Kim, D. Kaspar, JP. Vasseur || eunah.ietf@gmail.com, dokaspar.ietf@gmail.com, jpv@cisco.com +# RFC6569 || JM. Valin, S. Borilin, K. Vos, C. Montgomery, R. Chen || jmvalin@jmvalin.ca, borilin@spiritdsp.net, koen.vos@skype.net, xiphmont@xiph.org, rchen@broadcom.com +# RFC6570 || J. Gregorio, R. Fielding, M. Hadley, M. Nottingham, D. Orchard || joe@bitworking.org, fielding@gbiv.com, mhadley@mitre.org, mnot@mnot.net, orchard@pacificspirit.com +# RFC6571 || C. Filsfils, Ed., P. Francois, Ed., M. Shand, B. Decraene, J. Uttaro, N. Leymann, M. Horneffer || cf@cisco.com, pierre.francois@imdea.org, imc.shand@googlemail.com, bruno.decraene@orange.com, uttaro@att.com, N.Leymann@telekom.de, Martin.Horneffer@telekom.de +# RFC6572 || F. Xia, B. Sarikaya, J. Korhonen, Ed., S. Gundavelli, D. Damic || xiayangsong@huawei.com, sarikaya@ieee.org, jouni.nospam@gmail.com, sgundave@cisco.com, damjan.damic@siemens.com +# RFC6573 || M. Amundsen || mca@amundsen.com +# RFC6574 || H. Tschofenig, J. Arkko || Hannes.Tschofenig@gmx.net, jari.arkko@piuha.net +# RFC6575 || H. Shah, Ed., E. Rosen, Ed., G. Heron, Ed., V. Kompella, Ed. || hshah@ciena.com, erosen@cisco.com, giheron@cisco.com, vach.kompella@alcatel-lucent.com +# RFC6576 || R. Geib, Ed., A. Morton, R. Fardid, A. Steinmitz || Ruediger.Geib@telekom.de, acmorton@att.com, rfardid@cariden.com, Alexander.Steinmitz@telekom.de +# RFC6577 || M. Kucherawy || msk@cloudmark.com +# RFC6578 || C. Daboo, A. Quillaud || cyrus@daboo.name, arnaud.quillaud@oracle.com +# RFC6579 || M. Yevstifeyev || evnikita2@gmail.com +# RFC6580 || M. Ko, D. Black || mkosjc@gmail.com, david.black@emc.com +# RFC6581 || A. Kanevsky, Ed., C. Bestler, Ed., R. Sharp, S. Wise || arkady.kanevsky@gmail.com, Caitlin.Bestler@nexenta.com, robert.o.sharp@intel.com, swise@opengridcomputing.com +# RFC6582 || T. Henderson, S. Floyd, A. Gurtov, Y. Nishida || thomas.r.henderson@boeing.com, floyd@acm.org, gurtov@ee.oulu.fi, nishida@wide.ad.jp +# RFC6583 || I. Gashinsky, J. Jaeggli, W. Kumari || igor@yahoo-inc.com, jjaeggli@zynga.com, warren@kumari.net +# RFC6584 || V. Roca || vincent.roca@inria.fr +# RFC6585 || M. Nottingham, R. Fielding || mnot@mnot.net, fielding@gbiv.com +# RFC6586 || J. Arkko, A. Keranen || jari.arkko@piuha.net, ari.keranen@ericsson.com +# RFC6587 || R. Gerhards, C. Lonvick || rgerhards@adiscon.com, clonvick@cisco.com +# RFC6588 || C. Ishikawa || chiaki.ishikawa@ubin.jp +# RFC6589 || J. Livingood || jason_livingood@cable.comcast.com +# RFC6590 || J. Falk, Ed., M. Kucherawy, Ed. || ietf@cybernothing.org, msk@cloudmark.com +# RFC6591 || H. Fontana || hilda@hfontana.com +# RFC6592 || C. Pignataro || cpignata@cisco.com +# RFC6593 || C. Pignataro, J. Clarke, G. Salgueiro || cpignata@cisco.com, jclarke@cisco.com, gsalguei@cisco.com +# RFC6594 || O. Sury || ondrej.sury@nic.cz +# RFC6595 || K. Wierenga, E. Lear, S. Josefsson || klaas@cisco.com, lear@cisco.com, simon@josefsson.org +# RFC6596 || M. Ohye, J. Kupke || maileohye@gmail.com, joachim@kupke.za.net +# RFC6597 || J. Downs, Ed., J. Arbeiter, Ed. || jeff_downs@partech.com, jimsgti@gmail.com +# RFC6598 || J. Weil, V. Kuarsingh, C. Donley, C. Liljenstolpe, M. Azinger || jason.weil@twcable.com, victor.kuarsingh@gmail.com, c.donley@cablelabs.com, cdl@asgaard.org, marla.azinger@frontiercorp.com +# RFC6601 || G. Ash, Ed., D. McDysan || gash5107@yahoo.com, dave.mcdysan@verizon.com +# RFC6602 || F. Abinader, Ed., S. Gundavelli, Ed., K. Leung, S. Krishnan, D. Premec || fabinader@gmail.com, sgundave@cisco.com, kleung@cisco.com, suresh.krishnan@ericsson.com, domagoj.premec@gmail.com +# RFC6603 || J. Korhonen, Ed., T. Savolainen, S. Krishnan, O. Troan || jouni.nospam@gmail.com, teemu.savolainen@nokia.com, suresh.krishnan@ericsson.com, ot@cisco.com +# RFC6604 || D. Eastlake 3rd || d3e3e3@gmail.com +# RFC6605 || P. Hoffman, W.C.A. Wijngaards || paul.hoffman@vpnc.org, wouter@nlnetlabs.nl +# RFC6606 || E. Kim, D. Kaspar, C. Gomez, C. Bormann || eunah.ietf@gmail.com, dokaspar.ietf@gmail.com, carlesgo@entel.upc.edu, cabo@tzi.org +# RFC6607 || K. Kinnear, R. Johnson, M. Stapp || kkinnear@cisco.com, raj@cisco.com, mjs@cisco.com +# RFC6608 || J. Dong, M. Chen, A. Suryanarayana || jie.dong@huawei.com, mach.chen@huawei.com, asuryana@cisco.com +# RFC6609 || C. Daboo, A. Stone || cyrus@daboo.name, aaron@serendipity.cx +# RFC6610 || H. Jang, A. Yegin, K. Chowdhury, J. Choi, T. Lemon || heejin.jang@gmail.com, alper.yegin@yegin.org, kc@radiomobiles.com, jinchoe@gmail.com, Ted.Lemon@nominum.com +# RFC6611 || K. Chowdhury, Ed., A. Yegin || kc@radiomobiles.com, alper.yegin@yegin.org +# RFC6612 || G. Giaretta, Ed. || gerardog@qualcomm.com +# RFC6613 || A. DeKok || aland@freeradius.org +# RFC6614 || S. Winter, M. McCauley, S. Venaas, K. Wierenga || stefan.winter@restena.lu, mikem@open.com.au, stig@cisco.com, klaas@cisco.com +# RFC6615 || T. Dietz, Ed., A. Kobayashi, B. Claise, G. Muenz || Thomas.Dietz@neclab.eu, akoba@nttv6.net, bclaise@cisco.com, muenz@net.in.tum.de +# RFC6616 || E. Lear, H. Tschofenig, H. Mauldin, S. Josefsson || lear@cisco.com, Hannes.Tschofenig@gmx.net, hmauldin@cisco.com, simon@josefsson.org +# RFC6617 || D. Harkins || dharkins@arubanetworks.com +# RFC6618 || J. Korhonen, Ed., B. Patil, H. Tschofenig, D. Kroeselberg || jouni.nospam@gmail.com, basavaraj.patil@nokia.com, Hannes.Tschofenig@gmx.net, dirk.kroeselberg@siemens.com +# RFC6619 || J. Arkko, L. Eggert, M. Townsley || jari.arkko@piuha.net, lars@netapp.com, townsley@cisco.com +# RFC6620 || E. Nordmark, M. Bagnulo, E. Levy-Abegnoli || nordmark@acm.org, marcelo@it.uc3m.es, elevyabe@cisco.com +# RFC6621 || J. Macker, Ed. || macker@itd.nrl.navy.mil +# RFC6622 || U. Herberg, T. Clausen || ulrich@herberg.name, T.Clausen@computer.org +# RFC6623 || E. Burger || eburger@standardstrack.com +# RFC6624 || K. Kompella, B. Kothari, R. Cherukuri || kireeti@juniper.net, bhupesh@cisco.com, cherukuri@juniper.net +# RFC6625 || E. Rosen, Ed., Y. Rekhter, Ed., W. Hendrickx, R. Qiu || erosen@cisco.com, yakov@juniper.net, wim.henderickx@alcatel-lucent.be, rayq@huawei.com +# RFC6626 || G. Tsirtsis, V. Park, V. Narayanan, K. Leung || tsirtsis@googlemail.com, vpark@qualcomm.com, vidyan@qualcomm.com, kleung@cisco.com +# RFC6627 || G. Karagiannis, K. Chan, T. Moncaster, M. Menth, P. Eardley, B. Briscoe || g.karagiannis@utwente.nl, khchan.work@gmail.com, Toby.Moncaster@cl.cam.ac.uk, menth@informatik.uni-tuebingen.de, philip.eardley@bt.com, bob.briscoe@bt.com +# RFC6628 || S. Shin, K. Kobara || seonghan.shin@aist.go.jp, kobara_conf@m.aist.go.jp +# RFC6629 || J. Abley, M. Bagnulo, A. Garcia-Martinez || joe.abley@icann.org, marcelo@it.uc3m.es, alberto@it.uc3m.es +# RFC6630 || Z. Cao, H. Deng, Q. Wu, G. Zorn, Ed. || zehn.cao@gmail.com, denghui02@gmail.com, sunseawq@huawei.com, glenzorn@gmail.com +# RFC6631 || D. Kuegler, Y. Sheffer || dennis.kuegler@bsi.bund.de, yaronf.ietf@gmail.com +# RFC6632 || M. Ersue, Ed., B. Claise || mehmet.ersue@nsn.com, bclaise@cisco.com +# RFC6633 || F. Gont || fgont@si6networks.com +# RFC6635 || O. Kolkman, Ed., J. Halpern, Ed., IAB || olaf@nlnetlabs.nl, joel.halpern@ericsson.com, iab@iab.org +# RFC6636 || H. Asaeda, H. Liu, Q. Wu || asaeda@wide.ad.jp, helen.liu@huawei.com, bill.wu@huawei.com +# RFC6637 || A. Jivsov || Andrey_Jivsov@symantec.com +# RFC6638 || C. Daboo, B. Desruisseaux || cyrus@daboo.name, bernard.desruisseaux@oracle.com +# RFC6639 || D. King, Ed., M. Venkatesan, Ed. || daniel@olddog.co.uk, venkat.mahalingams@gmail.com +# RFC6640 || W. George || wesley.george@twcable.com +# RFC6641 || C. Everhart, W. Adamson, J. Zhang || everhart@netapp.com, andros@netapp.com, jiayingz@google.com +# RFC6642 || Q. Wu, Ed., F. Xia, R. Even || sunseawq@huawei.com, xiayangsong@huawei.com, even.roni@huawei.com +# RFC6643 || J. Schoenwaelder || j.schoenwaelder@jacobs-university.de +# RFC6644 || D. Evans, R. Droms, S. Jiang || N7DR@ipfonix.com, rdroms@cisco.com, jiangsheng@huawei.com +# RFC6645 || J. Novak || janovak@cisco.com +# RFC6646 || H. Song, N. Zong, Y. Yang, R. Alimi || haibin.song@huawei.com, zongning@huawei.com, yry@cs.yale.edu, ralimi@google.com +# RFC6647 || M. Kucherawy, D. Crocker || superuser@gmail.com, dcrocker@bbiw.net +# RFC6648 || P. Saint-Andre, D. Crocker, M. Nottingham || ietf@stpeter.im, dcrocker@bbiw.net, mnot@mnot.net +# RFC6649 || L. Hornquist Astrand, T. Yu || lha@apple.com, tlyu@mit.edu +# RFC6650 || J. Falk, M. Kucherawy, Ed. || none, superuser@gmail.com +# RFC6651 || M. Kucherawy || superuser@gmail.com +# RFC6652 || S. Kitterman || scott@kitterman.com +# RFC6653 || B. Sarikaya, F. Xia, T. Lemon || sarikaya@ieee.org, xiayangsong@huawei.com, mellon@nominum.com +# RFC6654 || T. Tsou, C. Zhou, T. Taylor, Q. Chen || Tina.Tsou.Zouting@huawei.com, cathy.zhou@huawei.com, tom.taylor.stds@gmail.com, chenqi.0819@gmail.com +# RFC6655 || D. McGrew, D. Bailey || mcgrew@cisco.com, dbailey@rsa.com +# RFC6656 || R. Johnson, K. Kinnear, M. Stapp || raj@cisco.com, kkinnear@cisco.com, mjs@cisco.com +# RFC6657 || A. Melnikov, J. Reschke || Alexey.Melnikov@isode.com, julian.reschke@greenbytes.de +# RFC6658 || S. Bryant, Ed., L. Martini, G. Swallow, A. Malis || stbryant@cisco.com, lmartini@cisco.com, swallow@cisco.com, andrew.g.malis@verizon.com +# RFC6659 || A. Begen || abegen@cisco.com +# RFC6660 || B. Briscoe, T. Moncaster, M. Menth || bob.briscoe@bt.com, toby.moncaster@cl.cam.ac.uk, menth@uni-tuebingen.de +# RFC6661 || A. Charny, F. Huang, G. Karagiannis, M. Menth, T. Taylor, Ed. || anna@mwsm.com, huangfuqing@huawei.com, g.karagiannis@utwente.nl, menth@uni-tuebingen.de, tom.taylor.stds@gmail.com +# RFC6662 || A. Charny, J. Zhang, G. Karagiannis, M. Menth, T. Taylor, Ed. || anna@mwsm.com, joyzhang@cisco.com, g.karagiannis@utwente.nl, menth@uni-tuebingen.de, tom.taylor.stds@gmail.com +# RFC6663 || G. Karagiannis, T. Taylor, K. Chan, M. Menth, P. Eardley || g.karagiannis@utwente.nl, tom.taylor.stds@gmail.com, khchan.work@gmail.com, menth@uni-tuebingen.de, philip.eardley@bt.com +# RFC6664 || J. Schaad || ietf@augustcellars.com +# RFC6665 || A.B. Roach || adam@nostrum.com +# RFC6666 || N. Hilliard, D. Freedman || nick@inex.ie, david.freedman@uk.clara.net +# RFC6667 || K. Raza, S. Boutros, C. Pignataro || skraza@cisco.com, sboutros@cisco.com, cpignata@cisco.com +# RFC6668 || D. Bider, M. Baushke || ietf-ssh2@denisbider.com, mdb@juniper.net +# RFC6669 || N. Sprecher, L. Fang || nurit.sprecher@nsn.com, lufang@cisco.com +# RFC6670 || N. Sprecher, KY. Hong || nurit.sprecher@nsn.com, hongk@cisco.com +# RFC6671 || M. Betts || malcolm.betts@zte.com.cn +# RFC6672 || S. Rose, W. Wijngaards || scott.rose@nist.gov, wouter@nlnetlabs.nl +# RFC6673 || A. Morton || acmorton@att.com +# RFC6674 || F. Brockners, S. Gundavelli, S. Speicher, D. Ward || fbrockne@cisco.com, sgundave@cisco.com, sebastian.speicher@telekom.de, wardd@cisco.com +# RFC6675 || E. Blanton, M. Allman, L. Wang, I. Jarvinen, M. Kojo, Y. Nishida || elb@psg.com, mallman@icir.org, liliw@juniper.net, ilpo.jarvinen@helsinki.fi, kojo@cs.helsinki.fi, nishida@wide.ad.jp +# RFC6676 || S. Venaas, R. Parekh, G. Van de Velde, T. Chown, M. Eubanks || stig@cisco.com, riparekh@cisco.com, gvandeve@cisco.com, tjc@ecs.soton.ac.uk, marshall.eubanks@iformata.com +# RFC6677 || S. Hartman, Ed., T. Clancy, K. Hoeper || hartmans-ietf@mit.edu, tcc@vt.edu, khoeper@motorolasolutions.com +# RFC6678 || K. Hoeper, S. Hanna, H. Zhou, J. Salowey, Ed. || khoeper@motorolasolutions.com, shanna@juniper.net, hzhou@cisco.com, jsalowey@cisco.com +# RFC6679 || M. Westerlund, I. Johansson, C. Perkins, P. O'Hanlon, K. Carlberg || magnus.westerlund@ericsson.com, ingemar.s.johansson@ericsson.com, csp@csperkins.org, piers.ohanlon@oii.ox.ac.uk, carlberg@g11.org.uk +# RFC6680 || N. Williams, L. Johansson, S. Hartman, S. Josefsson || nico@cryptonector.com, leifj@sunet.se, hartmans-ietf@mit.edu, simon@josefsson.org +# RFC6681 || M. Watson, T. Stockhammer, M. Luby || watsonm@netflix.com, stockhammer@nomor.de, luby@qti.qualcomm.com +# RFC6682 || M. Watson, T. Stockhammer, M. Luby || watsonm@netflix.com, stockhammer@nomor.de, luby@qti.qualcomm.com +# RFC6683 || A. Begen, T. Stockhammer || abegen@cisco.com, stockhammer@nomor.de +# RFC6684 || B. Trammell || trammell@tik.ee.ethz.ch +# RFC6685 || B. Trammell || trammell@tik.ee.ethz.ch +# RFC6686 || M. Kucherawy || superuser@gmail.com +# RFC6687 || J. Tripathi, Ed., J. de Oliveira, Ed., JP. Vasseur, Ed. || jt369@drexel.edu, jau@coe.drexel.edu, jpv@cisco.com +# RFC6688 || D. Black, Ed., J. Glasgow, S. Faibish || david.black@emc.com, jglasgow@google.com, sfaibish@emc.com +# RFC6689 || L. Berger || lberger@labn.net +# RFC6690 || Z. Shelby || zach@sensinode.com +# RFC6691 || D. Borman || david.borman@quantum.com +# RFC6692 || R. Clayton, M. Kucherawy || richard.clayton@cl.cam.ac.uk, superuser@gmail.com +# RFC6693 || A. Lindgren, A. Doria, E. Davies, S. Grasic || andersl@sics.se, avri@acm.org, elwynd@folly.org.uk, samo.grasic@ltu.se +# RFC6694 || S. Moonesamy, Ed. || sm+ietf@elandsys.com +# RFC6695 || R. Asati || rajiva@cisco.com +# RFC6696 || Z. Cao, B. He, Y. Shi, Q. Wu, Ed., G. Zorn, Ed. || caozhen@chinamobile.com, hebaohong@catr.cn, shiyang1@huawei.com, bill.wu@huawei.com, glenzorn@gmail.com +# RFC6697 || G. Zorn, Ed., Q. Wu, T. Taylor, Y. Nir, K. Hoeper, S. Decugis || glenzorn@gmail.com, bill.wu@huawei.com, tom.taylor.stds@gmail.com, ynir@checkpoint.com, khoeper@motorolasolutions.com, sdecugis@freediameter.net +# RFC6698 || P. Hoffman, J. Schlyter || paul.hoffman@vpnc.org, jakob@kirei.se +# RFC6701 || A. Farrel, P. Resnick || adrian@olddog.co.uk, presnick@qti.qualcomm.com +# RFC6702 || T. Polk, P. Saint-Andre || tim.polk@nist.gov, ietf@stpeter.im +# RFC6703 || A. Morton, G. Ramachandran, G. Maguluri || acmorton@att.com, gomathi@att.com, gmaguluri@att.com +# RFC6704 || D. Miles, W. Dec, J. Bristow, R. Maglione || davidmiles@google.com, wdec@cisco.com, James.Bristow@swisscom.com, roberta.maglione@telecomitalia.it +# RFC6705 || S. Krishnan, R. Koodli, P. Loureiro, Q. Wu, A. Dutta || suresh.krishnan@ericsson.com, rkoodli@cisco.com, loureiro@neclab.eu, Sunseawq@huawei.com, adutta@niksun.com +# RFC6706 || F. Templin, Ed. || fltemplin@acm.org +# RFC6707 || B. Niven-Jenkins, F. Le Faucheur, N. Bitar || ben@velocix.com, flefauch@cisco.com, nabil.n.bitar@verizon.com +# RFC6708 || S. Kiesel, Ed., S. Previdi, M. Stiemerling, R. Woundy, Y. Yang || ietf-alto@skiesel.de, sprevidi@cisco.com, martin.stiemerling@neclab.eu, Richard_Woundy@cable.comcast.com, yry@cs.yale.edu +# RFC6709 || B. Carpenter, B. Aboba, Ed., S. Cheshire || brian.e.carpenter@gmail.com, bernard_aboba@hotmail.com, cheshire@apple.com +# RFC6710 || A. Melnikov, K. Carlberg || Alexey.Melnikov@isode.com, carlberg@g11.org.uk +# RFC6711 || L. Johansson || leifj@nordu.net +# RFC6712 || T. Kause, M. Peylo || toka@ssh.com, martin.peylo@nsn.com +# RFC6713 || J. Levine || standards@taugh.com +# RFC6714 || C. Holmberg, S. Blau, E. Burger || christer.holmberg@ericsson.com, staffan.blau@ericsson.com, eburger@standardstrack.com +# RFC6715 || D. Cauchie, B. Leiba, K. Li || dany.cauchie@orange.com, barryleiba@computer.org, likepeng@huawei.com +# RFC6716 || JM. Valin, K. Vos, T. Terriberry || jmvalin@jmvalin.ca, koenvos74@gmail.com, tterriberry@mozilla.com +# RFC6717 || H. Hotz, R. Allbery || hotz@jpl.nasa.gov, rra@stanford.edu +# RFC6718 || P. Muley, M. Aissaoui, M. Bocci || praveen.muley@alcatel-lucent.com, mustapha.aissaoui@alcatel-lucent.com, matthew.bocci@alcatel-lucent.com +# RFC6719 || O. Gnawali, P. Levis || gnawali@cs.uh.edu, pal@cs.stanford.edu +# RFC6720 || C. Pignataro, R. Asati || cpignata@cisco.com, rajiva@cisco.com +# RFC6721 || J. Snell || jasnell@us.ibm.com +# RFC6722 || P. Hoffman, Ed. || paul.hoffman@vpnc.org +# RFC6723 || L. Jin, Ed., R. Key, Ed., S. Delord, T. Nadeau, S. Boutros || lizhong.jin@zte.com.cn, raymond.key@ieee.org, simon.delord@gmail.com, tnadeau@juniper.net, sboutros@cisco.com +# RFC6724 || D. Thaler, Ed., R. Draves, A. Matsumoto, T. Chown || dthaler@microsoft.com, richdr@microsoft.com, arifumi@nttv6.net, tjc@ecs.soton.ac.uk +# RFC6725 || S. Rose || scottr.nist@gmail.com +# RFC6726 || T. Paila, R. Walsh, M. Luby, V. Roca, R. Lehtonen || toni.paila@gmail.com, roderick.walsh@tut.fi, luby@qti.qualcomm.com, vincent.roca@inria.fr, rami.lehtonen@teliasonera.com +# RFC6727 || T. Dietz, Ed., B. Claise, J. Quittek || dietz@neclab.eu, bclaise@cisco.com, quittek@neclab.eu +# RFC6728 || G. Muenz, B. Claise, P. Aitken || muenz@net.in.tum.de, bclaise@cisco.com, paitken@cisco.com +# RFC6729 || D. Crocker, M. Kucherawy || dcrocker@bbiw.net, superuser@gmail.com +# RFC6730 || S. Krishnan, J. Halpern || suresh.krishnan@ericsson.com, joel.halpern@ericsson.com +# RFC6731 || T. Savolainen, J. Kato, T. Lemon || teemu.savolainen@nokia.com, kato@syce.net, Ted.Lemon@nominum.com +# RFC6732 || V. Kuarsingh, Ed., Y. Lee, O. Vautrin || victor.kuarsingh@gmail.com, yiu_lee@cable.comcast.com, olivier@juniper.net +# RFC6733 || V. Fajardo, Ed., J. Arkko, J. Loughney, G. Zorn, Ed. || vf0213@gmail.com, jari.arkko@ericsson.com, john.loughney@nokia.com, glenzorn@gmail.com +# RFC6734 || G. Zorn, Q. Wu, V. Cakulev || glenzorn@gmail.com, sunseawq@huawei.com, violeta.cakulev@alcatel-lucent.com +# RFC6735 || K. Carlberg, Ed., T. Taylor || carlberg@g11.org.uk, tom.taylor.stds@gmail.com +# RFC6736 || F. Brockners, S. Bhandari, V. Singh, V. Fajardo || fbrockne@cisco.com, shwethab@cisco.com, vaneeta.singh@gmail.com, vf0213@gmail.com +# RFC6737 || K. Jiao, G. Zorn || kangjiao@huawei.com, gwz@net-zen.net +# RFC6738 || V. Cakulev, A. Lior, S. Mizikovsky || violeta.cakulev@alcatel-lucent.com, avi.ietf@lior.org, Simon.Mizikovsky@alcatel-lucent.com +# RFC6739 || H. Schulzrinne, H. Tschofenig || hgs+ecrit@cs.columbia.edu, Hannes.Tschofenig@gmx.net +# RFC6740 || RJ Atkinson, SN Bhatti || rja.lists@gmail.com, saleem@cs.st-andrews.ac.uk +# RFC6741 || RJ Atkinson, SN Bhatti || rja.lists@gmail.com, saleem@cs.st-andrews.ac.uk +# RFC6742 || RJ Atkinson, SN Bhatti, S. Rose || rja.lists@gmail.com, saleem@cs.st-andrews.ac.uk, scottr.nist@gmail.com +# RFC6743 || RJ Atkinson, SN Bhatti || rja.lists@gmail.com, saleem@cs.st-andrews.ac.uk +# RFC6744 || RJ Atkinson, SN Bhatti || rja.lists@gmail.com, saleem@cs.st-andrews.ac.uk +# RFC6745 || RJ Atkinson, SN Bhatti || rja.lists@gmail.com, saleem@cs.st-andrews.ac.uk +# RFC6746 || RJ Atkinson, SN Bhatti || rja.lists@gmail.com, saleem@cs.st-andrews.ac.uk +# RFC6747 || RJ Atkinson, SN Bhatti || rja.lists@gmail.com, saleem@cs.st-andrews.ac.uk +# RFC6748 || RJ Atkinson, SN Bhatti || rja.lists@gmail.com, saleem@cs.st-andrews.ac.uk +# RFC6749 || D. Hardt, Ed. || dick.hardt@gmail.com +# RFC6750 || M. Jones, D. Hardt || mbj@microsoft.com, dick.hardt@gmail.com +# RFC6751 || R. Despres, Ed., B. Carpenter, D. Wing, S. Jiang || despres.remi@laposte.net, brian.e.carpenter@gmail.com, dwing-ietf@fuggles.com, shengjiang@huawei.com +# RFC6752 || A. Kirkham || tkirkham@paloaltonetworks.com +# RFC6753 || J. Winterbottom, H. Tschofenig, H. Schulzrinne, M. Thomson || james.winterbottom@commscope.com, Hannes.Tschofenig@gmx.net, hgs@cs.columbia.edu, martin.thomson@skype.net +# RFC6754 || Y. Cai, L. Wei, H. Ou, V. Arya, S. Jethwani || yiqunc@microsoft.com, lwei@cisco.com, hou@cisco.com, varya@directv.com, sjethwani@directv.com +# RFC6755 || B. Campbell, H. Tschofenig || brian.d.campbell@gmail.com, hannes.tschofenig@gmx.net +# RFC6756 || S. Trowbridge, Ed., E. Lear, Ed., G. Fishman, Ed., S. Bradner, Ed. || steve.trowbridge@alcatel-lucent.com, lear@cisco.com, gryfishman@aol.com, sob@harvard.edu +# RFC6757 || S. Gundavelli, Ed., J. Korhonen, Ed., M. Grayson, K. Leung, R. Pazhyannur || sgundave@cisco.com, jouni.nospam@gmail.com, mgrayson@cisco.com, kleung@cisco.com, rpazhyan@cisco.com +# RFC6758 || A. Melnikov, K. Carlberg || Alexey.Melnikov@isode.com, carlberg@g11.org.uk +# RFC6759 || B. Claise, P. Aitken, N. Ben-Dvora || bclaise@cisco.com, paitken@cisco.com, nirbd@cisco.com +# RFC6760 || S. Cheshire, M. Krochmal || cheshire@apple.com, marc@apple.com +# RFC6761 || S. Cheshire, M. Krochmal || cheshire@apple.com, marc@apple.com +# RFC6762 || S. Cheshire, M. Krochmal || cheshire@apple.com, marc@apple.com +# RFC6763 || S. Cheshire, M. Krochmal || cheshire@apple.com, marc@apple.com +# RFC6764 || C. Daboo || cyrus@daboo.name +# RFC6765 || E. Beili, M. Morgenstern || edward.beili@actelis.com, moti.morgenstern@ecitele.com +# RFC6766 || E. Beili || edward.beili@actelis.com +# RFC6767 || E. Beili, M. Morgenstern || edward.beili@actelis.com, moti.morgenstern@ecitele.com +# RFC6768 || E. Beili || edward.beili@actelis.com +# RFC6769 || R. Raszuk, J. Heitz, A. Lo, L. Zhang, X. Xu || robert@raszuk.net, jakob.heitz@ericsson.com, altonlo@aristanetworks.com, lixia@cs.ucla.edu, xuxh@huawei.com +# RFC6770 || G. Bertrand, Ed., E. Stephan, T. Burbridge, P. Eardley, K. Ma, G. Watson || gilles.bertrand@orange.com, emile.stephan@orange.com, trevor.burbridge@bt.com, philip.eardley@bt.com, kevin.ma@azukisystems.com, gwatson@velocix.com +# RFC6771 || L. Eggert, G. Camarillo || lars@netapp.com, Gonzalo.Camarillo@ericsson.com +# RFC6772 || H. Schulzrinne, Ed., H. Tschofenig, Ed., J. Cuellar, J. Polk, J. Morris, M. Thomson || schulzrinne@cs.columbia.edu, Hannes.Tschofenig@gmx.net, Jorge.Cuellar@siemens.com, jmpolk@cisco.com, ietf@jmorris.org, martin.thomson@gmail.com +# RFC6773 || T. Phelan, G. Fairhurst, C. Perkins || tphelan@sonusnet.com, gorry@erg.abdn.ac.uk, csp@csperkins.org +# RFC6774 || R. Raszuk, Ed., R. Fernando, K. Patel, D. McPherson, K. Kumaki || robert@raszuk.net, rex@cisco.com, keyupate@cisco.com, dmcpherson@verisign.com, ke-kumaki@kddi.com +# RFC6775 || Z. Shelby, Ed., S. Chakrabarti, E. Nordmark, C. Bormann || zach@sensinode.com, samita.chakrabarti@ericsson.com, nordmark@cisco.com, cabo@tzi.org +# RFC6776 || A. Clark, Q. Wu || alan.d.clark@telchemy.com, sunseawq@huawei.com +# RFC6777 || W. Sun, Ed., G. Zhang, Ed., J. Gao, G. Xie, R. Papneja || sun.weiqiang@gmail.com, zhangguoying@catr.cn, gjhhit@huawei.com, xieg@cs.ucr.edu, rajiv.papneja@huawei.com +# RFC6778 || R. Sparks || RjS@nostrum.com +# RFC6779 || U. Herberg, R. Cole, I. Chakeres || ulrich@herberg.name, robert.g.cole@us.army.mil, ian.chakeres@gmail.com +# RFC6780 || L. Berger, F. Le Faucheur, A. Narayanan || lberger@labn.net, flefauch@cisco.com, ashokn@cisco.com +# RFC6781 || O. Kolkman, W. Mekking, R. Gieben || olaf@nlnetlabs.nl, matthijs@nlnetlabs.nl, miek.gieben@sidn.nl +# RFC6782 || V. Kuarsingh, Ed., L. Howard || victor.kuarsingh@gmail.com, lee.howard@twcable.com +# RFC6783 || J. Levine, R. Gellens || standards@taugh.com, rg+ietf@qti.qualcomm.com +# RFC6784 || S. Sakane, M. Ishiyama || ssakane@cisco.com, masahiro.ishiyama@toshiba.co.jp +# RFC6785 || B. Leiba || barryleiba@computer.org +# RFC6786 || A. Yegin, R. Cragie || alper.yegin@yegin.org, robert.cragie@gridmerge.com +# RFC6787 || D. Burnett, S. Shanmugham || dburnett@voxeo.com, sarvi@cisco.com +# RFC6788 || S. Krishnan, A. Kavanagh, B. Varga, S. Ooghe, E. Nordmark || suresh.krishnan@ericsson.com, alan.kavanagh@ericsson.com, balazs.a.varga@ericsson.com, sven.ooghe@alcatel-lucent.com, nordmark@cisco.com +# RFC6789 || B. Briscoe, Ed., R. Woundy, Ed., A. Cooper, Ed. || bob.briscoe@bt.com, richard_woundy@cable.comcast.com, acooper@cdt.org +# RFC6790 || K. Kompella, J. Drake, S. Amante, W. Henderickx, L. Yong || kireeti.kompella@gmail.com, jdrake@juniper.net, shane@level3.net, wim.henderickx@alcatel-lucent.com, lucy.yong@huawei.com +# RFC6791 || X. Li, C. Bao, D. Wing, R. Vaithianathan, G. Huston || xing@cernet.edu.cn, congxiao@cernet.edu.cn, dwing-ietf@fuggles.com, rvaithia@cisco.com, gih@apnic.net +# RFC6792 || Q. Wu, Ed., G. Hunt, P. Arden || sunseawq@huawei.com, r.geoff.hunt@gmail.com, philip.arden@bt.com +# RFC6793 || Q. Vohra, E. Chen || quaizar.vohra@gmail.com, enkechen@cisco.com +# RFC6794 || V. Hilt, G. Camarillo, J. Rosenberg || volker.hilt@bell-labs.com, Gonzalo.Camarillo@ericsson.com, jdrosen@jdrosen.net +# RFC6795 || V. Hilt, G. Camarillo || volker.hilt@bell-labs.com, Gonzalo.Camarillo@ericsson.com +# RFC6796 || V. Hilt, G. Camarillo, J. Rosenberg, D. Worley || volker.hilt@bell-labs.com, Gonzalo.Camarillo@ericsson.com, jdrosen@jdrosen.net, worley@ariadne.com +# RFC6797 || J. Hodges, C. Jackson, A. Barth || Jeff.Hodges@PayPal.com, collin.jackson@sv.cmu.edu, ietf@adambarth.com +# RFC6798 || A. Clark, Q. Wu || alan.d.clark@telchemy.com, sunseawq@huawei.com +# RFC6801 || U. Kozat, A. Begen || kozat@docomolabs-usa.com, abegen@cisco.com +# RFC6802 || S. Baillargeon, C. Flinta, A. Johnsson || steve.baillargeon@ericsson.com, christofer.flinta@ericsson.com, andreas.a.johnsson@ericsson.com +# RFC6803 || G. Hudson || ghudson@mit.edu +# RFC6804 || B. Manning || bmanning@sfc.keio.ac.jp +# RFC6805 || D. King, Ed., A. Farrel, Ed. || daniel@olddog.co.uk, adrian@olddog.co.uk +# RFC6806 || S. Hartman, Ed., K. Raeburn, L. Zhu || hartmans-ietf@mit.edu, raeburn@mit.edu, lzhu@microsoft.com +# RFC6807 || D. Farinacci, G. Shepherd, S. Venaas, Y. Cai || dino@cisco.com, gjshep@gmail.com, stig@cisco.com, yiqunc@microsoft.com +# RFC6808 || L. Ciavattone, R. Geib, A. Morton, M. Wieser || lencia@att.com, Ruediger.Geib@telekom.de, acmorton@att.com, matthias_michael.wieser@stud.tu-darmstadt.de +# RFC6809 || C. Holmberg, I. Sedlacek, H. Kaplan || christer.holmberg@ericsson.com, ivo.sedlacek@ericsson.com, hkaplan@acmepacket.com +# RFC6810 || R. Bush, R. Austein || randy@psg.com, sra@hactrn.net +# RFC6811 || P. Mohapatra, J. Scudder, D. Ward, R. Bush, R. Austein || pmohapat@cisco.com, jgs@juniper.net, dward@cisco.com, randy@psg.com, sra@hactrn.net +# RFC6812 || M. Chiba, A. Clemm, S. Medley, J. Salowey, S. Thombare, E. Yedavalli || mchiba@cisco.com, alex@cisco.com, stmedley@cisco.com, jsalowey@cisco.com, thombare@cisco.com, eshwar@cisco.com +# RFC6813 || J. Salowey, S. Hanna || jsalowey@cisco.com, shanna@juniper.net +# RFC6814 || C. Pignataro, F. Gont || cpignata@cisco.com, fgont@si6networks.com +# RFC6815 || S. Bradner, K. Dubray, J. McQuaid, A. Morton || sob@harvard.edu, kdubray@juniper.net, jim@turnipvideo.com, acmorton@att.com +# RFC6816 || V. Roca, M. Cunche, J. Lacan || vincent.roca@inria.fr, mathieu.cunche@inria.fr, jerome.lacan@isae.fr +# RFC6817 || S. Shalunov, G. Hazel, J. Iyengar, M. Kuehlewind || shalunov@shlang.com, greg@bittorrent.com, jiyengar@fandm.edu, mirja.kuehlewind@ikr.uni-stuttgart.de +# RFC6818 || P. Yee || peter@akayla.com +# RFC6819 || T. Lodderstedt, Ed., M. McGloin, P. Hunt || torsten@lodderstedt.net, mark.mcgloin@ie.ibm.com, phil.hunt@yahoo.com +# RFC6820 || T. Narten, M. Karir, I. Foo || narten@us.ibm.com, mkarir@merit.edu, Ian.Foo@huawei.com +# RFC6821 || E. Marocco, A. Fusco, I. Rimac, V. Gurbani || enrico.marocco@telecomitalia.it, antonio2.fusco@telecomitalia.it, rimac@bell-labs.com, vkg@bell-labs.com +# RFC6822 || S. Previdi, Ed., L. Ginsberg, M. Shand, A. Roy, D. Ward || sprevidi@cisco.com, ginsberg@cisco.com, imc.shand@gmail.com, akr@cisco.com, wardd@cisco.com +# RFC6823 || L. Ginsberg, S. Previdi, M. Shand || ginsberg@cisco.com, sprevidi@cisco.com, imc.shand@gmail.com +# RFC6824 || A. Ford, C. Raiciu, M. Handley, O. Bonaventure || alanford@cisco.com, costin.raiciu@cs.pub.ro, m.handley@cs.ucl.ac.uk, olivier.bonaventure@uclouvain.be +# RFC6825 || M. Miyazawa, T. Otani, K. Kumaki, T. Nadeau || ma-miyazawa@kddilabs.jp, Tm-otani@kddi.com, ke-kumaki@kddi.com, tnadeau@juniper.net +# RFC6826 || IJ. Wijnands, Ed., T. Eckert, N. Leymann, M. Napierala || ice@cisco.com, eckert@cisco.com, n.leymann@telekom.de, mnapierala@att.com +# RFC6827 || A. Malis, Ed., A. Lindem, Ed., D. Papadimitriou, Ed. || andrew.g.malis@verizon.com, acee.lindem@ericsson.com, dimitri.papadimitriou@alcatel-lucent.com +# RFC6828 || J. Xia || xiajinwei@huawei.com +# RFC6829 || M. Chen, P. Pan, C. Pignataro, R. Asati || mach@huawei.com, ppan@infinera.com, cpignata@cisco.com, rajiva@cisco.com +# RFC6830 || D. Farinacci, V. Fuller, D. Meyer, D. Lewis || farinacci@gmail.com, vaf@vaf.net, dmm@1-4-5.net, darlewis@cisco.com +# RFC6831 || D. Farinacci, D. Meyer, J. Zwiebel, S. Venaas || farinacci@gmail.com, dmm@cisco.com, jzwiebel@cruzio.com, stig@cisco.com +# RFC6832 || D. Lewis, D. Meyer, D. Farinacci, V. Fuller || darlewis@cisco.com, dmm@1-4-5.net, farinacci@gmail.com, vaf@vaf.net +# RFC6833 || V. Fuller, D. Farinacci || vaf@vaf.net, farinacci@gmail.com +# RFC6834 || L. Iannone, D. Saucez, O. Bonaventure || luigi.iannone@telecom-paristech.fr, damien.saucez@inria.fr, olivier.bonaventure@uclouvain.be +# RFC6835 || D. Farinacci, D. Meyer || farinacci@gmail.com, dmm@cisco.com +# RFC6836 || V. Fuller, D. Farinacci, D. Meyer, D. Lewis || vaf@vaf.net, farinacci@gmail.com, dmm@1-4-5.net, darlewis@cisco.com +# RFC6837 || E. Lear || lear@cisco.com +# RFC6838 || N. Freed, J. Klensin, T. Hansen || ned+ietf@mrochek.com, john+ietf@jck.com, tony+mtsuffix@maillennium.att.com +# RFC6839 || T. Hansen, A. Melnikov || tony+sss@maillennium.att.com, Alexey.Melnikov@isode.com +# RFC6840 || S. Weiler, Ed., D. Blacka, Ed. || weiler@tislabs.com, davidb@verisign.com +# RFC6841 || F. Ljunggren, AM. Eklund Lowinder, T. Okubo || fredrik@kirei.se, amel@iis.se, tomofumi.okubo@icann.org +# RFC6842 || N. Swamy, G. Halwasia, P. Jhingran || nn.swamy@samsung.com, ghalwasi@cisco.com, pjhingra@cisco.com +# RFC6843 || A. Clark, K. Gross, Q. Wu || alan.d.clark@telchemy.com, kevin.gross@avanw.com, sunseawq@huawei.com +# RFC6844 || P. Hallam-Baker, R. Stradling || philliph@comodo.com, rob.stradling@comodo.com +# RFC6845 || N. Sheth, L. Wang, J. Zhang || nsheth@contrailsystems.com, liliw@juniper.net, zzhang@juniper.net +# RFC6846 || G. Pelletier, K. Sandlund, L-E. Jonsson, M. West || ghyslain.pelletier@interdigital.com, kristofer.sandlund@ericsson.com, lars-erik@lejonsson.com, mark.a.west@roke.co.uk +# RFC6847 || D. Melman, T. Mizrahi, D. Eastlake 3rd || davidme@marvell.com, talmi@marvell.com, d3e3e3@gmail.com +# RFC6848 || J. Winterbottom, M. Thomson, R. Barnes, B. Rosen, R. George || james.winterbottom@commscope.com, martin.thomson@gmail.com, rbarnes@bbn.com, br@brianrosen.net, robinsgv@gmail.com +# RFC6849 || H. Kaplan, Ed., K. Hedayat, N. Venna, P. Jones, N. Stratton || hkaplan@acmepacket.com, kh274@cornell.edu, vnagarjuna@saperix.com, paulej@packetizer.com, nathan@robotics.net +# RFC6850 || A. Rijhsinghani, K. Zebrose || anil@charter.net, zebrose@alum.mit.edu +# RFC6851 || A. Gulbrandsen, N. Freed, Ed. || arnt@gulbrandsen.priv.no, ned+ietf@mrochek.com +# RFC6852 || R. Housley, S. Mills, J. Jaffe, B. Aboba, L. St.Amour || housley@vigilsec.com, s.mills@ieee.org, jeff@w3.org, bernard_aboba@hotmail.com, st.amour@isoc.org +# RFC6853 || J. Brzozowski, J. Tremblay, J. Chen, T. Mrugalski || john_brzozowski@cable.comcast.com, jf@jftremblay.com, jack.chen@twcable.com, tomasz.mrugalski@gmail.com +# RFC6854 || B. Leiba || barryleiba@computer.org +# RFC6855 || P. Resnick, Ed., C. Newman, Ed., S. Shen, Ed. || presnick@qti.qualcomm.com, chris.newman@oracle.com, shenshuo@cnnic.cn +# RFC6856 || R. Gellens, C. Newman, J. Yao, K. Fujiwara || rg+ietf@qualcomm.com, chris.newman@oracle.com, yaojk@cnnic.cn, fujiwara@jprs.co.jp +# RFC6857 || K. Fujiwara || fujiwara@jprs.co.jp +# RFC6858 || A. Gulbrandsen || arnt@gulbrandsen.priv.no +# RFC6859 || B. Leiba || barryleiba@computer.org +# RFC6860 || Y. Yang, A. Retana, A. Roy || yiya@cisco.com, aretana@cisco.com, akr@cisco.com +# RFC6861 || I. Dzmanashvili || ioseb.dzmanashvili@gmail.com +# RFC6862 || G. Lebovitz, M. Bhatia, B. Weis || gregory.ietf@gmail.com, manav.bhatia@alcatel-lucent.com, bew@cisco.com +# RFC6863 || S. Hartman, D. Zhang || hartmans-ietf@mit.edu, zhangdacheng@huawei.com +# RFC6864 || J. Touch || touch@isi.edu +# RFC6865 || V. Roca, M. Cunche, J. Lacan, A. Bouabdallah, K. Matsuzono || vincent.roca@inria.fr, mathieu.cunche@inria.fr, jerome.lacan@isae.fr, abouabdallah@cdta.dz, kazuhisa@sfc.wide.ad.jp +# RFC6866 || B. Carpenter, S. Jiang || brian.e.carpenter@gmail.com, jiangsheng@huawei.com +# RFC6867 || Y. Nir, Q. Wu || ynir@checkpoint.com, sunseawq@huawei.com +# RFC6868 || C. Daboo || cyrus@daboo.name +# RFC6869 || G. Salgueiro, J. Clarke, P. Saint-Andre || gsalguei@cisco.com, jclarke@cisco.com, ietf@stpeter.im +# RFC6870 || P. Muley, Ed., M. Aissaoui, Ed. || praveen.muley@alcatel-lucent.com, mustapha.aissaoui@alcatel-lucent.com +# RFC6871 || R. Gilman, R. Even, F. Andreasen || bob_gilman@comcast.net, roni.even@mail01.huawei.com, fandreas@cisco.com +# RFC6872 || V. Gurbani, Ed., E. Burger, Ed., T. Anjali, H. Abdelnur, O. Festor || vkg@bell-labs.com, eburger@standardstrack.com, tricha@ece.iit.edu, humbol@gmail.com, Olivier.Festor@loria.fr +# RFC6873 || G. Salgueiro, V. Gurbani, A. B. Roach || gsalguei@cisco.com, vkg@bell-labs.com, adam@nostrum.com +# RFC6874 || B. Carpenter, S. Cheshire, R. Hinden || brian.e.carpenter@gmail.com, cheshire@apple.com, bob.hinden@gmail.com +# RFC6875 || S. Kamei, T. Momose, T. Inoue, T. Nishitani || skame@nttv6.jp, tmomose@cisco.com, inoue@jp.ntt.net, tomohiro.nishitani@ntt.com +# RFC6876 || P. Sangster, N. Cam-Winget, J. Salowey || paul_sangster@symantec.com, ncamwing@cisco.com, jsalowey@cisco.com +# RFC6877 || M. Mawatari, M. Kawashima, C. Byrne || mawatari@jpix.ad.jp, kawashimam@vx.jp.nec.com, cameron.byrne@t-mobile.com +# RFC6878 || A.B. Roach || adam@nostrum.com +# RFC6879 || S. Jiang, B. Liu, B. Carpenter || jiangsheng@huawei.com, leo.liubing@huawei.com, brian.e.carpenter@gmail.com +# RFC6880 || L. Johansson || leifj@sunet.se +# RFC6881 || B. Rosen, J. Polk || br@brianrosen.net, jmpolk@cisco.com +# RFC6882 || K. Kumaki, Ed., T. Murai, D. Cheng, S. Matsushima, P. Jiang || ke-kumaki@kddi.com, murai@fnsc.co.jp, dean.cheng@huawei.com, satoru.matsushima@g.softbank.co.jp, pe-jiang@kddi.com +# RFC6883 || B. Carpenter, S. Jiang || brian.e.carpenter@gmail.com, jiangsheng@huawei.com +# RFC6884 || Z. Fang || zfang@qualcomm.com +# RFC6885 || M. Blanchet, A. Sullivan || Marc.Blanchet@viagenie.ca, asullivan@dyn.com +# RFC6886 || S. Cheshire, M. Krochmal || cheshire@apple.com, marc@apple.com +# RFC6887 || D. Wing, Ed., S. Cheshire, M. Boucadair, R. Penno, P. Selkirk || dwing-ietf@fuggles.com, cheshire@apple.com, mohamed.boucadair@orange.com, repenno@cisco.com, pselkirk@isc.org +# RFC6888 || S. Perreault, Ed., I. Yamagata, S. Miyakawa, A. Nakagawa, H. Ashida || simon.perreault@viagenie.ca, ikuhei@nttv6.jp, miyakawa@nttv6.jp, a-nakagawa@jpix.ad.jp, hiashida@cisco.com +# RFC6889 || R. Penno, T. Saxena, M. Boucadair, S. Sivakumar || rpenno@juniper.net, tasaxena@cisco.com, mohamed.boucadair@orange.com, ssenthil@cisco.com +# RFC6890 || M. Cotton, L. Vegoda, R. Bonica, Ed., B. Haberman || michelle.cotton@icann.org, leo.vegoda@icann.org, rbonica@juniper.net, brian@innovationslab.net +# RFC6891 || J. Damas, M. Graff, P. Vixie || joao@bondis.org, explorer@flame.org, vixie@isc.org +# RFC6892 || E. Wilde || erik.wilde@emc.com +# RFC6893 || P. Higgs, P. Szucs || paul.higgs@ericsson.com, paul.szucs@eu.sony.com +# RFC6894 || R. Papneja, S. Vapiwala, J. Karthik, S. Poretsky, S. Rao, JL. Le Roux || rajiv.papneja@huawei.com, svapiwal@cisco.com, jkarthik@cisco.com, sporetsky@allot.com, shankar.rao@du.edu, jeanlouis.leroux@orange.com +# RFC6895 || D. Eastlake 3rd || d3e3e3@gmail.com +# RFC6896 || S. Barbato, S. Dorigotti, T. Fossati, Ed. || tat@koanlogic.com, stewy@koanlogic.com, tho@koanlogic.com +# RFC6897 || M. Scharf, A. Ford || michael.scharf@alcatel-lucent.com, alanford@cisco.com +# RFC6898 || D. Li, D. Ceccarelli, L. Berger || huawei.danli@huawei.com, daniele.ceccarelli@ericsson.com, lberger@labn.net +# RFC6901 || P. Bryan, Ed., K. Zyp, M. Nottingham, Ed. || pbryan@anode.ca, kris@sitepen.com, mnot@mnot.net +# RFC6902 || P. Bryan, Ed., M. Nottingham, Ed. || pbryan@anode.ca, mnot@mnot.net +# RFC6903 || J. Snell || jasnell@gmail.com +# RFC6904 || J. Lennox || jonathan@vidyo.com +# RFC6905 || T. Senevirathne, D. Bond, S. Aldrin, Y. Li, R. Watve || tsenevir@cisco.com, mokon@mokon.net, aldrin.ietf@gmail.com, liyizhou@huawei.com, rwatve@cisco.com +# RFC6906 || E. Wilde || erik.wilde@emc.com +# RFC6907 || T. Manderson, K. Sriram, R. White || terry.manderson@icann.org, ksriram@nist.gov, russ@riw.us +# RFC6908 || Y. Lee, R. Maglione, C. Williams, C. Jacquenet, M. Boucadair || yiu_lee@cable.comcast.com, robmgl@cisco.com, carlw@mcsr-labs.org, christian.jacquenet@orange.com, mohamed.boucadair@orange.com +# RFC6909 || S. Gundavelli, Ed., X. Zhou, J. Korhonen, G. Feige, R. Koodli || sgundave@cisco.com, zhou.xingyue@zte.com.cn, jouni.nospam@gmail.com, gfeige@cisco.com, rkoodli@cisco.com +# RFC6910 || D. Worley, M. Huelsemann, R. Jesske, D. Alexeitsev || worley@ariadne.com, martin.huelsemann@telekom.de, r.jesske@telekom.de, alexeitsev@teleflash.com +# RFC6911 || W. Dec, Ed., B. Sarikaya, G. Zorn, Ed., D. Miles, B. Lourdelet || wdec@cisco.com, sarikaya@ieee.org, glenzorn@gmail.com, davidmiles@google.com, blourdel@juniper.net +# RFC6912 || A. Sullivan, D. Thaler, J. Klensin, O. Kolkman || asullivan@dyn.com, dthaler@microsoft.com, john-ietf@jck.com, olaf@NLnetLabs.nl +# RFC6913 || D. Hanes, G. Salgueiro, K. Fleming || dhanes@cisco.com, gsalguei@cisco.com, kevin@kpfleming.us +# RFC6914 || J. Rosenberg || jdrosen@jdrosen.net +# RFC6915 || R. Bellis || ray.bellis@nominet.org.uk +# RFC6916 || R. Gagliano, S. Kent, S. Turner || rogaglia@cisco.com, kent@bbn.com, turners@ieca.com +# RFC6917 || C. Boulton, L. Miniero, G. Munson || chris@ns-technologies.com, lorenzo@meetecho.com, gamunson@gmail.com +# RFC6918 || F. Gont, C. Pignataro || fgont@si6networks.com, cpignata@cisco.com +# RFC6919 || R. Barnes, S. Kent, E. Rescorla || rlb@ipv.sx, kent@bbn.com, ekr@rtfm.com +# RFC6920 || S. Farrell, D. Kutscher, C. Dannewitz, B. Ohlman, A. Keranen, P. Hallam-Baker || stephen.farrell@cs.tcd.ie, kutscher@neclab.eu, cdannewitz@googlemail.com, Borje.Ohlman@ericsson.com, ari.keranen@ericsson.com, philliph@comodo.com +# RFC6921 || R. Hinden || bob.hinden@gmail.com +# RFC6922 || Y. Shafranovich || ietf@shaftek.org +# RFC6923 || R. Winter, E. Gray, H. van Helvoort, M. Betts || rolf.winter@neclab.eu, eric.gray@ericsson.com, huub.van.helvoort@huawei.com, malcolm.betts@zte.com.cn +# RFC6924 || B. Leiba || barryleiba@computer.org +# RFC6925 || B. Joshi, R. Desetti, M. Stapp || bharat_joshi@infosys.com, ramakrishnadtv@infosys.com, mjs@cisco.com +# RFC6926 || K. Kinnear, M. Stapp, R. Desetti, B. Joshi, N. Russell, P. Kurapati, B. Volz || kkinnear@cisco.com, mjs@cisco.com, ramakrishnadtv@infosys.com, bharat_joshi@infosys.com, neil.e.russell@gmail.com, kurapati@juniper.net, volz@cisco.com +# RFC6927 || J. Levine, P. Hoffman || standards@taugh.com, paul.hoffman@vpnc.org +# RFC6928 || J. Chu, N. Dukkipati, Y. Cheng, M. Mathis || hkchu@google.com, nanditad@google.com, ycheng@google.com, mattmathis@google.com +# RFC6929 || A. DeKok, A. Lior || aland@networkradius.com, avi.ietf@lior.org +# RFC6930 || D. Guo, S. Jiang, Ed., R. Despres, R. Maglione || guoseu@huawei.com, jiangsheng@huawei.com, despres.remi@laposte.net, robmgl@cisco.com +# RFC6931 || D. Eastlake 3rd || d3e3e3@gmail.com +# RFC6932 || D. Harkins, Ed. || dharkins@arubanetworks.com +# RFC6933 || A. Bierman, D. Romascanu, J. Quittek, M. Chandramouli || andy@yumaworks.com, dromasca@gmail.com , quittek@neclab.eu, moulchan@cisco.com +# RFC6934 || N. Bitar, Ed., S. Wadhwa, Ed., T. Haag, H. Li || nabil.n.bitar@verizon.com, sanjay.wadhwa@alcatel-lucent.com, HaagT@telekom.de, hongyu.lihongyu@huawei.com +# RFC6935 || M. Eubanks, P. Chimento, M. Westerlund || marshall.eubanks@gmail.com, Philip.Chimento@jhuapl.edu, magnus.westerlund@ericsson.com +# RFC6936 || G. Fairhurst, M. Westerlund || gorry@erg.abdn.ac.uk, magnus.westerlund@ericsson.com +# RFC6937 || M. Mathis, N. Dukkipati, Y. Cheng || mattmathis@google.com, nanditad@google.com, ycheng@google.com +# RFC6938 || J. Scudder || jgs@juniper.net +# RFC6939 || G. Halwasia, S. Bhandari, W. Dec || ghalwasi@cisco.com, shwethab@cisco.com, wdec@cisco.com +# RFC6940 || C. Jennings, B. Lowekamp, Ed., E. Rescorla, S. Baset, H. Schulzrinne || fluffy@cisco.com, bbl@lowekamp.net, ekr@rtfm.com, salman@cs.columbia.edu, hgs@cs.columbia.edu +# RFC6941 || L. Fang, Ed., B. Niven-Jenkins, Ed., S. Mansfield, Ed., R. Graveman, Ed. || lufang@cisco.com, ben@niven-jenkins.co.uk, scott.mansfield@ericsson.com, rfg@acm.org +# RFC6942 || J. Bournelle, L. Morand, S. Decugis, Q. Wu, G. Zorn || julien.bournelle@orange.com, lionel.morand@orange.com, sdecugis@freediameter.net, sunseawq@huawei.com, glenzorn@gmail.com +# RFC6943 || D. Thaler, Ed. || dthaler@microsoft.com +# RFC6944 || S. Rose || scottr.nist@gmail.com +# RFC6945 || R. Bush, B. Wijnen, K. Patel, M. Baer || randy@psg.com, bertietf@bwijnen.net, keyupate@cisco.com, baerm@tislabs.com +# RFC6946 || F. Gont || fgont@si6networks.com +# RFC6947 || M. Boucadair, H. Kaplan, R. Gilman, S. Veikkolainen || mohamed.boucadair@orange.com, hkaplan@acmepacket.com, bob_gilman@comcast.net, Simo.Veikkolainen@nokia.com +# RFC6948 || A. Keranen, J. Arkko || ari.keranen@ericsson.com, jari.arkko@piuha.net +# RFC6949 || H. Flanagan, N. Brownlee || rse@rfc-editor.org, rfc-ise@rfc-editor.org +# RFC6950 || J. Peterson, O. Kolkman, H. Tschofenig, B. Aboba || jon.peterson@neustar.biz, olaf@nlnetlabs.nl, Hannes.Tschofenig@gmx.net, Bernard_aboba@hotmail.com +# RFC6951 || M. Tuexen, R. Stewart || tuexen@fh-muenster.de, randall@lakerest.net +# RFC6952 || M. Jethanandani, K. Patel, L. Zheng || mjethanandani@gmail.com, keyupate@cisco.com, vero.zheng@huawei.com +# RFC6953 || A. Mancuso, Ed., S. Probasco, B. Patil || amancuso@google.com, scott@probasco.me, basavpat@cisco.com +# RFC6954 || J. Merkle, M. Lochter || johannes.merkle@secunet.com, manfred.lochter@bsi.bund.de +# RFC6955 || J. Schaad, H. Prafullchandra || ietf@augustcellars.com, HPrafullchandra@hytrust.com +# RFC6956 || W. Wang, E. Haleplidis, K. Ogawa, C. Li, J. Halpern || wmwang@zjsu.edu.cn, ehalep@ece.upatras.gr, ogawa.kentaro@lab.ntt.co.jp, chuanhuang_li@zjsu.edu.cn, joel.halpern@ericsson.com +# RFC6957 || F. Costa, J-M. Combes, Ed., X. Pougnard, H. Li || fabio.costa@orange.com, jeanmichel.combes@orange.com, xavier.pougnard@orange.com, lihy@huawei.com +# RFC6958 || A. Clark, S. Zhang, J. Zhao, Q. Wu, Ed. || alan.d.clark@telchemy.com, zhangyx@sttri.com.cn, zhaojing@sttri.com.cn, sunseawq@huawei.com +# RFC6959 || D. McPherson, F. Baker, J. Halpern || dmcpherson@verisign.com, fred@cisco.com, joel.halpern@ericsson.com +# RFC6960 || S. Santesson, M. Myers, R. Ankney, A. Malpani, S. Galperin, C. Adams || sts@aaa-sec.com, mmyers@fastq.com, none, ambarish@gmail.com, slava.galperin@gmail.com, cadams@eecs.uottawa.ca +# RFC6961 || Y. Pettersen || yngve@spec-work.net +# RFC6962 || B. Laurie, A. Langley, E. Kasper || benl@google.com, agl@google.com, ekasper@google.com +# RFC6963 || P. Saint-Andre || ietf@stpeter.im +# RFC6964 || F. Templin || fltemplin@acm.org +# RFC6965 || L. Fang, Ed., N. Bitar, R. Zhang, M. Daikoku, P. Pan || lufang@cisco.com, nabil.bitar@verizon.com, raymond.zhang@alcatel-lucent.com, ms-daikoku@kddi.com, ppan@infinera.com +# RFC6967 || M. Boucadair, J. Touch, P. Levis, R. Penno || mohamed.boucadair@orange.com, touch@isi.edu, pierre.levis@orange.com, repenno@cisco.com +# RFC6968 || V. Roca, B. Adamson || vincent.roca@inria.fr, adamson@itd.nrl.navy.mil +# RFC6969 || A. Retana, D. Cheng || aretana@cisco.com, dean.cheng@huawei.com +# RFC6970 || M. Boucadair, R. Penno, D. Wing || mohamed.boucadair@orange.com, repenno@cisco.com, dwing-ietf@fuggles.com +# RFC6971 || U. Herberg, Ed., A. Cardenas, T. Iwao, M. Dow, S. Cespedes || ulrich.herberg@us.fujitsu.com, alvaro.cardenas@me.com, smartnetpro-iwao_std@ml.css.fujitsu.com, m.dow@freescale.com, scespedes@icesi.edu.co +# RFC6972 || Y. Zhang, N. Zong || hishigh@gmail.com, zongning@huawei.com +# RFC6973 || A. Cooper, H. Tschofenig, B. Aboba, J. Peterson, J. Morris, M. Hansen, R. Smith || acooper@cdt.org, Hannes.Tschofenig@gmx.net, bernard_aboba@hotmail.com, jon.peterson@neustar.biz, ietf@jmorris.org, marit.hansen@datenschutzzentrum.de, rhys.smith@ja.net +# RFC6974 || Y. Weingarten, S. Bryant, D. Ceccarelli, D. Caviglia, F. Fondelli, M. Corsi, B. Wu, X. Dai || wyaacov@gmail.com, stbryant@cisco.com, daniele.ceccarelli@ericsson.com, diego.caviglia@ericsson.com, francesco.fondelli@ericsson.com, corsi.marco@gmail.com, wu.bo@zte.com.cn, dai.xuehui@zte.com.cn +# RFC6975 || S. Crocker, S. Rose || steve@shinkuro.com, scottr.nist@gmail.com +# RFC6976 || M. Shand, S. Bryant, S. Previdi, C. Filsfils, P. Francois, O. Bonaventure || imc.shand@googlemail.com, stbryant@cisco.com, sprevidi@cisco.com, cfilsfil@cisco.com, pierre.francois@imdea.org, Olivier.Bonaventure@uclouvain.be +# RFC6977 || M. Boucadair, X. Pougnard || mohamed.boucadair@orange.com, xavier.pougnard@orange.com +# RFC6978 || J. Touch || touch@isi.edu +# RFC6979 || T. Pornin || pornin@bolet.org +# RFC6980 || F. Gont || fgont@si6networks.com +# RFC6981 || S. Bryant, S. Previdi, M. Shand || stbryant@cisco.com, sprevidi@cisco.com, imc.shand@googlemail.com +# RFC6982 || Y. Sheffer, A. Farrel || yaronf.ietf@gmail.com, adrian@olddog.co.uk +# RFC6983 || R. van Brandenburg, O. van Deventer, F. Le Faucheur, K. Leung || ray.vanbrandenburg@tno.nl, oskar.vandeventer@tno.nl, flefauch@cisco.com, kleung@cisco.com +# RFC6984 || W. Wang, K. Ogawa, E. Haleplidis, M. Gao, J. Hadi Salim || wmwang@zjsu.edu.cn, ogawa.kentaro@lab.ntt.co.jp, ehalep@ece.upatras.gr, gaoming@mail.zjgsu.edu.cn, hadi@mojatatu.com +# RFC6985 || A. Morton || acmorton@att.com +# RFC6986 || V. Dolmatov, Ed., A. Degtyarev || dol@cryptocom.ru, alexey@renatasystems.org +# RFC6987 || A. Retana, L. Nguyen, A. Zinin, R. White, D. McPherson || aretana@cisco.com, lhnguyen@cisco.com, alex.zinin@gmail.com, Russ.White@vce.com, dmcpherson@verisign.com +# RFC6988 || J. Quittek, Ed., M. Chandramouli, R. Winter, T. Dietz, B. Claise || quittek@neclab.eu, moulchan@cisco.com, Rolf.Winter@neclab.eu, Thomas.Dietz@neclab.eu, bclaise@cisco.com +# RFC6989 || Y. Sheffer, S. Fluhrer || yaronf.ietf@gmail.com, sfluhrer@cisco.com +# RFC6990 || R. Huang, Q. Wu, H. Asaeda, G. Zorn || rachel.huang@huawei.com, bill.wu@huawei.com, asaeda@nict.go.jp, glenzorn@gmail.com +# RFC6991 || J. Schoenwaelder, Ed. || j.schoenwaelder@jacobs-university.de +# RFC6992 || D. Cheng, M. Boucadair, A. Retana || dean.cheng@huawei.com, mohamed.boucadair@orange.com, aretana@cisco.com +# RFC6993 || P. Saint-Andre || ietf@stpeter.im +# RFC6994 || J. Touch || touch@isi.edu +# RFC6996 || J. Mitchell || Jon.Mitchell@microsoft.com +# RFC6997 || M. Goyal, Ed., E. Baccelli, M. Philipp, A. Brandt, J. Martocci || mukul@uwm.edu, Emmanuel.Baccelli@inria.fr, matthias-philipp@gmx.de, abr@sdesigns.dk, jerald.p.martocci@jci.com +# RFC6998 || M. Goyal, Ed., E. Baccelli, A. Brandt, J. Martocci || mukul@uwm.edu, Emmanuel.Baccelli@inria.fr, abr@sdesigns.dk, jerald.p.martocci@jci.com +# RFC7001 || M. Kucherawy || superuser@gmail.com +# RFC7002 || A. Clark, G. Zorn, Q. Wu || alan.d.clark@telchemy.com, glenzorn@gmail.com, sunseawq@huawei.com +# RFC7003 || A. Clark, R. Huang, Q. Wu, Ed. || alan.d.clark@telchemy.com, Rachel@huawei.com, sunseawq@huawei.com +# RFC7004 || G. Zorn, R. Schott, Q. Wu, Ed., R. Huang || glenzorn@gmail.com, Roland.Schott@telekom.de, sunseawq@huawei.com, Rachel@huawei.com +# RFC7005 || A. Clark, V. Singh, Q. Wu || alan.d.clark@telchemy.com, varun@comnet.tkk.fi, sunseawq@huawei.com +# RFC7006 || M. Garcia-Martin, S. Veikkolainen, R. Gilman || miguel.a.garcia@ericsson.com, simo.veikkolainen@nokia.com, bob_gilman@comcast.net +# RFC7007 || T. Terriberry || tterribe@xiph.org +# RFC7008 || S. Kiyomoto, W. Shin || kiyomoto@kddilabs.jp, ohpato@hanmail.net +# RFC7009 || T. Lodderstedt, Ed., S. Dronia, M. Scurtescu || torsten@lodderstedt.net, sdronia@gmx.de, mscurtescu@google.com +# RFC7010 || B. Liu, S. Jiang, B. Carpenter, S. Venaas, W. George || leo.liubing@huawei.com, jiangsheng@huawei.com, brian.e.carpenter@gmail.com, stig@cisco.com, wesley.george@twcable.com +# RFC7011 || B. Claise, Ed., B. Trammell, Ed., P. Aitken || bclaise@cisco.com, trammell@tik.ee.ethz.ch, paitken@cisco.com +# RFC7012 || B. Claise, Ed., B. Trammell, Ed. || bclaise@cisco.com, trammell@tik.ee.ethz.ch +# RFC7013 || B. Trammell, B. Claise || trammell@tik.ee.ethz.ch, bclaise@cisco.com +# RFC7014 || S. D'Antonio, T. Zseby, C. Henke, L. Peluso || salvatore.dantonio@uniparthenope.it, tanja@caida.org, christian.henke@tektronix.com, lorenzo.peluso@unina.it +# RFC7015 || B. Trammell, A. Wagner, B. Claise || trammell@tik.ee.ethz.ch, arno@wagner.name, bclaise@cisco.com +# RFC7016 || M. Thornburgh || mthornbu@adobe.com +# RFC7017 || R. Sparks || rjsparks@nostrum.com +# RFC7018 || V. Manral, S. Hanna || vishwas.manral@hp.com, shanna@juniper.net +# RFC7019 || J. Buford, M. Kolberg, Ed. || buford@avaya.com, mkolberg@ieee.org +# RFC7020 || R. Housley, J. Curran, G. Huston, D. Conrad || housley@vigilsec.com, jcurran@arin.net, gih@apnic.net, drc@virtualized.org +# RFC7021 || C. Donley, Ed., L. Howard, V. Kuarsingh, J. Berg, J. Doshi || c.donley@cablelabs.com, william.howard@twcable.com, victor@jvknet.com, j.berg@cablelabs.com, jineshd@juniper.net +# RFC7022 || A. Begen, C. Perkins, D. Wing, E. Rescorla || abegen@cisco.com, csp@csperkins.org, dwing-ietf@fuggles.com, ekr@rtfm.com +# RFC7023 || D. Mohan, Ed., N. Bitar, Ed., A. Sajassi, Ed., S. DeLord, P. Niger, R. Qiu || dinmohan@hotmail.com, nabil.n.bitar@verizon.com, sajassi@cisco.com, simon.delord@gmail.com, philippe.niger@orange.com, rqiu@juniper.net +# RFC7024 || H. Jeng, J. Uttaro, L. Jalil, B. Decraene, Y. Rekhter, R. Aggarwal || hj2387@att.com, ju1738@att.com, luay.jalil@verizon.com, bruno.decraene@orange.com, yakov@juniper.net, raggarwa_1@yahoo.com +# RFC7025 || T. Otani, K. Ogaki, D. Caviglia, F. Zhang, C. Margaria || tm-otani@kddi.com, ke-oogaki@kddi.com, diego.caviglia@ericsson.com, zhangfatai@huawei.com, cyril.margaria@coriant.com +# RFC7026 || A. Farrel, S. Bryant || adrian@olddog.co.uk, stbryant@cisco.com +# RFC7027 || J. Merkle, M. Lochter || johannes.merkle@secunet.com, manfred.lochter@bsi.bund.de +# RFC7028 || JC. Zuniga, LM. Contreras, CJ. Bernardos, S. Jeon, Y. Kim || JuanCarlos.Zuniga@InterDigital.com, lmcm@tid.es, cjbc@it.uc3m.es, seiljeon@av.it.pt, yhkim@dcn.ssu.ac.kr +# RFC7029 || S. Hartman, M. Wasserman, D. Zhang || hartmans-ietf@mit.edu, mrw@painless-security.com, zhangdacheng@huawei.com +# RFC7030 || M. Pritikin, Ed., P. Yee, Ed., D. Harkins, Ed. || pritikin@cisco.com, peter@akayla.com, dharkins@arubanetworks.com +# RFC7031 || T. Mrugalski, K. Kinnear || tomasz.mrugalski@gmail.com, kkinnear@cisco.com +# RFC7032 || T. Beckhaus, Ed., B. Decraene, K. Tiruveedhula, M. Konstantynowicz, Ed., L. Martini || thomas.beckhaus@telekom.de, bruno.decraene@orange.com, kishoret@juniper.net, maciek@cisco.com, lmartini@cisco.com +# RFC7033 || P. Jones, G. Salgueiro, M. Jones, J. Smarr || paulej@packetizer.com, gsalguei@cisco.com, mbj@microsoft.com, jsmarr@google.com +# RFC7034 || D. Ross, T. Gondrom || dross@microsoft.com, tobias.gondrom@gondrom.org +# RFC7035 || M. Thomson, B. Rosen, D. Stanley, G. Bajko, A. Thomson || martin.thomson@skype.net, br@brianrosen.net, dstanley@arubanetworks.com, Gabor.Bajko@nokia.com, athomson@lgscout.com +# RFC7036 || R. Housley || housley@vigilsec.com +# RFC7037 || L. Yeh, M. Boucadair || leaf.yeh.sdo@gmail.com, mohamed.boucadair@orange.com +# RFC7038 || R. Ogier || ogier@earthlink.net +# RFC7039 || J. Wu, J. Bi, M. Bagnulo, F. Baker, C. Vogt, Ed. || jianping@cernet.edu.cn, junbi@tsinghua.edu.cn, marcelo@it.uc3m.es, fred@cisco.com, mail@christianvogt.net +# RFC7040 || Y. Cui, J. Wu, P. Wu, O. Vautrin, Y. Lee || yong@csnet1.cs.tsinghua.edu.cn, jianping@cernet.edu.cn, pengwu.thu@gmail.com, Olivier@juniper.net, yiu_lee@cable.comcast.com +# RFC7041 || F. Balus, Ed., A. Sajassi, Ed., N. Bitar, Ed. || florin.balus@alcatel-lucent.com, sajassi@cisco.com, nabil.n.bitar@verizon.com +# RFC7042 || D. Eastlake 3rd, J. Abley || d3e3e3@gmail.com, jabley@dyn.com +# RFC7043 || J. Abley || jabley@dyn.com +# RFC7044 || M. Barnes, F. Audet, S. Schubert, J. van Elburg, C. Holmberg || mary.ietf.barnes@gmail.com, francois.audet@skype.net, shida@ntt-at.com, ietf.hanserik@gmail.com, christer.holmberg@ericsson.com +# RFC7045 || B. Carpenter, S. Jiang || brian.e.carpenter@gmail.com, jiangsheng@huawei.com +# RFC7046 || M. Waehlisch, T. Schmidt, S. Venaas || mw@link-lab.net, Schmidt@informatik.haw-hamburg.de, stig@cisco.com +# RFC7047 || B. Pfaff, B. Davie, Ed. || blp@nicira.com, bsd@nicira.com +# RFC7048 || E. Nordmark, I. Gashinsky || nordmark@acm.org, igor@yahoo-inc.com +# RFC7049 || C. Bormann, P. Hoffman || cabo@tzi.org, paul.hoffman@vpnc.org +# RFC7050 || T. Savolainen, J. Korhonen, D. Wing || teemu.savolainen@nokia.com, jouni.nospam@gmail.com, dwing-ietf@fuggles.com +# RFC7051 || J. Korhonen, Ed., T. Savolainen, Ed. || jouni.nospam@gmail.com, teemu.savolainen@nokia.com +# RFC7052 || G. Schudel, A. Jain, V. Moreno || gschudel@cisco.com, atjain@juniper.net, vimoreno@cisco.com +# RFC7053 || M. Tuexen, I. Ruengeler, R. Stewart || tuexen@fh-muenster.de, i.ruengeler@fh-muenster.de, randall@lakerest.net +# RFC7054 || A. Farrel, H. Endo, R. Winter, Y. Koike, M. Paul || adrian@olddog.co.uk, hideki.endo.es@hitachi.com, rolf.winter@neclab.eu, koike.yoshinori@lab.ntt.co.jp, Manuel.Paul@telekom.de +# RFC7055 || S. Hartman, Ed., J. Howlett || hartmans-ietf@mit.edu, josh.howlett@ja.net +# RFC7056 || S. Hartman, J. Howlett || hartmans-ietf@mit.edu, josh.howlett@ja.net +# RFC7057 || S. Winter, J. Salowey || stefan.winter@restena.lu, jsalowey@cisco.com +# RFC7058 || A. Amirante, T. Castaldi, L. Miniero, S P. Romano || alessandro.amirante@unina.it, tcastaldi@meetecho.com, lorenzo@meetecho.com, spromano@unina.it +# RFC7059 || S. Steffann, I. van Beijnum, R. van Rein || sander@steffann.nl, iljitsch@muada.com, rick@openfortress.nl +# RFC7060 || M. Napierala, E. Rosen, IJ. Wijnands || mnapierala@att.com, erosen@cisco.com, ice@cisco.com +# RFC7061 || R. Sinnema, E. Wilde || remon.sinnema@emc.com, erik.wilde@emc.com +# RFC7062 || F. Zhang, Ed., D. Li, H. Li, S. Belotti, D. Ceccarelli || zhangfatai@huawei.com, huawei.danli@huawei.com, lihan@chinamobile.com, sergio.belotti@alcatel-lucent.it, daniele.ceccarelli@ericsson.com +# RFC7063 || L. Zheng, J. Zhang, R. Parekh || vero.zheng@huawei.com, zzhang@juniper.net, riparekh@cisco.com +# RFC7064 || S. Nandakumar, G. Salgueiro, P. Jones, M. Petit-Huguenin || snandaku@cisco.com, gsalguei@cisco.com, paulej@packetizer.com, petithug@acm.org +# RFC7065 || M. Petit-Huguenin, S. Nandakumar, G. Salgueiro, P. Jones || petithug@acm.org, snandaku@cisco.com, gsalguei@cisco.com, paulej@packetizer.com +# RFC7066 || J. Korhonen, Ed., J. Arkko, Ed., T. Savolainen, S. Krishnan || jouni.nospam@gmail.com, jari.arkko@piuha.net, teemu.savolainen@nokia.com, suresh.krishnan@ericsson.com +# RFC7067 || L. Dunbar, D. Eastlake 3rd, R. Perlman, I. Gashinsky || ldunbar@huawei.com, d3e3e3@gmail.com, Radia@alum.mit.edu, igor@yahoo-inc.com +# RFC7068 || E. McMurry, B. Campbell || emcmurry@computer.org, ben@nostrum.com +# RFC7069 || R. Alimi, A. Rahman, D. Kutscher, Y. Yang, H. Song, K. Pentikousis || ralimi@google.com, Akbar.Rahman@InterDigital.com, dirk.kutscher@neclab.eu, yry@cs.yale.edu, haibin.song@huawei.com, k.pentikousis@eict.de +# RFC7070 || N. Borenstein, M. Kucherawy || nsb@guppylake.com, superuser@gmail.com +# RFC7071 || N. Borenstein, M. Kucherawy || nsb@guppylake.com, superuser@gmail.com +# RFC7072 || N. Borenstein, M. Kucherawy || nsb@guppylake.com, superuser@gmail.com +# RFC7073 || N. Borenstein, M. Kucherawy || nsb@guppylake.com, superuser@gmail.com +# RFC7074 || L. Berger, J. Meuric || lberger@labn.net, julien.meuric@orange.com +# RFC7075 || T. Tsou, R. Hao, T. Taylor, Ed. || tina.tsou.zouting@huawei.com, ruibing_hao@cable.comcast.com, tom.taylor.stds@gmail.com +# RFC7076 || M. Joseph, J. Susoy || mark@p6r.com, jim@p6r.com +# RFC7077 || S. Krishnan, S. Gundavelli, M. Liebsch, H. Yokota, J. Korhonen || suresh.krishnan@ericsson.com, sgundave@cisco.com, marco.liebsch@neclab.eu, yokota@kddilabs.jp, jouni.nospam@gmail.com +# RFC7078 || A. Matsumoto, T. Fujisaki, T. Chown || arifumi@nttv6.net, fujisaki@nttv6.net, tjc@ecs.soton.ac.uk +# RFC7079 || N. Del Regno, Ed., A. Malis, Ed. || nick.delregno@verizon.com, amalis@gmail.com +# RFC7080 || A. Sajassi, S. Salam, N. Bitar, F. Balus || sajassi@cisco.com, ssalam@cisco.com, nabil.n.bitar@verizon.com, florin.balus@nuagenetworks.net +# RFC7081 || E. Ivov, P. Saint-Andre, E. Marocco || emcho@jitsi.org, ietf@stpeter.im, enrico.marocco@telecomitalia.it +# RFC7082 || R. Shekh-Yusef, M. Barnes || rifaat.ietf@gmail.com, mary.ietf.barnes@gmail.com +# RFC7083 || R. Droms || rdroms@cisco.com +# RFC7084 || H. Singh, W. Beebee, C. Donley, B. Stark || shemant@cisco.com, wbeebee@cisco.com, c.donley@cablelabs.com, barbara.stark@att.com +# RFC7085 || J. Levine, P. Hoffman || standards@taugh.com, paul.hoffman@cybersecurity.org +# RFC7086 || A. Keranen, G. Camarillo, J. Maenpaa || Ari.Keranen@ericsson.com, Gonzalo.Camarillo@ericsson.com, Jouni.Maenpaa@ericsson.com +# RFC7087 || H. van Helvoort, Ed., L. Andersson, Ed., N. Sprecher, Ed. || Huub.van.Helvoort@huawei.com, loa@mail01.huawei.com, nurit.sprecher@nsn.com +# RFC7088 || D. Worley || worley@ariadne.com +# RFC7089 || H. Van de Sompel, M. Nelson, R. Sanderson || hvdsomp@gmail.com, mln@cs.odu.edu, azaroth42@gmail.com +# RFC7090 || H. Schulzrinne, H. Tschofenig, C. Holmberg, M. Patel || hgs+ecrit@cs.columbia.edu, Hannes.Tschofenig@gmx.net, christer.holmberg@ericsson.com, Milan.Patel@huawei.com +# RFC7091 || V. Dolmatov, Ed., A. Degtyarev || dol@cryptocom.ru, alexey@renatasystems.org +# RFC7092 || H. Kaplan, V. Pascual || hadriel.kaplan@oracle.com, victor.pascual@quobis.com +# RFC7093 || S. Turner, S. Kent, J. Manger || turners@ieca.com, kent@bbn.com, james.h.manger@team.telstra.com +# RFC7094 || D. McPherson, D. Oran, D. Thaler, E. Osterweil || dmcpherson@verisign.com, oran@cisco.com, dthaler@microsoft.com, eosterweil@verisign.com +# RFC7095 || P. Kewisch || mozilla@kewis.ch +# RFC7096 || S. Belotti, Ed., P. Grandi, D. Ceccarelli, Ed., D. Caviglia, F. Zhang, D. Li || sergio.belotti@alcatel-lucent.com, pietro_vittorio.grandi@alcatel-lucent.com, daniele.ceccarelli@ericsson.com, diego.caviglia@ericsson.com, zhangfatai@huawei.com, danli@huawei.com +# RFC7097 || J. Ott, V. Singh, Ed., I. Curcio || jo@comnet.tkk.fi, varun@comnet.tkk.fi, igor.curcio@nokia.com +# RFC7098 || B. Carpenter, S. Jiang, W. Tarreau || brian.e.carpenter@gmail.com, jiangsheng@huawei.com, willy@haproxy.com +# RFC7100 || P. Resnick || presnick@qti.qualcomm.com +# RFC7101 || S. Ginoza || sginoza@amsl.com +# RFC7102 || JP. Vasseur || jpv@cisco.com +# RFC7103 || M. Kucherawy, G. Shapiro, N. Freed || superuser@gmail.com, gshapiro@proofpoint.com, ned.freed@mrochek.com +# RFC7104 || A. Begen, Y. Cai, H. Ou || abegen@cisco.com, yiqunc@microsoft.com, hou@cisco.com +# RFC7105 || M. Thomson, J. Winterbottom || martin.thomson@gmail.com, a.james.winterbottom@gmail.com +# RFC7106 || E. Ivov || emcho@jitsi.org +# RFC7107 || R. Housley || housley@vigilsec.com +# RFC7108 || J. Abley, T. Manderson || jabley@dyn.com, terry.manderson@icann.org +# RFC7109 || H. Yokota, D. Kim, B. Sarikaya, F. Xia || yokota@kddilabs.jp, dskim@jejutp.or.kr, sarikaya@ieee.org, xiayangsong@huawei.com +# RFC7110 || M. Chen, W. Cao, S. Ning, F. Jounay, S. Delord || mach.chen@huawei.com, wayne.caowei@huawei.com, ning.so@tatacommunications.com, frederic.jounay@orange.ch, simon.delord@alcatel-lucent.com +# RFC7111 || M. Hausenblas, E. Wilde, J. Tennison || mhausenblas@maprtech.com, dret@berkeley.edu, jeni@jenitennison.com +# RFC7112 || F. Gont, V. Manral, R. Bonica || fgont@si6networks.com, vishwas@ionosnetworks.com, rbonica@juniper.net +# RFC7113 || F. Gont || fgont@si6networks.com +# RFC7114 || B. Leiba || barryleiba@computer.org +# RFC7115 || R. Bush || randy@psg.com +# RFC7116 || K. Scott, M. Blanchet || kscott@mitre.org, marc.blanchet@viagenie.ca +# RFC7117 || R. Aggarwal, Ed., Y. Kamite, L. Fang, Y. Rekhter, C. Kodeboniya || raggarwa_1@yahoo.com, y.kamite@ntt.com, lufang@microsoft.com, yakov@juniper.net, chaitk@yahoo.com +# RFC7118 || I. Baz Castillo, J. Millan Villegas, V. Pascual || ibc@aliax.net, jmillan@aliax.net, victor.pascual@quobis.com +# RFC7119 || B. Claise, A. Kobayashi, B. Trammell || bclaise@cisco.com, akoba@nttv6.net, trammell@tik.ee.ethz.ch +# RFC7120 || M. Cotton || michelle.cotton@icann.org +# RFC7121 || K. Ogawa, W. Wang, E. Haleplidis, J. Hadi Salim || k.ogawa@ntt.com, wmwang@mail.zjgsu.edu.cn, ehalep@ece.upatras.gr, hadi@mojatatu.com +# RFC7122 || H. Kruse, S. Jero, S. Ostermann || kruse@ohio.edu, sjero@purdue.edu, ostermann@eecs.ohiou.edu +# RFC7123 || F. Gont, W. Liu || fgont@si6networks.com, liushucheng@huawei.com +# RFC7124 || E. Beili || edward.beili@actelis.com +# RFC7125 || B. Trammell, P. Aitken || trammell@tik.ee.ethz.ch, paitken@cisco.com +# RFC7126 || F. Gont, R. Atkinson, C. Pignataro || fgont@si6networks.com, rja.lists@gmail.com, cpignata@cisco.com +# RFC7127 || O. Kolkman, S. Bradner, S. Turner || olaf@nlnetlabs.nl, sob@harvard.edu, turners@ieca.com +# RFC7128 || R. Bush, R. Austein, K. Patel, H. Gredler, M. Waehlisch || randy@psg.com, sra@hactrn.net, keyupate@cisco.com, hannes@juniper.net, waehlisch@ieee.org +# RFC7129 || R. Gieben, W. Mekking || miek@google.com, matthijs@nlnetlabs.nl +# RFC7130 || M. Bhatia, Ed., M. Chen, Ed., S. Boutros, Ed., M. Binderberger, Ed., J. Haas, Ed. || manav.bhatia@alcatel-lucent.com, mach@huawei.com, sboutros@cisco.com, mbinderb@cisco.com, jhaas@juniper.net +# RFC7131 || M. Barnes, F. Audet, S. Schubert, H. van Elburg, C. Holmberg || mary.ietf.barnes@gmail.com, francois.audet@skype.net, shida@ntt-at.com, ietf.hanserik@gmail.com, christer.holmberg@ericsson.com +# RFC7132 || S. Kent, A. Chi || kent@bbn.com, achi@cs.unc.edu +# RFC7133 || S. Kashima, A. Kobayashi, Ed., P. Aitken || kashima@nttv6.net, akoba@nttv6.net, paitken@cisco.com +# RFC7134 || B. Rosen || br@brianrosen.net +# RFC7135 || J. Polk || jmpolk@cisco.com +# RFC7136 || B. Carpenter, S. Jiang || brian.e.carpenter@gmail.com, jiangsheng@huawei.com +# RFC7137 || A. Retana, S. Ratliff || aretana@cisco.com, sratliff@cisco.com +# RFC7138 || D. Ceccarelli, Ed., F. Zhang, S. Belotti, R. Rao, J. Drake || daniele.ceccarelli@ericsson.com, zhangfatai@huawei.com, sergio.belotti@alcatel-lucent.com, rrao@infinera.com, jdrake@juniper.net +# RFC7139 || F. Zhang, Ed., G. Zhang, S. Belotti, D. Ceccarelli, K. Pithewan || zhangfatai@huawei.com, zhangguoying@mail.ritt.com.cn, sergio.belotti@alcatel-lucent.it, daniele.ceccarelli@ericsson.com, kpithewan@infinera.com +# RFC7140 || L. Jin, F. Jounay, IJ. Wijnands, N. Leymann || lizho.jin@gmail.com, frederic.jounay@orange.ch, ice@cisco.com, n.leymann@telekom.de +# RFC7141 || B. Briscoe, J. Manner || bob.briscoe@bt.com, jukka.manner@aalto.fi +# RFC7142 || M. Shand, L. Ginsberg || imc.shand@googlemail.com, ginsberg@cisco.com +# RFC7143 || M. Chadalapaka, J. Satran, K. Meth, D. Black || cbm@chadalapaka.com, julians@infinidat.com, meth@il.ibm.com, david.black@emc.com +# RFC7144 || F. Knight, M. Chadalapaka || knight@netapp.com, cbm@chadalapaka.com +# RFC7145 || M. Ko, A. Nezhinsky || mkosjc@gmail.com, alexandern@mellanox.com +# RFC7146 || D. Black, P. Koning || david.black@emc.com, paul_koning@Dell.com +# RFC7147 || M. Bakke, P. Venkatesen || mark_bakke@dell.com, prakashvn@hcl.com +# RFC7148 || X. Zhou, J. Korhonen, C. Williams, S. Gundavelli, CJ. Bernardos || zhou.xingyue@zte.com.cn, jouni.nospam@gmail.com, carlw@mcsr-labs.org, sgundave@cisco.com, cjbc@it.uc3m.es +# RFC7149 || M. Boucadair, C. Jacquenet || mohamed.boucadair@orange.com, christian.jacquenet@orange.com +# RFC7150 || F. Zhang, A. Farrel || zhangfatai@huawei.com, adrian@olddog.co.uk +# RFC7151 || P. Hethmon, R. McMurray || phethmon@hethmon.com, robmcm@microsoft.com +# RFC7152 || R. Key, Ed., S. DeLord, F. Jounay, L. Huang, Z. Liu, M. Paul || raymond.key@ieee.org, simon.delord@gmail.com, frederic.jounay@orange.ch, huanglu@chinamobile.com, zhliu@gsta.com, manuel.paul@telekom.de +# RFC7153 || E. Rosen, Y. Rekhter || erosen@cisco.com, yakov@juniper.net +# RFC7154 || S. Moonesamy, Ed. || sm+ietf@elandsys.com +# RFC7155 || G. Zorn, Ed. || glenzorn@gmail.com +# RFC7156 || G. Zorn, Q. Wu, J. Korhonen || glenzorn@gmail.com, bill.wu@huawei.com, jouni.nospam@gmail.com +# RFC7157 || O. Troan, Ed., D. Miles, S. Matsushima, T. Okimoto, D. Wing || ot@cisco.com, davidmiles@google.com, satoru.matsushima@g.softbank.co.jp, t.okimoto@west.ntt.co.jp, dwing-ietf@fuggles.com +# RFC7158 || T. Bray, Ed. || tbray@textuality.com +# RFC7159 || T. Bray, Ed. || tbray@textuality.com +# RFC7160 || M. Petit-Huguenin, G. Zorn, Ed. || petithug@acm.org, glenzorn@gmail.com +# RFC7161 || LM. Contreras, CJ. Bernardos, I. Soto || lmcm@tid.es, cjbc@it.uc3m.es, isoto@it.uc3m.es +# RFC7162 || A. Melnikov, D. Cridland || Alexey.Melnikov@isode.com, dave.cridland@surevine.com +# RFC7163 || C. Holmberg, I. Sedlacek || christer.holmberg@ericsson.com, ivo.sedlacek@ericsson.com +# RFC7164 || K. Gross, R. Brandenburg || kevin.gross@avanw.com, ray.vanbrandenburg@tno.nl +# RFC7165 || R. Barnes || rlb@ipv.sx +# RFC7166 || M. Bhatia, V. Manral, A. Lindem || manav.bhatia@alcatel-lucent.com, vishwas@ionosnetworks.com, acee.lindem@ericsson.com +# RFC7167 || D. Frost, S. Bryant, M. Bocci, L. Berger || frost@mm.st, stbryant@cisco.com, matthew.bocci@alcatel-lucent.com, lberger@labn.net +# RFC7168 || I. Nazar || inazar@deviantart.com +# RFC7169 || S. Turner || turners@ieca.com +# RFC7170 || H. Zhou, N. Cam-Winget, J. Salowey, S. Hanna || hzhou@cisco.com, ncamwing@cisco.com, jsalowey@cisco.com, steve.hanna@infineon.com +# RFC7171 || N. Cam-Winget, P. Sangster || ncamwing@cisco.com, paul_sangster@symantec.com +# RFC7172 || D. Eastlake 3rd, M. Zhang, P. Agarwal, R. Perlman, D. Dutt || d3e3e3@gmail.com, zhangmingui@huawei.com, pagarwal@broadcom.com, Radia@alum.mit.edu, ddutt.ietf@hobbesdutt.com +# RFC7173 || L. Yong, D. Eastlake 3rd, S. Aldrin, J. Hudson || lucy.yong@huawei.com, d3e3e3@gmail.com, sam.aldrin@huawei.com, jon.hudson@gmail.com +# RFC7174 || S. Salam, T. Senevirathne, S. Aldrin, D. Eastlake 3rd || ssalam@cisco.com, tsenevir@cisco.com, sam.aldrin@gmail.com, d3e3e3@gmail.com +# RFC7175 || V. Manral, D. Eastlake 3rd, D. Ward, A. Banerjee || vishwas@ionosnetworks.com, d3e3e3@gmail.com, dward@cisco.com, ayabaner@gmail.com +# RFC7176 || D. Eastlake 3rd, T. Senevirathne, A. Ghanwani, D. Dutt, A. Banerjee || d3e3e3@gmail.com, tsenevir@cisco.com, anoop@alumni.duke.edu, ddutt.ietf@hobbesdutt.com, ayabaner@gmail.com +# RFC7177 || D. Eastlake 3rd, R. Perlman, A. Ghanwani, H. Yang, V. Manral || d3e3e3@gmail.com, radia@alum.mit.edu, anoop@alumni.duke.edu, howardy@cisco.com, vishwas@ionosnetworks.com +# RFC7178 || D. Eastlake 3rd, V. Manral, Y. Li, S. Aldrin, D. Ward || d3e3e3@gmail.com, vishwas@ionosnetworks.com, liyizhou@huawei.com, sam.aldrin@huawei.com, dward@cisco.com +# RFC7179 || D. Eastlake 3rd, A. Ghanwani, V. Manral, Y. Li, C. Bestler || d3e3e3@gmail.com, anoop@alumni.duke.edu, vishwas@ionosnetworks.com, liyizhou@huawei.com, caitlin.bestler@nexenta.com +# RFC7180 || D. Eastlake 3rd, M. Zhang, A. Ghanwani, V. Manral, A. Banerjee || d3e3e3@gmail.com, zhangmingui@huawei.com, anoop@alumni.duke.edu, vishwas@ionosnetworks.com, ayabaner@gmail.com +# RFC7181 || T. Clausen, C. Dearlove, P. Jacquet, U. Herberg || T.Clausen@computer.org, chris.dearlove@baesystems.com, philippe.jacquet@alcatel-lucent.com, ulrich@herberg.name +# RFC7182 || U. Herberg, T. Clausen, C. Dearlove || ulrich@herberg.name, T.Clausen@computer.org, chris.dearlove@baesystems.com +# RFC7183 || U. Herberg, C. Dearlove, T. Clausen || ulrich@herberg.name, chris.dearlove@baesystems.com, T.Clausen@computer.org +# RFC7184 || U. Herberg, R. Cole, T. Clausen || ulrich@herberg.name, robert.g.cole@us.army.mil, T.Clausen@computer.org +# RFC7185 || C. Dearlove, T. Clausen, P. Jacquet || chris.dearlove@baesystems.com, T.Clausen@computer.org, philippe.jacquet@alcatel-lucent.com +# RFC7186 || J. Yi, U. Herberg, T. Clausen || jiazi@jiaziyi.com, ulrich@herberg.name, T.Clausen@computer.org +# RFC7187 || C. Dearlove, T. Clausen || chris.dearlove@baesystems.com, T.Clausen@computer.org +# RFC7188 || C. Dearlove, T. Clausen || chris.dearlove@baesystems.com, T.Clausen@computer.org +# RFC7189 || G. Mirsky || gregory.mirsky@ericsson.com +# RFC7190 || C. Villamizar || curtis@occnc.com +# RFC7191 || R. Housley || housley@vigilsec.com +# RFC7192 || S. Turner || turners@ieca.com +# RFC7193 || S. Turner, R. Housley, J. Schaad || turners@ieca.com, housley@vigilsec.com, ietf@augustcellars.com +# RFC7194 || R. Hartmann || richih.mailinglist@gmail.com +# RFC7195 || M. Garcia-Martin, S. Veikkolainen || miguel.a.garcia@ericsson.com, simo.veikkolainen@nokia.com +# RFC7196 || C. Pelsser, R. Bush, K. Patel, P. Mohapatra, O. Maennel || cristel@iij.ad.jp, randy@psg.com, keyupate@cisco.com, mpradosh@yahoo.com, o@maennel.net +# RFC7197 || A. Begen, Y. Cai, H. Ou || abegen@cisco.com, yiqunc@microsoft.com, hou@cisco.com +# RFC7198 || A. Begen, C. Perkins || abegen@cisco.com, csp@csperkins.org +# RFC7199 || R. Barnes, M. Thomson, J. Winterbottom, H. Tschofenig || rlb@ipv.sx, martin.thomson@gmail.com, a.james.winterbottom@gmail.com, Hannes.Tschofenig@gmx.net +# RFC7200 || C. Shen, H. Schulzrinne, A. Koike || charles@cs.columbia.edu, schulzrinne@cs.columbia.edu, koike.arata@lab.ntt.co.jp +# RFC7201 || M. Westerlund, C. Perkins || magnus.westerlund@ericsson.com, csp@csperkins.org +# RFC7202 || C. Perkins, M. Westerlund || csp@csperkins.org, magnus.westerlund@ericsson.com +# RFC7203 || T. Takahashi, K. Landfield, Y. Kadobayashi || takeshi_takahashi@nict.go.jp, kent_landfield@mcafee.com, youki-k@is.aist-nara.ac.jp +# RFC7204 || T. Haynes || tdh@excfb.com +# RFC7205 || A. Romanow, S. Botzko, M. Duckworth, R. Even, Ed. || allyn@cisco.com, stephen.botzko@polycom.com, mark.duckworth@polycom.com, roni.even@mail01.huawei.com +# RFC7206 || P. Jones, G. Salgueiro, J. Polk, L. Liess, H. Kaplan || paulej@packetizer.com, gsalguei@cisco.com, jmpolk@cisco.com, laura.liess.dt@gmail.com, hadriel.kaplan@oracle.com +# RFC7207 || M. Ortseifen, G. Dickfeld || iso20022@bundesbank.de, iso20022@bundesbank.de +# RFC7208 || S. Kitterman || scott@kitterman.com +# RFC7209 || A. Sajassi, R. Aggarwal, J. Uttaro, N. Bitar, W. Henderickx, A. Isaac || sajassi@cisco.com, raggarwa_1@yahoo.com, uttaro@att.com, nabil.n.bitar@verizon.com, wim.henderickx@alcatel-lucent.com, aisaac71@bloomberg.net +# RFC7210 || R. Housley, T. Polk, S. Hartman, D. Zhang || housley@vigilsec.com, tim.polk@nist.gov, hartmans-ietf@mit.edu, zhangdacheng@huawei.com +# RFC7211 || S. Hartman, D. Zhang || hartmans-ietf@mit.edu, zhangdacheng@huawei.com +# RFC7212 || D. Frost, S. Bryant, M. Bocci || frost@mm.st, stbryant@cisco.com, matthew.bocci@alcatel-lucent.com +# RFC7213 || D. Frost, S. Bryant, M. Bocci || frost@mm.st, stbryant@cisco.com, matthew.bocci@alcatel-lucent.com +# RFC7214 || L. Andersson, C. Pignataro || loa@mail01.huawei.com, cpignata@cisco.com +# RFC7215 || L. Jakab, A. Cabellos-Aparicio, F. Coras, J. Domingo-Pascual, D. Lewis || lojakab@cisco.com, acabello@ac.upc.edu, fcoras@ac.upc.edu, jordi.domingo@ac.upc.edu, darlewis@cisco.com +# RFC7216 || M. Thomson, R. Bellis || martin.thomson@gmail.com, ray.bellis@nominet.org.uk +# RFC7217 || F. Gont || fgont@si6networks.com +# RFC7218 || O. Gudmundsson || ogud@ogud.com +# RFC7219 || M. Bagnulo, A. Garcia-Martinez || marcelo@it.uc3m.es, alberto@it.uc3m.es +# RFC7220 || M. Boucadair, R. Penno, D. Wing || mohamed.boucadair@orange.com, repenno@cisco.com, dwing-ietf@fuggles.com +# RFC7221 || A. Farrel, D. Crocker, Ed. || adrian@olddog.co.uk, dcrocker@bbiw.net +# RFC7222 || M. Liebsch, P. Seite, H. Yokota, J. Korhonen, S. Gundavelli || liebsch@neclab.eu, pierrick.seite@orange.com, yokota@kddilabs.jp, jouni.nospam@gmail.com, sgundave@cisco.com +# RFC7223 || M. Bjorklund || mbj@tail-f.com +# RFC7224 || M. Bjorklund || mbj@tail-f.com +# RFC7225 || M. Boucadair || mohamed.boucadair@orange.com +# RFC7226 || C. Villamizar, Ed., D. McDysan, Ed., S. Ning, A. Malis, L. Yong || curtis@occnc.com, dave.mcdysan@verizon.com, ning.so@tatacommunications.com, agmalis@gmail.com, lucy.yong@huawei.com +# RFC7227 || D. Hankins, T. Mrugalski, M. Siodelski, S. Jiang, S. Krishnan || dhankins@google.com, tomasz.mrugalski@gmail.com, msiodelski@gmail.com, jiangsheng@huawei.com, suresh.krishnan@ericsson.com +# RFC7228 || C. Bormann, M. Ersue, A. Keranen || cabo@tzi.org, mehmet.ersue@nsn.com, ari.keranen@ericsson.com +# RFC7229 || R. Housley || housley@vigilsec.com +# RFC7230 || R. Fielding, Ed., J. Reschke, Ed. || fielding@gbiv.com, julian.reschke@greenbytes.de +# RFC7231 || R. Fielding, Ed., J. Reschke, Ed. || fielding@gbiv.com, julian.reschke@greenbytes.de +# RFC7232 || R. Fielding, Ed., J. Reschke, Ed. || fielding@gbiv.com, julian.reschke@greenbytes.de +# RFC7233 || R. Fielding, Ed., Y. Lafon, Ed., J. Reschke, Ed. || fielding@gbiv.com, ylafon@w3.org, julian.reschke@greenbytes.de +# RFC7234 || R. Fielding, Ed., M. Nottingham, Ed., J. Reschke, Ed. || fielding@gbiv.com, mnot@mnot.net, julian.reschke@greenbytes.de +# RFC7235 || R. Fielding, Ed., J. Reschke, Ed. || fielding@gbiv.com, julian.reschke@greenbytes.de +# RFC7236 || J. Reschke || julian.reschke@greenbytes.de +# RFC7237 || J. Reschke || julian.reschke@greenbytes.de +# RFC7238 || J. Reschke || julian.reschke@greenbytes.de +# RFC7239 || A. Petersson, M. Nilsson || andreas@sbin.se, nilsson@opera.com +# RFC7240 || J. Snell || jasnell@gmail.com +# RFC7241 || S. Dawkins, P. Thaler, D. Romascanu, B. Aboba, Ed. || spencerdawkins.ietf@gmail.com, pthaler@broadcom.com, dromasca@gmail.com , bernard_aboba@hotmail.com +# RFC7242 || M. Demmer, J. Ott, S. Perreault || demmer@cs.berkeley.edu, jo@netlab.tkk.fi, simon@per.reau.lt +# RFC7243 || V. Singh, Ed., J. Ott, I. Curcio || varun@comnet.tkk.fi, jo@comnet.tkk.fi, igor.curcio@nokia.com +# RFC7244 || H. Asaeda, Q. Wu, R. Huang || asaeda@nict.go.jp, bill.wu@huawei.com, Rachel@huawei.com +# RFC7245 || A. Hutton, Ed., L. Portman, Ed., R. Jain, K. Rehor || andrew.hutton@unify.com, leon.portman@gmail.com, rajnish.jain@outlook.com, krehor@cisco.com +# RFC7246 || IJ. Wijnands, Ed., P. Hitchen, N. Leymann, W. Henderickx, A. Gulko, J. Tantsura || ice@cisco.com, paul.hitchen@bt.com, n.leymann@telekom.de, wim.henderickx@alcatel-lucent.com, arkadiy.gulko@thomsonreuters.com, jeff.tantsura@ericsson.com +# RFC7247 || P. Saint-Andre, A. Houri, J. Hildebrand || ietf@stpeter.im, avshalom@il.ibm.com, jhildebr@cisco.com +# RFC7248 || P. Saint-Andre, A. Houri, J. Hildebrand || ietf@stpeter.im, avshalom@il.ibm.com, jhildebr@cisco.com +# RFC7249 || R. Housley || housley@vigilsec.com +# RFC7250 || P. Wouters, Ed., H. Tschofenig, Ed., J. Gilmore, S. Weiler, T. Kivinen || pwouters@redhat.com, Hannes.Tschofenig@gmx.net, gnu@toad.com, weiler@tislabs.com, kivinen@iki.fi +# RFC7251 || D. McGrew, D. Bailey, M. Campagna, R. Dugal || mcgrew@cisco.com, danbailey@sth.rub.de, mcampagna@gmail.com, rdugal@certicom.com +# RFC7252 || Z. Shelby, K. Hartke, C. Bormann || zach.shelby@arm.com, hartke@tzi.org, cabo@tzi.org +# RFC7253 || T. Krovetz, P. Rogaway || ted@krovetz.net, rogaway@cs.ucdavis.edu +# RFC7254 || M. Montemurro, Ed., A. Allen, D. McDonald, P. Gosden || mmontemurro@blackberry.com, aallen@blackberry.com, david.mcdonald@meteor.ie, pgosden@gsma.com +# RFC7255 || A. Allen, Ed. || aallen@blackberry.com +# RFC7256 || F. Le Faucheur, R. Maglione, T. Taylor || flefauch@cisco.com, robmgl@cisco.com, tom.taylor.stds@gmail.com +# RFC7257 || T. Nadeau, Ed., A. Kiran Koushik, Ed., R. Mediratta, Ed. || tnadeau@lucidvision.com, kkoushik@brocade.com, romedira@cisco.com +# RFC7258 || S. Farrell, H. Tschofenig || stephen.farrell@cs.tcd.ie, Hannes.Tschofenig@gmx.net +# RFC7259 || P. Saint-Andre || ietf@stpeter.im +# RFC7260 || A. Takacs, D. Fedyk, J. He || attila.takacs@ericsson.com, don.fedyk@hp.com, hejia@huawei.com +# RFC7261 || M. Perumal, P. Ravindran || mperumal@cisco.com, partha@parthasarathi.co.in +# RFC7262 || A. Romanow, S. Botzko, M. Barnes || allyn@cisco.com, stephen.botzko@polycom.com, mary.ietf.barnes@gmail.com +# RFC7263 || N. Zong, X. Jiang, R. Even, Y. Zhang || zongning@huawei.com, jiang.x.f@huawei.com, roni.even@mail01.huawei.com, hishigh@gmail.com +# RFC7264 || N. Zong, X. Jiang, R. Even, Y. Zhang || zongning@huawei.com, jiang.x.f@huawei.com, roni.even@mail01.huawei.com, hishigh@gmail.com +# RFC7265 || P. Kewisch, C. Daboo, M. Douglass || mozilla@kewis.ch, cyrus@daboo.name, douglm@rpi.edu +# RFC7266 || A. Clark, Q. Wu, R. Schott, G. Zorn || alan.d.clark@telchemy.com, sunseawq@huawei.com, Roland.Schott@telekom.de, gwz@net-zen.net +# RFC7267 || L. Martini, Ed., M. Bocci, Ed., F. Balus, Ed. || lmartini@cisco.com, matthew.bocci@alcatel-lucent.com, florin@nuagenetworks.net +# RFC7268 || B. Aboba, J. Malinen, P. Congdon, J. Salowey, M. Jones || Bernard_Aboba@hotmail.com, j@w1.fi, paul.congdon@tallac.com, jsalowey@cisco.com, mark@azu.ca +# RFC7269 || G. Chen, Z. Cao, C. Xie, D. Binet || chengang@chinamobile.com, caozhen@chinamobile.com, xiechf@ctbri.com.cn, david.binet@orange.com +# RFC7270 || A. Yourtchenko, P. Aitken, B. Claise || ayourtch@cisco.com, paitken@cisco.com, bclaise@cisco.com +# RFC7271 || J. Ryoo, Ed., E. Gray, Ed., H. van Helvoort, A. D'Alessandro, T. Cheung, E. Osborne || ryoo@etri.re.kr, eric.gray@ericsson.com, huub.van.helvoort@huawei.com, alessandro.dalessandro@telecomitalia.it, cts@etri.re.kr, eric.osborne@notcom.com +# RFC7272 || R. van Brandenburg, H. Stokking, O. van Deventer, F. Boronat, M. Montagud, K. Gross || ray.vanbrandenburg@tno.nl, hans.stokking@tno.nl, oskar.vandeventer@tno.nl, fboronat@dcom.upv.es, mamontor@posgrado.upv.es, kevin.gross@avanw.com +# RFC7273 || A. Williams, K. Gross, R. van Brandenburg, H. Stokking || aidan.williams@audinate.com, kevin.gross@avanw.com, ray.vanbrandenburg@tno.nl, hans.stokking@tno.nl +# RFC7274 || K. Kompella, L. Andersson, A. Farrel || kireeti.kompella@gmail.com, loa@mail01.huawei.com, adrian@olddog.co.uk +# RFC7275 || L. Martini, S. Salam, A. Sajassi, M. Bocci, S. Matsushima, T. Nadeau || lmartini@cisco.com, ssalam@cisco.com, sajassi@cisco.com, matthew.bocci@alcatel-lucent.com, satoru.matsushima@gmail.com, tnadeau@brocade.com +# RFC7276 || T. Mizrahi, N. Sprecher, E. Bellagamba, Y. Weingarten || talmi@marvell.com, nurit.sprecher@nsn.com, elisa.bellagamba@ericsson.com, wyaacov@gmail.com +# RFC7277 || M. Bjorklund || mbj@tail-f.com +# RFC7278 || C. Byrne, D. Drown, A. Vizdal || cameron.byrne@t-mobile.com, dan@drown.org, ales.vizdal@t-mobile.cz +# RFC7279 || M. Shore, C. Pignataro || melinda.shore@nomountain.net, cpignata@cisco.com +# RFC7280 || G. Fairhurst || gorry@erg.abdn.ac.uk +# RFC7281 || A. Melnikov || Alexey.Melnikov@isode.com +# RFC7282 || P. Resnick || presnick@qti.qualcomm.com +# RFC7283 || Y. Cui, Q. Sun, T. Lemon || yong@csnet1.cs.tsinghua.edu.cn, sunqi@csnet1.cs.tsinghua.edu.cn, Ted.Lemon@nominum.com +# RFC7284 || M. Lanthaler || mail@markus-lanthaler.com +# RFC7285 || R. Alimi, Ed., R. Penno, Ed., Y. Yang, Ed., S. Kiesel, S. Previdi, W. Roome, S. Shalunov, R. Woundy || ralimi@google.com, repenno@cisco.com, yry@cs.yale.edu, ietf-alto@skiesel.de, sprevidi@cisco.com, w.roome@alcatel-lucent.com, shalunov@shlang.com, Richard_Woundy@cable.comcast.com +# RFC7286 || S. Kiesel, M. Stiemerling, N. Schwan, M. Scharf, H. Song || ietf-alto@skiesel.de, mls.ietf@gmail.com, ietf@nico-schwan.de, michael.scharf@alcatel-lucent.com, haibin.song@huawei.com +# RFC7287 || T. Schmidt, Ed., S. Gao, H. Zhang, M. Waehlisch || Schmidt@informatik.haw-hamburg.de, shgao@bjtu.edu.cn, hkzhang@bjtu.edu.cn, mw@link-lab.net +# RFC7288 || D. Thaler || dthaler@microsoft.com +# RFC7289 || V. Kuarsingh, Ed., J. Cianfarani || victor@jvknet.com, john.cianfarani@rci.rogers.com +# RFC7290 || L. Ciavattone, R. Geib, A. Morton, M. Wieser || lencia@att.com, Ruediger.Geib@telekom.de, acmorton@att.com, matthias_michael.wieser@stud.tu-darmstadt.de +# RFC7291 || M. Boucadair, R. Penno, D. Wing || mohamed.boucadair@orange.com, repenno@cisco.com, dwing-ietf@fuggles.com +# RFC7292 || K. Moriarty, Ed., M. Nystrom, S. Parkinson, A. Rusch, M. Scott || Kathleen.Moriarty@emc.com, mnystrom@microsoft.com, sean.parkinson@rsa.com, andreas.rusch@rsa.com, michael2.scott@rsa.com +# RFC7293 || W. Mills, M. Kucherawy || wmills_92105@yahoo.com, msk@fb.com +# RFC7294 || A. Clark, G. Zorn, C. Bi, Q. Wu || alan.d.clark@telchemy.com, gwz@net-zen.net, bijy@sttri.com.cn, sunseawq@huawei.com +# RFC7295 || H. Tschofenig, L. Eggert, Z. Sarker || Hannes.Tschofenig@gmx.net, lars@netapp.com, Zaheduzzaman.Sarker@ericsson.com +# RFC7296 || C. Kaufman, P. Hoffman, Y. Nir, P. Eronen, T. Kivinen || charliekaufman@outlook.com, paul.hoffman@vpnc.org, nir.ietf@gmail.com, pe@iki.fi, kivinen@iki.fi +# RFC7297 || M. Boucadair, C. Jacquenet, N. Wang || mohamed.boucadair@orange.com, christian.jacquenet@orange.com, n.wang@surrey.ac.uk +# RFC7298 || D. Ovsienko || infrastation@yandex.ru +# RFC7299 || R. Housley || housley@vigilsec.com +# RFC7300 || J. Haas, J. Mitchell || jhaas@juniper.net, jon.mitchell@microsoft.com +# RFC7301 || S. Friedl, A. Popov, A. Langley, E. Stephan || sfriedl@cisco.com, andreipo@microsoft.com, agl@google.com, emile.stephan@orange.com +# RFC7302 || P. Lemieux || pal@sandflow.com +# RFC7303 || H. Thompson, C. Lilley || ht@inf.ed.ac.uk, chris@w3.org +# RFC7304 || W. Kumari || warren@kumari.net +# RFC7305 || E. Lear, Ed. || lear@cisco.com +# RFC7306 || H. Shah, F. Marti, W. Noureddine, A. Eiriksson, R. Sharp || hemal@broadcom.com, felix@chelsio.com, asgeir@chelsio.com, wael@chelsio.com, robert.o.sharp@intel.com +# RFC7307 || Q. Zhao, K. Raza, C. Zhou, L. Fang, L. Li, D. King || quintin.zhao@huawei.com, skraza@cisco.com, czhou@cisco.com, lufang@microsoft.com, lilianyuan@chinamobile.com, daniel@olddog.co.uk +# RFC7308 || E. Osborne || none +# RFC7309 || Z. Liu, L. Jin, R. Chen, D. Cai, S. Salam || zhliu@gsta.com, lizho.jin@gmail.com, chen.ran@zte.com.cn, dcai@cisco.com, ssalam@cisco.com +# RFC7310 || J. Lindsay, H. Foerster || lindsay@worldcastsystems.com, foerster@worldcastsystems.com +# RFC7311 || P. Mohapatra, R. Fernando, E. Rosen, J. Uttaro || mpradosh@yahoo.com, rex@cisco.com, erosen@cisco.com, uttaro@att.com +# RFC7312 || J. Fabini, A. Morton || joachim.fabini@tuwien.ac.at, acmorton@att.com +# RFC7313 || K. Patel, E. Chen, B. Venkatachalapathy || keyupate@cisco.com, enkechen@cisco.com, balaji_pv@hotmail.com +# RFC7314 || M. Andrews || marka@isc.org +# RFC7315 || R. Jesske, K. Drage, C. Holmberg || r.jesske@telekom.de, drage@alcatel-lucent.com, christer.holmberg@ericsson.com +# RFC7316 || J. van Elburg, K. Drage, M. Ohsugi, S. Schubert, K. Arai || ietf.hanserik@gmail.com, drage@alcatel-lucent.com, mayumi.ohsugi@ntt-at.co.jp, shida@ntt-at.com, arai.kenjiro@lab.ntt.co.jp +# RFC7317 || A. Bierman, M. Bjorklund || andy@yumaworks.com, mbj@tail-f.com +# RFC7318 || A. Newton, G. Huston || andy@arin.net, gih@apnic.net +# RFC7319 || D. Eastlake 3rd || d3e3e3@gmail.com +# RFC7320 || M. Nottingham || mnot@mnot.net +# RFC7321 || D. McGrew, P. Hoffman || mcgrew@cisco.com, paul.hoffman@vpnc.org +# RFC7322 || H. Flanagan, S. Ginoza || rse@rfc-editor.org, rfc-editor@rfc-editor.org +# RFC7323 || D. Borman, B. Braden, V. Jacobson, R. Scheffenegger, Ed. || david.borman@quantum.com, braden@isi.edu, vanj@google.com, rs@netapp.com +# RFC7324 || E. Osborne || eric.osborne@notcom.com +# RFC7325 || C. Villamizar, Ed., K. Kompella, S. Amante, A. Malis, C. Pignataro || curtis@occnc.com, kireeti@juniper.net, amante@apple.com, agmalis@gmail.com, cpignata@cisco.com +# RFC7326 || J. Parello, B. Claise, B. Schoening, J. Quittek || jparello@cisco.com, bclaise@cisco.com, brad.schoening@verizon.net, quittek@netlab.nec.de +# RFC7328 || R. Gieben || miek@google.com +# RFC7329 || H. Kaplan || hadrielk@yahoo.com +# RFC7330 || T. Nadeau, Z. Ali, N. Akiya || tnadeau@lucidvision.com, zali@cisco.com, nobo@cisco.com +# RFC7331 || T. Nadeau, Z. Ali, N. Akiya || tnadeau@lucidvision.com, zali@cisco.com, nobo@cisco.com +# RFC7332 || H. Kaplan, V. Pascual || hadrielk@yahoo.com, victor.pascual@quobis.com +# RFC7333 || H. Chan, Ed., D. Liu, P. Seite, H. Yokota, J. Korhonen || h.a.chan@ieee.org, liudapeng@chinamobile.com, pierrick.seite@orange.com, hidetoshi.yokota@landisgyr.com, jouni.nospam@gmail.com +# RFC7334 || Q. Zhao, D. Dhody, D. King, Z. Ali, R. Casellas || quintin.zhao@huawei.com, dhruv.dhody@huawei.com, daniel@olddog.co.uk, zali@cisco.com, ramon.casellas@cttc.es +# RFC7335 || C. Byrne || cameron.byrne@t-mobile.com +# RFC7336 || L. Peterson, B. Davie, R. van Brandenburg, Ed. || lapeters@akamai.com, bdavie@vmware.com, ray.vanbrandenburg@tno.nl +# RFC7337 || K. Leung, Ed., Y. Lee, Ed. || kleung@cisco.com, yiu_lee@cable.comcast.com +# RFC7338 || F. Jounay, Ed., Y. Kamite, Ed., G. Heron, M. Bocci || frederic.jounay@orange.ch, y.kamite@ntt.com, giheron@cisco.com, Matthew.Bocci@alcatel-lucent.com +# RFC7339 || V. Gurbani, Ed., V. Hilt, H. Schulzrinne || vkg@bell-labs.com, volker.hilt@bell-labs.com, hgs@cs.columbia.edu +# RFC7340 || J. Peterson, H. Schulzrinne, H. Tschofenig || jon.peterson@neustar.biz, hgs@cs.columbia.edu, Hannes.Tschofenig@gmx.net +# RFC7341 || Q. Sun, Y. Cui, M. Siodelski, S. Krishnan, I. Farrer || sunqi@csnet1.cs.tsinghua.edu.cn, yong@csnet1.cs.tsinghua.edu.cn, msiodelski@gmail.com, suresh.krishnan@ericsson.com, ian.farrer@telekom.de +# RFC7342 || L. Dunbar, W. Kumari, I. Gashinsky || ldunbar@huawei.com, warren@kumari.net, igor@yahoo-inc.com +# RFC7343 || J. Laganier, F. Dupont || julien.ietf@gmail.com, fdupont@isc.org +# RFC7344 || W. Kumari, O. Gudmundsson, G. Barwood || warren@kumari.net, ogud@ogud.com, george.barwood@blueyonder.co.uk +# RFC7345 || C. Holmberg, I. Sedlacek, G. Salgueiro || christer.holmberg@ericsson.com, ivo.sedlacek@ericsson.com, gsalguei@cisco.com +# RFC7346 || R. Droms || rdroms.ietf@gmail.com +# RFC7347 || H. van Helvoort, Ed., J. Ryoo, Ed., H. Zhang, F. Huang, H. Li, A. D'Alessandro || huub@van-helvoort.eu, ryoo@etri.re.kr, zhanghaiyan@huawei.com, feng.huang@philips.com, lihan@chinamobile.com, alessandro.dalessandro@telecomitalia.it +# RFC7348 || M. Mahalingam, D. Dutt, K. Duda, P. Agarwal, L. Kreeger, T. Sridhar, M. Bursell, C. Wright || mallik_mahalingam@yahoo.com, ddutt.ietf@hobbesdutt.com, kduda@arista.com, pagarwal@broadcom.com, kreeger@cisco.com, tsridhar@vmware.com, mike.bursell@intel.com, chrisw@redhat.com +# RFC7349 || L. Zheng, M. Chen, M. Bhatia || vero.zheng@huawei.com, mach.chen@huawei.com, manav@ionosnetworks.com +# RFC7350 || M. Petit-Huguenin, G. Salgueiro || marcph@getjive.com, gsalguei@cisco.com +# RFC7351 || E. Wilde || dret@berkeley.edu +# RFC7352 || S. Bosch || stephan@rename-it.nl +# RFC7353 || S. Bellovin, R. Bush, D. Ward || bellovin@acm.org, randy@psg.com, dward@cisco.com +# RFC7354 || A. Adolf, P. Siebert || alexander.adolf@condition-alpha.com, dvb@dvb.org +# RFC7355 || G. Salgueiro, V. Pascual, A. Roman, S. Garcia || gsalguei@cisco.com, victor.pascual@quobis.com, anton.roman@quobis.com, sergio.garcia@quobis.com +# RFC7356 || L. Ginsberg, S. Previdi, Y. Yang || ginsberg@cisco.com, sprevidi@cisco.com, yiya@cisco.com +# RFC7357 || H. Zhai, F. Hu, R. Perlman, D. Eastlake 3rd, O. Stokes || zhai.hongjun@zte.com.cn, hu.fangwei@zte.com.cn, Radia@alum.mit.edu, d3e3e3@gmail.com, ostokes@extremenetworks.com +# RFC7358 || K. Raza, S. Boutros, L. Martini, N. Leymann || skraza@cisco.com, sboutros@cisco.com, lmartini@cisco.com, n.leymann@telekom.de +# RFC7359 || F. Gont || fgont@si6networks.com +# RFC7360 || A. DeKok || aland@freeradius.org +# RFC7361 || P. Dutta, F. Balus, O. Stokes, G. Calvignac, D. Fedyk || pranjal.dutta@alcatel-lucent.com, florin.balus@alcatel-lucent.com, ostokes@extremenetworks.com, geraldine.calvignac@orange.com, don.fedyk@hp.com +# RFC7362 || E. Ivov, H. Kaplan, D. Wing || emcho@jitsi.org, hadrielk@yahoo.com, dwing-ietf@fuggles.com +# RFC7363 || J. Maenpaa, G. Camarillo || Jouni.Maenpaa@ericsson.com, Gonzalo.Camarillo@ericsson.com +# RFC7364 || T. Narten, Ed., E. Gray, Ed., D. Black, L. Fang, L. Kreeger, M. Napierala || narten@us.ibm.com, Eric.Gray@Ericsson.com, david.black@emc.com, lufang@microsoft.com, kreeger@cisco.com, mnapierala@att.com +# RFC7365 || M. Lasserre, F. Balus, T. Morin, N. Bitar, Y. Rekhter || marc.lasserre@alcatel-lucent.com, florin.balus@alcatel-lucent.com, thomas.morin@orange.com, nabil.n.bitar@verizon.com, yakov@juniper.net +# RFC7366 || P. Gutmann || pgut001@cs.auckland.ac.nz +# RFC7367 || R. Cole, J. Macker, B. Adamson || robert.g.cole@us.army.mil, macker@itd.nrl.navy.mil, adamson@itd.nrl.navy.mil +# RFC7368 || T. Chown, Ed., J. Arkko, A. Brandt, O. Troan, J. Weil || tjc@ecs.soton.ac.uk, jari.arkko@piuha.net, Anders_Brandt@sigmadesigns.com, ot@cisco.com, jason.weil@twcable.com +# RFC7369 || A. Takacs, B. Gero, H. Long || attila.takacs@ericsson.com, balazs.peter.gero@ericsson.com, lonho@huawei.com +# RFC7370 || L. Ginsberg || ginsberg@cisco.com +# RFC7371 || M. Boucadair, S. Venaas || mohamed.boucadair@orange.com, stig@cisco.com +# RFC7372 || M. Kucherawy || superuser@gmail.com +# RFC7373 || B. Trammell || ietf@trammell.ch +# RFC7374 || J. Maenpaa, G. Camarillo || Jouni.Maenpaa@ericsson.com, gonzalo.camarillo@ericsson.com +# RFC7375 || J. Peterson || jon.peterson@neustar.biz +# RFC7376 || T. Reddy, R. Ravindranath, M. Perumal, A. Yegin || tireddy@cisco.com, rmohanr@cisco.com, muthu.arul@gmail.com, alper.yegin@yegin.org +# RFC7377 || B. Leiba, A. Melnikov || barryleiba@computer.org, alexey.melnikov@isode.com +# RFC7378 || H. Tschofenig, H. Schulzrinne, B. Aboba, Ed. || Hannes.Tschofenig@gmx.net, hgs@cs.columbia.edu, Bernard_Aboba@hotmail.com +# RFC7379 || Y. Li, W. Hao, R. Perlman, J. Hudson, H. Zhai || liyizhou@huawei.com, haoweiguo@huawei.com, radia@alum.mit.edu, jon.hudson@gmail.com, honjun.zhai@tom.com +# RFC7380 || J. Tong, C. Bi, Ed., R. Even, Q. Wu, Ed., R. Huang || tongjg@sttri.com.cn, bijy@sttri.com.cn, roni.even@mail01.huawei.com, bill.wu@huawei.com, rachel.huang@huawei.com +# RFC7381 || K. Chittimaneni, T. Chown, L. Howard, V. Kuarsingh, Y. Pouffary, E. Vyncke || kk@dropbox.com, tjc@ecs.soton.ac.uk, lee.howard@twcable.com, victor@jvknet.com, yanick.pouffary@hp.com, evyncke@cisco.com +# RFC7382 || S. Kent, D. Kong, K. Seo || skent@bbn.com, dkong@bbn.com, kseo@bbn.com +# RFC7383 || V. Smyslov || svan@elvis.ru +# RFC7384 || T. Mizrahi || talmi@marvell.com +# RFC7385 || L. Andersson, G. Swallow || loa@mail01.huawei.com, swallow@cisco.com +# RFC7386 || P. Hoffman, J. Snell || paul.hoffman@vpnc.org, jasnell@gmail.com +# RFC7387 || R. Key, Ed., L. Yong, Ed., S. Delord, F. Jounay, L. Jin || raymond.key@ieee.org, lucy.yong@huawei.com, simon.delord@gmail.com, frederic.jounay@orange.ch, lizho.jin@gmail.com +# RFC7388 || J. Schoenwaelder, A. Sehgal, T. Tsou, C. Zhou || j.schoenwaelder@jacobs-university.de, s.anuj@jacobs-university.de, tina.tsou.zouting@huawei.com, cathyzhou@huawei.com +# RFC7389 || R. Wakikawa, R. Pazhyannur, S. Gundavelli, C. Perkins || ryuji.wakikawa@gmail.com, rpazhyan@cisco.com, sgundave@cisco.com, charliep@computer.org +# RFC7390 || A. Rahman, Ed., E. Dijk, Ed. || Akbar.Rahman@InterDigital.com, esko.dijk@philips.com +# RFC7391 || J. Hadi Salim || hadi@mojatatu.com +# RFC7392 || P. Dutta, M. Bocci, L. Martini || pranjal.dutta@alcatel-lucent.com, matthew.bocci@alcatel-lucent.com, lmartini@cisco.com +# RFC7393 || X. Deng, M. Boucadair, Q. Zhao, J. Huang, C. Zhou || dxhbupt@gmail.com, mohamed.boucadair@orange.com, zhaoqin.bupt@gmail.com, james.huang@huawei.com, cathy.zhou@huawei.com +# RFC7394 || S. Boutros, S. Sivabalan, G. Swallow, S. Saxena, V. Manral, S. Aldrin || sboutros@cisco.com, msiva@cisco.com, swallow@cisco.com, ssaxena@cisco.com, vishwas@ionosnetworks.com, aldrin.ietf@gmail.com +# RFC7395 || L. Stout, Ed., J. Moffitt, E. Cestari || lance@andyet.net, jack@metajack.im, eric@cstar.io +# RFC7396 || P. Hoffman, J. Snell || paul.hoffman@vpnc.org, jasnell@gmail.com +# RFC7397 || J. Gilger, H. Tschofenig || Gilger@ITSec.RWTH-Aachen.de, Hannes.tschofenig@gmx.net +# RFC7398 || M. Bagnulo, T. Burbridge, S. Crawford, P. Eardley, A. Morton || marcelo@it.uc3m.es, trevor.burbridge@bt.com, sam@samknows.com, philip.eardley@bt.com, acmorton@att.com +# RFC7399 || A. Farrel, D. King || adrian@olddog.co.uk, daniel@olddog.co.uk +# RFC7400 || C. Bormann || cabo@tzi.org +# RFC7401 || R. Moskowitz, Ed., T. Heer, P. Jokela, T. Henderson || rgm@labs.htt-consult.com, tobias.heer@belden.com, petri.jokela@nomadiclab.com, tomhend@u.washington.edu +# RFC7402 || P. Jokela, R. Moskowitz, J. Melen || petri.jokela@nomadiclab.com, rgm@labs.htt-consult.com, jan.melen@nomadiclab.com +# RFC7403 || H. Kaplan || hadrielk@yahoo.com +# RFC7404 || M. Behringer, E. Vyncke || mbehring@cisco.com, evyncke@cisco.com +# RFC7405 || P. Kyzivat || pkyzivat@alum.mit.edu +# RFC7406 || H. Schulzrinne, S. McCann, G. Bajko, H. Tschofenig, D. Kroeselberg || hgs+ecrit@cs.columbia.edu, smccann@blackberry.com, gabor.bajko@mediatek.com, Hannes.Tschofenig@gmx.net, dirk.kroeselberg@siemens.com +# RFC7407 || M. Bjorklund, J. Schoenwaelder || mbj@tail-f.com, j.schoenwaelder@jacobs-university.de +# RFC7408 || E. Haleplidis || ehalep@ece.upatras.gr +# RFC7409 || E. Haleplidis, J. Halpern || ehalep@ece.upatras.gr, joel.halpern@ericsson.com +# RFC7410 || M. Kucherawy || superuser@gmail.com +# RFC7411 || T. Schmidt, Ed., M. Waehlisch, R. Koodli, G. Fairhurst, D. Liu || t.schmidt@haw-hamburg.de, mw@link-lab.net, rajeev.koodli@intel.com, gorry@erg.abdn.ac.uk, liudapeng@chinamobile.com +# RFC7412 || Y. Weingarten, S. Aldrin, P. Pan, J. Ryoo, G. Mirsky || wyaacov@gmail.com, aldrin.ietf@gmail.com, ppan@infinera.com, ryoo@etri.re.kr, gregory.mirsky@ericsson.com +# RFC7413 || Y. Cheng, J. Chu, S. Radhakrishnan, A. Jain || ycheng@google.com, hkchu@google.com, sivasankar@cs.ucsd.edu, arvind@google.com +# RFC7414 || M. Duke, R. Braden, W. Eddy, E. Blanton, A. Zimmermann || m.duke@f5.com, braden@isi.edu, wes@mti-systems.com, elb@interruptsciences.com, alexander.zimmermann@netapp.com +# RFC7415 || E. Noel, P. Williams || ecnoel@att.com, phil.m.williams@bt.com +# RFC7416 || T. Tsao, R. Alexander, M. Dohler, V. Daza, A. Lozano, M. Richardson, Ed. || tzetatsao@eaton.com, rogeralexander@eaton.com, mischa.dohler@kcl.ac.uk, vanesa.daza@upf.edu, angel.lozano@upf.edu, mcr+ietf@sandelman.ca +# RFC7417 || G. Karagiannis, A. Bhargava || georgios.karagiannis@huawei.com, anuragb@cisco.com +# RFC7418 || S. Dawkins, Ed. || spencerdawkins.ietf@gmail.com +# RFC7419 || N. Akiya, M. Binderberger, G. Mirsky || nobo@cisco.com, mbinderb@cisco.com, gregory.mirsky@ericsson.com +# RFC7420 || A. Koushik, E. Stephan, Q. Zhao, D. King, J. Hardwick || kkoushik@brocade.com, emile.stephan@orange.com, qzhao@huawei.com, daniel@olddog.co.uk, jonathan.hardwick@metaswitch.com +# RFC7421 || B. Carpenter, Ed., T. Chown, F. Gont, S. Jiang, A. Petrescu, A. Yourtchenko || brian.e.carpenter@gmail.com, tjc@ecs.soton.ac.uk, fgont@si6networks.com, jiangsheng@huawei.com, alexandru.petrescu@cea.fr, ayourtch@cisco.com +# RFC7422 || C. Donley, C. Grundemann, V. Sarawat, K. Sundaresan, O. Vautrin || c.donley@cablelabs.com, cgrundemann@gmail.com, v.sarawat@cablelabs.com, k.sundaresan@cablelabs.com, Olivier@juniper.net +# RFC7423 || L. Morand, Ed., V. Fajardo, H. Tschofenig || lionel.morand@orange.com, vf0213@gmail.com, Hannes.Tschofenig@gmx.net +# RFC7424 || R. Krishnan, L. Yong, A. Ghanwani, N. So, B. Khasnabish || ramkri123@gmail.com, lucy.yong@huawei.com, anoop@alumni.duke.edu, ningso@yahoo.com, vumip1@gmail.com +# RFC7425 || M. Thornburgh || mthornbu@adobe.com +# RFC7426 || E. Haleplidis, Ed., K. Pentikousis, Ed., S. Denazis, J. Hadi Salim, D. Meyer, O. Koufopavlou || ehalep@ece.upatras.gr, k.pentikousis@eict.de, sdena@upatras.gr, hadi@mojatatu.com, dmm@1-4-5.net, odysseas@ece.upatras.gr +# RFC7427 || T. Kivinen, J. Snyder || kivinen@iki.fi, jms@opus1.com +# RFC7428 || A. Brandt, J. Buron || anders_brandt@sigmadesigns.com, jakob_buron@sigmadesigns.com +# RFC7429 || D. Liu, Ed., JC. Zuniga, Ed., P. Seite, H. Chan, CJ. Bernardos || liudapeng@chinamobile.com, JuanCarlos.Zuniga@InterDigital.com, pierrick.seite@orange.com, h.a.chan@ieee.org, cjbc@it.uc3m.es +# RFC7430 || M. Bagnulo, C. Paasch, F. Gont, O. Bonaventure, C. Raiciu || marcelo@it.uc3m.es, christoph.paasch@gmail.com, fgont@si6networks.com, Olivier.Bonaventure@uclouvain.be, costin.raiciu@cs.pub.ro +# RFC7431 || A. Karan, C. Filsfils, IJ. Wijnands, Ed., B. Decraene || apoorva@cisco.com, cfilsfil@cisco.com, ice@cisco.com, bruno.decraene@orange.com +# RFC7432 || A. Sajassi, Ed., R. Aggarwal, N. Bitar, A. Isaac, J. Uttaro, J. Drake, W. Henderickx || sajassi@cisco.com, raggarwa_1@yahoo.com, nabil.n.bitar@verizon.com, aisaac71@bloomberg.net, uttaro@att.com, jdrake@juniper.net, wim.henderickx@alcatel-lucent.com +# RFC7433 || A. Johnston, J. Rafferty || alan.b.johnston@gmail.com, jay@humancomm.com +# RFC7434 || K. Drage, Ed., A. Johnston || keith.drage@alcatel-lucent.com, alan.b.johnston@gmail.com +# RFC7435 || V. Dukhovni || ietf-dane@dukhovni.org +# RFC7436 || H. Shah, E. Rosen, F. Le Faucheur, G. Heron || hshah@ciena.com, erosen@juniper.net, flefauch@cisco.com, giheron@cisco.com +# RFC7437 || M. Kucherawy, Ed. || superuser@gmail.com +# RFC7438 || IJ. Wijnands, Ed., E. Rosen, A. Gulko, U. Joorde, J. Tantsura || ice@cisco.com, erosen@juniper.net, arkadiy.gulko@thomsonreuters.com, uwe.joorde@telekom.de, jeff.tantsura@ericsson.com +# RFC7439 || W. George, Ed., C. Pignataro, Ed. || wesley.george@twcable.com, cpignata@cisco.com +# RFC7440 || P. Masotta || patrick.masotta.ietf@vercot.com +# RFC7441 || IJ. Wijnands, E. Rosen, U. Joorde || ice@cisco.com, erosen@juniper.net, uwe.joorde@telekom.de +# RFC7442 || Y. Rekhter, R. Aggarwal, N. Leymann, W. Henderickx, Q. Zhao, R. Li || yakov@juniper.net, raggarwa_1@yahoo.com, N.Leymann@telekom.de, wim.henderickx@alcatel-lucent.com, quintin.zhao@huawei.com, renwei.li@huawei.com +# RFC7443 || P. Patil, T. Reddy, G. Salgueiro, M. Petit-Huguenin || praspati@cisco.com, tireddy@cisco.com, gsalguei@cisco.com, marc@petit-huguenin.org +# RFC7444 || K. Zeilenga, A. Melnikov || kurt.zeilenga@isode.com, alexey.melnikov@isode.com +# RFC7445 || G. Chen, H. Deng, D. Michaud, J. Korhonen, M. Boucadair || phdgang@gmail.com, denghui@chinamobile.com, dave.michaud@rci.rogers.com, jouni.nospam@gmail.com, mohamed.boucadair@orange.com +# RFC7446 || Y. Lee, Ed., G. Bernstein, Ed., D. Li, W. Imajuku || leeyoung@huawei.com, gregb@grotto-networking.com, danli@huawei.com, imajuku.wataru@lab.ntt.co.jp +# RFC7447 || J. Scudder, K. Kompella || jgs@juniper.net, kireeti@juniper.net +# RFC7448 || T. Taylor, Ed., D. Romascanu || tom.taylor.stds@gmail.com, dromasca@gmail.com +# RFC7449 || Y. Lee, Ed., G. Bernstein, Ed., J. Martensson, T. Takeda, T. Tsuritani, O. Gonzalez de Dios || leeyoung@huawei.com, gregb@grotto-networking.com, jonas.martensson@acreo.se, tomonori.takeda@ntt.com, tsuri@kddilabs.jp, oscar.gonzalezdedios@telefonica.com +# RFC7450 || G. Bumgardner || gbumgard@gmail.com +# RFC7451 || S. Hollenbeck || shollenbeck@verisign.com +# RFC7452 || H. Tschofenig, J. Arkko, D. Thaler, D. McPherson || Hannes.Tschofenig@gmx.net, jari.arkko@piuha.net, dthaler@microsoft.com, dmcpherson@verisign.com +# RFC7453 || V. Mahalingam, K. Sampath, S. Aldrin, T. Nadeau || venkat.mahalingams@gmail.com, kannankvs@gmail.com, aldrin.ietf@gmail.com, tnadeau@lucidvision.com +# RFC7454 || J. Durand, I. Pepelnjak, G. Doering || jerduran@cisco.com, ip@ipspace.net, gert@space.net +# RFC7455 || T. Senevirathne, N. Finn, S. Salam, D. Kumar, D. Eastlake 3rd, S. Aldrin, Y. Li || tsenevir@cisco.com, nfinn@cisco.com, ssalam@cisco.com, dekumar@cisco.com, d3e3e3@gmail.com, aldrin.ietf@gmail.com, liyizhou@huawei.com +# RFC7456 || T. Mizrahi, T. Senevirathne, S. Salam, D. Kumar, D. Eastlake 3rd || talmi@marvell.com, tsenevir@cisco.com, ssalam@cisco.com, dekumar@cisco.com, d3e3e3@gmail.com +# RFC7457 || Y. Sheffer, R. Holz, P. Saint-Andre || yaronf.ietf@gmail.com, holz@net.in.tum.de, peter@andyet.com +# RFC7458 || R. Valmikam, R. Koodli || valmikam@gmail.com, rajeev.koodli@intel.com +# RFC7459 || M. Thomson, J. Winterbottom || martin.thomson@gmail.com, a.james.winterbottom@gmail.com +# RFC7460 || M. Chandramouli, B. Claise, B. Schoening, J. Quittek, T. Dietz || moulchan@cisco.com, bclaise@cisco.com, brad.schoening@verizon.net, quittek@neclab.eu, Thomas.Dietz@neclab.eu +# RFC7461 || J. Parello, B. Claise, M. Chandramouli || jparello@cisco.com, bclaise@cisco.com, moulchan@cisco.com +# RFC7462 || L. Liess, Ed., R. Jesske, A. Johnston, D. Worley, P. Kyzivat || laura.liess.dt@gmail.com, r.jesske@telekom.de, alan.b.johnston@gmail.com, worley@ariadne.com, pkyzivat@alum.mit.edu +# RFC7463 || A. Johnston, Ed., M. Soroushnejad, Ed., V. Venkataramanan || alan.b.johnston@gmail.com, msoroush@gmail.com, vvenkatar@gmail.com +# RFC7464 || N. Williams || nico@cryptonector.com +# RFC7465 || A. Popov || andreipo@microsoft.com +# RFC7466 || C. Dearlove, T. Clausen || chris.dearlove@baesystems.com, T.Clausen@computer.org +# RFC7467 || A. Murdock || Aidan.murdock@ncia.nato.int +# RFC7468 || S. Josefsson, S. Leonard || simon@josefsson.org, dev+ietf@seantek.com +# RFC7469 || C. Evans, C. Palmer, R. Sleevi || cevans@google.com, palmer@google.com, sleevi@google.com +# RFC7470 || F. Zhang, A. Farrel || zhangfatai@huawei.com, adrian@olddog.co.uk +# RFC7471 || S. Giacalone, D. Ward, J. Drake, A. Atlas, S. Previdi || spencer.giacalone@gmail.com, dward@cisco.com, jdrake@juniper.net, akatlas@juniper.net, sprevidi@cisco.com +# RFC7472 || I. McDonald, M. Sweet || blueroofmusic@gmail.com, msweet@apple.com +# RFC7473 || K. Raza, S. Boutros || skraza@cisco.com, sboutros@cisco.com +# RFC7474 || M. Bhatia, S. Hartman, D. Zhang, A. Lindem, Ed. || manav@ionosnetworks.com, hartmans-ietf@mit.edu, dacheng.zhang@gmail.com, acee@cisco.com +# RFC7475 || S. Dawkins || spencerdawkins.ietf@gmail.com +# RFC7476 || K. Pentikousis, Ed., B. Ohlman, D. Corujo, G. Boggia, G. Tyson, E. Davies, A. Molinaro, S. Eum || k.pentikousis@eict.de, Borje.Ohlman@ericsson.com, dcorujo@av.it.pt, g.boggia@poliba.it, gareth.tyson@eecs.qmul.ac.uk, davieseb@scss.tcd.ie, antonella.molinaro@unirc.it, suyong@nict.go.jp +# RFC7477 || W. Hardaker || ietf@hardakers.net +# RFC7478 || C. Holmberg, S. Hakansson, G. Eriksson || christer.holmberg@ericsson.com, stefan.lk.hakansson@ericsson.com, goran.ap.eriksson@ericsson.com +# RFC7479 || S. Moonesamy || sm+ietf@elandsys.com +# RFC7480 || A. Newton, B. Ellacott, N. Kong || andy@arin.net, bje@apnic.net, nkong@cnnic.cn +# RFC7481 || S. Hollenbeck, N. Kong || shollenbeck@verisign.com, nkong@cnnic.cn +# RFC7482 || A. Newton, S. Hollenbeck || andy@arin.net, shollenbeck@verisign.com +# RFC7483 || A. Newton, S. Hollenbeck || andy@arin.net, shollenbeck@verisign.com +# RFC7484 || M. Blanchet || Marc.Blanchet@viagenie.ca +# RFC7485 || L. Zhou, N. Kong, S. Shen, S. Sheng, A. Servin || zhoulinlin@cnnic.cn, nkong@cnnic.cn, shenshuo@cnnic.cn, steve.sheng@icann.org, arturo.servin@gmail.com +# RFC7486 || S. Farrell, P. Hoffman, M. Thomas || stephen.farrell@cs.tcd.ie, paul.hoffman@vpnc.org, mike@phresheez.com +# RFC7487 || E. Bellagamba, A. Takacs, G. Mirsky, L. Andersson, P. Skoldstrom, D. Ward || elisa.bellagamba@ericsson.com, attila.takacs@ericsson.com, gregory.mirsky@ericsson.com, loa@mail01.huawei.com, pontus.skoldstrom@acreo.se, dward@cisco.com +# RFC7488 || M. Boucadair, R. Penno, D. Wing, P. Patil, T. Reddy || mohamed.boucadair@orange.com, repenno@cisco.com, dwing-ietf@fuggles.com, praspati@cisco.com, tireddy@cisco.com +# RFC7489 || M. Kucherawy, Ed., E. Zwicky, Ed. || superuser@gmail.com, zwicky@yahoo-inc.com +# RFC7490 || S. Bryant, C. Filsfils, S. Previdi, M. Shand, N. So || stbryant@cisco.com, cfilsfil@cisco.com, sprevidi@cisco.com, imc.shand@gmail.com, ningso@vinci-systems.com +# RFC7491 || D. King, A. Farrel || daniel@olddog.co.uk, adrian@olddog.co.uk +# RFC7492 || M. Bhatia, D. Zhang, M. Jethanandani || manav@ionosnetworks.com, dacheng.zhang@gmail.com, mjethanandani@gmail.com +# RFC7493 || T. Bray, Ed. || tbray@textuality.com +# RFC7494 || C. Shao, H. Deng, R. Pazhyannur, F. Bari, R. Zhang, S. Matsushima || shaochunju@chinamobile.com, denghui@chinamobile.com, rpazhyan@cisco.com, farooq.bari@att.com, zhangr@gsta.com, satoru.matsushima@g.softbank.co.jp +# RFC7495 || A. Montville, D. Black || adam.w.montville@gmail.com, david.black@emc.com +# RFC7496 || M. Tuexen, R. Seggelmann, R. Stewart, S. Loreto || tuexen@fh-muenster.de, rfc@robin-seggelmann.com, randall@lakerest.net, Salvatore.Loreto@ericsson.com +# RFC7497 || A. Morton || acmorton@att.com +# RFC7498 || P. Quinn, Ed., T. Nadeau, Ed. || paulq@cisco.com, tnadeau@lucidvision.com +# RFC7499 || A. Perez-Mendez, Ed., R. Marin-Lopez, F. Pereniguez-Garcia, G. Lopez-Millan, D. Lopez, A. DeKok || alex@um.es, rafa@um.es, pereniguez@um.es, gabilm@um.es, diego@tid.es, aland@networkradius.com +# RFC7500 || R. Housley, Ed., O. Kolkman, Ed. || housley@vigilsec.com, kolkman@isoc.org +# RFC7501 || C. Davids, V. Gurbani, S. Poretsky || davids@iit.edu, vkg@bell-labs.com, sporetsky@allot.com +# RFC7502 || C. Davids, V. Gurbani, S. Poretsky || davids@iit.edu, vkg@bell-labs.com, sporetsky@allot.com +# RFC7503 || A. Lindem, J. Arkko || acee@cisco.com, jari.arkko@piuha.net +# RFC7504 || J. Klensin || john-ietf@jck.com +# RFC7505 || J. Levine, M. Delany || standards@taugh.com, mx0dot@yahoo.com +# RFC7506 || K. Raza, N. Akiya, C. Pignataro || skraza@cisco.com, nobo.akiya.dev@gmail.com, cpignata@cisco.com +# RFC7507 || B. Moeller, A. Langley || bmoeller@acm.org, agl@google.com +# RFC7508 || L. Cailleux, C. Bonatti || laurent.cailleux@intradef.gouv.fr, bonatti252@ieca.com +# RFC7509 || R. Huang, V. Singh || rachel.huang@huawei.com, varun@comnet.tkk.fi +# RFC7510 || X. Xu, N. Sheth, L. Yong, R. Callon, D. Black || xuxiaohu@huawei.com, nsheth@juniper.net, lucy.yong@huawei.com, rcallon@juniper.net, david.black@emc.com +# RFC7511 || M. Wilhelm || max@rfc2324.org +# RFC7512 || J. Pechanec, D. Moffat || Jan.Pechanec@Oracle.COM, Darren.Moffat@Oracle.COM +# RFC7513 || J. Bi, J. Wu, G. Yao, F. Baker || junbi@tsinghua.edu.cn, jianping@cernet.edu.cn, yaoguang@cernet.edu.cn, fred@cisco.com +# RFC7514 || M. Luckie || mjl@caida.org +# RFC7515 || M. Jones, J. Bradley, N. Sakimura || mbj@microsoft.com, ve7jtb@ve7jtb.com, n-sakimura@nri.co.jp +# RFC7516 || M. Jones, J. Hildebrand || mbj@microsoft.com, jhildebr@cisco.com +# RFC7517 || M. Jones || mbj@microsoft.com +# RFC7518 || M. Jones || mbj@microsoft.com +# RFC7519 || M. Jones, J. Bradley, N. Sakimura || mbj@microsoft.com, ve7jtb@ve7jtb.com, n-sakimura@nri.co.jp +# RFC7520 || M. Miller || mamille2@cisco.com +# RFC7521 || B. Campbell, C. Mortimore, M. Jones, Y. Goland || brian.d.campbell@gmail.com, cmortimore@salesforce.com, mbj@microsoft.com, yarong@microsoft.com +# RFC7522 || B. Campbell, C. Mortimore, M. Jones || brian.d.campbell@gmail.com, cmortimore@salesforce.com, mbj@microsoft.com +# RFC7523 || M. Jones, B. Campbell, C. Mortimore || mbj@microsoft.com, brian.d.campbell@gmail.com, cmortimore@salesforce.com +# RFC7524 || Y. Rekhter, E. Rosen, R. Aggarwal, T. Morin, I. Grosclaude, N. Leymann, S. Saad || yakov@juniper.net, erosen@juniper.net, raggarwa_1@yahoo.com, thomas.morin@orange.com, irene.grosclaude@orange.com, n.leymann@telekom.de, samirsaad1@outlook.com +# RFC7525 || Y. Sheffer, R. Holz, P. Saint-Andre || yaronf.ietf@gmail.com, ralph.ietf@gmail.com, peter@andyet.com +# RFC7526 || O. Troan, B. Carpenter, Ed. || ot@cisco.com, brian.e.carpenter@gmail.com +# RFC7527 || R. Asati, H. Singh, W. Beebee, C. Pignataro, E. Dart, W. George || rajiva@cisco.com, shemant@cisco.com, wbeebee@cisco.com, cpignata@cisco.com, dart@es.net, wesley.george@twcable.com +# RFC7528 || P. Higgs, J. Piesing || paul.higgs@ericsson.com, jon.piesing@tpvision.com +# RFC7529 || C. Daboo, G. Yakushev || cyrus@daboo.name, gyakushev@yahoo.com +# RFC7530 || T. Haynes, Ed., D. Noveck, Ed. || thomas.haynes@primarydata.com, dave_noveck@dell.com +# RFC7531 || T. Haynes, Ed., D. Noveck, Ed. || thomas.haynes@primarydata.com, dave_noveck@dell.com +# RFC7532 || J. Lentini, R. Tewari, C. Lever, Ed. || jlentini@netapp.com, tewarir@us.ibm.com, chuck.lever@oracle.com +# RFC7533 || J. Lentini, R. Tewari, C. Lever, Ed. || jlentini@netapp.com, tewarir@us.ibm.com, chuck.lever@oracle.com +# RFC7534 || J. Abley, W. Sotomayor || jabley@dyn.com, wfms@ottix.net +# RFC7535 || J. Abley, B. Dickson, W. Kumari, G. Michaelson || jabley@dyn.com, bdickson@twitter.com, warren@kumari.net, ggm@apnic.net +# RFC7536 || M. Linsner, P. Eardley, T. Burbridge, F. Sorensen || mlinsner@cisco.com, philip.eardley@bt.com, trevor.burbridge@bt.com, frode.sorensen@nkom.no +# RFC7537 || B. Decraene, N. Akiya, C. Pignataro, L. Andersson, S. Aldrin || bruno.decraene@orange.com, nobo.akiya.dev@gmail.com, cpignata@cisco.com, loa@mail01.huawei.com, aldrin.ietf@gmail.com +# RFC7538 || J. Reschke || julian.reschke@greenbytes.de +# RFC7539 || Y. Nir, A. Langley || ynir.ietf@gmail.com, agl@google.com +# RFC7540 || M. Belshe, R. Peon, M. Thomson, Ed. || mike@belshe.com, fenix@google.com, martin.thomson@gmail.com +# RFC7541 || R. Peon, H. Ruellan || fenix@google.com, herve.ruellan@crf.canon.fr +# RFC7542 || A. DeKok || aland@freeradius.org +# RFC7543 || H. Jeng, L. Jalil, R. Bonica, K. Patel, L. Yong || hj2387@att.com, luay.jalil@verizon.com, rbonica@juniper.net, keyupate@cisco.com, lucy.yong@huawei.com +# RFC7544 || M. Mohali || marianne.mohali@orange.com +# RFC7545 || V. Chen, Ed., S. Das, L. Zhu, J. Malyar, P. McCann || vchen@google.com, sdas@appcomsci.com, lei.zhu@huawei.com, jmalyar@iconectiv.com, peter.mccann@huawei.com +# RFC7546 || B. Kaduk || kaduk@mit.edu +# RFC7547 || M. Ersue, Ed., D. Romascanu, J. Schoenwaelder, U. Herberg || mehmet.ersue@nsn.com, dromasca@gmail.com , j.schoenwaelder@jacobs-university.de, ulrich@herberg.name +# RFC7548 || M. Ersue, Ed., D. Romascanu, J. Schoenwaelder, A. Sehgal || mehmet.ersue@nsn.com, dromasca@gmail.com , j.schoenwaelder@jacobs-university.de, s.anuj@jacobs-university.de +# RFC7549 || C. Holmberg, J. Holm, R. Jesske, M. Dolly || christer.holmberg@ericsson.com, jan.holm@ericsson.com, r.jesske@telekom.de, md3135@att.com +# RFC7550 || O. Troan, B. Volz, M. Siodelski || ot@cisco.com, volz@cisco.com, msiodelski@gmail.com +# RFC7551 || F. Zhang, Ed., R. Jing, R. Gandhi, Ed. || zhangfei7@huawei.com, jingrq@ctbri.com.cn, rgandhi@cisco.com +# RFC7552 || R. Asati, C. Pignataro, K. Raza, V. Manral, R. Papneja || rajiva@cisco.com, cpignata@cisco.com, skraza@cisco.com, vishwas@ionosnetworks.com, rajiv.papneja@huawei.com +# RFC7553 || P. Faltstrom, O. Kolkman || paf@netnod.se, kolkman@isoc.org +# RFC7554 || T. Watteyne, Ed., M. Palattella, L. Grieco || twatteyne@linear.com, maria-rita.palattella@uni.lu, a.grieco@poliba.it +# RFC7555 || G. Swallow, V. Lim, S. Aldrin || swallow@cisco.com, vlim@cisco.com, aldrin.ietf@gmail.com +# RFC7556 || D. Anipko, Ed. || dmitry.anipko@gmail.com +# RFC7557 || J. Chroboczek || jch@pps.univ-paris-diderot.fr +# RFC7558 || K. Lynn, S. Cheshire, M. Blanchet, D. Migault || kerry.lynn@verizon.com, cheshire@apple.com, Marc.Blanchet@viagenie.ca, daniel.migault@ericsson.com +# RFC7559 || S. Krishnan, D. Anipko, D. Thaler || suresh.krishnan@ericsson.com, dmitry.anipko@gmail.com, dthaler@microsoft.com +# RFC7560 || M. Kuehlewind, Ed., R. Scheffenegger, B. Briscoe || mirja.kuehlewind@tik.ee.ethz.ch, rs@netapp.com, ietf@bobbriscoe.net +# RFC7561 || J. Kaippallimalil, R. Pazhyannur, P. Yegani || john.kaippallimalil@huawei.com, rpazhyan@cisco.com, pyegani@juniper.net +# RFC7562 || D. Thakore || d.thakore@cablelabs.com +# RFC7563 || R. Pazhyannur, S. Speicher, S. Gundavelli, J. Korhonen, J. Kaippallimalil || rpazhyan@cisco.com, sespeich@cisco.com, sgundave@cisco.com, jouni.nospam@gmail.com, john.kaippallimalil@huawei.com +# RFC7564 || P. Saint-Andre, M. Blanchet || peter@andyet.com, Marc.Blanchet@viagenie.ca +# RFC7565 || P. Saint-Andre || peter@andyet.com +# RFC7566 || L. Goix, K. Li || laurent.goix@econocom-osiatis.com, kepeng.likp@gmail.com +# RFC7567 || F. Baker, Ed., G. Fairhurst, Ed. || fred@cisco.com, gorry@erg.abdn.ac.uk +# RFC7568 || R. Barnes, M. Thomson, A. Pironti, A. Langley || rlb@ipv.sx, martin.thomson@gmail.com, alfredo@pironti.eu, agl@google.com +# RFC7569 || D. Quigley, J. Lu, T. Haynes || dpquigl@davequigley.com, Jarrett.Lu@oracle.com, thomas.haynes@primarydata.com +# RFC7570 || C. Margaria, Ed., G. Martinelli, S. Balls, B. Wright || cmargaria@juniper.net, giomarti@cisco.com, steve.balls@metaswitch.com, ben.wright@metaswitch.com +# RFC7571 || J. Dong, M. Chen, Z. Li, D. Ceccarelli || jie.dong@huawei.com, mach.chen@huawei.com, lizhenqiang@chinamobile.com, daniele.ceccarelli@ericsson.com +# RFC7572 || P. Saint-Andre, A. Houri, J. Hildebrand || peter@andyet.com, avshalom@il.ibm.com, jhildebr@cisco.com +# RFC7573 || P. Saint-Andre, S. Loreto || peter@andyet.com, Salvatore.Loreto@ericsson.com +# RFC7574 || A. Bakker, R. Petrocco, V. Grishchenko || arno@cs.vu.nl, r.petrocco@gmail.com, victor.grishchenko@gmail.com +# RFC7575 || M. Behringer, M. Pritikin, S. Bjarnason, A. Clemm, B. Carpenter, S. Jiang, L. Ciavaglia || mbehring@cisco.com, pritikin@cisco.com, sbjarnas@cisco.com, alex@cisco.com, brian.e.carpenter@gmail.com, jiangsheng@huawei.com, Laurent.Ciavaglia@alcatel-lucent.com +# RFC7576 || S. Jiang, B. Carpenter, M. Behringer || jiangsheng@huawei.com, brian.e.carpenter@gmail.com, mbehring@cisco.com +# RFC7577 || J. Quittek, R. Winter, T. Dietz || quittek@neclab.eu, rolf.winter@neclab.eu, Thomas.Dietz@neclab.eu +# RFC7578 || L. Masinter || masinter@adobe.com +# RFC7579 || G. Bernstein, Ed., Y. Lee, Ed., D. Li, W. Imajuku, J. Han || gregb@grotto-networking.com, ylee@huawei.com, danli@huawei.com, imajuku.wataru@lab.ntt.co.jp, hanjianrui@huawei.com +# RFC7580 || F. Zhang, Y. Lee, J. Han, G. Bernstein, Y. Xu || zhangfatai@huawei.com, leeyoung@huawei.com, hanjianrui@huawei.com, gregb@grotto-networking.com, xuyunbin@mail.ritt.com.cn +# RFC7581 || G. Bernstein, Ed., Y. Lee, Ed., D. Li, W. Imajuku, J. Han || gregb@grotto-networking.com, leeyoung@huawei.com, danli@huawei.com, imajuku.wataru@lab.ntt.co.jp, hanjianrui@huawei.com +# RFC7582 || E. Rosen, IJ. Wijnands, Y. Cai, A. Boers || erosen@juniper.net, ice@cisco.com, yiqunc@microsoft.com, arjen@boers.com +# RFC7583 || S. Morris, J. Ihren, J. Dickinson, W. Mekking || stephen@isc.org, johani@netnod.se, jad@sinodun.com, mmekking@dyn.com +# RFC7584 || R. Ravindranath, T. Reddy, G. Salgueiro || rmohanr@cisco.com, tireddy@cisco.com, gsalguei@cisco.com +# RFC7585 || S. Winter, M. McCauley || stefan.winter@restena.lu, mikem@airspayce.com +# RFC7586 || Y. Nachum, L. Dunbar, I. Yerushalmi, T. Mizrahi || youval.nachum@gmail.com, ldunbar@huawei.com, yilan@marvell.com, talmi@marvell.com +# RFC7587 || J. Spittka, K. Vos, JM. Valin || jspittka@gmail.com, koenvos74@gmail.com, jmvalin@jmvalin.ca +# RFC7588 || R. Bonica, C. Pignataro, J. Touch || rbonica@juniper.net, cpignata@cisco.com, touch@isi.edu +# RFC7589 || M. Badra, A. Luchuk, J. Schoenwaelder || mohamad.badra@zu.ac.ae, luchuk@snmp.com, j.schoenwaelder@jacobs-university.de +# RFC7590 || P. Saint-Andre, T. Alkemade || peter@andyet.com, me@thijsalkema.de +# RFC7591 || J. Richer, Ed., M. Jones, J. Bradley, M. Machulak, P. Hunt || ietf@justin.richer.org, mbj@microsoft.com, ve7jtb@ve7jtb.com, maciej.machulak@gmail.com, phil.hunt@yahoo.com +# RFC7592 || J. Richer, Ed., M. Jones, J. Bradley, M. Machulak || ietf@justin.richer.org, mbj@microsoft.com, ve7jtb@ve7jtb.com, maciej.machulak@gmail.com +# RFC7593 || K. Wierenga, S. Winter, T. Wolniewicz || klaas@cisco.com, stefan.winter@restena.lu, twoln@umk.pl +# RFC7594 || P. Eardley, A. Morton, M. Bagnulo, T. Burbridge, P. Aitken, A. Akhter || philip.eardley@bt.com, acmorton@att.com, marcelo@it.uc3m.es, trevor.burbridge@bt.com, paitken@brocade.com, aakhter@gmail.com +# RFC7595 || D. Thaler, Ed., T. Hansen, T. Hardie || dthaler@microsoft.com, tony+urireg@maillennium.att.com, ted.ietf@gmail.com +# RFC7596 || Y. Cui, Q. Sun, M. Boucadair, T. Tsou, Y. Lee, I. Farrer || yong@csnet1.cs.tsinghua.edu.cn, sunqiong@ctbri.com.cn, mohamed.boucadair@orange.com, tena@huawei.com, yiu_lee@cable.comcast.com, ian.farrer@telekom.de +# RFC7597 || O. Troan, Ed., W. Dec, X. Li, C. Bao, S. Matsushima, T. Murakami, T. Taylor, Ed. || ot@cisco.com, wdec@cisco.com, xing@cernet.edu.cn, congxiao@cernet.edu.cn, satoru.matsushima@g.softbank.co.jp, tetsuya@ipinfusion.com, tom.taylor.stds@gmail.com +# RFC7598 || T. Mrugalski, O. Troan, I. Farrer, S. Perreault, W. Dec, C. Bao, L. Yeh, X. Deng || tomasz.mrugalski@gmail.com, ot@cisco.com, ian.farrer@telekom.de, sperreault@jive.com, wdec@cisco.com, congxiao@cernet.edu.cn, leaf.y.yeh@hotmail.com, dxhbupt@gmail.com +# RFC7599 || X. Li, C. Bao, W. Dec, Ed., O. Troan, S. Matsushima, T. Murakami || xing@cernet.edu.cn, congxiao@cernet.edu.cn, wdec@cisco.com, ot@cisco.com, satoru.matsushima@g.softbank.co.jp, tetsuya@ipinfusion.com +# RFC7600 || R. Despres, S. Jiang, Ed., R. Penno, Y. Lee, G. Chen, M. Chen || despres.remi@laposte.net, jiangsheng@huawei.com, repenno@cisco.com, yiu_lee@cable.comcast.com, phdgang@gmail.com, maoke@bbix.net +# RFC7601 || M. Kucherawy || superuser@gmail.com +# RFC7602 || U. Chunduri, W. Lu, A. Tian, N. Shen || uma.chunduri@ericsson.com, wenhu.lu@ericsson.com, albert.tian@ericsson.com, naiming@cisco.com +# RFC7603 || B. Schoening, M. Chandramouli, B. Nordman || brad.schoening@verizon.net, moulchan@cisco.com, bnordman@lbl.gov +# RFC7604 || M. Westerlund, T. Zeng || magnus.westerlund@ericsson.com, thomas.zeng@gmail.com +# RFC7605 || J. Touch || touch@isi.edu +# RFC7606 || E. Chen, Ed., J. Scudder, Ed., P. Mohapatra, K. Patel || enkechen@cisco.com, jgs@juniper.net, mpradosh@yahoo.com, keyupate@cisco.com +# RFC7607 || W. Kumari, R. Bush, H. Schiller, K. Patel || warren@kumari.net, randy@psg.com, has@google.com, keyupate@cisco.com +# RFC7608 || M. Boucadair, A. Petrescu, F. Baker || mohamed.boucadair@orange.com, alexandre.petrescu@cea.fr, fred@cisco.com +# RFC7609 || M. Fox, C. Kassimis, J. Stevens || mjfox@us.ibm.com, kassimis@us.ibm.com, sjerry@us.ibm.com +# RFC7610 || F. Gont, W. Liu, G. Van de Velde || fgont@si6networks.com, liushucheng@huawei.com, gunter.van_de_velde@alcatel-lucent.com +# RFC7611 || J. Uttaro, P. Mohapatra, D. Smith, R. Raszuk, J. Scudder || uttaro@att.com, mpradosh@yahoo.com, djsmith@cisco.com, robert@raszuk.net, jgs@juniper.net +# RFC7612 || P. Fleming, I. McDonald || patfleminghtc@gmail.com, blueroofmusic@gmail.com +# RFC7613 || P. Saint-Andre, A. Melnikov || peter@andyet.com, alexey.melnikov@isode.com +# RFC7614 || R. Sparks || rjsparks@nostrum.com +# RFC7615 || J. Reschke || julian.reschke@greenbytes.de +# RFC7616 || R. Shekh-Yusef, Ed., D. Ahrens, S. Bremer || rifaat.ietf@gmail.com, ahrensdc@gmail.com, sophie.bremer@netzkonform.de +# RFC7617 || J. Reschke || julian.reschke@greenbytes.de +# RFC7618 || Y. Cui, Q. Sun, I. Farrer, Y. Lee, Q. Sun, M. Boucadair || yong@csnet1.cs.tsinghua.edu.cn, sunqi.ietf@gmail.com, ian.farrer@telekom.de, yiu_lee@cable.comcast.com, sunqiong@ctbri.com.cn, mohamed.boucadair@orange.com +# RFC7619 || V. Smyslov, P. Wouters || svan@elvis.ru, pwouters@redhat.com +# RFC7620 || M. Boucadair, Ed., B. Chatras, T. Reddy, B. Williams, B. Sarikaya || mohamed.boucadair@orange.com, bruno.chatras@orange.com, tireddy@cisco.com, brandon.williams@akamai.com, sarikaya@ieee.org +# RFC7621 || A.B. Roach || adam@nostrum.com +# RFC7622 || P. Saint-Andre || peter@andyet.com +# RFC7623 || A. Sajassi, Ed., S. Salam, N. Bitar, A. Isaac, W. Henderickx || sajassi@cisco.com, ssalam@cisco.com, nabil.n.bitar@verizon.com, aisaac@juniper.net, wim.henderickx@alcatel-lucent.com +# RFC7624 || R. Barnes, B. Schneier, C. Jennings, T. Hardie, B. Trammell, C. Huitema, D. Borkmann || rlb@ipv.sx, schneier@schneier.com, fluffy@cisco.com, ted.ietf@gmail.com, ietf@trammell.ch, huitema@huitema.net, daniel@iogearbox.net +# RFC7625 || J. T. Hao, P. Maheshwari, R. Huang, L. Andersson, M. Chen || haojiangtao@huawei.com, praveen.maheshwari@in.airtel.com, river.huang@huawei.com, loa@mail01.huawei.com, mach.chen@huawei.com +# RFC7626 || S. Bortzmeyer || bortzmeyer+ietf@nic.fr +# RFC7627 || K. Bhargavan, Ed., A. Delignat-Lavaud, A. Pironti, A. Langley, M. Ray || karthikeyan.bhargavan@inria.fr, antoine.delignat-lavaud@inria.fr, alfredo.pironti@inria.fr, agl@google.com, maray@microsoft.com +# RFC7628 || W. Mills, T. Showalter, H. Tschofenig || wmills_92105@yahoo.com, tjs@psaux.com, Hannes.Tschofenig@gmx.net +# RFC7629 || S. Gundavelli, Ed., K. Leung, G. Tsirtsis, A. Petrescu || sgundave@cisco.com, kleung@cisco.com, tsirtsis@qualcomm.com, alexandru.petrescu@cea.fr +# RFC7630 || J. Merkle, Ed., M. Lochter || johannes.merkle@secunet.com, manfred.lochter@bsi.bund.de +# RFC7631 || C. Dearlove, T. Clausen || chris.dearlove@baesystems.com, T.Clausen@computer.org +# RFC7632 || D. Waltermire, D. Harrington || david.waltermire@nist.gov, ietfdbh@gmail.com +# RFC7633 || P. Hallam-Baker || philliph@comodo.com +# RFC7634 || Y. Nir || ynir.ietf@gmail.com +# RFC7635 || T. Reddy, P. Patil, R. Ravindranath, J. Uberti || tireddy@cisco.com, praspati@cisco.com, rmohanr@cisco.com, justin@uberti.name +# RFC7636 || N. Sakimura, Ed., J. Bradley, N. Agarwal || n-sakimura@nri.co.jp, ve7jtb@ve7jtb.com, naa@google.com +# RFC7637 || P. Garg, Ed., Y. Wang, Ed. || pankajg@microsoft.com, yushwang@microsoft.com +# RFC7638 || M. Jones, N. Sakimura || mbj@microsoft.com, n-sakimura@nri.co.jp +# RFC7639 || A. Hutton, J. Uberti, M. Thomson || andrew.hutton@unify.com, justin@uberti.name, martin.thomson@gmail.com +# RFC7640 || B. Constantine, R. Krishnan || barry.constantine@jdsu.com, ramkri123@gmail.com +# RFC7641 || K. Hartke || hartke@tzi.org +# RFC7642 || K. LI, Ed., P. Hunt, B. Khasnabish, A. Nadalin, Z. Zeltsan || kepeng.lkp@alibaba-inc.com, phil.hunt@oracle.com, vumip1@gmail.com, tonynad@microsoft.com, zachary.zeltsan@gmail.com +# RFC7643 || P. Hunt, Ed., K. Grizzle, E. Wahlstroem, C. Mortimore || phil.hunt@yahoo.com, kelly.grizzle@sailpoint.com, erik.wahlstrom@nexusgroup.com, cmortimore@salesforce.com +# RFC7644 || P. Hunt, Ed., K. Grizzle, M. Ansari, E. Wahlstroem, C. Mortimore || phil.hunt@yahoo.com, kelly.grizzle@sailpoint.com, morteza.ansari@cisco.com, erik.wahlstrom@nexusgroup.com, cmortimore@salesforce.com +# RFC7645 || U. Chunduri, A. Tian, W. Lu || uma.chunduri@ericsson.com, albert.tian@ericsson.com, wenhu.lu@ericsson.com +# RFC7646 || P. Ebersman, W. Kumari, C. Griffiths, J. Livingood, R. Weber || ebersman-ietf@dragon.net, warren@kumari.net, cgriffiths@gmail.com, jason_livingood@cable.comcast.com, ralf.weber@nominum.com +# RFC7647 || R. Sparks, A.B. Roach || rjsparks@nostrum.com, adam@nostrum.com +# RFC7648 || S. Perreault, M. Boucadair, R. Penno, D. Wing, S. Cheshire || sperreault@jive.com, mohamed.boucadair@orange.com, repenno@cisco.com, dwing-ietf@fuggles.com, cheshire@apple.com +# RFC7649 || P. Saint-Andre, D. York || peter@andyet.com, york@isoc.org +# RFC7650 || J. Jimenez, J. Lopez-Vega, J. Maenpaa, G. Camarillo || jaime.jimenez@ericsson.com, jmlvega@ugr.es, jouni.maenpaa@ericsson.com, gonzalo.camarillo@ericsson.com +# RFC7651 || A. Dodd-Noble, S. Gundavelli, J. Korhonen, F. Baboescu, B. Weis || noblea@cisco.com, sgundave@cisco.com, jouni.nospam@gmail.com, baboescu@broadcom.com, bew@cisco.com +# RFC7652 || M. Cullen, S. Hartman, D. Zhang, T. Reddy || margaret@painless-security.com, hartmans@painless-security.com, zhang_dacheng@hotmail.com, tireddy@cisco.com +# RFC7653 || D. Raghuvanshi, K. Kinnear, D. Kukrety || draghuva@cisco.com, kkinnear@cisco.com, dkukrety@cisco.com +# RFC7654 || S. Banks, F. Calabria, G. Czirjak, R. Machat || sbanks@encrypted.net, fcalabri@cisco.com, gczirjak@juniper.net, rmachat@juniper.net +# RFC7655 || M. Ramalho, Ed., P. Jones, N. Harada, M. Perumal, L. Miao || mramalho@cisco.com, paulej@packetizer.com, harada.noboru@lab.ntt.co.jp, muthu.arul@gmail.com, lei.miao@huawei.com +# RFC7656 || J. Lennox, K. Gross, S. Nandakumar, G. Salgueiro, B. Burman, Ed. || jonathan@vidyo.com, kevin.gross@avanw.com, snandaku@cisco.com, gsalguei@cisco.com, bo.burman@ericsson.com +# RFC7657 || D. Black, Ed., P. Jones || david.black@emc.com, paulej@packetizer.com +# RFC7658 || S. Perreault, T. Tsou, S. Sivakumar, T. Taylor || sperreault@jive.com, tina.tsou.zouting@huawei.com, ssenthil@cisco.com, tom.taylor.stds@gmail.com +# RFC7659 || S. Perreault, T. Tsou, S. Sivakumar, T. Taylor || sperreault@jive.com, tina.tsou.zouting@huawei.com, ssenthil@cisco.com, tom.taylor.stds@gmail.com +# RFC7660 || L. Bertz, S. Manning, B. Hirschman || lyleb551144@gmail.com, sergem913@gmail.com, Brent.Hirschman@gmail.com +# RFC7661 || G. Fairhurst, A. Sathiaseelan, R. Secchi || gorry@erg.abdn.ac.uk, arjuna@erg.abdn.ac.uk, raffaello@erg.abdn.ac.uk +# RFC7662 || J. Richer, Ed. || ietf@justin.richer.org +# RFC7663 || B. Trammell, Ed., M. Kuehlewind, Ed. || ietf@trammell.ch, mirja.kuehlewind@tik.ee.ethz.ch +# RFC7664 || D. Harkins, Ed. || dharkins@arubanetworks.com +# RFC7665 || J. Halpern, Ed., C. Pignataro, Ed. || jmh@joelhalpern.com, cpignata@cisco.com +# RFC7666 || H. Asai, M. MacFaden, J. Schoenwaelder, K. Shima, T. Tsou || panda@hongo.wide.ad.jp, mrm@vmware.com, j.schoenwaelder@jacobs-university.de, keiichi@iijlab.net, tina.tsou.zouting@huawei.com +# RFC7667 || M. Westerlund, S. Wenger || magnus.westerlund@ericsson.com, stewe@stewe.org +# RFC7668 || J. Nieminen, T. Savolainen, M. Isomaki, B. Patil, Z. Shelby, C. Gomez || johannamaria.nieminen@gmail.com, teemu.savolainen@nokia.com, markus.isomaki@nokia.com, basavaraj.patil@att.com, zach.shelby@arm.com, carlesgo@entel.upc.edu +# RFC7669 || J. Levine || standards@taugh.com +# RFC7670 || T. Kivinen, P. Wouters, H. Tschofenig || kivinen@iki.fi, pwouters@redhat.com, Hannes.Tschofenig@gmx.net +# RFC7671 || V. Dukhovni, W. Hardaker || ietf-dane@dukhovni.org, ietf@hardakers.net +# RFC7672 || V. Dukhovni, W. Hardaker || ietf-dane@dukhovni.org, ietf@hardakers.net +# RFC7673 || T. Finch, M. Miller, P. Saint-Andre || dot@dotat.at, mamille2@cisco.com, peter@andyet.com +# RFC7674 || J. Haas, Ed. || jhaas@juniper.net +# RFC7675 || M. Perumal, D. Wing, R. Ravindranath, T. Reddy, M. Thomson || muthu.arul@gmail.com, dwing-ietf@fuggles.com, rmohanr@cisco.com, tireddy@cisco.com, martin.thomson@gmail.com +# RFC7676 || C. Pignataro, R. Bonica, S. Krishnan || cpignata@cisco.com, rbonica@juniper.net, suresh.krishnan@ericsson.com +# RFC7677 || T. Hansen || tony+scramsha256@maillennium.att.com +# RFC7678 || C. Zhou, T. Taylor, Q. Sun, M. Boucadair || cathy.zhou@huawei.com, tom.taylor.stds@gmail.com, sunqiong@ctbri.com.cn, mohamed.boucadair@orange.com +# RFC7679 || G. Almes, S. Kalidindi, M. Zekauskas, A. Morton, Ed. || almes@acm.org, skalidindi@ixiacom.com, matt@internet2.edu, acmorton@att.com +# RFC7680 || G. Almes, S. Kalidindi, M. Zekauskas, A. Morton, Ed. || almes@acm.org, skalidindi@ixiacom.com, matt@internet2.edu, acmorton@att.com +# RFC7681 || J. Davin || info@eesst.org +# RFC7682 || D. McPherson, S. Amante, E. Osterweil, L. Blunk, D. Mitchell || dmcpherson@verisign.com, amante@apple.com, eosterweil@verisign.com, ljb@merit.edu, dave@singularity.cx +# RFC7683 || J. Korhonen, Ed., S. Donovan, Ed., B. Campbell, L. Morand || jouni.nospam@gmail.com, srdonovan@usdonovans.com, ben@nostrum.com, lionel.morand@orange.com +# RFC7684 || P. Psenak, H. Gredler, R. Shakir, W. Henderickx, J. Tantsura, A. Lindem || ppsenak@cisco.com, hannes@gredler.at, rjs@rob.sh, wim.henderickx@alcatel-lucent.com, jeff.tantsura@ericsson.com, acee@cisco.com +# RFC7685 || A. Langley || agl@google.com +# RFC7686 || J. Appelbaum, A. Muffett || jacob@appelbaum.net, alecm@fb.com +# RFC7687 || S. Farrell, R. Wenning, B. Bos, M. Blanchet, H. Tschofenig || stephen.farrell@cs.tcd.ie, rigo@w3.org, bert@w3.org, Marc.Blanchet@viagenie.ca, Hannes.Tschofenig@gmx.net +# RFC7688 || Y. Lee, Ed., G. Bernstein, Ed. || leeyoung@huawei.com, gregb@grotto-networking.com +# RFC7689 || G. Bernstein, Ed., S. Xu, Y. Lee, Ed., G. Martinelli, H. Harai || gregb@grotto-networking.com, xsg@nict.go.jp, leeyoung@huawei.com, giomarti@cisco.com, harai@nict.go.jp +# RFC7690 || M. Byerly, M. Hite, J. Jaeggli || suckawha@gmail.com, mhite@hotmail.com, joelja@gmail.com +# RFC7691 || S. Bradner, Ed. || sob@harvard.edu +# RFC7692 || T. Yoshino || tyoshino@google.com +# RFC7693 || M-J. Saarinen, Ed., J-P. Aumasson || m.saarinen@qub.ac.uk, jean-philippe.aumasson@nagra.com +# RFC7694 || J. Reschke || julian.reschke@greenbytes.de +# RFC7695 || P. Pfister, B. Paterson, J. Arkko || pierre.pfister@darou.fr, paterson.b@gmail.com, jari.arkko@piuha.net +# RFC7696 || R. Housley || housley@vigilsec.com +# RFC7697 || P. Pan, S. Aldrin, M. Venkatesan, K. Sampath, T. Nadeau, S. Boutros || none, aldrin.ietf@gmail.com, venkat.mahalingams@gmail.com, kannankvs@gmail.com, tnadeau@lucidvision.com, sboutros@vmware.com +# RFC7698 || O. Gonzalez de Dios, Ed., R. Casellas, Ed., F. Zhang, X. Fu, D. Ceccarelli, I. Hussain || oscar.gonzalezdedios@telefonica.com, ramon.casellas@cttc.es, zhangfatai@huawei.com, fu.xihua@stairnote.com, daniele.ceccarelli@ericsson.com, ihussain@infinera.com +# RFC7699 || A. Farrel, D. King, Y. Li, F. Zhang || adrian@olddog.co.uk, daniel@olddog.co.uk, wsliguotou@hotmail.com, zhangfatai@huawei.com +# RFC7700 || P. Saint-Andre || peter@andyet.com +# RFC7701 || A. Niemi, M. Garcia-Martin, G. Sandbakken || aki.niemi@iki.fi, miguel.a.garcia@ericsson.com, geirsand@cisco.com +# RFC7702 || P. Saint-Andre, S. Ibarra, S. Loreto || peter@andyet.com, saul@ag-projects.com, Salvatore.Loreto@ericsson.com +# RFC7703 || E. Cordeiro, R. Carnier, A. Moreiras || edwin@scordeiro.net, rodrigocarnier@gmail.com, moreiras@nic.br +# RFC7704 || D. Crocker, N. Clark || dcrocker@bbiw.net, narelle.clark@pavonis.com.au +# RFC7705 || W. George, S. Amante || wesley.george@twcable.com, amante@apple.com +# RFC7706 || W. Kumari, P. Hoffman || warren@kumari.net, paul.hoffman@icann.org +# RFC7707 || F. Gont, T. Chown || fgont@si6networks.com, tim.chown@jisc.ac.uk +# RFC7708 || T. Nadeau, L. Martini, S. Bryant || tnadeau@lucidvision.com, lmartini@cisco.com, stewart.bryant@gmail.com +# RFC7709 || A. Malis, Ed., B. Wilson, G. Clapp, V. Shukla || agmalis@gmail.com, bwilson@appcomsci.com, clapp@research.att.com, vishnu.shukla@verizon.com +# RFC7710 || W. Kumari, O. Gudmundsson, P. Ebersman, S. Sheng || warren@kumari.net, olafur@cloudflare.com, ebersman-ietf@dragon.net, steve.sheng@icann.org +# RFC7711 || M. Miller, P. Saint-Andre || mamille2@cisco.com, peter@andyet.com +# RFC7712 || P. Saint-Andre, M. Miller, P. Hancke || peter@andyet.com, mamille2@cisco.com, fippo@andyet.com +# RFC7713 || M. Mathis, B. Briscoe || mattmathis@google.com, ietf@bobbriscoe.net +# RFC7714 || D. McGrew, K. Igoe || mcgrew@cisco.com, mythicalkevin@yahoo.com +# RFC7715 || IJ. Wijnands, Ed., K. Raza, A. Atlas, J. Tantsura, Q. Zhao || ice@cisco.com, skraza@cisco.com, akatlas@juniper.net, jeff.tantsura@ericsson.com, quintin.zhao@huawei.com +# RFC7716 || J. Zhang, L. Giuliano, E. Rosen, Ed., K. Subramanian, D. Pacella || zzhang@juniper.net, lenny@juniper.net, erosen@juniper.net, kartsubr@cisco.com, dante.j.pacella@verizonbusiness.com +# RFC7717 || K. Pentikousis, Ed., E. Zhang, Y. Cui || k.pentikousis@eict.de, emma.zhanglijia@huawei.com, cuiyang@huawei.com +# RFC7718 || A. Morton || acmorton@att.com +# RFC7719 || P. Hoffman, A. Sullivan, K. Fujiwara || paul.hoffman@icann.org, asullivan@dyn.com, fujiwara@jprs.co.jp +# RFC7720 || M. Blanchet, L-J. Liman || Marc.Blanchet@viagenie.ca, liman@netnod.se +# RFC7721 || A. Cooper, F. Gont, D. Thaler || alcoop@cisco.com, fgont@si6networks.com, dthaler@microsoft.com +# RFC7722 || C. Dearlove, T. Clausen || chris.dearlove@baesystems.com, T.Clausen@computer.org +# RFC7723 || S. Kiesel, R. Penno || ietf-pcp@skiesel.de, repenno@cisco.com +# RFC7724 || K. Kinnear, M. Stapp, B. Volz, N. Russell || kkinnear@cisco.com, mjs@cisco.com, volz@cisco.com, neil.e.russell@gmail.com +# RFC7725 || T. Bray || tbray@textuality.com +# RFC7726 || V. Govindan, K. Rajaraman, G. Mirsky, N. Akiya, S. Aldrin || venggovi@cisco.com, kalyanir@cisco.com, gregory.mirsky@ericsson.com, nobo.akiya.dev@gmail.com, aldrin.ietf@gmail.com +# RFC7727 || M. Zhang, H. Wen, J. Hu || zhangmingui@huawei.com, wenhuafeng@huawei.com, hujie@ctbri.com.cn +# RFC7728 || B. Burman, A. Akram, R. Even, M. Westerlund || bo.burman@ericsson.com, akram.muhammadazam@gmail.com, roni.even@mail01.huawei.com, magnus.westerlund@ericsson.com +# RFC7729 || B. Khasnabish, E. Haleplidis, J. Hadi Salim, Ed. || vumip1@gmail.com, ehalep@ece.upatras.gr, hadi@mojatatu.com +# RFC7730 || G. Huston, S. Weiler, G. Michaelson, S. Kent || gih@apnic.net, weiler@tislabs.com, ggm@apnic.net, kent@bbn.com +# RFC7731 || J. Hui, R. Kelsey || jonhui@nestlabs.com, richard.kelsey@silabs.com +# RFC7732 || P. van der Stok, R. Cragie || consultancy@vanderstok.org, robert.cragie@arm.com +# RFC7733 || A. Brandt, E. Baccelli, R. Cragie, P. van der Stok || anders_brandt@sigmadesigns.com, Emmanuel.Baccelli@inria.fr, robert.cragie@arm.com, consultancy@vanderstok.org +# RFC7734 || D. Allan, Ed., J. Tantsura, D. Fedyk, A. Sajassi || david.i.allan@ericsson.com, jeff.tantsura@ericsson.com, don.fedyk@hpe.com, sajassi@cisco.com +# RFC7735 || R. Sparks, T. Kivinen || rjsparks@nostrum.com, kivinen@iki.fi +# RFC7736 || K. Ma || kevin.j.ma@ericsson.com +# RFC7737 || N. Akiya, G. Swallow, C. Pignataro, L. Andersson, M. Chen || nobo.akiya.dev@gmail.com, swallow@cisco.com, cpignata@cisco.com, loa@mail01.huawei.com, mach.chen@huawei.com +# RFC7738 || M. Blanchet, A. Schiltknecht, P. Shames || Marc.Blanchet@viagenie.ca, audric.schiltknecht@viagenie.ca, peter.m.shames@jpl.nasa.gov +# RFC7739 || F. Gont || fgont@si6networks.com +# RFC7740 || Z. Zhang, Y. Rekhter, A. Dolganow || zzhang@juniper.net, none, andrew.dolganow@alcatel-lucent.com +# RFC7741 || P. Westin, H. Lundin, M. Glover, J. Uberti, F. Galligan || patrik.westin@gmail.com, hlundin@google.com, michaelglover262@gmail.com, justin@uberti.name, fgalligan@google.com +# RFC7742 || A.B. Roach || adam@nostrum.com +# RFC7743 || J. Luo, Ed., L. Jin, Ed., T. Nadeau, Ed., G. Swallow, Ed. || luo.jian@zte.com.cn, lizho.jin@gmail.com, tnadeau@lucidvision.com, swallow@cisco.com +# RFC7744 || L. Seitz, Ed., S. Gerdes, Ed., G. Selander, M. Mani, S. Kumar || ludwig@sics.se, gerdes@tzi.org, goran.selander@ericsson.com, mehdi.mani@itron.com, sandeep.kumar@philips.com +# RFC7745 || T. Manderson || terry.manderson@icann.org +# RFC7746 || R. Bonica, I. Minei, M. Conn, D. Pacella, L. Tomotaki || rbonica@juniper.net, inaminei@google.com, meconn26@gmail.com, dante.j.pacella@verizon.com, luis.tomotaki@verizon.com +# RFC7747 || R. Papneja, B. Parise, S. Hares, D. Lee, I. Varlashkin || rajiv.papneja@huawei.com, bparise@skyportsystems.com, shares@ndzh.com, dlee@ixiacom.com, ilya@nobulus.com +# RFC7748 || A. Langley, M. Hamburg, S. Turner || agl@google.com, mike@shiftleft.org, sean@sn3rd.com +# RFC7749 || J. Reschke || julian.reschke@greenbytes.de +# RFC7750 || J. Hedin, G. Mirsky, S. Baillargeon || jonas.hedin@ericsson.com, gregory.mirsky@ericsson.com, steve.baillargeon@ericsson.com +# RFC7751 || S. Sorce, T. Yu || ssorce@redhat.com, tlyu@mit.edu +# RFC7752 || H. Gredler, Ed., J. Medved, S. Previdi, A. Farrel, S. Ray || hannes@gredler.at, jmedved@cisco.com, sprevidi@cisco.com, adrian@olddog.co.uk, raysaikat@gmail.com +# RFC7753 || Q. Sun, M. Boucadair, S. Sivakumar, C. Zhou, T. Tsou, S. Perreault || sunqiong@ctbri.com.cn, mohamed.boucadair@orange.com, ssenthil@cisco.com, cathy.zhou@huawei.com, tina.tsou@philips.com, sperreault@jive.com +# RFC7754 || R. Barnes, A. Cooper, O. Kolkman, D. Thaler, E. Nordmark || rlb@ipv.sx, alcoop@cisco.com, kolkman@isoc.org, dthaler@microsoft.com, nordmark@arista.com +# RFC7755 || T. Anderson || tore@redpill-linpro.com +# RFC7756 || T. Anderson, S. Steffann || tore@redpill-linpro.com, sander@steffann.nl +# RFC7757 || T. Anderson, A. Leiva Popper || tore@redpill-linpro.com, ydahhrk@gmail.com +# RFC7758 || T. Mizrahi, Y. Moses || dew@tx.technion.ac.il, moses@ee.technion.ac.il +# RFC7759 || E. Bellagamba, G. Mirsky, L. Andersson, P. Skoldstrom, D. Ward, J. Drake || elisa.bellagamba@gmail.com, gregory.mirsky@ericsson.com, loa@mail01.huawei.com, pontus.skoldstrom@acreo.se, dward@cisco.com, jdrake@juniper.net +# RFC7760 || R. Housley || housley@vigilsec.com +# RFC7761 || B. Fenner, M. Handley, H. Holbrook, I. Kouvelas, R. Parekh, Z. Zhang, L. Zheng || fenner@arista.com, m.handley@cs.ucl.ac.uk, holbrook@arista.com, kouvelas@arista.com, riparekh@cisco.com, zzhang@juniper.net, vero.zheng@huawei.com +# RFC7762 || M. West || mkwst@google.com +# RFC7763 || S. Leonard || dev+ietf@seantek.com +# RFC7764 || S. Leonard || dev+ietf@seantek.com +# RFC7765 || P. Hurtig, A. Brunstrom, A. Petlund, M. Welzl || per.hurtig@kau.se, anna.brunstrom@kau.se, apetlund@simula.no, michawe@ifi.uio.no +# RFC7766 || J. Dickinson, S. Dickinson, R. Bellis, A. Mankin, D. Wessels || jad@sinodun.com, sara@sinodun.com, ray@isc.org, allison.mankin@gmail.com, dwessels@verisign.com +# RFC7767 || S. Vinapamula, S. Sivakumar, M. Boucadair, T. Reddy || sureshk@juniper.net, ssenthil@cisco.com, mohamed.boucadair@orange.com, tireddy@cisco.com +# RFC7768 || T. Tsou, W. Li, T. Taylor, J. Huang || tina.tsou@philips.com, mweiboli@gmail.com, tom.taylor.stds@gmail.com, james.huang@huawei.com +# RFC7769 || S. Sivabalan, S. Boutros, H. Shah, S. Aldrin, M. Venkatesan || msiva@cisco.com, sboutros@cisco.com, hshah@ciena.com, aldrin.ietf@gmail.com, mannan_venkatesan@cable.comcast.com +# RFC7770 || A. Lindem, Ed., N. Shen, JP. Vasseur, R. Aggarwal, S. Shaffer || acee@cisco.com, naiming@cisco.com, jpv@cisco.com, raggarwa_1@yahoo.com, sshaffer@akamai.com +# RFC7771 || A. Malis, Ed., L. Andersson, H. van Helvoort, J. Shin, L. Wang, A. D'Alessandro || agmalis@gmail.com, loa@mail01.huawei.com, huubatwork@gmail.com, jongyoon.shin@sk.com, wangleiyj@chinamobile.com, alessandro.dalessandro@telecomitalia.it +# RFC7772 || A. Yourtchenko, L. Colitti || ayourtch@cisco.com, lorenzo@google.com +# RFC7773 || S. Santesson || sts@aaa-sec.com +# RFC7774 || Y. Doi, M. Gillmore || yusuke.doi@toshiba.co.jp, matthew.gillmore@itron.com +# RFC7775 || L. Ginsberg, S. Litkowski, S. Previdi || ginsberg@cisco.com, stephane.litkowski@orange.com, sprevidi@cisco.com +# RFC7776 || P. Resnick, A. Farrel || presnick@qti.qualcomm.com, adrian@olddog.co.uk +# RFC7777 || S. Hegde, R. Shakir, A. Smirnov, Z. Li, B. Decraene || shraddha@juniper.net, rjs@rob.sh, as@cisco.com, lizhenbin@huawei.com, bruno.decraene@orange.com +# RFC7778 || D. Kutscher, F. Mir, R. Winter, S. Krishnan, Y. Zhang, CJ. Bernardos || kutscher@neclab.eu, faisal.mir@gmail.com, rolf.winter@neclab.eu, suresh.krishnan@ericsson.com, ying.zhang13@hp.com, cjbc@it.uc3m.es +# RFC7779 || H. Rogge, E. Baccelli || henning.rogge@fkie.fraunhofer.de, Emmanuel.Baccelli@inria.fr +# RFC7780 || D. Eastlake 3rd, M. Zhang, R. Perlman, A. Banerjee, A. Ghanwani, S. Gupta || d3e3e3@gmail.com, zhangmingui@huawei.com, radia@alum.mit.edu, ayabaner@cisco.com, anoop@alumni.duke.edu, sujay.gupta@ipinfusion.com +# RFC7781 || H. Zhai, T. Senevirathne, R. Perlman, M. Zhang, Y. Li || honjun.zhai@tom.com, tsenevir@gmail.com, radia@alum.mit.edu, zhangmingui@huawei.com, liyizhou@huawei.com +# RFC7782 || M. Zhang, R. Perlman, H. Zhai, M. Durrani, S. Gupta || zhangmingui@huawei.com, radia@alum.mit.edu, honjun.zhai@tom.com, mdurrani@cisco.com, sujay.gupta@ipinfusion.com +# RFC7783 || T. Senevirathne, J. Pathangi, J. Hudson || tsenevir@gmail.com, pathangi_janardhanan@dell.com, jon.hudson@gmail.com +# RFC7784 || D. Kumar, S. Salam, T. Senevirathne || dekumar@cisco.com, ssalam@cisco.com, tsenevir@gmail.com +# RFC7785 || S. Vinapamula, M. Boucadair || sureshk@juniper.net, mohamed.boucadair@orange.com +# RFC7786 || M. Kuehlewind, Ed., R. Scheffenegger || mirja.kuehlewind@tik.ee.ethz.ch, rs.ietf@gmx.at +# RFC7787 || M. Stenberg, S. Barth || markus.stenberg@iki.fi, cyrus@openwrt.org +# RFC7788 || M. Stenberg, S. Barth, P. Pfister || markus.stenberg@iki.fi, cyrus@openwrt.org, pierre.pfister@darou.fr +# RFC7789 || C. Cardona, P. Francois, P. Lucente || juancamilo.cardona@imdea.org, pifranco@cisco.com, plucente@cisco.com +# RFC7790 || Y. Yoneya, T. Nemoto || yoshiro.yoneya@jprs.co.jp, t.nemo10@kmd.keio.ac.jp +# RFC7791 || D. Migault, Ed., V. Smyslov || daniel.migault@ericsson.com, svan@elvis.ru +# RFC7792 || F. Zhang, X. Zhang, A. Farrel, O. Gonzalez de Dios, D. Ceccarelli || zhangfatai@huawei.com, zhang.xian@huawei.com, adrian@olddog.co.uk, oscar.gonzalezdedios@telefonica.com, daniele.ceccarelli@ericsson.com +# RFC7793 || M. Andrews || marka@isc.org +# RFC7794 || L. Ginsberg, Ed., B. Decraene, S. Previdi, X. Xu, U. Chunduri || ginsberg@cisco.com, bruno.decraene@orange.com, sprevidi@cisco.com, xuxiaohu@huawei.com, uma.chunduri@ericsson.com +# RFC7795 || J. Dong, H. Wang || jie.dong@huawei.com, rainsword.wang@huawei.com +# RFC7796 || Y. Jiang, Ed., L. Yong, M. Paul || jiangyuanlong@huawei.com, lucyyong@huawei.com, Manuel.Paul@telekom.de +# RFC7797 || M. Jones || mbj@microsoft.com +# RFC7798 || Y.-K. Wang, Y. Sanchez, T. Schierl, S. Wenger, M. M. Hannuksela || yekui.wang@gmail.com, yago.sanchez@hhi.fraunhofer.de, thomas.schierl@hhi.fraunhofer.de, stewe@stewe.org, miska.hannuksela@nokia.com +# RFC7799 || A. Morton || acmorton@att.com +# RFC7800 || M. Jones, J. Bradley, H. Tschofenig || mbj@microsoft.com, ve7jtb@ve7jtb.com, Hannes.Tschofenig@gmx.net +# RFC7801 || V. Dolmatov, Ed. || dol@srcc.msu.ru +# RFC7802 || S. Emery, N. Williams || shawn.emery@oracle.com, nico@cryptonector.com +# RFC7803 || B. Leiba || barryleiba@computer.org +# RFC7804 || A. Melnikov || alexey.melnikov@isode.com +# RFC7805 || A. Zimmermann, W. Eddy, L. Eggert || alexander@zimmermann.eu.com, wes@mti-systems.com, lars@netapp.com +# RFC7806 || F. Baker, R. Pan || fred@cisco.com, ropan@cisco.com +# RFC7807 || M. Nottingham, E. Wilde || mnot@mnot.net, erik.wilde@dret.net +# RFC7808 || M. Douglass, C. Daboo || mdouglass@sphericalcowgroup.com, cyrus@daboo.name +# RFC7809 || C. Daboo || cyrus@daboo.name +# RFC7810 || S. Previdi, Ed., S. Giacalone, D. Ward, J. Drake, Q. Wu || sprevidi@cisco.com, spencer.giacalone@gmail.com, wardd@cisco.com, jdrake@juniper.net, sunseawq@huawei.com +# RFC7811 || G. Enyedi, A. Csaszar, A. Atlas, C. Bowers, A. Gopalan || Gabor.Sandor.Enyedi@ericsson.com, Andras.Csaszar@ericsson.com, akatlas@juniper.net, cbowers@juniper.net, abishek@ece.arizona.edu +# RFC7812 || A. Atlas, C. Bowers, G. Enyedi || akatlas@juniper.net, cbowers@juniper.net, Gabor.Sandor.Enyedi@ericsson.com +# RFC7813 || J. Farkas, Ed., N. Bragg, P. Unbehagen, G. Parsons, P. Ashwood-Smith, C. Bowers || janos.farkas@ericsson.com, nbragg@ciena.com, unbehagen@avaya.com, glenn.parsons@ericsson.com, Peter.AshwoodSmith@huawei.com, cbowers@juniper.net +# RFC7814 || X. Xu, C. Jacquenet, R. Raszuk, T. Boyes, B. Fee || xuxiaohu@huawei.com, christian.jacquenet@orange.com, robert@raszuk.net, tboyes@bloomberg.net, bfee@extremenetworks.com +# RFC7815 || T. Kivinen || kivinen@iki.fi +# RFC7816 || S. Bortzmeyer || bortzmeyer+ietf@nic.fr +# RFC7817 || A. Melnikov || alexey.melnikov@isode.com +# RFC7818 || M. Jethanandani || mjethanandani@gmail.com +# RFC7819 || S. Jiang, S. Krishnan, T. Mrugalski || jiangsheng@huawei.com, suresh.krishnan@ericsson.com, tomasz.mrugalski@gmail.com +# RFC7820 || T. Mizrahi || talmi@marvell.com +# RFC7821 || T. Mizrahi || talmi@marvell.com +# RFC7822 || T. Mizrahi, D. Mayer || talmi@marvell.com, mayer@ntp.org +# RFC7823 || A. Atlas, J. Drake, S. Giacalone, S. Previdi || akatlas@juniper.net, jdrake@juniper.net, spencer.giacalone@gmail.com, sprevidi@cisco.com +# RFC7824 || S. Krishnan, T. Mrugalski, S. Jiang || suresh.krishnan@ericsson.com, tomasz.mrugalski@gmail.com, jiangsheng@huawei.com +# RFC7825 || J. Goldberg, M. Westerlund, T. Zeng || jgoldber@cisco.com, magnus.westerlund@ericsson.com, thomas.zeng@gmail.com +# RFC7826 || H. Schulzrinne, A. Rao, R. Lanphier, M. Westerlund, M. Stiemerling, Ed. || schulzrinne@cs.columbia.edu, anrao@cisco.com, robla@robla.net, magnus.westerlund@ericsson.com, mls.ietf@gmail.com +# RFC7827 || L. Eggert || lars@netapp.com +# RFC7828 || P. Wouters, J. Abley, S. Dickinson, R. Bellis || pwouters@redhat.com, jabley@dyn.com, sara@sinodun.com, ray@isc.org +# RFC7829 || Y. Nishida, P. Natarajan, A. Caro, P. Amer, K. Nielsen || nishida@wide.ad.jp, prenatar@cisco.com, acaro@bbn.com, amer@udel.edu, karen.nielsen@tieto.com +# RFC7830 || A. Mayrhofer || alex.mayrhofer.ietf@gmail.com +# RFC7831 || J. Howlett, S. Hartman, H. Tschofenig, J. Schaad || josh.howlett@ja.net, hartmans-ietf@mit.edu, Hannes.Tschofenig@gmx.net, ietf@augustcellars.com +# RFC7832 || R. Smith, Ed. || rhys.smith@jisc.ac.uk +# RFC7833 || J. Howlett, S. Hartman, A. Perez-Mendez, Ed. || josh.howlett@ja.net, hartmans-ietf@mit.edu, alex@um.es +# RFC7834 || D. Saucez, L. Iannone, A. Cabellos, F. Coras || damien.saucez@inria.fr, ggx@gigix.net, acabello@ac.upc.edu, fcoras@ac.upc.edu +# RFC7835 || D. Saucez, L. Iannone, O. Bonaventure || damien.saucez@inria.fr, ggx@gigix.net, Olivier.Bonaventure@uclouvain.be +# RFC7836 || S. Smyshlyaev, Ed., E. Alekseev, I. Oshkin, V. Popov, S. Leontiev, V. Podobaev, D. Belyavsky || svs@cryptopro.ru, alekseev@cryptopro.ru, oshkin@cryptopro.ru, vpopov@cryptopro.ru, lse@CryptoPro.ru, v_podobaev@factor-ts.ru, beldmit@gmail.com +# RFC7837 || S. Krishnan, M. Kuehlewind, B. Briscoe, C. Ralli || suresh.krishnan@ericsson.com, mirja.kuehlewind@tik.ee.ethz.ch, ietf@bobbriscoe.net, ralli@tid.es +# RFC7838 || M. Nottingham, P. McManus, J. Reschke || mnot@mnot.net, mcmanus@ducksong.com, julian.reschke@greenbytes.de +# RFC7839 || S. Bhandari, S. Gundavelli, M. Grayson, B. Volz, J. Korhonen || shwethab@cisco.com, sgundave@cisco.com, mgrayson@cisco.com, volz@cisco.com, jouni.nospam@gmail.com +# RFC7840 || J. Winterbottom, H. Tschofenig, L. Liess || a.james.winterbottom@gmail.com, Hannes.Tschofenig@gmx.net, L.Liess@telekom.de +# RFC7841 || J. Halpern, Ed., L. Daigle, Ed., O. Kolkman, Ed. || jmh@joelhalpern.com, ldaigle@thinkingcat.com, kolkman@isoc.org +# RFC7842 || R. Sparks || rjsparks@nostrum.com +# RFC7843 || A. Ripke, R. Winter, T. Dietz, J. Quittek, R. da Silva || ripke@neclab.eu, winter@neclab.eu, dietz@neclab.eu, quittek@neclab.eu, rafaelalejandro.lopezdasilva@telefonica.com +# RFC7844 || C. Huitema, T. Mrugalski, S. Krishnan || huitema@microsoft.com, tomasz.mrugalski@gmail.com, suresh.krishnan@ericsson.com +# RFC7845 || T. Terriberry, R. Lee, R. Giles || tterribe@xiph.org, ron@debian.org, giles@xiph.org +# RFC7846 || R. Cruz, M. Nunes, J. Xia, R. Huang, Ed., J. Taveira, D. Lingli || rui.cruz@ieee.org, mario.nunes@inov.pt, xiajinwei@huawei.com, rachel.huang@huawei.com, joao.silva@inov.pt, denglingli@chinamobile.com +# RFC7847 || T. Melia, Ed., S. Gundavelli, Ed. || telemaco.melia@gmail.com, sgundave@cisco.com +# RFC7848 || G. Lozano || gustavo.lozano@icann.org +# RFC7849 || D. Binet, M. Boucadair, A. Vizdal, G. Chen, N. Heatley, R. Chandler, D. Michaud, D. Lopez, W. Haeffner || david.binet@orange.com, mohamed.boucadair@orange.com, Ales.Vizdal@T-Mobile.cz, phdgang@gmail.com, nick.heatley@ee.co.uk, ross@eircom.net, dave.michaud@rci.rogers.com, diego.r.lopez@telefonica.com, walter.haeffner@vodafone.com +# RFC7850 || S. Nandakumar || snandaku@cisco.com +# RFC7851 || H. Song, X. Jiang, R. Even, D. Bryan, Y. Sun || haibin.song@huawei.com, jiangxingfeng@huawei.com, ron.even.tlv@gmail.com, dbryan@ethernot.org, sunyi@ict.ac.cn +# RFC7852 || R. Gellens, B. Rosen, H. Tschofenig, R. Marshall, J. Winterbottom || rg+ietf@randy.pensive.org, br@brianrosen.net, Hannes.Tschofenig@gmx.net, rmarshall@telecomsys.com, a.james.winterbottom@gmail.com +# RFC7853 || S. Martin, S. Tuecke, B. McCollam, M. Lidman || sjmartin@uchicago.edu, tuecke@globus.org, bmccollam@uchicago.edu, mattias@uchicago.edu< +# RFC7854 || J. Scudder, Ed., R. Fernando, S. Stuart || jgs@juniper.net, rex@cisco.com, sstuart@google.com +# RFC7855 || S. Previdi, Ed., C. Filsfils, Ed., B. Decraene, S. Litkowski, M. Horneffer, R. Shakir || sprevidi@cisco.com, cfilsfil@cisco.com, bruno.decraene@orange.com, stephane.litkowski@orange.com, Martin.Horneffer@telekom.de, rjs@rob.sh +# RFC7856 || Y. Cui, J. Dong, P. Wu, M. Xu, A. Yla-Jaaski || yong@csnet1.cs.tsinghua.edu.cn, knight.dongjiang@gmail.com, weapon9@gmail.com, xmw@cernet.edu.cn, antti.yla-jaaski@aalto.fi +# RFC7857 || R. Penno, S. Perreault, M. Boucadair, Ed., S. Sivakumar, K. Naito || repenno@cisco.com, sperreault@jive.com, mohamed.boucadair@orange.com, ssenthil@cisco.com, k.naito@nttv6.jp +# RFC7858 || Z. Hu, L. Zhu, J. Heidemann, A. Mankin, D. Wessels, P. Hoffman || zihu@outlook.com, liangzhu@usc.edu, johnh@isi.edu, allison.mankin@gmail.com, dwessels@verisign.com, paul.hoffman@icann.org +# RFC7859 || C. Dearlove || chris.dearlove@baesystems.com +# RFC7860 || J. Merkle, Ed., M. Lochter || johannes.merkle@secunet.com, manfred.lochter@bsi.bund.de +# RFC7861 || A. Adamson, N. Williams || andros@netapp.com, nico@cryptonector.com +# RFC7862 || T. Haynes || thomas.haynes@primarydata.com +# RFC7863 || T. Haynes || thomas.haynes@primarydata.com +# RFC7864 || CJ. Bernardos, Ed. || cjbc@it.uc3m.es +# RFC7865 || R. Ravindranath, P. Ravindran, P. Kyzivat || rmohanr@cisco.com, partha@parthasarathi.co.in, pkyzivat@alum.mit.edu +# RFC7866 || L. Portman, H. Lum, Ed., C. Eckel, A. Johnston, A. Hutton || leon.portman@gmail.com, henry.lum@genesyslab.com, eckelcu@cisco.com, alan.b.johnston@gmail.com, andrew.hutton@unify.com +# RFC7867 || R. Huang || rachel.huang@huawei.com +# RFC7868 || D. Savage, J. Ng, S. Moore, D. Slice, P. Paluch, R. White || dsavage@cisco.com, jamng@cisco.com, smoore@cisco.com, dslice@cumulusnetworks.com, peter.paluch@fri.uniza.sk, russ@riw.us +# RFC7869 || D. Warden, I. Iordanov || david_warden@dell.com, iiordanov@gmail.com +# RFC7870 || Y. Fu, S. Jiang, J. Dong, Y. Chen || fuyu@cnnic.cn, jiangsheng@huawei.com, knight.dongjiang@gmail.com, flashfoxmx@gmail.com +# RFC7871 || C. Contavalli, W. van der Gaast, D. Lawrence, W. Kumari || ccontavalli@google.com, wilmer@google.com, tale@akamai.com, warren@kumari.net +# RFC7872 || F. Gont, J. Linkova, T. Chown, W. Liu || fgont@si6networks.com, furry@google.com, tim.chown@jisc.ac.uk, liushucheng@huawei.com +# RFC7873 || D. Eastlake 3rd, M. Andrews || d3e3e3@gmail.com, marka@isc.org +# RFC7874 || JM. Valin, C. Bran || jmvalin@jmvalin.ca, cary.bran@plantronics.com +# RFC7875 || S. Proust, Ed. || stephane.proust@orange.com +# RFC7876 || S. Bryant, S. Sivabalan, S. Soni || stewart.bryant@gmail.com, msiva@cisco.com, sagsoni@cisco.com +# RFC7877 || K. Cartwright, V. Bhatia, S. Ali, D. Schwartz || kcartwright@tnsi.com, vbhatia@tnsi.com, syed.ali@neustar.biz, dschwartz@xconnect.net +# RFC7878 || K. Cartwright, V. Bhatia, J-F. Mule, A. Mayrhofer || kcartwright@tnsi.com, vbhatia@tnsi.com, jfmule@apple.com, alexander.mayrhofer@nic.at +# RFC7879 || R. Ravindranath, T. Reddy, G. Salgueiro, V. Pascual, P. Ravindran || rmohanr@cisco.com, tireddy@cisco.com, gsalguei@cisco.com, victor.pascual.avila@oracle.com, partha@parthasarathi.co.in +# RFC7880 || C. Pignataro, D. Ward, N. Akiya, M. Bhatia, S. Pallagatti || cpignata@cisco.com, wardd@cisco.com, nobo.akiya.dev@gmail.com, manav@ionosnetworks.com, santosh.pallagatti@gmail.com +# RFC7881 || C. Pignataro, D. Ward, N. Akiya || cpignata@cisco.com, wardd@cisco.com, nobo.akiya.dev@gmail.com +# RFC7882 || S. Aldrin, C. Pignataro, G. Mirsky, N. Kumar || aldrin.ietf@gmail.com, cpignata@cisco.com, gregory.mirsky@ericsson.com, naikumar@cisco.com +# RFC7883 || L. Ginsberg, N. Akiya, M. Chen || ginsberg@cisco.com, nobo.akiya.dev@gmail.com, mach.chen@huawei.com +# RFC7884 || C. Pignataro, M. Bhatia, S. Aldrin, T. Ranganath || cpignata@cisco.com, manav@ionosnetworks.com, aldrin.ietf@gmail.com, trilok.ranganatha@nokia.com +# RFC7885 || V. Govindan, C. Pignataro || venggovi@cisco.com, cpignata@cisco.com +# RFC7886 || V. Govindan, C. Pignataro || venggovi@cisco.com, cpignata@cisco.com +# RFC7887 || S. Venaas, J. Arango, I. Kouvelas || stig@cisco.com, jearango@cisco.com, kouvelas@arista.com +# RFC7888 || A. Melnikov, Ed. || alexey.melnikov@isode.com +# RFC7889 || J. SrimushnamBoovaraghamoorthy, N. Bisht || jayantheesh.sb@gmail.com, narendrasingh.bisht@gmail.com +# RFC7890 || D. Bryan, P. Matthews, E. Shim, D. Willis, S. Dawkins || dbryan@ethernot.org, philip_matthews@magma.ca, eunsooshim@gmail.com, dean.willis@softarmor.com, spencerdawkins.ietf@gmail.com +# RFC7891 || J. Asghar, IJ. Wijnands, Ed., S. Krishnaswamy, A. Karan, V. Arya || jasghar@cisco.com, ice@cisco.com, sowkrish@cisco.com, apoorva@cisco.com, varya@directv.com +# RFC7892 || Z. Ali, A. Bonfanti, M. Hartley, F. Zhang || zali@cisco.com, abonfant@cisco.com, mhartley@cisco.com, zhangfatai@huawei.com +# RFC7893 || Y(J) Stein, D. Black, B. Briscoe || yaakov_s@rad.com, david.black@emc.com, ietf@bobbriscoe.net +# RFC7894 || M. Pritikin, C. Wallace || pritikin@cisco.com, carl@redhoundsoftware.com +# RFC7895 || A. Bierman, M. Bjorklund, K. Watsen || andy@yumaworks.com, mbj@tail-f.com, kwatsen@juniper.net +# RFC7896 || D. Dhody || dhruv.ietf@gmail.com +# RFC7897 || D. Dhody, U. Palle, R. Casellas || dhruv.ietf@gmail.com, udayasree.palle@huawei.com, ramon.casellas@cttc.es +# RFC7898 || D. Dhody, U. Palle, V. Kondreddy, R. Casellas || dhruv.ietf@gmail.com, udayasree.palle@huawei.com, venugopalreddyk@huawei.com, ramon.casellas@cttc.es +# RFC7899 || T. Morin, Ed., S. Litkowski, K. Patel, Z. Zhang, R. Kebler, J. Haas || thomas.morin@orange.com, stephane.litkowski@orange.com, keyupate@cisco.com, zzhang@juniper.net, rkebler@juniper.net, jhaas@juniper.net +# RFC7900 || Y. Rekhter, Ed., E. Rosen, Ed., R. Aggarwal, Y. Cai, T. Morin || none, erosen@juniper.net, raggarwa_1@yahoo.com, yiqun.cai@alibaba-inc.com, thomas.morin@orange.com +# RFC7901 || P. Wouters || pwouters@redhat.com +# RFC7902 || E. Rosen, T. Morin || erosen@juniper.net, thomas.morin@orange.com +# RFC7903 || S. Leonard || dev+ietf@seantek.com +# RFC7904 || C. Jennings, B. Lowekamp, E. Rescorla, S. Baset, H. Schulzrinne, T. Schmidt, Ed. || fluffy@cisco.com, bbl@lowekamp.net, ekr@rtfm.com, sabaset@us.ibm.com, hgs@cs.columbia.edu, t.schmidt@haw-hamburg.de +# RFC7905 || A. Langley, W. Chang, N. Mavrogiannopoulos, J. Strombergson, S. Josefsson || agl@google.com, wtc@google.com, nmav@redhat.com, joachim@secworks.se, simon@josefsson.org +# RFC7906 || P. Timmel, R. Housley, S. Turner || pstimme@nsa.gov, housley@vigilsec.com, turners@ieca.com +# RFC7908 || K. Sriram, D. Montgomery, D. McPherson, E. Osterweil, B. Dickson || ksriram@nist.gov, dougm@nist.gov, dmcpherson@verisign.com, eosterweil@verisign.com, brian.peter.dickson@gmail.com +# RFC7909 || R. Kisteleki, B. Haberman || robert@ripe.net, brian@innovationslab.net +# RFC7910 || W. Zhou || zhouweiisu@gmail.com +# RFC7911 || D. Walton, A. Retana, E. Chen, J. Scudder || dwalton@cumulusnetworks.com, aretana@cisco.com, enkechen@cisco.com, jgs@juniper.net +# RFC7912 || A. Melnikov || alexey.melnikov@isode.com +# RFC7913 || C. Holmberg || christer.holmberg@ericsson.com +# RFC7914 || C. Percival, S. Josefsson || cperciva@tarsnap.com, simon@josefsson.org +# RFC7915 || C. Bao, X. Li, F. Baker, T. Anderson, F. Gont || congxiao@cernet.edu.cn, xing@cernet.edu.cn, fred@cisco.com, tore@redpill-linpro.com, fgont@si6networks.com +# RFC7916 || S. Litkowski, Ed., B. Decraene, C. Filsfils, K. Raza, M. Horneffer, P. Sarkar || stephane.litkowski@orange.com, bruno.decraene@orange.com, cfilsfil@cisco.com, skraza@cisco.com, Martin.Horneffer@telekom.de, pushpasis.ietf@gmail.com +# RFC7917 || P. Sarkar, Ed., H. Gredler, S. Hegde, S. Litkowski, B. Decraene || pushpasis.ietf@gmail.com, hannes@rtbrick.com, shraddha@juniper.net, stephane.litkowski@orange.com, bruno.decraene@orange.com +# RFC7918 || A. Langley, N. Modadugu, B. Moeller || agl@google.com, nagendra@cs.stanford.edu, bmoeller@acm.org +# RFC7919 || D. Gillmor || dkg@fifthhorseman.net +# RFC7920 || A. Atlas, Ed., T. Nadeau, Ed., D. Ward || akatlas@juniper.net, tnadeau@lucidvision.com, wardd@cisco.com +# RFC7921 || A. Atlas, J. Halpern, S. Hares, D. Ward, T. Nadeau || akatlas@juniper.net, joel.halpern@ericsson.com, shares@ndzh.com, wardd@cisco.com, tnadeau@lucidvision.com +# RFC7922 || J. Clarke, G. Salgueiro, C. Pignataro || jclarke@cisco.com, gsalguei@cisco.com, cpignata@cisco.com +# RFC7923 || E. Voit, A. Clemm, A. Gonzalez Prieto || evoit@cisco.com, alex@cisco.com, albertgo@cisco.com +# RFC7924 || S. Santesson, H. Tschofenig || sts@aaa-sec.com, Hannes.Tschofenig@gmx.net +# RFC7925 || H. Tschofenig, Ed., T. Fossati || Hannes.Tschofenig@gmx.net, thomas.fossati@nokia.com +# RFC7926 || A. Farrel, Ed., J. Drake, N. Bitar, G. Swallow, D. Ceccarelli, X. Zhang || adrian@olddog.co.uk, jdrake@juniper.net, nbitar40@gmail.com, swallow@cisco.com, daniele.ceccarelli@ericsson.com, zhang.xian@huawei.com +# RFC7927 || D. Kutscher, Ed., S. Eum, K. Pentikousis, I. Psaras, D. Corujo, D. Saucez, T. Schmidt, M. Waehlisch || kutscher@neclab.eu, suyong@ist.osaka-u.ac.jp, k.pentikousis@travelping.com, i.psaras@ucl.ac.uk, dcorujo@av.it.pt, damien.saucez@inria.fr, t.schmidt@haw-hamburg.de, waehlisch@ieee.org +# RFC7928 || N. Kuhn, Ed., P. Natarajan, Ed., N. Khademi, Ed., D. Ros || nicolas.kuhn@cnes.fr, prenatar@cisco.com, naeemk@ifi.uio.no, dros@simula.no +# RFC7929 || P. Wouters || pwouters@redhat.com +# RFC7930 || S. Hartman || hartmans-ietf@mit.edu +# RFC7931 || D. Noveck, Ed., P. Shivam, C. Lever, B. Baker || davenoveck@gmail.com, piyush.shivam@oracle.com, chuck.lever@oracle.com, bill.baker@oracle.com +# RFC7932 || J. Alakuijala, Z. Szabadka || jyrki@google.com, szabadka@google.com +# RFC7933 || C. Westphal, Ed., S. Lederer, D. Posch, C. Timmerer, A. Azgin, W. Liu, C. Mueller, A. Detti, D. Corujo, J. Wang, M. Montpetit, N. Murray || Cedric.Westphal@huawei.com, stefan.lederer@itec.aau.at, daniel.posch@itec.aau.at, christian.timmerer@itec.aau.at, aytac.azgin@huawei.com, liushucheng@huawei.com, christopher.mueller@bitmovin.net, andrea.detti@uniroma2.it, dcorujo@av.it.pt, jianwang@cityu.edu.hk, marie@mjmontpetit.com, nmurray@research.ait.ie +# RFC7934 || L. Colitti, V. Cerf, S. Cheshire, D. Schinazi || lorenzo@google.com, vint@google.com, cheshire@apple.com, dschinazi@apple.com +# RFC7935 || G. Huston, G. Michaelson, Ed. || gih@apnic.net, ggm@apnic.net +# RFC7936 || T. Hardie || ted.ietf@gmail.com +# RFC7937 || F. Le Faucheur, Ed., G. Bertrand, Ed., I. Oprescu, Ed., R. Peterkofsky || flefauch@gmail.com, gilbertrand@gmail.com, iuniana.oprescu@gmail.com, peterkofsky@google.com +# RFC7938 || P. Lapukhov, A. Premji, J. Mitchell, Ed. || petr@fb.com, ariff@arista.com, jrmitche@puck.nether.net +# RFC7939 || U. Herberg, R. Cole, I. Chakeres, T. Clausen || ulrich@herberg.name, rgcole01@comcast.net, ian.chakeres@gmail.com, T.Clausen@computer.org +# RFC7940 || K. Davies, A. Freytag || kim.davies@icann.org, asmus@unicode.org +# RFC7941 || M. Westerlund, B. Burman, R. Even, M. Zanaty || magnus.westerlund@ericsson.com, bo.burman@ericsson.com, roni.even@mail01.huawei.com, mzanaty@cisco.com +# RFC7942 || Y. Sheffer, A. Farrel || yaronf.ietf@gmail.com, adrian@olddog.co.uk +# RFC7943 || F. Gont, W. Liu || fgont@si6networks.com, liushucheng@huawei.com +# RFC7944 || S. Donovan || srdonovan@usdonovans.com +# RFC7945 || K. Pentikousis, Ed., B. Ohlman, E. Davies, S. Spirou, G. Boggia || k.pentikousis@travelping.com, Borje.Ohlman@ericsson.com, davieseb@scss.tcd.ie, spis@intracom-telecom.com, g.boggia@poliba.it +# RFC7946 || H. Butler, M. Daly, A. Doyle, S. Gillies, S. Hagen, T. Schaub || howard@hobu.co, martin.daly@cadcorp.com, adoyle@intl-interfaces.com, sean.gillies@gmail.com, stefan@hagen.link, tim.schaub@gmail.com +# RFC7947 || E. Jasinska, N. Hilliard, R. Raszuk, N. Bakker || elisa@bigwaveit.org, nick@inex.ie, robert@raszuk.net, nbakker@akamai.com +# RFC7948 || N. Hilliard, E. Jasinska, R. Raszuk, N. Bakker || nick@inex.ie, elisa@bigwaveit.org, robert@raszuk.net, nbakker@akamai.com +# RFC7949 || I. Chen, A. Lindem, R. Atkinson || ichen@kuatrotech.com, acee@cisco.com, rja.lists@gmail.com +# RFC7950 || M. Bjorklund, Ed. || mbj@tail-f.com +# RFC7951 || L. Lhotka || lhotka@nic.cz +# RFC7952 || L. Lhotka || lhotka@nic.cz +# RFC7953 || C. Daboo, M. Douglass || cyrus@daboo.name, mdouglass@sphericalcowgroup.com +# RFC7954 || L. Iannone, D. Lewis, D. Meyer, V. Fuller || ggx@gigix.net, darlewis@cisco.com, dmm@1-4-5.net, vaf@vaf.net +# RFC7955 || L. Iannone, R. Jorgensen, D. Conrad, G. Huston || ggx@gigix.net, rogerj@gmail.com, drc@virtualized.org, gih@apnic.net +# RFC7956 || W. Hao, Y. Li, A. Qu, M. Durrani, P. Sivamurugan || haoweiguo@huawei.com, liyizhou@huawei.com, laodulaodu@gmail.com, mdurrani@equinix.com, ponkarthick.sivamurugan@ipinfusion.com +# RFC7957 || B. Campbell, Ed., A. Cooper, B. Leiba || ben@nostrum.com, alcoop@cisco.com, barryleiba@computer.org +# RFC7958 || J. Abley, J. Schlyter, G. Bailey, P. Hoffman || jabley@dyn.com, jakob@kirei.se, guillaumebailey@outlook.com, paul.hoffman@icann.org +# RFC7959 || C. Bormann, Z. Shelby, Ed. || cabo@tzi.org, zach.shelby@arm.com +# RFC7960 || F. Martin, Ed., E. Lear, Ed., T. Draegen. Ed., E. Zwicky, Ed., K. Andersen, Ed. || fmartin@linkedin.com, lear@cisco.com, tim@dmarcian.com, zwicky@yahoo-inc.com, kandersen@linkedin.com +# RFC7961 || D. Eastlake 3rd, L. Yizhou || d3e3e3@gmail.com, liyizhou@huawei.com +# RFC7962 || J. Saldana, Ed., A. Arcia-Moret, B. Braem, E. Pietrosemoli, A. Sathiaseelan, M. Zennaro || jsaldana@unizar.es, andres.arcia@cl.cam.ac.uk, bart.braem@iminds.be, ermanno@ictp.it, arjuna.sathiaseelan@cl.cam.ac.uk, mzennaro@ictp.it +# RFC7963 || Z. Ali, A. Bonfanti, M. Hartley, F. Zhang || zali@cisco.com, abonfant@cisco.com, mhartley@cisco.com, zhangfatai@huawei.com +# RFC7964 || D. Walton, A. Retana, E. Chen, J. Scudder || dwalton@cumulusnetworks.com, aretana@cisco.com, enkechen@cisco.com, jgs@juniper.net +# RFC7965 || M. Chen, W. Cao, A. Takacs, P. Pan || mach.chen@huawei.com, wayne.caowei@huawei.com, attila.takacs@ericsson.com, none +# RFC7966 || H. Tschofenig, J. Korhonen, Ed., G. Zorn, K. Pillay || Hannes.tschofenig@gmx.net, jouni.nospam@gmail.com, glenzorn@gmail.com, kervin.pillay@gmail.com +# RFC7967 || A. Bhattacharyya, S. Bandyopadhyay, A. Pal, T. Bose || abhijan.bhattacharyya@tcs.com, soma.bandyopadhyay@tcs.com, arpan.pal@tcs.com, tulika.bose@tcs.com +# RFC7968 || Y. Li, D. Eastlake 3rd, W. Hao, H. Chen, S. Chatterjee || liyizhou@huawei.com, d3e3e3@gmail.com, haoweiguo@huawei.com, philips.chenhao@huawei.com, somnath.chatterjee01@gmail.com +# RFC7969 || T. Lemon, T. Mrugalski || ted.lemon@nominum.com, tomasz.mrugalski@gmail.com +# RFC7970 || R. Danyliw || rdd@cert.org +# RFC7971 || M. Stiemerling, S. Kiesel, M. Scharf, H. Seidel, S. Previdi || mls.ietf@gmail.com, ietf-alto@skiesel.de, michael.scharf@nokia.com, hseidel@benocs.com, sprevidi@cisco.com +# RFC7972 || P. Lemieux || pal@sandflow.com +# RFC7973 || R. Droms, P. Duffy || rdroms.ietf@gmail.com, paduffy@cisco.com +# RFC7974 || B. Williams, M. Boucadair, D. Wing || brandon.williams@akamai.com, mohamed.boucadair@orange.com, dwing-ietf@fuggles.com +# RFC7975 || B. Niven-Jenkins, Ed., R. van Brandenburg, Ed. || ben.niven-jenkins@nokia.com, ray.vanbrandenburg@tno.nl +# RFC7976 || C. Holmberg, N. Biondic, G. Salgueiro || christer.holmberg@ericsson.com, nevenka.biondic@ericsson.com, gsalguei@cisco.com +# RFC7977 || P. Dunkley, G. Llewellyn, V. Pascual, G. Salgueiro, R. Ravindranath || peter.dunkley@xura.com, gavin.llewellyn@xura.com, victor.pascual.avila@oracle.com, gsalguei@cisco.com, rmohanr@cisco.com +# RFC7978 || D. Eastlake 3rd, M. Umair, Y. Li || d3e3e3@gmail.com, mohammed.umair2@gmail.com, liyizhou@huawei.com +# RFC7979 || E. Lear, Ed., R. Housley, Ed. || lear@cisco.com, housley@vigilsec.com +# RFC7980 || M. Behringer, A. Retana, R. White, G. Huston || mbehring@cisco.com, aretana@cisco.com, russw@riw.us, gih@apnic.net +# RFC7981 || L. Ginsberg, S. Previdi, M. Chen || ginsberg@cisco.com, sprevidi@cisco.com, mach.chen@huawei.com +# RFC7982 || P. Martinsen, T. Reddy, D. Wing, V. Singh || palmarti@cisco.com, tireddy@cisco.com, dwing-ietf@fuggles.com, varun@callstats.io +# RFC7983 || M. Petit-Huguenin, G. Salgueiro || marc@petit-huguenin.org, gsalguei@cisco.com +# RFC7984 || O. Johansson, G. Salgueiro, V. Gurbani, D. Worley, Ed. || oej@edvina.net, gsalguei@cisco.com, vkg@bell-labs.com, worley@ariadne.com +# RFC7985 || J. Yi, T. Clausen, U. Herberg || jiazi@jiaziyi.com, T.Clausen@computer.org, ulrich@herberg.name +# RFC7986 || C. Daboo || cyrus@daboo.name +# RFC7987 || L. Ginsberg, P. Wells, B. Decraene, T. Przygienda, H. Gredler || ginsberg@cisco.com, pauwells@cisco.com, bruno.decraene@orange.com, prz@juniper.net, hannes@rtbrick.com +# RFC7988 || E. Rosen, Ed., K. Subramanian, Z. Zhang || erosen@juniper.net, karthik@sproute.com, zzhang@juniper.net +# RFC7989 || P. Jones, G. Salgueiro, C. Pearce, P. Giralt || paulej@packetizer.com, gsalguei@cisco.com, chrep@cisco.com, pgiralt@cisco.com +# RFC7990 || H. Flanagan || rse@rfc-editor.org +# RFC7991 || P. Hoffman || paul.hoffman@icann.org +# RFC7992 || J. Hildebrand, Ed., P. Hoffman || joe-ietf@cursive.net, paul.hoffman@icann.org +# RFC7993 || H. Flanagan || rse@rfc-editor.org +# RFC7994 || H. Flanagan || rse@rfc-editor.org +# RFC7995 || T. Hansen, Ed., L. Masinter, M. Hardy || tony@att.com, masinter@adobe.com, mahardy@adobe.com +# RFC7996 || N. Brownlee || n.brownlee@auckland.ac.nz +# RFC7997 || H. Flanagan, Ed. || rse@rfc-editor.org +# RFC7998 || P. Hoffman, J. Hildebrand || paul.hoffman@icann.org, joe-ietf@cursive.net +# RFC7999 || T. King, C. Dietzel, J. Snijders, G. Doering, G. Hankins || thomas.king@de-cix.net, christoph.dietzel@de-cix.net, job@ntt.net, gert@space.net, greg.hankins@nokia.com +# RFC8000 || A. Adamson, N. Williams || andros@netapp.com, nico@cryptonector.com +# RFC8001 || F. Zhang, Ed., O. Gonzalez de Dios, Ed., C. Margaria, M. Hartley, Z. Ali || zhangfatai@huawei.com, oscar.gonzalezdedios@telefonica.com, cmargaria@juniper.net, mhartley@cisco.com, zali@cisco.com +# RFC8002 || T. Heer, S. Varjonen || heer@hs-albsig.de, samu.varjonen@helsinki.fi +# RFC8003 || J. Laganier, L. Eggert || julien.ietf@gmail.com, lars@netapp.com +# RFC8004 || J. Laganier, L. Eggert || julien.ietf@gmail.com, lars@netapp.com +# RFC8005 || J. Laganier || julien.ietf@gmail.com +# RFC8006 || B. Niven-Jenkins, R. Murray, M. Caulfield, K. Ma || ben.niven-jenkins@nokia.com, rob.murray@nokia.com, mcaulfie@cisco.com, kevin.j.ma@ericsson.com +# RFC8007 || R. Murray, B. Niven-Jenkins || rob.murray@nokia.com, ben.niven-jenkins@nokia.com +# RFC8008 || J. Seedorf, J. Peterson, S. Previdi, R. van Brandenburg, K. Ma || jan.seedorf@hft-stuttgart.de, jon.peterson@neustar.biz, sprevidi@cisco.com, ray.vanbrandenburg@tno.nl, kevin.j.ma@ericsson.com +# RFC8009 || M. Jenkins, M. Peck, K. Burgin || mjjenki@tycho.ncsc.mil, mpeck@mitre.org, kelley.burgin@gmail.com +# RFC8010 || M. Sweet, I. McDonald || msweet@apple.com, blueroofmusic@gmail.com +# RFC8011 || M. Sweet, I. McDonald || msweet@apple.com, blueroofmusic@gmail.com +# RFC8012 || N. Akiya, G. Swallow, C. Pignataro, A. Malis, S. Aldrin || nobo.akiya.dev@gmail.com, swallow@cisco.com, cpignata@cisco.com, agmalis@gmail.com, aldrin.ietf@gmail.com +# RFC8013 || D. Joachimpillai, J. Hadi Salim || damascene.joachimpillai@verizon.com, hadi@mojatatu.com +# RFC8014 || D. Black, J. Hudson, L. Kreeger, M. Lasserre, T. Narten || david.black@dell.com, jon.hudson@gmail.com, lkreeger@gmail.com, mmlasserre@gmail.com, narten@us.ibm.com +# RFC8015 || V. Singh, C. Perkins, A. Clark, R. Huang || varun@callstats.io, csp@csperkins.org, alan.d.clark@telchemy.com, Rachel@huawei.com +# RFC8016 || T. Reddy, D. Wing, P. Patil, P. Martinsen || tireddy@cisco.com, dwing-ietf@fuggles.com, praspati@cisco.com, palmarti@cisco.com +# RFC8017 || K. Moriarty, Ed., B. Kaliski, J. Jonsson, A. Rusch || Kathleen.Moriarty@emc.com, bkaliski@verisign.com, jakob.jonsson@subset.se, andreas.rusch@rsa.com +# RFC8018 || K. Moriarty, Ed., B. Kaliski, A. Rusch || Kathleen.Moriarty@Dell.com, bkaliski@verisign.com, andreas.rusch@rsa.com +# RFC8019 || Y. Nir, V. Smyslov || ynir.ietf@gmail.com, svan@elvis.ru +# RFC8020 || S. Bortzmeyer, S. Huque || bortzmeyer+ietf@nic.fr, shuque@verisign.com +# RFC8021 || F. Gont, W. Liu, T. Anderson || fgont@si6networks.com, liushucheng@huawei.com, tore@redpill-linpro.com +# RFC8022 || L. Lhotka, A. Lindem || lhotka@nic.cz, acee@cisco.com +# RFC8023 || M. Thomas, A. Mankin, L. Zhang || mthomas@verisign.com, allison.mankin@gmail.com, lixia@cs.ucla.edu +# RFC8024 || Y. Jiang, Ed., Y. Luo, E. Mallette, Ed., Y. Shen, W. Cheng || jiangyuanlong@huawei.com, dennis.luoyong@huawei.com, edwin.mallette@gmail.com, yshen@juniper.net, chengweiqiang@chinamobile.com +# RFC8025 || P. Thubert, Ed., R. Cragie || pthubert@cisco.com, robert.cragie@gridmerge.com +# RFC8026 || M. Boucadair, I. Farrer || mohamed.boucadair@orange.com, ian.farrer@telekom.de +# RFC8027 || W. Hardaker, O. Gudmundsson, S. Krishnaswamy || ietf@hardakers.net, olafur+ietf@cloudflare.com, suresh@tislabs.com +# RFC8028 || F. Baker, B. Carpenter || fredbaker.ietf@gmail.com, brian.e.carpenter@gmail.com +# RFC8029 || K. Kompella, G. Swallow, C. Pignataro, Ed., N. Kumar, S. Aldrin, M. Chen || kireeti.kompella@gmail.com, swallow.ietf@gmail.com, cpignata@cisco.com, naikumar@cisco.com, aldrin.ietf@gmail.com, mach.chen@huawei.com +# RFC8030 || M. Thomson, E. Damaggio, B. Raymor, Ed. || martin.thomson@gmail.com, elioda@microsoft.com, brian.raymor@microsoft.com +# RFC8031 || Y. Nir, S. Josefsson || ynir.ietf@gmail.com, simon@josefsson.org +# RFC8032 || S. Josefsson, I. Liusvaara || simon@josefsson.org, ilariliusvaara@welho.com +# RFC8033 || R. Pan, P. Natarajan, F. Baker, G. White || ropan@cisco.com, prenatar@cisco.com, fredbaker.ietf@gmail.com, g.white@cablelabs.com +# RFC8034 || G. White, R. Pan || g.white@cablelabs.com, ropan@cisco.com +# RFC8035 || C. Holmberg || christer.holmberg@ericsson.com +# RFC8036 || N. Cam-Winget, Ed., J. Hui, D. Popa || ncamwing@cisco.com, jonhui@nestlabs.com, daniel.popa@itron.com +# RFC8037 || I. Liusvaara || ilariliusvaara@welho.com +# RFC8039 || A. Shpiner, R. Tse, C. Schelp, T. Mizrahi || alexshp@mellanox.com, Richard.Tse@microsemi.com, craig.schelp@oracle.com, talmi@marvell.com +# RFC8040 || A. Bierman, M. Bjorklund, K. Watsen || andy@yumaworks.com, mbj@tail-f.com, kwatsen@juniper.net +# RFC8041 || O. Bonaventure, C. Paasch, G. Detal || Olivier.Bonaventure@uclouvain.be, cpaasch@apple.com, gregory.detal@tessares.net +# RFC8042 || Z. Zhang, L. Wang, A. Lindem || zzhang@juniper.net, liliw@juniper.net, acee@cisco.com +# RFC8043 || B. Sarikaya, M. Boucadair || sarikaya@ieee.org, mohamed.boucadair@orange.com +# RFC8044 || A. DeKok || aland@freeradius.org +# RFC8045 || D. Cheng, J. Korhonen, M. Boucadair, S. Sivakumar || dean.cheng@huawei.com, jouni.nospam@gmail.com, mohamed.boucadair@orange.com, ssenthil@cisco.com +# RFC8046 || T. Henderson, Ed., C. Vogt, J. Arkko || tomhend@u.washington.edu, mail@christianvogt.net, jari.arkko@piuha.net +# RFC8047 || T. Henderson, Ed., C. Vogt, J. Arkko || tomhend@u.washington.edu, mail@christianvogt.net, jari.arkko@piuha.net +# RFC8048 || P. Saint-Andre || peter@filament.com +# RFC8049 || S. Litkowski, L. Tomotaki, K. Ogaki || stephane.litkowski@orange.com, luis.tomotaki@verizon.com, ke-oogaki@kddi.com +# RFC8051 || X. Zhang, Ed., I. Minei, Ed. || zhang.xian@huawei.com, inaminei@google.com +# RFC8053 || Y. Oiwa, H. Watanabe, H. Takagi, K. Maeda, T. Hayashi, Y. Ioku || y.oiwa@aist.go.jp, h-watanabe@aist.go.jp, takagi.hiromitsu@aist.go.jp, maeda@lepidum.co.jp, hayashi@lepidum.co.jp, mutual-work@ioku.org +# RFC8054 || K. Murchison, J. Elie || murch@andrew.cmu.edu, julien@trigofacile.com +# RFC8055 || C. Holmberg, Y. Jiang || christer.holmberg@ericsson.com, jiangyi@chinamobile.com +# RFC8056 || J. Gould || jgould@verisign.com +# RFC8057 || B. Stark, D. Sinicrope, W. Lupton || barbara.stark@att.com, david.sinicrope@ericsson.com, wlupton@broadband-forum.org +# RFC8058 || J. Levine, T. Herkula || standards@taugh.com, t.herkula@optivo.com +# RFC8059 || J. Arango, S. Venaas, I. Kouvelas, D. Farinacci || jearango@cisco.com, stig@cisco.com, kouvelas@arista.com, farinacci@gmail.com +# RFC8060 || D. Farinacci, D. Meyer, J. Snijders || farinacci@gmail.com, dmm@1-4-5.net, job@ntt.net +# RFC8061 || D. Farinacci, B. Weis || farinacci@gmail.com, bew@cisco.com +# RFC8062 || L. Zhu, P. Leach, S. Hartman, S. Emery, Ed. || larry.zhu@microsoft.com, pauljleach@msn.com, hartmans-ietf@mit.edu, shawn.emery@gmail.com +# RFC8063 || H.W. Ribbers, M.W. Groeneweg, R. Gieben, A.L.J. Verschuren || rik.ribbers@sidn.nl, marc.groeneweg@sidn.nl, miek@miek.nl, ietf@antoin.nl +# RFC8064 || F. Gont, A. Cooper, D. Thaler, W. Liu || fgont@si6networks.com, alcoop@cisco.com, dthaler@microsoft.com, liushucheng@huawei.com +# RFC8065 || D. Thaler || dthaler@microsoft.com +# RFC8066 || S. Chakrabarti, G. Montenegro, R. Droms, J. Woodyatt || samitac.ietf@gmail.com, Gabriel.Montenegro@microsoft.com, rdroms.ietf@gmail.com, jhw@google.com +# RFC8067 || B. Leiba || barryleiba@computer.org +# RFC8068 || R. Ravindranath, P. Ravindran, P. Kyzivat || rmohanr@cisco.com, partha@parthasarathi.co.in, pkyzivat@alum.mit.edu +# RFC8069 || A. Thomas || a.n.thomas@ieee.org +# RFC8070 || M. Short, Ed., S. Moore, P. Miller || michikos@microsoft.com, sethmo@microsoft.com, paumil@microsoft.com +# RFC8071 || K. Watsen || kwatsen@juniper.net +# RFC8072 || A. Bierman, M. Bjorklund, K. Watsen || andy@yumaworks.com, mbj@tail-f.com, kwatsen@juniper.net +# RFC8073 || K. Moriarty, M. Ford || Kathleen.Moriarty@dell.com, ford@isoc.org +# RFC8074 || J. Bi, G. Yao, J. Halpern, E. Levy-Abegnoli, Ed. || junbi@tsinghua.edu.cn, yaoguang.china@gmail.com, joel.halpern@ericsson.com, elevyabe@cisco.com +# RFC8075 || A. Castellani, S. Loreto, A. Rahman, T. Fossati, E. Dijk || angelo@castellani.net, Salvatore.Loreto@ericsson.com, Akbar.Rahman@InterDigital.com, thomas.fossati@nokia.com, esko.dijk@philips.com +# RFC8076 || A. Knauf, T. Schmidt, Ed., G. Hege, M. Waehlisch || alexanderknauf@gmail.com, t.schmidt@haw-hamburg.de, hege@daviko.com, mw@link-lab.net +# RFC8077 || L. Martini, Ed., G. Heron, Ed. || lmartini@monoski.com, giheron@cisco.com +# RFC8078 || O. Gudmundsson, P. Wouters || olafur+ietf@cloudflare.com, pwouters@redhat.com +# RFC8079 || L. Miniero, S. Garcia Murillo, V. Pascual || lorenzo@meetecho.com, sergio.garcia.murillo@gmail.com, victor.pascual.avila@oracle.com +# RFC8080 || O. Sury, R. Edmonds || ondrej.sury@nic.cz, edmonds@mycre.ws +# RFC8081 || C. Lilley || chris@w3.org +# RFC8082 || S. Wenger, J. Lennox, B. Burman, M. Westerlund || stewe@stewe.org, jonathan@vidyo.com, bo.burman@ericsson.com, magnus.westerlund@ericsson.com +# RFC8083 || C. Perkins, V. Singh || csp@csperkins.org, varun@callstats.io +# RFC8084 || G. Fairhurst || gorry@erg.abdn.ac.uk +# RFC8085 || L. Eggert, G. Fairhurst, G. Shepherd || lars@netapp.com, gorry@erg.abdn.ac.uk, gjshep@gmail.com +# RFC8086 || L. Yong, Ed., E. Crabbe, X. Xu, T. Herbert || lucy.yong@huawei.com, edward.crabbe@gmail.com, xuxiaohu@huawei.com, tom@herbertland.com +# RFC8087 || G. Fairhurst, M. Welzl || gorry@erg.abdn.ac.uk, michawe@ifi.uio.no +# RFC8089 || M. Kerwin || matthew.kerwin@qut.edu.au +# RFC8090 || R. Housley || housley@vigilsec.com +# RFC8091 || E. Wilde || erik.wilde@dret.net +# RFC8092 || J. Heitz, Ed., J. Snijders, Ed., K. Patel, I. Bagdonas, N. Hilliard || jheitz@cisco.com, job@ntt.net, keyur@arrcus.com, ibagdona.ietf@gmail.com, nick@inex.ie +# RFC8093 || J. Snijders || job@ntt.net +# RFC8094 || T. Reddy, D. Wing, P. Patil || tireddy@cisco.com, dwing-ietf@fuggles.com, praspati@cisco.com +# RFC8095 || G. Fairhurst, Ed., B. Trammell, Ed., M. Kuehlewind, Ed. || gorry@erg.abdn.ac.uk, ietf@trammell.ch, mirja.kuehlewind@tik.ee.ethz.ch +# RFC8096 || B. Fenner || fenner@fenron.com +# RFC8097 || P. Mohapatra, K. Patel, J. Scudder, D. Ward, R. Bush || mpradosh@yahoo.com, keyur@arrcus.com, jgs@juniper.net, dward@cisco.com, randy@psg.com +# RFC8098 || T. Hansen, Ed., A. Melnikov, Ed. || tony@att.com, alexey.melnikov@isode.com +# RFC8099 || H. Chen, R. Li, A. Retana, Y. Yang, Z. Liu || huaimo.chen@huawei.com, renwei.li@huawei.com, aretana@cisco.com, yyang1998@gmail.com, liu.cmri@gmail.com +# RFC8100 || R. Geib, Ed., D. Black || Ruediger.Geib@telekom.de, david.black@dell.com +# RFC8101 || C. Holmberg, J. Axell || christer.holmberg@ericsson.com, jorgen.axell@ericsson.com +# RFC8102 || P. Sarkar, Ed., S. Hegde, C. Bowers, H. Gredler, S. Litkowski || pushpasis.ietf@gmail.com, shraddha@juniper.net, cbowers@juniper.net, hannes@rtbrick.com, stephane.litkowski@orange.com +# RFC8103 || R. Housley || housley@vigilsec.com +# RFC8104 || Y. Shen, R. Aggarwal, W. Henderickx, Y. Jiang || yshen@juniper.net, raggarwa_1@yahoo.com, wim.henderickx@nokia.com, jiangyuanlong@huawei.com +# RFC8106 || J. Jeong, S. Park, L. Beloeil, S. Madanapalli || pauljeong@skku.edu, soohong.park@samsung.com, luc.beloeil@orange.com, smadanapalli@gmail.com +# RFC8107 || J. Wold || jwold@ad-id.org +# RFC8108 || J. Lennox, M. Westerlund, Q. Wu, C. Perkins || jonathan@vidyo.com, magnus.westerlund@ericsson.com, bill.wu@huawei.com, csp@csperkins.org +# RFC8109 || P. Koch, M. Larson, P. Hoffman || pk@DENIC.DE, matt.larson@icann.org, paul.hoffman@icann.org +# RFC8110 || D. Harkins, Ed., W. Kumari, Ed. || dharkins@arubanetworks.com, warren@kumari.net +# RFC8113 || M. Boucadair, C. Jacquenet || mohamed.boucadair@orange.com, christian.jacquenet@orange.com +# RFC8114 || M. Boucadair, C. Qin, C. Jacquenet, Y. Lee, Q. Wang || mohamed.boucadair@orange.com, jacni@jacni.com, christian.jacquenet@orange.com, yiu_lee@cable.comcast.com, 13301168516@189.cn +# RFC8115 || M. Boucadair, J. Qin, T. Tsou, X. Deng || mohamed.boucadair@orange.com, jacni@jacni.com, tina.tsou@philips.com, dxhbupt@gmail.com +# RFC8117 || C. Huitema, D. Thaler, R. Winter || huitema@huitema.net, dthaler@microsoft.com, rolf.winter@hs-augsburg.de +# RFC8118 || M. Hardy, L. Masinter, D. Markovic, D. Johnson, M. Bailey || mahardy@adobe.com, masinter@adobe.com, dmarkovi@adobe.com, duff.johnson@pdfa.org, martin.bailey@globalgraphics.com +# RFC8119 || M. Mohali, M. Barnes || marianne.mohali@orange.com, mary.ietf.barnes@gmail.com +# RFC8120 || Y. Oiwa, H. Watanabe, H. Takagi, K. Maeda, T. Hayashi, Y. Ioku || y.oiwa@aist.go.jp, h-watanabe@aist.go.jp, takagi.hiromitsu@aist.go.jp, kaorumaeda.ml@gmail.com, hayashi@lepidum.co.jp, mutual-work@ioku.org +# RFC8121 || Y. Oiwa, H. Watanabe, H. Takagi, K. Maeda, T. Hayashi, Y. Ioku || y.oiwa@aist.go.jp, h-watanabe@aist.go.jp, takagi.hiromitsu@aist.go.jp, kaorumaeda.ml@gmail.com, hayashi@lepidum.co.jp, mutual-work@ioku.org +# RFC8122 || J. Lennox, C. Holmberg || jonathan@vidyo.com, christer.holmberg@ericsson.com +# RFC8123 || P. Dawes, C. Arunachalam || peter.dawes@vodafone.com, carunach@cisco.com +# RFC8124 || R. Ravindranath, G. Salgueiro || rmohanr@cisco.com, gsalguei@cisco.com +# RFC8128 || C. Morgan || cmorgan@amsl.com +# RFC8129 || A. Jain, N. Kinder, N. McCallum || ajain323@gatech.edu, nkinder@redhat.com, npmccallum@redhat.com +# RFC8130 || V. Demjanenko, D. Satterlee || victor.demjanenko@vocal.com, david.satterlee@vocal.com +# RFC8131 || X. Zhang, H. Zheng, Ed., R. Gandhi, Ed., Z. Ali, P. Brzozowski || zhang.xian@huawei.com, zhenghaomian@huawei.com, rgandhi@cisco.com, zali@cisco.com, pbrzozowski@advaoptical.com +# RFC8132 || P. van der Stok, C. Bormann, A. Sehgal || consultancy@vanderstok.org, cabo@tzi.org, anuj.sehgal@navomi.com +# RFC8133 || S. Smyshlyaev. Ed., E. Alekseev, I. Oshkin, V. Popov || svs@cryptopro.ru, alekseev@cryptopro.ru, oshkin@cryptopro.ru, vpopov@cryptopro.ru +# RFC8135 || M. Danielson, M. Nilsson || magda@netinsight.net, mansaxel@besserwisser.org +# RFC8136 || B. Carpenter, R. Hinden || brian.e.carpenter@gmail.com, bob.hinden@gmail.com +# RFC8138 || P. Thubert, Ed., C. Bormann, L. Toutain, R. Cragie || pthubert@cisco.com, cabo@tzi.org, Laurent.Toutain@IMT-Atlantique.fr, robert.cragie@arm.com +# RFC8140 || A. Farrel || adrian@olddog.co.uk +# RFC8144 || K. Murchison || murch@andrew.cmu.edu +# RFC8145 || D. Wessels, W. Kumari, P. Hoffman || dwessels@verisign.com, warren@kumari.net, paul.hoffman@icann.org""".split('\n') + + +# # Many of these are addresses that the draft parser found incorrectly +# ignore_addresses =[ +# '0004454742@mcimail.com', +# '0006423401@mcimail.com', +# 'cabo@tzi.orgemail', +# 'california@san', +# 'cdl@rincon.com', +# 'hss@lando.hns.com', +# 'ietf-info@cnri.reston.va.us', +# 'illinois@urbana-champaign', +# 'jasdips@rwhois.net', +# 'labs@network', +# 'member@the', +# 'park@mit', +# 'research@icsi', +# 'research@isci', +# 'technopark@chaichee', +# 'texas@arlington', +# 'ura-bunyip@bunyip.com', +# ] + +# def get_rfc_data(): +# author_names = dict() +# author_emails = dict() +# for line in rfced_data: +# (rfc,names,emails) = line.split('||') +# rfc = int(rfc.lower().strip()[3:]) +# author_names[rfc] = [ x for x in map(str.lower,map(str.strip,names.split(','))) if x not in ['Ed.', 'ed.', '' ] ] +# author_emails[rfc] = [ x for x in map(unicode,map(str.lower,map(str.strip,emails.split(',')))) if x not in [ '', ] ] +# return author_names, author_emails + +# def get_all_the_email(): +# all_the_email = Email.objects.all() +# for e in all_the_email: +# e.l_address = e.address.lower() +# return all_the_email + +# def get_matching_emails(all_the_email,addrlist): +# """ Find Email objects with addresses case-insensitively matching things in the supplied list (for lack of __iin) """ +# l_addrlist = map(unicode.lower,addrlist) +# return [ e for e in all_the_email if e.l_address in l_addrlist ] + +# def show_verbose(rfc_num,*args): +# print "rfc%-4d :"%rfc_num,' '.join(map(str,args)) + +# ParsedAuthor = namedtuple('ParsedAuthor',['name','address']) + +# def get_parsed_authors(rfc_num): +# h,n = mkstemp() +# os.close(h) +# f = open(n,"w") +# f.write('%s/rfc%d.txt\n'%(settings.RFC_PATH,rfc_num)) +# f.close() +# lines = subprocess.check_output(['ietf/utils/draft.py','-a',n]) +# os.unlink(n) +# if not 'docauthors ' in lines: +# return [] +# authorline = [l for l in lines.split('\n') if l.startswith('docauthors ')][0] +# authstrings = authorline.split(':')[1].split(',') +# retval = [] +# for a in authstrings: +# if '<' in a: +# retval.append(ParsedAuthor(a[:a.find('<')].strip(),a[a.find('<'):a.find('>')][1:])) +# else: +# retval.append(ParsedAuthor(a.strip(),None)) + +# return retval + +# def calculate_changes(tracker_persons,tracker_emails,names,emails): +# adds = set() +# deletes = set() +# for email in emails: +# if email and email!='none' and email not in ignore_addresses: +# p = Person.objects.filter(email__address=email).first() +# if p: +# if not set(map(unicode.lower,p.email_set.values_list('address',flat=True))).intersection(tracker_emails): +# adds.add(email) +# else: +# #person_name = names[emails.index(email)] +# adds.add(email) +# for person in tracker_persons: +# if not set(map(unicode.lower,person.email_set.values_list('address',flat=True))).intersection(emails): +# match = False +# for index in [i for i,j in enumerate(emails) if j=='none' or not j]: +# if names[index].split()[-1].lower()==person.last_name().lower(): +# match = True +# if not match: +# deletes.add(person) +# return adds, deletes + +# def _main(): + +# parser = argparse.ArgumentParser(description="Recalculate RFC documentauthor_set"+'\n\n'+__doc__, +# formatter_class=argparse.RawDescriptionHelpFormatter,) +# parser.add_argument('-v','--verbose',help="Show the action taken for each RFC",action='store_true') +# parser.add_argument('--rfc',type=int, nargs='*',help="Only recalculate the given rfc numbers",dest='rfcnumberlist') +# args = parser.parse_args() + +# probable_email_match = set() +# probable_duplicates = [] + +# all_the_email = get_all_the_email() +# author_names, author_emails = get_rfc_data() + +# stats = { 'rfc not in tracker' :0, +# 'same addresses' :0, +# 'different addresses belonging to same people' :0, +# 'same names, rfced emails do not match' :0, +# 'rfced data is unusable' :0, +# "data doesn't match but no changes found" :0, +# 'changed authors' :0, } + +# for rfc_num in args.rfcnumberlist or sorted(author_names.keys()): + +# rfc = Document.objects.filter(docalias__name='rfc%s'%rfc_num).first() + +# if not rfc: +# if args.verbose: +# show_verbose(rfc_num,'rfc not in tracker') +# stats['rfc not in tracker'] += 1 +# continue + +# rfced_emails = set(author_emails[rfc_num]) +# tracker_emails = set(map(unicode.lower,rfc.authors.values_list('address',flat=True))) +# tracker_persons = set([x.person for x in rfc.authors.all()]) +# matching_emails = get_matching_emails(all_the_email,rfced_emails) +# rfced_persons = set([x.person for x in matching_emails]) +# known_emails = set([e.l_address for e in matching_emails]) +# unknown_emails = rfced_emails - known_emails +# unknown_persons = tracker_persons-rfced_persons + +# rfced_lastnames = sorted([n.split()[-1].lower() for n in author_names[rfc_num]]) +# tracker_lastnames = sorted([p.last_name().lower() for p in tracker_persons]) + +# if rfced_emails == tracker_emails: +# if args.verbose: +# show_verbose(rfc_num,'tracker and rfc editor have the same addresses') +# stats['same addresses'] += 1 +# continue + +# if len(rfced_emails)==len(tracker_emails) and not 'none' in author_emails[rfc_num]: +# if tracker_persons == rfced_persons: +# if args.verbose: +# show_verbose(rfc_num,'tracker and rfc editor have the different addresses belonging to same people') +# stats['different addresses belonging to same people'] += 1 +# continue +# else: +# if len(unknown_emails)==1 and len(tracker_persons-rfced_persons)==1: +# p = list(tracker_persons-rfced_persons)[0] +# probable_email_match.add(u"%s is probably %s (%s) : %s "%(list(unknown_emails)[0], p, p.pk, rfc_num)) +# elif len(unknown_emails)==len(unknown_persons): +# probable_email_match.add(u"%s are probably %s : %s"%(unknown_emails,[(p.ascii,p.pk) for p in unknown_persons],rfc_num)) +# else: +# probable_duplicates.append((tracker_persons^rfced_persons,rfc_num)) + +# if tracker_lastnames == rfced_lastnames: +# if args.verbose: +# show_verbose(rfc_num,"emails don't match up, but person names appear to be the same") +# stats[ 'same names, rfced emails do not match'] += 1 +# continue + +# use_rfc_data = bool(len(author_emails[rfc_num])==len(author_names[rfc_num])) +# if not use_rfc_data: +# if args.verbose: +# print 'Ignoring rfc database for rfc%d'%rfc_num +# stats[ 'rfced data is unusable'] += 1 + +# if use_rfc_data: +# adds, deletes = calculate_changes(tracker_persons,tracker_emails,author_names[rfc_num],author_emails[rfc_num]) +# parsed_authors=get_parsed_authors(rfc_num) +# parsed_adds, parsed_deletes = calculate_changes(tracker_persons,tracker_emails,[x.name for x in parsed_authors],[x.address for x in parsed_authors]) + +# for e in adds.union(parsed_adds) if use_rfc_data else parsed_adds: +# if not e or e in ignore_addresses: +# continue +# if not Person.objects.filter(email__address=e).exists(): +# if e not in parsed_adds: +# #print rfc_num,"Would add",e,"as",author_names[rfc_num][author_emails[rfc_num].index(e)],"(rfced database)" +# print "(address='%s',name='%s'),"%(e,author_names[rfc_num][author_emails[rfc_num].index(e)]),"# (rfced %d)"%rfc_num +# for p in Person.objects.filter(name__iendswith=author_names[rfc_num][author_emails[rfc_num].index(e)].split(' ')[-1]): +# print "\t", p.pk, p.ascii +# else: +# name = [x.name for x in parsed_authors if x.address==e][0] +# p = Person.objects.filter(name=name).first() +# if p: +# #print e,"is probably",p.pk,p +# print "'%s': %d, # %s (%d)"%(e,p.pk,p.ascii,rfc_num) + +# else: +# p = Person.objects.filter(ascii=name).first() +# if p: +# print e,"is probably",p.pk,p +# print "'%s': %d, # %s (%d)"%(e,p.pk,p.ascii,rfc_num) +# else: +# p = Person.objects.filter(ascii_short=name).first() +# if p: +# print e,"is probably",p.pk,p +# print "'%s': %d, # %s (%d)"%(e,p.pk,p.ascii,rfc_num) +# #print rfc_num,"Would add",e,"as",name,"(parsed)" +# print "(address='%s',name='%s'),"%(e,name),"# (parsed %d)"%rfc_num +# for p in Person.objects.filter(name__iendswith=name.split(' ')[-1]): +# print "\t", p.pk, p.ascii + +# if False: # This was a little useful, but the noise in the rfc_ed file keeps it from being completely useful +# for p in deletes: +# for n in author_names[rfc_num]: +# if p.last_name().lower()==n.split()[-1].lower(): +# email_candidate = author_emails[rfc_num][author_names[rfc_num].index(n)] +# email_found = Email.objects.filter(address=email_candidate).first() +# if email_found: +# probable_duplicates.append((set([p,email_found.person]),rfc_num)) +# else: +# probable_email_match.add(u"%s is probably %s (%s) : %s"%(email_candidate, p, p.pk, rfc_num)) + +# if args.verbose: +# if use_rfc_data: +# working_adds = parsed_adds +# seen_people = set(Email.objects.get(address=e).person for e in parsed_adds) +# for addr in adds: +# person = Email.objects.get(address=addr).person +# if person not in seen_people: +# working_adds.add(addr) +# seen_people.add(person) +# working_deletes = deletes.union(parsed_deletes) +# else: +# working_adds = parsed_adds +# working_deletes = parsed_deletes +# # unique_adds = set() # TODO don't add different addresses for the same person from the two sources +# if working_adds or working_deletes: +# show_verbose(rfc_num,"Changing original list",tracker_persons,"by adding",working_adds," and deleting",working_deletes) +# print "(",rfc_num,",",[e for e in working_adds],",",[p.pk for p in working_deletes],"), #",[p.ascii for p in working_deletes] +# else: +# stats["data doesn't match but no changes found"] += 1 +# show_verbose(rfc_num,"Couldn't figure out what to change") + +# if False: +# #if tracker_persons: +# #if any(['iab@' in e for e in adds]) or any(['iesg@' in e for e in adds]) or any(['IESG'==p.name for p in deletes]) or any(['IAB'==p.name for p in deletes]): +# print rfc_num +# print "tracker_persons",tracker_persons +# print "author_names",author_names[rfc_num] +# print "author_emails",author_emails[rfc_num] +# print "Adds:", adds +# print "Deletes:", deletes + +# stats['changed authors'] += 1 + +# if False: +# debug.show('rfc_num') +# debug.show('rfced_emails') +# debug.show('tracker_emails') +# debug.show('known_emails') +# debug.show('unknown_emails') +# debug.show('tracker_persons') +# debug.show('rfced_persons') +# debug.show('tracker_persons==rfced_persons') +# debug.show('[p.id for p in tracker_persons]') +# debug.show('[p.id for p in rfced_persons]') +# exit() + +# if True: +# for p in sorted(list(probable_email_match)): +# print p +# if True: +# print "Probable duplicate persons" +# for d,r in sorted(probable_duplicates): +# print [(p,p.pk) for p in d], r +# else: +# print len(probable_duplicates)," probable duplicate persons" + +# print stats + +# if __name__ == "__main__": +# _main() + diff --git a/dev/mq/Dockerfile b/dev/mq/Dockerfile index e8871c30a9..1738c4b3d2 100644 --- a/dev/mq/Dockerfile +++ b/dev/mq/Dockerfile @@ -1,6 +1,8 @@ # Dockerfile for RabbitMQ worker # -FROM rabbitmq:3-alpine +ARG RABBITMQ_VERSION=3.11-alpine + +FROM rabbitmq:${RABBITMQ_VERSION} LABEL maintainer="IETF Tools Team " # Copy the startup file diff --git a/dev/tests/debug.sh b/dev/tests/debug.sh index 37e7bc3ab2..d87c504bb9 100644 --- a/dev/tests/debug.sh +++ b/dev/tests/debug.sh @@ -3,7 +3,7 @@ # This script recreate the same environment used during tests on GitHub Actions # and drops you into a terminal at the point where the actual tests would be run. # -# Refer to https://github.com/ietf-tools/datatracker/blob/main/.github/workflows/build.yml#L141-L155 +# Refer to https://github.com/ietf-tools/datatracker/blob/main/.github/workflows/tests.yml#L47-L66 # for the commands to run next. # # Simply type "exit" + ENTER to exit and shutdown this test environment. @@ -12,13 +12,13 @@ echo "Fetching latest images..." docker pull ghcr.io/ietf-tools/datatracker-app-base:latest docker pull ghcr.io/ietf-tools/datatracker-db:latest echo "Starting containers..." -docker compose -f docker-compose.debug.yml -p dtdebug up -d +docker compose -f docker-compose.debug.yml -p dtdebug --compatibility up -d echo "Copying working directory into container..." docker compose -p dtdebug cp ../../. app:/__w/datatracker/datatracker/ echo "Run prepare script..." docker compose -p dtdebug exec app chmod +x ./dev/tests/prepare.sh docker compose -p dtdebug exec app sh ./dev/tests/prepare.sh -docker compose -p dtdebug exec app /usr/local/bin/wait-for db:3306 -- echo "DB ready" +docker compose -p dtdebug exec app /usr/local/bin/wait-for db:5432 -- echo "DB ready" echo "=================================================================" echo "Launching zsh terminal:" docker compose -p dtdebug exec app /bin/zsh diff --git a/dev/tests/docker-compose.debug.yml b/dev/tests/docker-compose.debug.yml index 44649c117b..8117b92375 100644 --- a/dev/tests/docker-compose.debug.yml +++ b/dev/tests/docker-compose.debug.yml @@ -1,33 +1,35 @@ -# This docker-compose replicates the test workflow happening on GitHub during a PR / build check. -# To be used from the debug.sh script. - -version: '3.8' - -services: - app: - image: ghcr.io/ietf-tools/datatracker-app-base:latest - command: -f /dev/null - working_dir: /__w/datatracker/datatracker - entrypoint: tail - hostname: app - volumes: - - /var/run/docker.sock:/var/run/docker.sock - environment: - CI: 'true' - GITHUB_ACTIONS: 'true' - HOME: /github/home - db: - image: ghcr.io/ietf-tools/datatracker-db:latest - restart: unless-stopped - volumes: - - mariadb-data:/var/lib/mysql - environment: - MYSQL_ROOT_PASSWORD: RkTkDPFnKpko - MYSQL_DATABASE: ietf_utf8 - MYSQL_USER: django - MYSQL_PASSWORD: RkTkDPFnKpko - CI: 'true' - GITHUB_ACTIONS: 'true' - -volumes: - mariadb-data: +# This docker-compose replicates the test workflow happening on GitHub during a PR / build check. +# To be used from the debug.sh script. + +version: '3.8' + +services: + app: + image: ghcr.io/ietf-tools/datatracker-app-base:latest + command: -f /dev/null + working_dir: /__w/datatracker/datatracker + entrypoint: tail + hostname: app + volumes: + - /var/run/docker.sock:/var/run/docker.sock + environment: + CI: 'true' + GITHUB_ACTIONS: 'true' + HOME: /github/home + deploy: + resources: + limits: + cpus: '2' + memory: '7GB' + + db: + image: ghcr.io/ietf-tools/datatracker-db:latest + restart: unless-stopped + volumes: + - postgresdb-data:/var/lib/postgresql/data + + blobstore: + image: ghcr.io/ietf-tools/datatracker-devblobstore:latest + +volumes: + postgresdb-data: diff --git a/dev/tests/prepare.sh b/dev/tests/prepare.sh index 2fb7cf7c95..dd8e3d6b4d 100644 --- a/dev/tests/prepare.sh +++ b/dev/tests/prepare.sh @@ -18,3 +18,4 @@ chmod +x ./docker/scripts/app-create-dirs.sh ./docker/scripts/app-create-dirs.sh echo "Fetching latest coverage results file..." curl -fsSL https://github.com/ietf-tools/datatracker/releases/download/baseline/coverage.json -o release-coverage.json +psql -U django -h db -d datatracker -v ON_ERROR_STOP=1 -c '\x' -c 'ALTER USER django set search_path=datatracker,public;' diff --git a/dev/tests/settings_local.py b/dev/tests/settings_local.py index 2b3d541387..e1ffd60edb 100644 --- a/dev/tests/settings_local.py +++ b/dev/tests/settings_local.py @@ -1,35 +1,24 @@ # Copyright The IETF Trust 2007-2019, All Rights Reserved # -*- coding: utf-8 -*- -from ietf.settings import * # pyflakes:ignore +from ietf.settings import * # pyflakes:ignore ALLOWED_HOSTS = ['*'] DATABASES = { 'default': { 'HOST': 'db', - 'PORT': 3306, - 'NAME': 'ietf_utf8', - 'ENGINE': 'django.db.backends.mysql', + 'PORT': 5432, + 'NAME': 'datatracker', + 'ENGINE': 'django.db.backends.postgresql', 'USER': 'django', 'PASSWORD': 'RkTkDPFnKpko', - 'OPTIONS': { - 'sql_mode': 'STRICT_TRANS_TABLES', - 'init_command': 'SET storage_engine=InnoDB; SET names "utf8"', - }, }, } -DATABASE_TEST_OPTIONS = { - 'init_command': 'SET storage_engine=InnoDB', -} - IDSUBMIT_IDNITS_BINARY = "/usr/local/bin/idnits" -IDSUBMIT_REPOSITORY_PATH = "test/id/" -IDSUBMIT_STAGING_PATH = "test/staging/" -INTERNET_DRAFT_ARCHIVE_DIR = "test/archive/" -INTERNET_ALL_DRAFTS_ARCHIVE_DIR = "test/archive/" -RFC_PATH = "test/rfc/" +IDSUBMIT_REPOSITORY_PATH = "/assets/ietfdata/doc/draft/repository" +IDSUBMIT_STAGING_PATH = "/assets/www6s/staging/" AGENDA_PATH = '/assets/www6s/proceedings/' MEETINGHOST_LOGO_PATH = AGENDA_PATH @@ -47,7 +36,6 @@ SUBMIT_YANG_CATALOG_MODEL_DIR = '/assets/ietf-ftp/yang/catalogmod/' SUBMIT_YANG_DRAFT_MODEL_DIR = '/assets/ietf-ftp/yang/draftmod/' -SUBMIT_YANG_INVAL_MODEL_DIR = '/assets/ietf-ftp/yang/invalmod/' SUBMIT_YANG_IANA_MODEL_DIR = '/assets/ietf-ftp/yang/ianamod/' SUBMIT_YANG_RFC_MODEL_DIR = '/assets/ietf-ftp/yang/rfcmod/' @@ -67,8 +55,11 @@ BOFREQ_PATH = '/assets/ietf-ftp/bofreq/' CONFLICT_REVIEW_PATH = '/assets/ietf-ftp/conflict-reviews/' STATUS_CHANGE_PATH = '/assets/ietf-ftp/status-changes/' -INTERNET_DRAFT_ARCHIVE_DIR = '/assets/ietf-ftp/internet-drafts/' +INTERNET_DRAFT_ARCHIVE_DIR = '/assets/collection/draft-archive' INTERNET_ALL_DRAFTS_ARCHIVE_DIR = '/assets/ietf-ftp/internet-drafts/' +BIBXML_BASE_PATH = '/assets/ietfdata/derived/bibxml' +FTP_DIR = '/assets/ftp' +NFS_METRICS_TMP_DIR = '/assets/tmp' NOMCOM_PUBLIC_KEYS_DIR = 'data/nomcom_keys/public_keys/' SLIDE_STAGING_PATH = 'test/staging/' diff --git a/docker-compose.yml b/docker-compose.yml index db8ea9152a..073d04b896 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -1,5 +1,3 @@ -version: '3.8' - services: app: build: @@ -15,6 +13,8 @@ services: # network_mode: service:db depends_on: + - blobdb + - blobstore - db - mq @@ -26,7 +26,7 @@ services: # UID: 1001 # GID: 1001 # DATADIR: data - # DJANGO_SETTINGS_MODULE: settings_sqlitetest + # DJANGO_SETTINGS_MODULE: settings_test # Uncomment the next line to use a non-root user for all processes. # user: dev @@ -37,55 +37,118 @@ services: db: image: ghcr.io/ietf-tools/datatracker-db:latest # build: - # context: .. + # context: . # dockerfile: docker/db.Dockerfile restart: unless-stopped volumes: - - mariadb-data:/var/lib/mysql - environment: - MYSQL_ROOT_PASSWORD: RkTkDPFnKpko - MYSQL_DATABASE: ietf_utf8 - MYSQL_USER: django - MYSQL_PASSWORD: RkTkDPFnKpko - command: - - '--character-set-server=utf8' - - '--collation-server=utf8_unicode_ci' - - '--innodb-buffer-pool-size=1G' - - '--innodb-log-buffer-size=128M' - - '--innodb-log-file-size=256M' - - '--innodb-write-io-threads=8' - - '--innodb-flush-log-at-trx-commit=0' - - '--performance-schema=1' + - postgresdb-data:/var/lib/postgresql/data # Add "forwardPorts": ["5432"] to **devcontainer.json** to forward PostgreSQL locally. # (Adding the "ports" property to this file will not forward from a Codespace.) + pgadmin: + image: dpage/pgadmin4:latest + restart: unless-stopped + environment: + - PGADMIN_DEFAULT_EMAIL=dev@ietf.org + - PGADMIN_DEFAULT_PASSWORD=dev + - PGADMIN_CONFIG_LOGIN_BANNER="Login with dev@ietf.org / dev" + - PGADMIN_DISABLE_POSTFIX=True + - PGADMIN_CONFIG_MASTER_PASSWORD_REQUIRED=False + - SCRIPT_NAME=/pgadmin + volumes: + - ./docker/configs/pgadmin-servers.json:/pgadmin4/servers.json + + static: + image: ghcr.io/ietf-tools/static:latest + restart: unless-stopped + mq: image: rabbitmq:3-alpine restart: unless-stopped celery: - image: ghcr.io/ietf-tools/datatracker-celery:latest + build: + context: . + dockerfile: docker/celery.Dockerfile + init: true + environment: + CELERY_APP: ietf + CELERY_ROLE: worker + UPDATE_REQUIREMENTS_FROM: requirements.txt + DEV_MODE: "yes" + command: + - '--loglevel=INFO' + depends_on: + - blobdb + - blobstore + - db + - mq + restart: unless-stopped + stop_grace_period: 1m + volumes: + - .:/workspace + - app-assets:/assets + + replicator: + build: + context: . + dockerfile: docker/celery.Dockerfile init: true environment: CELERY_APP: ietf CELERY_ROLE: worker UPDATE_REQUIREMENTS_FROM: requirements.txt + DEV_MODE: "yes" command: - '--loglevel=INFO' + - '--queues=blobdb' + - '--concurrency=1' + depends_on: + - blobdb + - blobstore - db + - mq restart: unless-stopped stop_grace_period: 1m volumes: - .:/workspace - app-assets:/assets + blobstore: + image: ghcr.io/ietf-tools/datatracker-devblobstore:latest + restart: unless-stopped + volumes: + - "minio-data:/data" + + blobdb: + image: postgres:17 + restart: unless-stopped + environment: + POSTGRES_DB: blob + POSTGRES_USER: dt + POSTGRES_PASSWORD: abcd1234 + volumes: + - blobdb-data:/var/lib/postgresql/data + +# typesense: +# image: typesense/typesense:30.1 +# restart: on-failure +# ports: +# - "8108:8108" +# volumes: +# - ./typesense-data:/data +# command: +# - '--data-dir=/data' +# - '--api-key=typesense-api-key' +# - '--enable-cors' + # Celery Beat is a periodic task runner. It is not normally needed for development, # but can be enabled by uncommenting the following. # # beat: -# image: ghcr.io/ietf-tools/datatracker-celery:latest +# image: "${COMPOSE_PROJECT_NAME}-celery" # init: true # environment: # CELERY_APP: ietf @@ -99,7 +162,10 @@ services: # stop_grace_period: 1m # volumes: # - .:/workspace +# - app-assets:/assets volumes: - mariadb-data: + postgresdb-data: app-assets: + minio-data: + blobdb-data: diff --git a/docker/README.md b/docker/README.md index b16687618f..0ca79a6e89 100644 --- a/docker/README.md +++ b/docker/README.md @@ -1,14 +1,28 @@ # Datatracker Development in Docker +- [Getting started](#getting-started) +- [Using Visual Studio Code](#using-visual-studio-code) + - [Initial Setup](#initial-setup) + - [Subsequent Launch](#subsequent-launch) + - [Usage](#usage) +- [Using Other Editors / Generic](#using-other-editors--generic) + - [Exit Environment](#exit-environment) + - [Accessing PostgreSQL Port](#accessing-postgresql-port) +- [Clean and Rebuild DB from latest image](#clean-and-rebuild-db-from-latest-image) +- [Clean all](#clean-all) +- [Updating an older environment](#updating-an-older-environment) +- [Notes / Troubleshooting](#notes--troubleshooting) + ## Getting started 1. [Set up Docker](https://docs.docker.com/get-started/) on your preferred platform. On Windows, it is highly recommended to use the [WSL 2 *(Windows Subsystem for Linux)*](https://docs.docker.com/desktop/windows/wsl/) backend. +> [!IMPORTANT] > See the [IETF Tools Windows Dev guide](https://github.com/ietf-tools/.github/blob/main/docs/windows-dev.md) on how to get started when using Windows. -2. On Linux, you must also install [Docker Compose](https://docs.docker.com/compose/install/). Docker Desktop for Mac and Windows already include Docker Compose. +2. On Linux, you must [install Docker Compose manually](https://docs.docker.com/compose/install/linux/#install-the-plugin-manually) and not install Docker Desktop. On Mac and Windows install Docker Desktop which already includes Docker Compose. -2. If you have a copy of the datatracker code checked out already, simply `cd` to the top-level directory. +3. If you have a copy of the datatracker code checked out already, simply `cd` to the top-level directory. If not, check out a datatracker branch as usual. We'll check out `main` below, but you can use any branch: @@ -18,7 +32,7 @@ git checkout main ``` -3. Follow the instructions for your preferred editor: +4. Follow the instructions for your preferred editor: - [Visual Studio Code](#using-visual-studio-code) - [Other Editors / Generic](#using-other-editors--generic) @@ -29,9 +43,9 @@ This project includes a devcontainer configuration which automates the setup of ### Initial Setup 1. Launch [VS Code](https://code.visualstudio.com/) -2. Under the **Extensions** tab, ensure you have the **Remote - Containers** ([ms-vscode-remote.remote-containers](https://marketplace.visualstudio.com/items?itemName=ms-vscode-remote.remote-containers)) extension installed. +2. Under the **Extensions** tab, ensure you have the **Dev Containers** ([ms-vscode-remote.remote-containers](https://marketplace.visualstudio.com/items?itemName=ms-vscode-remote.remote-containers)) extension installed. * On Linux, note that the Snap installation of VS Code is [incompatible with this plugin](https://code.visualstudio.com/docs/devcontainers/containers#_system-requirements:~:text=snap%20package%20is%20not%20supported). - * On Windows, you also need the **Remote - WSL** ([ms-vscode-remote.remote-wsl](https://marketplace.visualstudio.com/items?itemName=ms-vscode-remote.remote-wsl)) extension to take advantage of the WSL 2 *(Windows Subsystem for Linux)* native integration. + * On Windows, you also need the **WSL** ([ms-vscode-remote.remote-wsl](https://marketplace.visualstudio.com/items?itemName=ms-vscode-remote.remote-wsl)) extension to take advantage of the WSL 2 *(Windows Subsystem for Linux)* native integration. 2. Open the top-level directory of the datatracker code you fetched above. 3. A prompt inviting you to reopen the project in containers will appear in the bottom-right corner. Click the **Reopen in Container** button. If you missed the prompt, you can press `F1`, start typing `reopen in container` task and launch it. 4. VS Code will relaunch in the dev environment and create the containers automatically. @@ -45,7 +59,7 @@ You can also open the datatracker project folder and click the **Reopen in conta ### Usage -- Under the **Run and Debug** tab, you can run the server with the debugger attached using **Run Server** (F5). Once the server is ready to accept connections, you'll be prompted to open in a browser. You can also open [http://localhost:8000](http://localhost:8000) in a browser. +- Under the **Run and Debug** tab, you can run the server with the debugger attached using **Run Server** (F5). Once the server is ready to accept connections, you'll be prompted to open in a browser. Navigate to [http://localhost:8000](http://localhost:8000) in your preferred browser. > An alternate profile **Run Server with Debug Toolbar** is also available from the dropdown menu, which displays various tools on top of the webpage. However, note that this configuration has a significant performance impact. @@ -64,11 +78,7 @@ You can also open the datatracker project folder and click the **Reopen in conta ![](assets/vscode-terminal-new.png) -- Under the **SQL Tools** tab, a connection **Local Dev** is preconfigured to connect to the DB container. Using this tool, you can list tables, view records and execute SQL queries directly from VS Code. - - > The port `3306` is also exposed to the host automatically, should you prefer to use your own SQL tool. - - ![](assets/vscode-sqltools.png) +- The pgAdmin web interface, a PostgreSQL DB browser / management UI, is available at [http://localhost:8000/pgadmin/](http://localhost:8000/pgadmin/). - Under the **Task Explorer** tab, a list of available preconfigured tasks is displayed. *(You may need to expand the tree to `src > vscode` to see it.)* These are common scritps you can run *(e.g. run tests, fetch assets, etc.)*. @@ -87,11 +97,10 @@ You can also open the datatracker project folder and click the **Reopen in conta On Linux / macOS: ```sh - cd docker - ./run + ./docker/run # or whatever path you need ``` - > Note that you can pass the `-r` flag to `./run` to force a rebuild of the containers. This is useful if you switched branches and that the existing containers still contain configurations from the old branch. You should also use this if you don't regularly keep up with main and your containers reflect a much older version of the branch. + > Note that you can pass the `-r` flag to `run` to force a rebuild of the containers. This is useful if you switched branches and that the existing containers still contain configurations from the old branch. You should also use this if you don't regularly keep up with main and your containers reflect a much older version of the branch. On Windows *(using Powershell)*: ```sh @@ -104,7 +113,7 @@ You can also open the datatracker project folder and click the **Reopen in conta 2. Wait for the containers to initialize. Upon completion, you will be dropped into a shell from which you can start the datatracker and execute related commands as usual, for example ``` - ietf/manage.py runserver 0.0.0.0:8000 + ietf/manage.py runserver 8001 ``` to start the datatracker. @@ -127,7 +136,14 @@ docker compose down to terminate the containers. -### Clean and Rebuild DB from latest image +### Accessing PostgreSQL Port + +The port is exposed but not automatically mapped to `5432` to avoid potential conflicts with the host. To get the mapped port, run the command *(from the project `/docker` directory)*: +```sh +docker compose port db 5432 +``` + +## Clean and Rebuild DB from latest image To delete the active DB container, its volume and get the latest image / DB dump, simply run the following command: @@ -145,7 +161,7 @@ docker compose pull db docker compose build --no-cache db ``` -### Clean all +## Clean all To delete all containers for this project, its associated images and purge any remaining dangling images, simply run the following command: @@ -162,11 +178,19 @@ docker compose down -v --rmi all docker image prune ``` -### Accessing MariaDB Port +## Updating an older environment + +If you already have a clone, such as from a previous codesprint, and are updating that clone, before starting the datatracker from the updated image: +1. `rm ietf/settings_local.py` *(The startup script will put a new one, appropriate to the current release, in place)* +1. Execute the [Clean all](#clean-all) sequence above. + +If the dev environment fails to start, even after running the [Clean all](#clean-all) sequence above, you can fully purge all docker cache, containers, images and volumes by running the command below. + +> [!CAUTION] +> Note that this will delete everything docker-related, including non-datatracker docker resources you might have. -The port is exposed but not mapped to `3306` to avoid potential conflicts with the host. To get the mapped port, run the command *(from the project `/docker` directory)*: ```sh -docker compose port db 3306 +docker system prune -a --volumes ``` ## Notes / Troubleshooting @@ -188,3 +212,17 @@ The content of the source files will be copied into the target `.ics` files. Mak ### Missing assets in the data folder Because including all assets in the image would significantly increase the file size, they are not included by default. You can however fetch them by running the **Fetch assets via rsync** task in VS Code or run manually the script `docker/scripts/app-rsync-extras.sh` + +### Linux file permissions leaking to the host system + +If on the host filesystem you have permissions that look like this, + +```bash +$ ls -la +total 4624 +drwxrwxr-x 2 100999 100999 4096 May 25 07:56 bin +drwxrwxr-x 5 100999 100999 4096 May 25 07:56 client +(etc...) +``` + +Try uninstalling Docker Desktop and installing Docker Compose manually. The Docker Compose bundled with Docker Desktop is incompatible with our software. See also [Rootless Docker: file ownership changes #3343](https://github.com/lando/lando/issues/3343), [Docker context desktop-linux has container permission issues #75](https://github.com/docker/desktop-linux/issues/75). diff --git a/docker/app.Dockerfile b/docker/app.Dockerfile index 6d539b1a56..dd4cf72ffd 100644 --- a/docker/app.Dockerfile +++ b/docker/app.Dockerfile @@ -9,6 +9,7 @@ ARG USER_UID=1000 ARG USER_GID=$USER_UID COPY docker/scripts/app-setup-debian.sh /tmp/library-scripts/docker-setup-debian.sh RUN sed -i 's/\r$//' /tmp/library-scripts/docker-setup-debian.sh && chmod +x /tmp/library-scripts/docker-setup-debian.sh + RUN apt-get update && export DEBIAN_FRONTEND=noninteractive \ # Remove imagemagick due to https://security-tracker.debian.org/tracker/CVE-2019-10131 && apt-get purge -y imagemagick imagemagick-6-common \ @@ -24,12 +25,21 @@ COPY docker/scripts/app-setup-python.sh /tmp/library-scripts/docker-setup-python RUN sed -i 's/\r$//' /tmp/library-scripts/docker-setup-python.sh && chmod +x /tmp/library-scripts/docker-setup-python.sh RUN bash /tmp/library-scripts/docker-setup-python.sh "none" "/usr/local" "${PIPX_HOME}" "${USERNAME}" +# Setup nginx +COPY docker/scripts/app-setup-nginx.sh /tmp/library-scripts/docker-setup-nginx.sh +RUN sed -i 's/\r$//' /tmp/library-scripts/docker-setup-nginx.sh && chmod +x /tmp/library-scripts/docker-setup-nginx.sh +RUN bash /tmp/library-scripts/docker-setup-nginx.sh +COPY docker/configs/nginx-proxy.conf /etc/nginx/sites-available/default +COPY docker/configs/nginx-502.html /var/www/html/502.html + # Remove library scripts for final image RUN rm -rf /tmp/library-scripts # Copy the startup file COPY docker/scripts/app-init.sh /docker-init.sh -RUN sed -i 's/\r$//' /docker-init.sh && chmod +x /docker-init.sh +COPY docker/scripts/app-start.sh /docker-start.sh +RUN sed -i 's/\r$//' /docker-init.sh && chmod +rx /docker-init.sh +RUN sed -i 's/\r$//' /docker-start.sh && chmod +rx /docker-start.sh # Fix user UID / GID to match host RUN groupmod --gid $USER_GID $USERNAME \ diff --git a/docker/base.Dockerfile b/docker/base.Dockerfile index 97604e62f9..2501636049 100644 --- a/docker/base.Dockerfile +++ b/docker/base.Dockerfile @@ -1,132 +1,159 @@ -FROM python:3.9-bullseye -LABEL maintainer="IETF Tools Team " - -ENV DEBIAN_FRONTEND=noninteractive - -# Update system packages -RUN apt-get update \ - && apt-get -qy upgrade \ - && apt-get -y install --no-install-recommends apt-utils dialog 2>&1 - -# Add Node.js Source -RUN curl -fsSL https://deb.nodesource.com/setup_16.x | bash - - -# Add Docker Source -RUN curl -fsSL https://download.docker.com/linux/debian/gpg | gpg --dearmor -o /usr/share/keyrings/docker-archive-keyring.gpg -RUN echo "deb [arch=$(dpkg --print-architecture) signed-by=/usr/share/keyrings/docker-archive-keyring.gpg] https://download.docker.com/linux/debian \ - $(lsb_release -cs) stable" | tee /etc/apt/sources.list.d/docker.list > /dev/null - -# Install the packages we need -RUN apt-get update --fix-missing && apt-get install -qy \ - apache2-utils \ - apt-file \ - bash \ - build-essential \ - curl \ - default-jdk \ - docker-ce-cli \ - enscript \ - gawk \ - g++ \ - gcc \ - ghostscript \ - git \ - gnupg \ - jq \ - less \ - libcairo2-dev \ - libgtk2.0-0 \ - libgtk-3-0 \ - libnotify-dev \ - libgconf-2-4 \ - libgbm-dev \ - libnss3 \ - libxss1 \ - libasound2 \ - libxtst6 \ - libmagic-dev \ - libmariadb-dev \ - libmemcached-tools \ - locales \ - make \ - mariadb-client \ - memcached \ - nano \ - netcat \ - nodejs \ - pigz \ - pv \ - python3-ipython \ - ripgrep \ - rsync \ - rsyslog \ - ruby \ - ruby-rubygems \ - unzip \ - wget \ - xauth \ - xvfb \ - yang-tools \ - zsh - -# Install kramdown-rfc2629 (ruby) -RUN gem install kramdown-rfc2629 - -# Install chromedriver -COPY docker/scripts/app-install-chromedriver.sh /tmp/app-install-chromedriver.sh -RUN sed -i 's/\r$//' /tmp/app-install-chromedriver.sh && \ - chmod +x /tmp/app-install-chromedriver.sh -RUN /tmp/app-install-chromedriver.sh - -# Fix /dev/shm permissions for chromedriver -RUN chmod 1777 /dev/shm - -# Activate Yarn -RUN corepack enable - -# Get rid of installation files we don't need in the image, to reduce size -RUN apt-get autoremove -y && apt-get clean -y && rm -rf /var/lib/apt/lists/* - -# "fake" dbus address to prevent errors -# https://github.com/SeleniumHQ/docker-selenium/issues/87 -ENV DBUS_SESSION_BUS_ADDRESS=/dev/null - -# avoid million NPM install messages -ENV npm_config_loglevel warn -# allow installing when the main user is root -ENV npm_config_unsafe_perm true -# disable NPM funding messages -ENV npm_config_fund false - -# Set locale to en_US.UTF-8 -RUN echo "LC_ALL=en_US.UTF-8" >> /etc/environment && \ - echo "en_US.UTF-8 UTF-8" >> /etc/locale.gen && \ - echo "LANG=en_US.UTF-8" > /etc/locale.conf && \ - dpkg-reconfigure locales && \ - locale-gen en_US.UTF-8 && \ - update-locale LC_ALL en_US.UTF-8 -ENV LC_ALL en_US.UTF-8 - -# Install idnits -ADD https://raw.githubusercontent.com/ietf-tools/idnits-mirror/main/idnits /usr/local/bin/ -RUN chmod +rx /usr/local/bin/idnits - -# Turn off rsyslog kernel logging (doesn't work in Docker) -RUN sed -i '/imklog/s/^/#/' /etc/rsyslog.conf - -# Colorize the bash shell -RUN sed -i 's/#force_color_prompt=/force_color_prompt=/' /root/.bashrc - -# Turn off rsyslog kernel logging (doesn't work in Docker) -RUN sed -i '/imklog/s/^/#/' /etc/rsyslog.conf - -# Fetch wait-for utility -ADD https://raw.githubusercontent.com/eficode/wait-for/v2.1.3/wait-for /usr/local/bin/ -RUN chmod +rx /usr/local/bin/wait-for - -# Create assets directory -RUN mkdir -p /assets - -# Create workspace -RUN mkdir -p /workspace -WORKDIR /workspace +FROM python:3.12-bookworm +LABEL maintainer="IETF Tools Team " + +ENV DEBIAN_FRONTEND=noninteractive +ENV NODE_MAJOR=16 + +# Update system packages +RUN apt-get update \ + && apt-get -qy upgrade \ + && apt-get -y install --no-install-recommends apt-utils dialog 2>&1 + +# Add Node.js Source +RUN apt-get install -y --no-install-recommends ca-certificates curl gnupg \ + && mkdir -p /etc/apt/keyrings \ + && curl -fsSL https://deb.nodesource.com/gpgkey/nodesource-repo.gpg.key | gpg --dearmor -o /etc/apt/keyrings/nodesource.gpg \ + && echo "deb [signed-by=/etc/apt/keyrings/nodesource.gpg] https://deb.nodesource.com/node_$NODE_MAJOR.x nodistro main" | tee /etc/apt/sources.list.d/nodesource.list +RUN echo "Package: nodejs" >> /etc/apt/preferences.d/preferences \ + && echo "Pin: origin deb.nodesource.com" >> /etc/apt/preferences.d/preferences \ + && echo "Pin-Priority: 1001" >> /etc/apt/preferences.d/preferences + +# Add Docker Source +RUN mkdir -p /etc/apt/keyrings \ + && curl -fsSL https://download.docker.com/linux/debian/gpg | gpg --dearmor -o /etc/apt/keyrings/docker-archive-keyring.gpg \ + && echo "deb [arch=$(dpkg --print-architecture) signed-by=/etc/apt/keyrings/docker-archive-keyring.gpg] https://download.docker.com/linux/debian $(. /etc/os-release && echo "$VERSION_CODENAME") stable" | tee /etc/apt/sources.list.d/docker.list + +# Add PostgreSQL Source +RUN mkdir -p /etc/apt/keyrings \ + && curl -fsSL https://www.postgresql.org/media/keys/ACCC4CF8.asc | gpg --dearmor -o /etc/apt/keyrings/apt.postgresql.org.gpg \ + && echo "deb [signed-by=/etc/apt/keyrings/apt.postgresql.org.gpg] https://apt.postgresql.org/pub/repos/apt $(. /etc/os-release && echo "$VERSION_CODENAME")-pgdg main" | tee /etc/apt/sources.list.d/pgdg.list + +# Install the packages we need +RUN apt-get update --fix-missing && apt-get install -qy --no-install-recommends \ + apache2-utils \ + apt-file \ + bash \ + build-essential \ + curl \ + default-jdk \ + docker-ce-cli \ + enscript \ + firefox-esr \ + gawk \ + g++ \ + gcc \ + ghostscript \ + git \ + gnupg \ + jq \ + less \ + libcairo2-dev \ + libgtk2.0-0 \ + libgtk-3-0 \ + libnotify-dev \ + libgconf-2-4 \ + libgbm-dev \ + libnss3 \ + libxss1 \ + libasound2 \ + libxtst6 \ + libmagic-dev \ + libmariadb-dev \ + libmemcached-tools \ + libyang2-tools \ + locales \ + make \ + mariadb-client \ + memcached \ + nano \ + netcat-traditional \ + nodejs \ + pgloader \ + pigz \ + postgresql-client-17 \ + pv \ + python3-ipython \ + ripgrep \ + rsync \ + rsyslog \ + ruby \ + ruby-rubygems \ + unzip \ + wget \ + xauth \ + xvfb \ + zsh + +# Install kramdown-rfc2629 (ruby) +RUN gem install kramdown-rfc2629 + +# GeckoDriver +ARG GECKODRIVER_VERSION=latest +RUN GK_VERSION=$(if [ ${GECKODRIVER_VERSION:-latest} = "latest" ]; then echo "0.34.0"; else echo $GECKODRIVER_VERSION; fi) \ + && echo "Using GeckoDriver version: "$GK_VERSION \ + && wget --no-verbose -O /tmp/geckodriver.tar.gz https://github.com/mozilla/geckodriver/releases/download/v$GK_VERSION/geckodriver-v$GK_VERSION-linux64.tar.gz \ + && rm -rf /opt/geckodriver \ + && tar -C /opt -zxf /tmp/geckodriver.tar.gz \ + && rm /tmp/geckodriver.tar.gz \ + && mv /opt/geckodriver /opt/geckodriver-$GK_VERSION \ + && chmod 755 /opt/geckodriver-$GK_VERSION \ + && ln -fs /opt/geckodriver-$GK_VERSION /usr/bin/geckodriver + +# Activate Yarn +RUN corepack enable + +# Get rid of installation files we don't need in the image, to reduce size +RUN apt-get autoremove -y && apt-get clean -y && rm -rf /var/lib/apt/lists/* /var/cache/apt/* + +# "fake" dbus address to prevent errors +# https://github.com/SeleniumHQ/docker-selenium/issues/87 +ENV DBUS_SESSION_BUS_ADDRESS=/dev/null + +# avoid million NPM install messages +ENV npm_config_loglevel=warn +# allow installing when the main user is root +ENV npm_config_unsafe_perm=true +# disable NPM funding messages +ENV npm_config_fund=false + +# Set locale to en_US.UTF-8 +RUN echo "LC_ALL=en_US.UTF-8" >> /etc/environment && \ + echo "en_US.UTF-8 UTF-8" >> /etc/locale.gen && \ + echo "LANG=en_US.UTF-8" > /etc/locale.conf && \ + dpkg-reconfigure locales && \ + locale-gen en_US.UTF-8 && \ + update-locale LC_ALL en_US.UTF-8 +ENV LC_ALL=en_US.UTF-8 + +# Install idnits +ADD https://raw.githubusercontent.com/ietf-tools/idnits-mirror/main/idnits /usr/local/bin/ +RUN chmod +rx /usr/local/bin/idnits + +# Install required fonts +RUN mkdir -p /tmp/fonts && \ + wget -q -O /tmp/fonts.tar.gz https://github.com/ietf-tools/xml2rfc-fonts/archive/refs/tags/3.22.0.tar.gz && \ + tar zxf /tmp/fonts.tar.gz -C /tmp/fonts && \ + mv /tmp/fonts/*/noto/* /usr/local/share/fonts/ && \ + mv /tmp/fonts/*/roboto_mono/* /usr/local/share/fonts/ && \ + rm -rf /tmp/fonts.tar.gz /tmp/fonts/ && \ + fc-cache -f + +# Turn off rsyslog kernel logging (doesn't work in Docker) +RUN sed -i '/imklog/s/^/#/' /etc/rsyslog.conf + +# Colorize the bash shell +RUN sed -i 's/#force_color_prompt=/force_color_prompt=/' /root/.bashrc + +# Turn off rsyslog kernel logging (doesn't work in Docker) +RUN sed -i '/imklog/s/^/#/' /etc/rsyslog.conf + +# Fetch wait-for utility +ADD https://raw.githubusercontent.com/eficode/wait-for/v2.1.3/wait-for /usr/local/bin/ +RUN chmod +rx /usr/local/bin/wait-for + +# Create assets directory +RUN mkdir -p /assets + +# Create workspace +RUN mkdir -p /workspace +WORKDIR /workspace diff --git a/docker/celery.Dockerfile b/docker/celery.Dockerfile new file mode 100644 index 0000000000..e93ca3cf77 --- /dev/null +++ b/docker/celery.Dockerfile @@ -0,0 +1,55 @@ +FROM ghcr.io/ietf-tools/datatracker-app-base:latest +LABEL maintainer="IETF Tools Team " + +ENV DEBIAN_FRONTEND=noninteractive + +# Install needed packages and setup non-root user. +ARG USERNAME=dev +ARG USER_UID=1000 +ARG USER_GID=$USER_UID +COPY docker/scripts/app-setup-debian.sh /tmp/library-scripts/docker-setup-debian.sh +RUN sed -i 's/\r$//' /tmp/library-scripts/docker-setup-debian.sh && chmod +x /tmp/library-scripts/docker-setup-debian.sh + +RUN apt-get update && export DEBIAN_FRONTEND=noninteractive \ + # Remove imagemagick due to https://security-tracker.debian.org/tracker/CVE-2019-10131 + && apt-get purge -y imagemagick imagemagick-6-common \ + # Install common packages, non-root user + # Syntax: ./docker-setup-debian.sh [install zsh flag] [username] [user UID] [user GID] [upgrade packages flag] [install Oh My Zsh! flag] [Add non-free packages] + && bash /tmp/library-scripts/docker-setup-debian.sh "true" "${USERNAME}" "${USER_UID}" "${USER_GID}" "false" "true" "true" + +# Setup default python tools in a venv via pipx to avoid conflicts +ENV PIPX_HOME=/usr/local/py-utils \ + PIPX_BIN_DIR=/usr/local/py-utils/bin +ENV PATH=${PATH}:${PIPX_BIN_DIR} +COPY docker/scripts/app-setup-python.sh /tmp/library-scripts/docker-setup-python.sh +RUN sed -i 's/\r$//' /tmp/library-scripts/docker-setup-python.sh && chmod +x /tmp/library-scripts/docker-setup-python.sh +RUN bash /tmp/library-scripts/docker-setup-python.sh "none" "/usr/local" "${PIPX_HOME}" "${USERNAME}" + +# Remove library scripts for final image +RUN rm -rf /tmp/library-scripts + +# Copy the startup file +COPY docker/scripts/app-init-celery.sh /docker-init.sh +RUN sed -i 's/\r$//' /docker-init.sh && \ + chmod +x /docker-init.sh + +ENTRYPOINT [ "/docker-init.sh" ] + +# Fix user UID / GID to match host +RUN groupmod --gid $USER_GID $USERNAME \ + && usermod --uid $USER_UID --gid $USER_GID $USERNAME \ + && chown -R $USER_UID:$USER_GID /home/$USERNAME \ + || exit 0 + +# Switch to local dev user +USER dev:dev + +# Install current datatracker python dependencies +COPY requirements.txt /tmp/pip-tmp/ +RUN pip3 --disable-pip-version-check --no-cache-dir install --user --no-warn-script-location -r /tmp/pip-tmp/requirements.txt +RUN pip3 --disable-pip-version-check --no-cache-dir install --user --no-warn-script-location watchdog[watchmedo] + +RUN sudo rm -rf /tmp/pip-tmp + +VOLUME [ "/assets" ] + diff --git a/docker/cleanall b/docker/cleanall index 91eac1764b..c6104aaef9 100755 --- a/docker/cleanall +++ b/docker/cleanall @@ -1,5 +1,11 @@ #!/bin/bash +if test $(basename $PWD ) != "docker" +then + echo "Run this from the docker directory" 1>&2 + exit 1 +fi + read -p "Stop and remove all containers, volumes and images for this project? [y/N] " -n 1 -r echo if [[ $REPLY =~ ^[Yy]$ ]] @@ -7,6 +13,5 @@ then cd .. echo "Shutting down any instance still running and purge images..." docker compose down -v --rmi all - cd docker echo "Done!" fi diff --git a/docker/cleandb b/docker/cleandb index f772ebce39..c881503eae 100755 --- a/docker/cleandb +++ b/docker/cleandb @@ -1,13 +1,19 @@ #!/bin/bash +if test $(basename $PWD ) != "docker" +then + echo "Run this from the docker directory" 1>&2 + exit 1 +fi + cd .. echo "Shutting down any instance still running..." docker compose down echo "Removing DB volume..." PROJNAME=$(basename $PWD) -docker volume rm -f "${PROJNAME}_mariadb-data" +docker volume rm -f "${PROJNAME}_postgresdb-data" echo "Rebuilding the DB image..." docker compose pull db docker compose build --no-cache db -cd docker + echo "Done!" diff --git a/docker/configs/nginx-502.html b/docker/configs/nginx-502.html new file mode 100644 index 0000000000..b5577d3e17 --- /dev/null +++ b/docker/configs/nginx-502.html @@ -0,0 +1,61 @@ + + + + + + Datatracker DEV + + + + IETF +

Datatracker

+

Could not connect to dev server.

+
+

Is the datatracker server running?

+

Using VS Code, open the Run and Debug tab on the left and click the symbol (Run Server) to start the server.

+

Otherwise, run the command ietf/manage.py runserver 8001 from the terminal.

+
+
+

You can manage the database at /pgadmin.

+
+

For more information, check out the Datatracker Development in Docker guide.

+ + diff --git a/docker/configs/nginx-proxy.conf b/docker/configs/nginx-proxy.conf new file mode 100644 index 0000000000..5a9ae31ad0 --- /dev/null +++ b/docker/configs/nginx-proxy.conf @@ -0,0 +1,35 @@ +server { + listen 8000 default_server; + listen [::]:8000 default_server; + + proxy_read_timeout 1d; + proxy_send_timeout 1d; + client_max_body_size 0; # disable checking + + root /var/www/html; + index index.html index.htm index.nginx-debian.html; + + server_name _; + + location /_static/ { + proxy_pass http://static/; + } + + location /pgadmin/ { + proxy_set_header X-Script-Name /pgadmin; + proxy_set_header Host $host; + proxy_pass http://pgadmin; + proxy_redirect off; + } + + location / { + error_page 502 /502.html; + proxy_pass http://localhost:8001/; + proxy_set_header Host localhost:8000; + } + + location /502.html { + root /var/www/html; + internal; + } +} diff --git a/docker/configs/pgadmin-servers.json b/docker/configs/pgadmin-servers.json new file mode 100644 index 0000000000..b4458af923 --- /dev/null +++ b/docker/configs/pgadmin-servers.json @@ -0,0 +1,22 @@ +{ + "Servers": { + "1": { + "Name": "Local Dev", + "Group": "Servers", + "Host": "db", + "Port": 5432, + "MaintenanceDB": "postgres", + "Username": "django", + "UseSSHTunnel": 0, + "TunnelPort": "22", + "TunnelAuthentication": 0, + "KerberosAuthentication": false, + "ConnectionParameters": { + "sslmode": "prefer", + "connect_timeout": 10, + "sslcert": "/.postgresql/postgresql.crt", + "sslkey": "/.postgresql/postgresql.key" + } + } + } +} diff --git a/docker/configs/settings_local.py b/docker/configs/settings_local.py index afcb5b202e..94adc516a4 100644 --- a/docker/configs/settings_local.py +++ b/docker/configs/settings_local.py @@ -1,35 +1,30 @@ -# Copyright The IETF Trust 2007-2019, All Rights Reserved +# Copyright The IETF Trust 2007-2025, All Rights Reserved # -*- coding: utf-8 -*- -from ietf.settings import * # pyflakes:ignore +from ietf.settings import * # pyflakes:ignore +from ietf.settings import ( + ARTIFACT_STORAGE_NAMES, + STORAGES, + BLOBSTORAGE_MAX_ATTEMPTS, + BLOBSTORAGE_READ_TIMEOUT, + BLOBSTORAGE_CONNECT_TIMEOUT, +) ALLOWED_HOSTS = ['*'] -DATABASES = { - 'default': { - 'HOST': 'db', - 'PORT': 3306, - 'NAME': 'ietf_utf8', - 'ENGINE': 'django.db.backends.mysql', - 'USER': 'django', - 'PASSWORD': 'RkTkDPFnKpko', - 'OPTIONS': { - 'sql_mode': 'STRICT_TRANS_TABLES', - 'init_command': 'SET storage_engine=InnoDB; SET names "utf8"', - }, - }, -} - -DATABASE_TEST_OPTIONS = { - 'init_command': 'SET storage_engine=InnoDB', +from ietf.settings_postgresqldb import DATABASES # pyflakes:ignore +DATABASE_ROUTERS = ["ietf.blobdb.routers.BlobdbStorageRouter"] +BLOBDB_DATABASE = "blobdb" +BLOBDB_REPLICATION = { + "ENABLED": True, + "DEST_STORAGE_PATTERN": "r2-{bucket}", + "INCLUDE_BUCKETS": ARTIFACT_STORAGE_NAMES, + "EXCLUDE_BUCKETS": ["staging"], + "VERBOSE_LOGGING": True, } IDSUBMIT_IDNITS_BINARY = "/usr/local/bin/idnits" -IDSUBMIT_REPOSITORY_PATH = "test/id/" -IDSUBMIT_STAGING_PATH = "test/staging/" -INTERNET_DRAFT_ARCHIVE_DIR = "test/archive/" -INTERNET_ALL_DRAFTS_ARCHIVE_DIR = "test/archive/" -RFC_PATH = "test/rfc/" +IDSUBMIT_STAGING_PATH = "/assets/www6s/staging/" AGENDA_PATH = '/assets/www6s/proceedings/' MEETINGHOST_LOGO_PATH = AGENDA_PATH @@ -47,7 +42,6 @@ SUBMIT_YANG_CATALOG_MODEL_DIR = '/assets/ietf-ftp/yang/catalogmod/' SUBMIT_YANG_DRAFT_MODEL_DIR = '/assets/ietf-ftp/yang/draftmod/' -SUBMIT_YANG_INVAL_MODEL_DIR = '/assets/ietf-ftp/yang/invalmod/' SUBMIT_YANG_IANA_MODEL_DIR = '/assets/ietf-ftp/yang/ianamod/' SUBMIT_YANG_RFC_MODEL_DIR = '/assets/ietf-ftp/yang/rfcmod/' @@ -60,17 +54,69 @@ # 'ietf.context_processors.sql_debug', # ] -DOCUMENT_PATH_PATTERN = '/assets/ietf-ftp/{doc.type_id}/' +DOCUMENT_PATH_PATTERN = '/assets/ietfdata/doc/{doc.type_id}/' INTERNET_DRAFT_PATH = '/assets/ietf-ftp/internet-drafts/' RFC_PATH = '/assets/ietf-ftp/rfc/' CHARTER_PATH = '/assets/ietf-ftp/charter/' BOFREQ_PATH = '/assets/ietf-ftp/bofreq/' CONFLICT_REVIEW_PATH = '/assets/ietf-ftp/conflict-reviews/' STATUS_CHANGE_PATH = '/assets/ietf-ftp/status-changes/' -INTERNET_DRAFT_ARCHIVE_DIR = '/assets/ietf-ftp/internet-drafts/' -INTERNET_ALL_DRAFTS_ARCHIVE_DIR = '/assets/ietf-ftp/internet-drafts/' +INTERNET_DRAFT_ARCHIVE_DIR = '/assets/collection/draft-archive' +INTERNET_ALL_DRAFTS_ARCHIVE_DIR = '/assets/archive/id' +BIBXML_BASE_PATH = '/assets/ietfdata/derived/bibxml' +IDSUBMIT_REPOSITORY_PATH = INTERNET_DRAFT_PATH +FTP_DIR = '/assets/ftp' +NFS_METRICS_TMP_DIR = '/assets/tmp' NOMCOM_PUBLIC_KEYS_DIR = 'data/nomcom_keys/public_keys/' -SLIDE_STAGING_PATH = 'test/staging/' +SLIDE_STAGING_PATH = '/assets/www6s/staging/' DE_GFM_BINARY = '/usr/local/bin/de-gfm' + +STATIC_IETF_ORG = "/_static" +STATIC_IETF_ORG_INTERNAL = "http://static" + + +# Blob replication storage for dev +import botocore.config +for storagename in ARTIFACT_STORAGE_NAMES: + replica_storagename = f"r2-{storagename}" + STORAGES[replica_storagename] = { + "BACKEND": "ietf.doc.storage.MetadataS3Storage", + "OPTIONS": dict( + endpoint_url="http://blobstore:9000", + access_key="minio_root", + secret_key="minio_pass", + security_token=None, + client_config=botocore.config.Config( + request_checksum_calculation="when_required", + response_checksum_validation="when_required", + signature_version="s3v4", + connect_timeout=BLOBSTORAGE_CONNECT_TIMEOUT, + read_timeout=BLOBSTORAGE_READ_TIMEOUT, + retries={"total_max_attempts": BLOBSTORAGE_MAX_ATTEMPTS}, + ), + verify=False, + bucket_name=f"{storagename}", + ), + } + +# For dev on rfc-index generation, create a red_bucket/ directory in the project root +# and uncomment these settings. Generated files will appear in this directory. To +# generate an accurate index, put up-to-date copies of unusable-rfc-numbers.json, +# april-first-rfc-numbers.json, and publication-std-levels.json in this directory +# before generating the index. +# +# STORAGES["red_bucket"] = { +# "BACKEND": "django.core.files.storage.FileSystemStorage", +# "OPTIONS": {"location": "red_bucket"}, +# } + +APP_API_TOKENS = { + "ietf.api.red_api" : ["devtoken", "redtoken"], # Not a real secret + "ietf.api.views_rpc" : ["devtoken"], # Not a real secret +} + +# Errata system api configuration +ERRATA_METADATA_NOTIFICATION_URL = "http://host.docker.internal:8808/api/rfc_metadata_update/" +ERRATA_METADATA_NOTIFICATION_API_KEY = "not a real secret" diff --git a/docker/configs/settings_local_sqlitetest.py b/docker/configs/settings_local_sqlitetest.py deleted file mode 100644 index 268fe6ec1d..0000000000 --- a/docker/configs/settings_local_sqlitetest.py +++ /dev/null @@ -1,81 +0,0 @@ -# Copyright The IETF Trust 2010-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -# Standard settings except we use SQLite and skip migrations, this is -# useful for speeding up tests that depend on the test database, try -# for instance: -# -# ./manage.py test --settings=settings_sqlitetest doc.ChangeStateTestCase -# - -import os -from ietf.settings import * # pyflakes:ignore -from ietf.settings import TEST_CODE_COVERAGE_CHECKER -import debug # pyflakes:ignore -debug.debug = True - -# Workaround to avoid spending minutes stepping through the migrations in -# every test run. The result of this is to use the 'syncdb' way of creating -# the test database instead of doing it through the migrations. Taken from -# https://gist.github.com/NotSqrt/5f3c76cd15e40ef62d09 - -class DisableMigrations(object): - - def __contains__(self, item): - return True - - def __getitem__(self, item): - return None - -MIGRATION_MODULES = DisableMigrations() - -DATABASES = { - 'default': { - 'NAME': 'test.db', - 'ENGINE': 'django.db.backends.sqlite3', - }, - } - -if TEST_CODE_COVERAGE_CHECKER and not TEST_CODE_COVERAGE_CHECKER._started: # pyflakes:ignore - TEST_CODE_COVERAGE_CHECKER.start() # pyflakes:ignore - -NOMCOM_PUBLIC_KEYS_DIR=os.path.abspath("tmp-nomcom-public-keys-dir") - -# Undo any developer-dependent middleware when running the tests -MIDDLEWARE = [ c for c in MIDDLEWARE if not c in DEV_MIDDLEWARE ] # pyflakes:ignore - -TEMPLATES[0]['OPTIONS']['context_processors'] = [ p for p in TEMPLATES[0]['OPTIONS']['context_processors'] if not p in DEV_TEMPLATE_CONTEXT_PROCESSORS ] # pyflakes:ignore - -REQUEST_PROFILE_STORE_ANONYMOUS_SESSIONS = False - -IDSUBMIT_IDNITS_BINARY = "/usr/local/bin/idnits" -IDSUBMIT_REPOSITORY_PATH = "test/id/" -IDSUBMIT_STAGING_PATH = "test/staging/" -INTERNET_DRAFT_ARCHIVE_DIR = "test/archive/" -INTERNET_ALL_DRAFTS_ARCHIVE_DIR = "test/archive/" -RFC_PATH = "test/rfc/" - -AGENDA_PATH = '/assets/www6s/proceedings/' -MEETINGHOST_LOGO_PATH = AGENDA_PATH - -USING_DEBUG_EMAIL_SERVER=True -EMAIL_HOST='localhost' -EMAIL_PORT=2025 - -MEDIA_BASE_DIR = 'test' -MEDIA_ROOT = MEDIA_BASE_DIR + '/media/' -MEDIA_URL = '/media/' - -PHOTOS_DIRNAME = 'photo' -PHOTOS_DIR = MEDIA_ROOT + PHOTOS_DIRNAME - -DOCUMENT_PATH_PATTERN = '/assets/ietf-ftp/{doc.type_id}/' - -SUBMIT_YANG_CATALOG_MODEL_DIR = '/assets/ietf-ftp/yang/catalogmod/' -SUBMIT_YANG_DRAFT_MODEL_DIR = '/assets/ietf-ftp/yang/draftmod/' -SUBMIT_YANG_INVAL_MODEL_DIR = '/assets/ietf-ftp/yang/invalmod/' -SUBMIT_YANG_IANA_MODEL_DIR = '/assets/ietf-ftp/yang/ianamod/' -SUBMIT_YANG_RFC_MODEL_DIR = '/assets/ietf-ftp/yang/rfcmod/' - -DE_GFM_BINARY = '/usr/local/bin/de-gfm' diff --git a/docker/configs/settings_local_vite.py b/docker/configs/settings_local_vite.py index 7fb12a003d..9116905b12 100644 --- a/docker/configs/settings_local_vite.py +++ b/docker/configs/settings_local_vite.py @@ -2,5 +2,9 @@ # -*- coding: utf-8 -*- from ietf.settings_local import * # pyflakes:ignore +from ietf.settings_local import DJANGO_VITE -DJANGO_VITE_DEV_MODE = True +DJANGO_VITE["default"] |= { + "dev_mode": True, + "dev_server_port": 3000, +} diff --git a/docker/configs/settings_postgresqldb.py b/docker/configs/settings_postgresqldb.py new file mode 100644 index 0000000000..9b98586658 --- /dev/null +++ b/docker/configs/settings_postgresqldb.py @@ -0,0 +1,18 @@ +DATABASES = { + 'default': { + 'HOST': 'db', + 'PORT': 5432, + 'NAME': 'datatracker', + 'ENGINE': 'django.db.backends.postgresql', + 'USER': 'django', + 'PASSWORD': 'RkTkDPFnKpko', + }, + 'blobdb': { + 'HOST': 'blobdb', + 'PORT': 5432, + 'NAME': 'blob', + 'ENGINE': 'django.db.backends.postgresql', + 'USER': 'dt', + 'PASSWORD': 'abcd1234', + }, +} diff --git a/docker/db.Dockerfile b/docker/db.Dockerfile index 5f55401c76..48ab298780 100644 --- a/docker/db.Dockerfile +++ b/docker/db.Dockerfile @@ -1,32 +1,37 @@ # ===================== # --- Builder Stage --- # ===================== -FROM mariadb:10 AS builder +FROM postgres:17 AS builder -# That file does the DB initialization but also runs mysql daemon, by removing the last line it will only init -RUN ["sed", "-i", "s/exec \"$@\"/echo \"not running $@\"/", "/usr/local/bin/docker-entrypoint.sh"] +ENV POSTGRES_PASSWORD=hk2j22sfiv +ENV POSTGRES_USER=django +ENV POSTGRES_DB=datatracker +ENV POSTGRES_HOST_AUTH_METHOD=trust +ENV PGDATA=/data -# needed for intialization -ENV MARIADB_ROOT_PASSWORD=RkTkDPFnKpko -ENV MARIADB_DATABASE=ietf_utf8 -ENV MARIADB_USER=django -ENV MARIADB_PASSWORD=RkTkDPFnKpko +COPY docker/scripts/db-load-default-extensions.sh /docker-entrypoint-initdb.d/ +COPY docker/scripts/db-import.sh /docker-entrypoint-initdb.d/ +COPY datatracker.dump / -# Import the latest database dump -ADD https://www.ietf.org/lib/dt/sprint/ietf_utf8.sql.gz /docker-entrypoint-initdb.d/ -RUN chmod 0777 /docker-entrypoint-initdb.d/ietf_utf8.sql.gz - -# Need to change the datadir to something else that /var/lib/mysql because the parent docker file defines it as a volume. -# https://docs.docker.com/engine/reference/builder/#volume : -# Changing the volume from within the Dockerfile: If any build steps change the data within the volume after -# it has been declared, those changes will be discarded. -RUN ["/usr/local/bin/docker-entrypoint.sh", "mysqld", "--datadir", "/initialized-db", "--aria-log-dir-path", "/initialized-db"] +RUN ["sed", "-i", "s/exec \"$@\"/echo \"skipping...\"/", "/usr/local/bin/docker-entrypoint.sh"] +RUN ["/usr/local/bin/docker-entrypoint.sh", "postgres"] # =================== # --- Final Image --- # =================== -FROM mariadb:10 +FROM postgres:17 LABEL maintainer="IETF Tools Team " -# Copy the mysql data folder from the builder stage -COPY --from=builder /initialized-db /var/lib/mysql +COPY --from=builder /data $PGDATA + +ENV POSTGRES_PASSWORD=hk2j22sfiv +ENV POSTGRES_USER=django +ENV POSTGRES_DB=datatracker +ENV POSTGRES_HOST_AUTH_METHOD=trust + +# build-args for db dump tagging - exposed in the environment and +# in image metadata +ARG datatracker_dumpinfo_date="" +ENV DATATRACKER_DUMPINFO_DATE=$datatracker_dumpinfo_date +ARG datatracker_snapshot="" +ENV DATATRACKER_SNAPSHOT=$datatracker_snapshot diff --git a/docker/devblobstore.Dockerfile b/docker/devblobstore.Dockerfile new file mode 100644 index 0000000000..40bfbd0e96 --- /dev/null +++ b/docker/devblobstore.Dockerfile @@ -0,0 +1,9 @@ +ARG MINIO_VERSION=latest +FROM quay.io/minio/minio:${MINIO_VERSION} +LABEL maintainer="IETF Tools Team " + +ENV MINIO_ROOT_USER=minio_root +ENV MINIO_ROOT_PASSWORD=minio_pass +ENV MINIO_DEFAULT_BUCKETS=defaultbucket + +CMD ["server", "--console-address", ":9001", "/data"] diff --git a/docker/docker-compose.celery.yml b/docker/docker-compose.celery.yml deleted file mode 100644 index dedae2d004..0000000000 --- a/docker/docker-compose.celery.yml +++ /dev/null @@ -1,51 +0,0 @@ -version: '2.4' -# Use version 2.4 for mem_limit setting. Version 3+ uses deploy.resources.limits.memory -# instead, but that only works for swarm with docker-compose 1.25.1. - -services: - mq: - image: rabbitmq:3-alpine - user: '${RABBITMQ_UID:-499:499}' - hostname: datatracker-mq -# deploy: -# resources: -# limits: -# memory: 1gb # coordinate with settings in rabbitmq.conf -# reservations: -# memory: 512mb - mem_limit: 1gb # coordinate with settings in rabbitmq.conf - ports: - - '${MQ_PORT:-5672}:5672' - volumes: - - ./lib.rabbitmq:/var/lib/rabbitmq - - ./rabbitmq.conf:/etc/rabbitmq/conf.d/90-ietf.conf - - ./definitions.json:/ietf-conf/definitions.json - restart: unless-stopped - logging: - driver: "syslog" - options: - syslog-address: 'unixgram:///dev/log' - tag: 'docker/{{.Name}}' -# syslog-address: "tcp://ietfa.amsl.com:514" - - celery: - image: ghcr.io/ietf-tools/datatracker-celery:latest - environment: - CELERY_APP: ietf - # UPDATE_REQUIREMENTS: 1 # uncomment to update Python requirements on startup - command: - - '--loglevel=INFO' - user: '${CELERY_UID:-499:499}' - volumes: - - '${DATATRACKER_PATH:-..}:/workspace' - - '${MYSQL_SOCKET_PATH:-/run/mysql}:/run/mysql' - depends_on: - - mq - network_mode: 'service:mq' - restart: unless-stopped - logging: - driver: "syslog" - options: - syslog-address: 'unixgram:///dev/log' - tag: 'docker/{{.Name}}' -# syslog-address: "tcp://ietfa.amsl.com:514" diff --git a/docker/docker-compose.extend.yml b/docker/docker-compose.extend.yml index 06e47bbb08..12ebe447d5 100644 --- a/docker/docker-compose.extend.yml +++ b/docker/docker-compose.extend.yml @@ -1,5 +1,3 @@ -version: '3.8' - services: app: ports: @@ -14,7 +12,14 @@ services: - app-assets:/assets db: ports: - - '3306' + - '5432' + pgadmin: + ports: + - '5433' + blobstore: + ports: + - '9000:9000' + - '9001:9001' celery: volumes: - .:/workspace diff --git a/docker/misc/build b/docker/misc/build deleted file mode 100644 index b3cc97e52c..0000000000 --- a/docker/misc/build +++ /dev/null @@ -1,96 +0,0 @@ -#!/bin/bash - -version=0.20 -program=${0##*/} -progdir=${0%/*} -if [ "$progdir" = "$program" ]; then progdir="."; fi -if [ "$progdir" = "." ]; then progdir="$PWD"; fi -parent=$(dirname "$progdir") -if [ "$parent" = "." ]; then parent="$PWD"; fi -if [[ $(uname) =~ CYGWIN.* ]]; then parent=$(echo "$parent" | sed -e 's/^\/cygdrive\/\(.\)/\1:/'); fi - - -function usage() { - cat < - Lars Eggert, - -COPYRIGHT - Copyright (c) 2016 IETF Trust and the persons identified as authors of - the code. All rights reserved. Redistribution and use in source and - binary forms, with or without modification, is permitted pursuant to, - and subject to the license terms contained in, the Revised BSD - License set forth in Section 4.c of the IETF Trust’s Legal Provisions - Relating to IETF Documents(https://trustee.ietf.org/license-info). - -EOF -} - - -function die() { - echo -e "\n$program: error: $*" >&2 - exit 1 -} - - -function version() { - echo -e "$program $version" -} - - -trap 'echo "$program($LINENO): Command failed with error code $? ([$$] $0 $*)"; exit 1' ERR - -# Default values -IMAGE=ietf/datatracker-environment -TAG=$(basename "$(svn info "$parent" | grep ^URL | awk '{print $2}' | tr -d '\r')") -LOCAL=1 - -# Option parsing -shortopts=hut:V -args=$(getopt -o$shortopts $*) -if [ $? != 0 ] ; then die "Terminating..." >&2 ; exit 1 ; fi -set -- $args - -while true ; do - case "$1" in - -h) usage; exit;; # Show this help, then exit - -u) LOCAL=0;; # Upload image to repository after build - -t) TAG=$2; shift;; # Use this docker image tag - -V) version; exit;; # Show program version, then exit - --) shift; break;; - *) die "Internal error, inconsistent option specification: '$1'";; - esac - shift -done - -# The program itself -docker rmi -f $IMAGE:trunk 2>/dev/null || true -docker build --progress plain -t "$IMAGE-app:$TAG" -f docker/app.Dockerfile . -docker build --progress plain -t "$IMAGE-db:$TAG" -f docker/db.Dockerfile . -docker tag "$(docker images -q $IMAGE-app | head -n 1)" $IMAGE-app:latest -docker tag "$(docker images -q $IMAGE-db | head -n 1)" $IMAGE-db:latest -if [ -z "$LOCAL" ]; then - docker push $IMAGE-app:latest - docker push "$IMAGE-app:$TAG" - docker push $IMAGE-db:latest - docker push "$IMAGE-db:$TAG" -fi \ No newline at end of file diff --git a/docker/misc/copydb b/docker/misc/copydb deleted file mode 100644 index 748189bd6a..0000000000 --- a/docker/misc/copydb +++ /dev/null @@ -1,97 +0,0 @@ -#!/bin/bash - -version=0.11 -program=${0##*/} -progdir=${0%/*} -if [ "$progdir" = "$program" ]; then progdir="."; fi -if [ "$progdir" = "." ]; then progdir="$PWD"; fi -parent=$(dirname "$progdir") -if [ "$parent" = "." ]; then parent="$PWD"; fi -if [[ $(uname) =~ CYGWIN.* ]]; then parent=$(echo "$parent" | sed -e 's/^\/cygdrive\/\(.\)/\1:/'); fi - - -function usage() { - cat < - Lars Eggert, - -COPYRIGHT - Copyright (c) 2016 IETF Trust and the persons identified as authors of - the code. All rights reserved. Redistribution and use in source and - binary forms, with or without modification, is permitted pursuant to, - and subject to the license terms contained in, the Revised BSD - License set forth in Section 4.c of the IETF Trust’s Legal Provisions - Relating to IETF Documents(https://trustee.ietf.org/license-info). - -EOF -} - - -function die() { - echo -e "\n$program: error: $*" >&2 - exit 1 -} - - -function version() { - echo -e "$program $version" -} - - -trap 'echo "$program($LINENO): Command failed with error code $? ([$$] $0 $*)"; exit 1' ERR - -# Option parsing -shortopts=hV -args=$(getopt -o$shortopts $*) -if [ $? != 0 ] ; then die "Terminating..." >&2 ; exit 1 ; fi -set -- $args - -while true ; do - case "$1" in - -h) usage; exit;; # Show this help, then exit - -V) version; exit;; # Show program version, then exit - --) shift; break;; - *) die "Internal error, inconsistent option specification: '$1'";; - esac - shift -done - -# The program itself -if [ -e "/.dockerenv" -o -n "$(grep -s '/docker/' /proc/self/cgroup)" ]; then - die "It looks as if you're running inside docker -- please quit docker first." -fi - -workdir=$(realpath $progdir/../data/mysql/..) -echo "Working directory: $workdir" -cd $workdir -echo "Building tarfile ..." -tar cjf ietf_utf8.bin.tar.bz2 mysql -echo "Copying tarfile to ietfa.amsl.com ..." -scp ietf_utf8.bin.tar.bz2 ietfa.amsl.com:/a/www/www6s/lib/dt/sprint/ \ No newline at end of file diff --git a/docker/run b/docker/run index cc0a19941f..71e794aaa9 100755 --- a/docker/run +++ b/docker/run @@ -37,12 +37,22 @@ while getopts "hp:r" opt; do esac done +# Ensure the run script is in our directory. +dirname=$(dirname $0) +if [ "$dirname" = "." -o "$dirname" = "" ] ; then + : +else + echo "Changing to directory $dirname" + cd $dirname || exit 1 +fi + # Remove mounted temp directories rm -rf .parcel-cache __pycache__ # Create extended docker-compose definition -cp docker-compose.extend.yml docker-compose.extend-custom.yml -sed -i -r -e "s/CUSTOM_PORT/$CUSTOM_PORT/" docker-compose.extend-custom.yml +sed -e "s/CUSTOM_PORT/$CUSTOM_PORT/" \ + docker-compose.extend-custom.yml cd .. # Set UID/GID mappings diff --git a/docker/scripts/app-configure-blobstore.py b/docker/scripts/app-configure-blobstore.py new file mode 100755 index 0000000000..9ae64e0041 --- /dev/null +++ b/docker/scripts/app-configure-blobstore.py @@ -0,0 +1,44 @@ +#!/usr/bin/env python +# Copyright The IETF Trust 2024, All Rights Reserved + +import boto3 +import botocore.config +import botocore.exceptions +import os +import sys + +from ietf.settings import ARTIFACT_STORAGE_NAMES + + +def init_blobstore(): + blobstore = boto3.resource( + "s3", + endpoint_url=os.environ.get("BLOB_STORE_ENDPOINT_URL", "http://blobstore:9000"), + aws_access_key_id=os.environ.get("BLOB_STORE_ACCESS_KEY", "minio_root"), + aws_secret_access_key=os.environ.get("BLOB_STORE_SECRET_KEY", "minio_pass"), + aws_session_token=None, + config=botocore.config.Config( + request_checksum_calculation="when_required", + response_checksum_validation="when_required", + signature_version="s3v4", + ), + ) + for bucketname in ARTIFACT_STORAGE_NAMES: + adjusted_bucket_name = ( + os.environ.get("BLOB_STORE_BUCKET_PREFIX", "") + + bucketname + + os.environ.get("BLOB_STORE_BUCKET_SUFFIX", "") + ).strip() + try: + blobstore.create_bucket(Bucket=adjusted_bucket_name) + except botocore.exceptions.ClientError as err: + if err.response["Error"]["Code"] == "BucketAlreadyExists": + print(f"Bucket {bucketname} already exists") + else: + print(f"Error creating {bucketname}: {err.response['Error']['Code']}") + else: + print(f"Bucket {bucketname} created") + + +if __name__ == "__main__": + sys.exit(init_blobstore()) diff --git a/docker/scripts/app-create-dirs.sh b/docker/scripts/app-create-dirs.sh index 807522e25b..3eb328a280 100755 --- a/docker/scripts/app-create-dirs.sh +++ b/docker/scripts/app-create-dirs.sh @@ -1,13 +1,9 @@ #!/bin/bash for sub in \ - test/id \ - test/staging \ - test/archive \ - test/rfc \ - test/media \ - test/wiki/ietf \ - data/nomcom_keys/public_keys \ + /assets/archive/id \ + /assets/collection \ + /assets/collection/draft-archive \ /assets/ietf-ftp \ /assets/ietf-ftp/bofreq \ /assets/ietf-ftp/charter \ @@ -20,6 +16,11 @@ for sub in \ /assets/ietf-ftp/yang/ianamod \ /assets/ietf-ftp/yang/invalmod \ /assets/ietf-ftp/yang/rfcmod \ + /assets/ietfdata \ + /assets/ietfdata/derived \ + /assets/ietfdata/derived/bibxml \ + /assets/ietfdata/derived/bibxml/bibxml-ids \ + /assets/ietfdata/doc/draft/repository \ /assets/www6s \ /assets/www6s/staging \ /assets/www6s/wg-descriptions \ @@ -28,6 +29,11 @@ for sub in \ /assets/www6/iesg \ /assets/www6/iesg/evaluation \ /assets/media/photo \ + /assets/tmp \ + /assets/ftp \ + /assets/ftp/charter \ + /assets/ftp/internet-drafts \ + /assets/ftp/review \ ; do if [ ! -d "$sub" ]; then echo "Creating dir $sub" diff --git a/docker/scripts/app-cypress.sh b/docker/scripts/app-cypress.sh deleted file mode 100755 index 81e510592b..0000000000 --- a/docker/scripts/app-cypress.sh +++ /dev/null @@ -1,23 +0,0 @@ -#!/bin/bash - -WORKSPACEDIR="/workspace" - -pushd . -cd $WORKSPACEDIR - -echo "Starting datatracker server..." -ietf/manage.py runserver 0.0.0.0:8000 --settings=settings_local > /dev/null 2>&1 & -serverPID=$! - -echo "Waiting for server to come online ..." -/usr/local/bin/wait-for localhost:8000 -- echo "Server ready" - -echo "Run dbus process to silence warnings..." -sudo mkdir -p /run/dbus -sudo dbus-daemon --system &> /dev/null - -echo "Starting JS tests..." -yarn cypress - -kill $serverPID -popd diff --git a/docker/scripts/app-init-celery.sh b/docker/scripts/app-init-celery.sh new file mode 100755 index 0000000000..17925633d2 --- /dev/null +++ b/docker/scripts/app-init-celery.sh @@ -0,0 +1,119 @@ +#!/bin/bash -e +# +# Environment parameters: +# +# CELERY_APP - name of application to pass to celery (defaults to ietf) +# +# CELERY_ROLE - 'worker' or 'beat' (defaults to 'worker') +# +# CELERY_UID - numeric uid for the celery worker process +# +# CELERY_GID - numeric gid for the celery worker process +# +# UPDATES_REQUIREMENTS_FROM - path, relative to /workspace mount, to a pip requirements +# file that should be installed at container startup. Default is no package install/update. +# +# DEBUG_TERM_TIMING - if non-empty, writes debug messages during shutdown after a TERM signal +# +# DEV_MODE - if non-empty, restart celery worker on Python file change +# +WORKSPACEDIR="/workspace" +CELERY_ROLE="${CELERY_ROLE:-worker}" + +cd "$WORKSPACEDIR" || exit 255 + +if [[ -n "${UPDATE_REQUIREMENTS_FROM}" ]]; then + # Need to run as root in the container for this + reqs_file="${WORKSPACEDIR}/${UPDATE_REQUIREMENTS_FROM}" + echo "Updating requirements from ${reqs_file}..." + pip install --upgrade -r "${reqs_file}" +fi + +CELERY_OPTS=( "${CELERY_ROLE}" ) +if [[ -n "${CELERY_UID}" ]]; then + # ensure that a user with the necessary UID exists in container + if ! id "${CELERY_UID}" ; then + adduser --system --uid "${CELERY_UID}" --no-create-home --disabled-login "celery-user-${CELERY_UID}" + fi + CELERY_OPTS+=("--uid=${CELERY_UID}") + CELERY_USERNAME="$(id -nu ${CELERY_UID})" +fi + +if [[ -n "${CELERY_GID}" ]]; then + # ensure that some group with the necessary GID exists in container + if ! getent group "${CELERY_GID}" ; then + addgroup --gid "${CELERY_GID}" "celery-group-${CELERY_GID}" + fi + CELERY_OPTS+=("--gid=${CELERY_GID}") + CELERY_GROUP="$(getent group ${CELERY_GID} | awk -F: '{print $1}')" +fi + +run_as_celery_uid () { + IAM=$(whoami) + if [ "${IAM}" = "${CELERY_USERNAME:-root}" ]; then + SU_OPTS=() + if [[ -n "${CELERY_GROUP}" ]]; then + SU_OPTS+=("-g" "${CELERY_GROUP}") + fi + su "${SU_OPTS[@]}" "${CELERY_USERNAME:-root}" -s /bin/sh -c "$*" + else + /bin/sh -c "$*" + fi +} + +log_term_timing_msgs () { + # output periodic debug message + while true; do + echo "Waiting for celery worker shutdown ($(date --utc --iso-8601=ns))" + sleep 0.5s + done +} + +cleanup () { + # Cleanly terminate the celery app by sending it a TERM, then waiting for it to exit. + if [[ -n "${celery_pid}" ]]; then + echo "Gracefully terminating celery worker. This may take a few minutes if tasks are in progress..." + kill -TERM "${celery_pid}" + if [[ -n "${DEBUG_TERM_TIMING}" ]]; then + log_term_timing_msgs & + fi + wait "${celery_pid}" + fi +} + +echo "Running checks as root to apply patches..." +/usr/local/bin/python $WORKSPACEDIR/ietf/manage.py check + +if [[ "${CELERY_ROLE}" == "worker" ]]; then + echo "Running initial checks..." + # Run checks as celery worker if one was specified + run_as_celery_uid /usr/local/bin/python $WORKSPACEDIR/ietf/manage.py check +fi + +USER_BIN_PATH="/home/dev/.local/bin" +WATCHMEDO="$USER_BIN_PATH/watchmedo" +# Find a celery that works +if [[ -x "$USER_BIN_PATH/celery" ]]; then + # This branch is used for dev + CELERY="$USER_BIN_PATH/celery" +else + # This branch is used for sandbox instances + CELERY="/usr/local/bin/celery" +fi +trap 'trap "" TERM; cleanup' TERM +# start celery in the background so we can trap the TERM signal +if [[ -n "${DEV_MODE}" && -x "${WATCHMEDO}" ]]; then + $WATCHMEDO auto-restart \ + --patterns '*.py' \ + --directory 'ietf' \ + --recursive \ + --debounce-interval 5 \ + -- \ + $CELERY --app="${CELERY_APP:-ietf}" "${CELERY_OPTS[@]}" $@ & + celery_pid=$! +else + $CELERY --app="${CELERY_APP:-ietf}" "${CELERY_OPTS[@]}" "$@" & + celery_pid=$! +fi + +wait "${celery_pid}" diff --git a/docker/scripts/app-init.sh b/docker/scripts/app-init.sh index 6d06bb4429..1d895cdf53 100755 --- a/docker/scripts/app-init.sh +++ b/docker/scripts/app-init.sh @@ -2,8 +2,17 @@ WORKSPACEDIR="/workspace" +# Handle Linux host mounting the workspace dir as root +if [ ! -O "${WORKSPACEDIR}/ietf" ]; then + sudo chown -R dev:dev $WORKSPACEDIR +fi + +# Start rsyslog service sudo service rsyslog start &>/dev/null +# Add /workspace as a safe git directory +git config --global --add safe.directory /workspace + # Turn off git info in zsh prompt (causes slowdowns) git config oh-my-zsh.hide-info 1 @@ -15,63 +24,48 @@ sudo chown -R dev:dev "$WORKSPACEDIR/.vite" sudo chown -R dev:dev "$WORKSPACEDIR/.yarn/unplugged" sudo chown dev:dev "/assets" -echo "Fix chromedriver /dev/shm permissions..." -sudo chmod 1777 /dev/shm +# Run nginx +echo "Starting nginx..." +sudo nginx # Build node packages that requrie native compilation echo "Compiling native node packages..." yarn rebuild +# Silence Browserlist warnings +export BROWSERSLIST_IGNORE_OLD_DATA=1 + # Generate static assets echo "Building static assets... (this could take a minute or two)" yarn build yarn legacy:build # Copy config files if needed +cp $WORKSPACEDIR/docker/configs/settings_postgresqldb.py $WORKSPACEDIR/ietf/settings_postgresqldb.py if [ ! -f "$WORKSPACEDIR/ietf/settings_local.py" ]; then echo "Setting up a default settings_local.py ..." - cp $WORKSPACEDIR/docker/configs/settings_local.py $WORKSPACEDIR/ietf/settings_local.py else - echo "Using existing ietf/settings_local.py file" - if ! cmp -s $WORKSPACEDIR/docker/configs/settings_local.py $WORKSPACEDIR/ietf/settings_local.py; then - echo "NOTE: Differences detected compared to docker/configs/settings_local.py!" - echo "We'll assume you made these deliberately." - fi + echo "Renaming existing ietf/settings_local.py to ietf/settings_local.py.bak" + mv -f $WORKSPACEDIR/ietf/settings_local.py $WORKSPACEDIR/ietf/settings_local.py.bak fi +cp $WORKSPACEDIR/docker/configs/settings_local.py $WORKSPACEDIR/ietf/settings_local.py if [ ! -f "$WORKSPACEDIR/ietf/settings_local_debug.py" ]; then echo "Setting up a default settings_local_debug.py ..." - cp $WORKSPACEDIR/docker/configs/settings_local_debug.py $WORKSPACEDIR/ietf/settings_local_debug.py else - echo "Using existing ietf/settings_local_debug.py file" - if ! cmp -s $WORKSPACEDIR/docker/configs/settings_local_debug.py $WORKSPACEDIR/ietf/settings_local_debug.py; then - echo "NOTE: Differences detected compared to docker/configs/settings_local_debug.py!" - echo "We'll assume you made these deliberately." - fi -fi - -if [ ! -f "$WORKSPACEDIR/ietf/settings_local_sqlitetest.py" ]; then - echo "Setting up a default settings_local_sqlitetest.py ..." - cp $WORKSPACEDIR/docker/configs/settings_local_sqlitetest.py $WORKSPACEDIR/ietf/settings_local_sqlitetest.py -else - echo "Using existing ietf/settings_local_sqlitetest.py file" - if ! cmp -s $WORKSPACEDIR/docker/configs/settings_local_sqlitetest.py $WORKSPACEDIR/ietf/settings_local_sqlitetest.py; then - echo "NOTE: Differences detected compared to docker/configs/settings_local_sqlitetest.py!" - echo "We'll assume you made these deliberately." - fi + echo "Renaming existing ietf/settings_local_debug.py to ietf/settings_local_debug.py.bak" + mv -f $WORKSPACEDIR/ietf/settings_local_debug.py $WORKSPACEDIR/ietf/settings_local_debug.py.bak fi +cp $WORKSPACEDIR/docker/configs/settings_local_debug.py $WORKSPACEDIR/ietf/settings_local_debug.py if [ ! -f "$WORKSPACEDIR/ietf/settings_local_vite.py" ]; then echo "Setting up a default settings_local_vite.py ..." - cp $WORKSPACEDIR/docker/configs/settings_local_vite.py $WORKSPACEDIR/ietf/settings_local_vite.py else - echo "Using existing ietf/settings_local_vite.py file" - if ! cmp -s $WORKSPACEDIR/docker/configs/settings_local_vite.py $WORKSPACEDIR/ietf/settings_local_vite.py; then - echo "NOTE: Differences detected compared to docker/configs/settings_local_vite.py!" - echo "We'll assume you made these deliberately." - fi + echo "Renaming existing ietf/settings_local_vite.py to ietf/settings_local_vite.py.bak" + mv -f $WORKSPACEDIR/ietf/settings_local_vite.py $WORKSPACEDIR/ietf/settings_local_vite.py.bak fi +cp $WORKSPACEDIR/docker/configs/settings_local_vite.py $WORKSPACEDIR/ietf/settings_local_vite.py # Create data directories @@ -79,6 +73,11 @@ echo "Creating data directories..." chmod +x ./docker/scripts/app-create-dirs.sh ./docker/scripts/app-create-dirs.sh +# Configure the development blobstore + +echo "Configuring blobstore..." +PYTHONPATH=/workspace python ./docker/scripts/app-configure-blobstore.py + # Download latest coverage results file echo "Downloading latest coverage results file..." @@ -88,7 +87,7 @@ curl -fsSL https://github.com/ietf-tools/datatracker/releases/download/baseline/ if [ -n "$EDITOR_VSCODE" ]; then echo "Waiting for DB container to come online ..." - /usr/local/bin/wait-for localhost:3306 -- echo "DB ready" + /usr/local/bin/wait-for db:5432 -- echo "PostgreSQL ready" fi # Run memcached @@ -102,26 +101,26 @@ echo "Running initial checks..." /usr/local/bin/python $WORKSPACEDIR/ietf/manage.py check --settings=settings_local # Migrate, adjusting to what the current state of the underlying database might be: +/usr/local/bin/python $WORKSPACEDIR/ietf/manage.py migrate --fake-initial --settings=settings_local -/usr/local/bin/python $WORKSPACEDIR/ietf/manage.py migrate --settings=settings_local - - -echo "-----------------------------------------------------------------" -echo "Done!" -echo "-----------------------------------------------------------------" +# Apply migrations to the blobdb database as well (most are skipped) +/usr/local/bin/python $WORKSPACEDIR/ietf/manage.py migrate --settings=settings_local --database=blobdb if [ -z "$EDITOR_VSCODE" ]; then CODE=0 - python -m smtpd -n -c DebuggingServer localhost:2025 & + python -m aiosmtpd -n -c ietf.utils.aiosmtpd.DevDebuggingHandler -l localhost:2025 & if [ -z "$*" ]; then + echo "-----------------------------------------------------------------" + echo "Ready!" + echo "-----------------------------------------------------------------" echo echo "You can execute arbitrary commands now, e.g.," echo - echo " ietf/manage.py runserver 0.0.0.0:8000" + echo " ietf/manage.py runserver 8001" echo echo "to start a development instance of the Datatracker." echo - echo " ietf/manage.py test --settings=settings_sqlitetest" + echo " ietf/manage.py test --settings=settings_test" echo echo "to run all the python tests." echo diff --git a/docker/scripts/app-install-chromedriver.sh b/docker/scripts/app-install-chromedriver.sh deleted file mode 100755 index 43532a1cf6..0000000000 --- a/docker/scripts/app-install-chromedriver.sh +++ /dev/null @@ -1,18 +0,0 @@ -#!/bin/bash - -HOSTARCH=$(arch) -if [ $HOSTARCH == "x86_64" ]; then - echo "Installing chrome driver..." - wget -q -O - https://dl-ssl.google.com/linux/linux_signing_key.pub | apt-key add - - echo "deb http://dl.google.com/linux/chrome/deb/ stable main" >> /etc/apt/sources.list.d/google.list - apt-get update -y - apt-get install -y google-chrome-stable - CHROMEVER=$(google-chrome --product-version | grep -o "[^\.]*\.[^\.]*\.[^\.]*") - DRIVERVER=$(curl -s "https://chromedriver.storage.googleapis.com/LATEST_RELEASE_$CHROMEVER") - wget -q --continue -P /chromedriver "http://chromedriver.storage.googleapis.com/$DRIVERVER/chromedriver_linux64.zip" - unzip /chromedriver/chromedriver* -d /chromedriver - ln -s /chromedriver/chromedriver /usr/local/bin/chromedriver - ln -s /chromedriver/chromedriver /usr/bin/chromedriver -else - echo "This architecture doesn't support chromedriver. Skipping installation..." -fi \ No newline at end of file diff --git a/docker/scripts/app-setup-debian.sh b/docker/scripts/app-setup-debian.sh index ddfc351995..ea9cc3fb87 100644 --- a/docker/scripts/app-setup-debian.sh +++ b/docker/scripts/app-setup-debian.sh @@ -10,7 +10,6 @@ # Syntax: ./common-debian.sh [install zsh flag] [username] [user UID] [user GID] [upgrade packages flag] [install Oh My Zsh! flag] [Add non-free packages] set -e - INSTALL_ZSH=${1:-"true"} USERNAME=${2:-"automatic"} USER_UID=${3:-"automatic"} @@ -116,18 +115,9 @@ if [ "${PACKAGES_ALREADY_INSTALLED}" != "true" ]; then # Needed for adding manpages-posix and manpages-posix-dev which are non-free packages in Debian if [ "${ADD_NON_FREE_PACKAGES}" = "true" ]; then # Bring in variables from /etc/os-release like VERSION_CODENAME - . /etc/os-release - sed -i -E "s/deb http:\/\/(deb|httpredir)\.debian\.org\/debian ${VERSION_CODENAME} main/deb http:\/\/\1\.debian\.org\/debian ${VERSION_CODENAME} main contrib non-free/" /etc/apt/sources.list - sed -i -E "s/deb-src http:\/\/(deb|httredir)\.debian\.org\/debian ${VERSION_CODENAME} main/deb http:\/\/\1\.debian\.org\/debian ${VERSION_CODENAME} main contrib non-free/" /etc/apt/sources.list - sed -i -E "s/deb http:\/\/(deb|httpredir)\.debian\.org\/debian ${VERSION_CODENAME}-updates main/deb http:\/\/\1\.debian\.org\/debian ${VERSION_CODENAME}-updates main contrib non-free/" /etc/apt/sources.list - sed -i -E "s/deb-src http:\/\/(deb|httpredir)\.debian\.org\/debian ${VERSION_CODENAME}-updates main/deb http:\/\/\1\.debian\.org\/debian ${VERSION_CODENAME}-updates main contrib non-free/" /etc/apt/sources.list - sed -i "s/deb http:\/\/security\.debian\.org\/debian-security ${VERSION_CODENAME}\/updates main/deb http:\/\/security\.debian\.org\/debian-security ${VERSION_CODENAME}\/updates main contrib non-free/" /etc/apt/sources.list - sed -i "s/deb-src http:\/\/security\.debian\.org\/debian-security ${VERSION_CODENAME}\/updates main/deb http:\/\/security\.debian\.org\/debian-security ${VERSION_CODENAME}\/updates main contrib non-free/" /etc/apt/sources.list - sed -i "s/deb http:\/\/deb\.debian\.org\/debian ${VERSION_CODENAME}-backports main/deb http:\/\/deb\.debian\.org\/debian ${VERSION_CODENAME}-backports main contrib non-free/" /etc/apt/sources.list - sed -i "s/deb-src http:\/\/deb\.debian\.org\/debian ${VERSION_CODENAME}-backports main/deb http:\/\/deb\.debian\.org\/debian ${VERSION_CODENAME}-backports main contrib non-free/" /etc/apt/sources.list - # Handle bullseye location for security https://www.debian.org/releases/bullseye/amd64/release-notes/ch-information.en.html - sed -i "s/deb http:\/\/security\.debian\.org\/debian-security ${VERSION_CODENAME}-security main/deb http:\/\/security\.debian\.org\/debian-security ${VERSION_CODENAME}-security main contrib non-free/" /etc/apt/sources.list - sed -i "s/deb-src http:\/\/security\.debian\.org\/debian-security ${VERSION_CODENAME}-security main/deb http:\/\/security\.debian\.org\/debian-security ${VERSION_CODENAME}-security main contrib non-free/" /etc/apt/sources.list + . /etc/os-release + sed -i -E "s/Components: main/Components: main contrib non-free/" /etc/apt/sources.list.d/debian.sources + echo "Running apt-get update..." apt-get update package_list="${package_list} manpages-posix manpages-posix-dev" diff --git a/docker/scripts/app-setup-nginx.sh b/docker/scripts/app-setup-nginx.sh new file mode 100644 index 0000000000..cdeb2ea4cb --- /dev/null +++ b/docker/scripts/app-setup-nginx.sh @@ -0,0 +1,4 @@ +#!/usr/bin/env bash + +apt-get update -y +apt-get install -y nginx diff --git a/docker/scripts/app-start.sh b/docker/scripts/app-start.sh new file mode 100644 index 0000000000..c3369bab93 --- /dev/null +++ b/docker/scripts/app-start.sh @@ -0,0 +1,17 @@ +#!/bin/bash + +sudo service rsyslog start &>/dev/null + +# Run nginx + +echo "Starting nginx..." +pidof nginx >/dev/null && echo "nginx is already running [ OK ]" || sudo nginx + +# Run memcached + +echo "Starting memcached..." +pidof memcached >/dev/null && echo "memcached is already running [ OK ]" || /usr/bin/memcached -u dev -d + +echo "-----------------------------------------------------------------" +echo "Ready!" +echo "-----------------------------------------------------------------" diff --git a/docker/scripts/db-import.sh b/docker/scripts/db-import.sh new file mode 100644 index 0000000000..a0f22cd8fc --- /dev/null +++ b/docker/scripts/db-import.sh @@ -0,0 +1,12 @@ +#!/bin/bash +set -e + +echo "Drop dummy datatracker DB if it exists..." +dropdb -U django --if-exists datatracker + +# Extensions and search paths will be loaded from the dump +echo "Import DB dump into datatracker..." +pg_restore --clean --if-exists --create --no-owner -U django -d postgres datatracker.dump +echo "alter role django set search_path=datatracker,django,public;" | psql -U django -d datatracker + +echo "Done!" diff --git a/docker/scripts/db-load-default-extensions.sh b/docker/scripts/db-load-default-extensions.sh new file mode 100644 index 0000000000..efb64b75d0 --- /dev/null +++ b/docker/scripts/db-load-default-extensions.sh @@ -0,0 +1,10 @@ +#!/bin/bash +set -e + +# Adding the extension to the default template is needed to allow the test-suite +# to be run on postgres (see ietf.settings_test). The test runner always +# makes a fresh test database instance, and since we are bypassing the migration +# framework and using a fixture to set the database structure, there's no reaonable +# way to install the extension as part of the test run. +psql -U django -d template1 -v ON_ERROR_STOP=1 -c 'CREATE EXTENSION IF NOT EXISTS citext;' + diff --git a/docker/scripts/updatedb b/docker/scripts/updatedb deleted file mode 100644 index 85386daa4a..0000000000 --- a/docker/scripts/updatedb +++ /dev/null @@ -1,5 +0,0 @@ -#!/bin/bash - -echo "This script is deprecated. Please use the `cleandb` script in the parent folder instead." - -# Modified on 2021-12-20, remove this file after a while \ No newline at end of file diff --git a/ietf/.gitignore b/ietf/.gitignore index 322cc214d0..af738db98c 100644 --- a/ietf/.gitignore +++ b/ietf/.gitignore @@ -1,7 +1,8 @@ /*.pyc /settings_local.py -/settings_local.py.bak /settings_local_debug.py -/settings_local_sqlitetest.py /settings_local_vite.py +/settings_mysqldb.py +/settings_postgresqldb.py +/settings_*.py.bak /ietfdb.sql.gz diff --git a/ietf/__init__.py b/ietf/__init__.py index 2338d0428b..26124c3c67 100644 --- a/ietf/__init__.py +++ b/ietf/__init__.py @@ -6,7 +6,7 @@ # Version must stay in single quotes for automatic CI replace # Don't add patch number here: -__version__ = '8.0.0-dev' +__version__ = '1.0.0-dev' # Release hash must stay in single quotes for automatic CI replace __release_hash__ = '' @@ -17,6 +17,24 @@ # set this to ".p1", ".p2", etc. after patching __patch__ = "" +if __version__ == '1.0.0-dev' and __release_hash__ == '' and __release_branch__ == '': + import subprocess + branch = subprocess.run( + ["/usr/bin/git", "branch", "--show-current"], + capture_output=True, + ).stdout.decode().strip() + git_hash = subprocess.run( + ["/usr/bin/git", "rev-parse", "head"], + capture_output=True, + ).stdout.decode().strip() + rev = subprocess.run( + ["/usr/bin/git", "describe", "--tags", git_hash], + capture_output=True, + ).stdout.decode().strip().split('-', 1)[0] + __version__ = f"{rev}-dev" + __release_branch__ = branch + __release_hash__ = git_hash + # This will make sure the app is always imported when # Django starts so that shared_task will use this app. diff --git a/ietf/ietfauth/management/__init__.py b/ietf/admin/__init__.py similarity index 100% rename from ietf/ietfauth/management/__init__.py rename to ietf/admin/__init__.py diff --git a/ietf/admin/apps.py b/ietf/admin/apps.py new file mode 100644 index 0000000000..20b762cfec --- /dev/null +++ b/ietf/admin/apps.py @@ -0,0 +1,6 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +from django.contrib.admin import apps as admin_apps + + +class AdminConfig(admin_apps.AdminConfig): + default_site = "ietf.admin.sites.AdminSite" diff --git a/ietf/admin/sites.py b/ietf/admin/sites.py new file mode 100644 index 0000000000..69cb62ae20 --- /dev/null +++ b/ietf/admin/sites.py @@ -0,0 +1,15 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +from django.contrib.admin import AdminSite as _AdminSite +from django.conf import settings +from django.utils.safestring import mark_safe + + +class AdminSite(_AdminSite): + site_title = "Datatracker admin" + + @staticmethod + def site_header(): + if settings.SERVER_MODE == "production": + return "Datatracker administration" + else: + return mark_safe('Datatracker administration δ') diff --git a/ietf/api/__init__.py b/ietf/api/__init__.py index b4c6203d98..d4562f97dd 100644 --- a/ietf/api/__init__.py +++ b/ietf/api/__init__.py @@ -4,56 +4,50 @@ import datetime import re +import sys from urllib.parse import urlencode -from django.conf import settings +from django.apps import apps as django_apps from django.core.exceptions import ObjectDoesNotExist +from django.utils.module_loading import autodiscover_modules + import debug # pyflakes:ignore -import tastypie import tastypie.resources +import tastypie.serializers from tastypie.api import Api from tastypie.bundle import Bundle from tastypie.exceptions import ApiFieldError -from tastypie.serializers import Serializer # pyflakes:ignore (we're re-exporting this) from tastypie.fields import ApiField _api_list = [] -for _app in settings.INSTALLED_APPS: +OMITTED_APPS_APIS = ["ietf.status"] + +# Pre-py3.11, fromisoformat() does not handle Z or +HH tz offsets +HAVE_BROKEN_FROMISOFORMAT = sys.version_info < (3, 11, 0, "", 0) + +def populate_api_list(): _module_dict = globals() - if '.' in _app: - _root, _name = _app.split('.', 1) - if _root == 'ietf': - if not '.' in _name: - _api = Api(api_name=_name) - _module_dict[_name] = _api - _api_list.append((_name, _api)) + for app_config in django_apps.get_app_configs(): + if '.' in app_config.name and app_config.name not in OMITTED_APPS_APIS: + _root, _name = app_config.name.split('.', 1) + if _root == 'ietf': + if not '.' in _name: + _api = Api(api_name=_name) + _module_dict[_name] = _api + _api_list.append((_name, _api)) def autodiscover(): """ Auto-discover INSTALLED_APPS resources.py modules and fail silently when - not present. This forces an import on them to register any admin bits they + not present. This forces an import on them to register any resources they may want. """ + autodiscover_modules("resources") - from importlib import import_module - from django.conf import settings - from django.utils.module_loading import module_has_submodule - - for app in settings.INSTALLED_APPS: - mod = import_module(app) - # Attempt to import the app's admin module. - try: - import_module('%s.resources' % (app, )) - except: - # Decide whether to bubble up this error. If the app just - # doesn't have an admin module, we can ignore the error - # attempting to import it, otherwise we want it to bubble up. - if module_has_submodule(mod, "resources"): - raise class ModelResource(tastypie.resources.ModelResource): def generate_cache_key(self, *args, **kwargs): @@ -68,6 +62,35 @@ def generate_cache_key(self, *args, **kwargs): # Use a list plus a ``.join()`` because it's faster than concatenation. return "%s:%s:%s:%s" % (self._meta.api_name, self._meta.resource_name, ':'.join(args), smooshed) + def _z_aware_fromisoformat(self, value: str) -> datetime.datetime: + """datetime.datetime.fromisoformat replacement that works with python < 3.11""" + if HAVE_BROKEN_FROMISOFORMAT: + if value.upper().endswith("Z"): + value = value[:-1] + "+00:00" # Z -> UTC + elif re.match(r"[+-][0-9][0-9]$", value[-3:]): + value = value + ":00" # -04 -> -04:00 + return datetime.datetime.fromisoformat(value) + + def filter_value_to_python( + self, value, field_name, filters, filter_expr, filter_type + ): + py_value = super().filter_value_to_python( + value, field_name, filters, filter_expr, filter_type + ) + if isinstance( + self.fields[field_name], tastypie.fields.DateTimeField + ) and isinstance(py_value, str): + # Ensure datetime values are TZ-aware, using UTC by default + try: + dt = self._z_aware_fromisoformat(py_value) + except ValueError: + pass # let tastypie deal with the original value + else: + if dt.tzinfo is None: + dt = dt.replace(tzinfo=datetime.timezone.utc) + py_value = dt.isoformat() + return py_value + TIMEDELTA_REGEX = re.compile(r'^(?P\d+d)?\s?(?P\d+h)?\s?(?P\d+m)?\s?(?P\d+s?)$') @@ -152,3 +175,29 @@ def dehydrate(self, bundle, for_list=True): dehydrated = self.dehydrate_related(fk_bundle, fk_resource, for_list=for_list) fk_resource._meta.cache.set(cache_key, dehydrated) return dehydrated + + +class Serializer(tastypie.serializers.Serializer): + OPTION_ESCAPE_NULLS = "datatracker-escape-nulls" + + def format_datetime(self, data): + return data.astimezone(datetime.UTC).replace(tzinfo=None).isoformat(timespec="seconds") + "Z" + + def to_simple(self, data, options): + options = options or {} + simple_data = super().to_simple(data, options) + if ( + options.get(self.OPTION_ESCAPE_NULLS, False) + and isinstance(simple_data, str) + ): + # replace nulls with unicode "symbol for null character", \u2400 + simple_data = simple_data.replace("\x00", "\u2400") + return simple_data + + def to_etree(self, data, options=None, name=None, depth=0): + # lxml does not escape nulls on its own, so ask to_simple() to do it. + # This is mostly (only?) an issue when generating errors responses for + # fuzzers. + options = options or {} + options[self.OPTION_ESCAPE_NULLS] = True + return super().to_etree(data, options, name, depth) diff --git a/ietf/api/__init__.pyi b/ietf/api/__init__.pyi index 63d9bc513b..ededea90a7 100644 --- a/ietf/api/__init__.pyi +++ b/ietf/api/__init__.pyi @@ -30,4 +30,5 @@ class Serializer(): ... class ToOneField(tastypie.fields.ToOneField): ... class TimedeltaField(tastypie.fields.ApiField): ... +def populate_api_list() -> None: ... def autodiscover() -> None: ... diff --git a/ietf/api/apps.py b/ietf/api/apps.py new file mode 100644 index 0000000000..4549e0d7f2 --- /dev/null +++ b/ietf/api/apps.py @@ -0,0 +1,19 @@ +from django.apps import AppConfig +from . import populate_api_list + + +class ApiConfig(AppConfig): + name = "ietf.api" + + def ready(self): + """Hook to do init after the app registry is fully populated + + Importing models or accessing the app registry is ok here, but do not + interact with the database. See + https://docs.djangoproject.com/en/4.2/ref/applications/#django.apps.AppConfig.ready + """ + # Populate our API list now that the app registry is set up + populate_api_list() + + # Import drf-spectacular extensions + import ietf.api.schema # pyflakes: ignore diff --git a/ietf/api/authentication.py b/ietf/api/authentication.py new file mode 100644 index 0000000000..dfab0d72b8 --- /dev/null +++ b/ietf/api/authentication.py @@ -0,0 +1,19 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +# +from rest_framework import authentication +from django.contrib.auth.models import AnonymousUser + + +class ApiKeyAuthentication(authentication.BaseAuthentication): + """API-Key header authentication""" + + def authenticate(self, request): + """Extract the authentication token, if present + + This does not validate the token, it just arranges for it to be available in request.auth. + It's up to a Permissions class to validate it for the appropriate endpoint. + """ + token = request.META.get("HTTP_X_API_KEY", None) + if token is None: + return None + return AnonymousUser(), token # available as request.user and request.auth diff --git a/ietf/api/ietf_utils.py b/ietf/api/ietf_utils.py new file mode 100644 index 0000000000..50767a5afd --- /dev/null +++ b/ietf/api/ietf_utils.py @@ -0,0 +1,76 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +# This is not utils.py because Tastypie implicitly consumes ietf.api.utils. +# See ietf.api.__init__.py for details. +from functools import wraps +from typing import Callable, Optional, Union + +from django.conf import settings +from django.http import HttpResponseForbidden + + +def is_valid_token(endpoint, token): + # This is where we would consider integration with vault + # Settings implementation for now. + if hasattr(settings, "APP_API_TOKENS"): + token_store = settings.APP_API_TOKENS + if endpoint in token_store: + endpoint_tokens = token_store[endpoint] + # Be sure endpoints is a list or tuple so we don't accidentally use substring matching! + if not isinstance(endpoint_tokens, (list, tuple)): + endpoint_tokens = [endpoint_tokens] + if token in endpoint_tokens: + return True + return False + + +def requires_api_token(func_or_endpoint: Optional[Union[Callable, str]] = None): + """Validate API token before executing the wrapped method + + Usage: + * Basic: endpoint defaults to the qualified name of the wrapped method. E.g., in ietf.api.views, + + @requires_api_token + def my_view(request): + ... + + will require a token for "ietf.api.views.my_view" + + * Custom endpoint: specify the endpoint explicitly + + @requires_api_token("ietf.api.views.some_other_thing") + def my_view(request): + ... + + will require a token for "ietf.api.views.some_other_thing" + """ + + def decorate(f): + if _endpoint is None: + fname = getattr(f, "__qualname__", None) + if fname is None: + raise TypeError( + "Cannot automatically decorate function that does not support __qualname__. " + "Explicitly set the endpoint." + ) + endpoint = "{}.{}".format(f.__module__, fname) + else: + endpoint = _endpoint + + @wraps(f) + def wrapped(request, *args, **kwargs): + authtoken = request.META.get("HTTP_X_API_KEY", None) + if authtoken is None or not is_valid_token(endpoint, authtoken): + return HttpResponseForbidden() + return f(request, *args, **kwargs) + + return wrapped + + # Magic to allow decorator to be used with or without parentheses + if callable(func_or_endpoint): + func = func_or_endpoint + _endpoint = None + return decorate(func) + else: + _endpoint = func_or_endpoint + return decorate diff --git a/ietf/api/permissions.py b/ietf/api/permissions.py new file mode 100644 index 0000000000..8f7fdd026f --- /dev/null +++ b/ietf/api/permissions.py @@ -0,0 +1,39 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +# +from rest_framework import permissions +from ietf.api.ietf_utils import is_valid_token + + +class HasApiKey(permissions.BasePermission): + """Permissions class that validates a token using is_valid_token + + The view class must indicate the relevant endpoint by setting `api_key_endpoint`. + Must be used with an Authentication class that puts a token in request.auth. + """ + def has_permission(self, request, view): + endpoint = getattr(view, "api_key_endpoint", None) + auth_token = getattr(request, "auth", None) + if endpoint is not None and auth_token is not None: + return is_valid_token(endpoint, auth_token) + return False + + +class IsOwnPerson(permissions.BasePermission): + """Permission to access own Person object""" + def has_object_permission(self, request, view, obj): + if not (request.user.is_authenticated and hasattr(request.user, "person")): + return False + return obj == request.user.person + + +class BelongsToOwnPerson(permissions.BasePermission): + """Permission to access objects associated with own Person + + Requires that the object have a "person" field that indicates ownership. + """ + def has_object_permission(self, request, view, obj): + if not (request.user.is_authenticated and hasattr(request.user, "person")): + return False + return ( + hasattr(obj, "person") and obj.person == request.user.person + ) diff --git a/ietf/api/routers.py b/ietf/api/routers.py new file mode 100644 index 0000000000..99afdb242a --- /dev/null +++ b/ietf/api/routers.py @@ -0,0 +1,31 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +"""Custom django-rest-framework routers""" +from django.core.exceptions import ImproperlyConfigured +from rest_framework import routers + + +class PrefixedBasenameMixin: + """Mixin to add a prefix to the basename of a rest_framework BaseRouter""" + def __init__(self, name_prefix="", *args, **kwargs): + self.name_prefix = name_prefix + if len(self.name_prefix) == 0 or self.name_prefix[-1] == ".": + raise ImproperlyConfigured("Cannot use a name_prefix that is empty or ends with '.'") + super().__init__(*args, **kwargs) + + def register(self, prefix, viewset, basename=None): + # Get the superclass "register" method from the class this is mixed-in with. + # This avoids typing issues with calling super().register() directly in a + # mixin class. + super_register = getattr(super(), "register") + if not super_register or not callable(super_register): + raise TypeError("Must mixin with superclass that has register() method") + super_register(prefix, viewset, basename=f"{self.name_prefix}.{basename}") + + +class PrefixedSimpleRouter(PrefixedBasenameMixin, routers.SimpleRouter): + """SimpleRouter that adds a dot-separated prefix to its basename""" + + +class PrefixedDefaultRouter(PrefixedBasenameMixin, routers.DefaultRouter): + """DefaultRouter that adds a dot-separated prefix to its basename""" + diff --git a/ietf/api/schema.py b/ietf/api/schema.py new file mode 100644 index 0000000000..7340149685 --- /dev/null +++ b/ietf/api/schema.py @@ -0,0 +1,20 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +# +from drf_spectacular.extensions import OpenApiAuthenticationExtension + + +class ApiKeyAuthenticationScheme(OpenApiAuthenticationExtension): + """Authentication scheme extension for the ApiKeyAuthentication + + Used by drf-spectacular when rendering the OpenAPI schema + """ + target_class = "ietf.api.authentication.ApiKeyAuthentication" + name = "apiKeyAuth" + + def get_security_definition(self, auto_schema): + return { + "type": "apiKey", + "description": "Shared secret in the X-Api-Key header", + "name": "X-Api-Key", + "in": "header", + } diff --git a/ietf/api/serializer.py b/ietf/api/serializer.py index 9d6cf1ebb8..d5bca430e0 100644 --- a/ietf/api/serializer.py +++ b/ietf/api/serializer.py @@ -1,6 +1,9 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved +# Copyright The IETF Trust 2018-2024, All Rights Reserved # -*- coding: utf-8 -*- +"""Serialization utilities +This is _not_ for django-rest-framework! +""" import hashlib import json @@ -9,11 +12,12 @@ from django.core.exceptions import ObjectDoesNotExist, FieldError from django.core.serializers.json import Serializer from django.http import HttpResponse -from django.utils.encoding import smart_text +from django.utils.encoding import smart_str from django.db.models import Field -from django.db.models.query import QuerySet from django.db.models.signals import post_save, post_delete, m2m_changed +from django_stubs_ext import QuerySetAny + import debug # pyflakes:ignore @@ -121,7 +125,7 @@ def end_object(self, obj): for name in expansions: try: field = getattr(obj, name) - #self._current["_"+name] = smart_text(field) + #self._current["_"+name] = smart_str(field) if not isinstance(field, Field): options = self.options.copy() options["expand"] = [ v[len(name)+2:] for v in options["expand"] if v.startswith(name+"__") ] @@ -145,7 +149,7 @@ def end_object(self, obj): field_value = None else: field_value = field - if isinstance(field_value, QuerySet) or isinstance(field_value, list): + if isinstance(field_value, QuerySetAny) or isinstance(field_value, list): self._current[name] = dict([ (rel.pk, self.expand_related(rel, name)) for rel in field_value ]) else: if hasattr(field_value, "_meta"): @@ -188,10 +192,10 @@ def handle_fk_field(self, obj, field): related = related.natural_key() elif field.remote_field.field_name == related._meta.pk.name: # Related to remote object via primary key - related = smart_text(related._get_pk_val(), strings_only=True) + related = smart_str(related._get_pk_val(), strings_only=True) else: # Related to remote object via other field - related = smart_text(getattr(related, field.remote_field.field_name), strings_only=True) + related = smart_str(getattr(related, field.remote_field.field_name), strings_only=True) self._current[field.name] = related def handle_m2m_field(self, obj, field): @@ -201,7 +205,7 @@ def handle_m2m_field(self, obj, field): elif self.use_natural_keys and hasattr(field.remote_field.to, 'natural_key'): m2m_value = lambda value: value.natural_key() else: - m2m_value = lambda value: smart_text(value._get_pk_val(), strings_only=True) + m2m_value = lambda value: smart_str(value._get_pk_val(), strings_only=True) self._current[field.name] = [m2m_value(related) for related in getattr(obj, field.name).iterator()] @@ -221,7 +225,7 @@ class JsonExportMixin(object): # obj = None # # if obj is None: -# raise Http404(_('%(name)s object with primary key %(key)r does not exist.') % {'name': force_text(self.model._meta.verbose_name), 'key': escape(object_id)}) +# raise Http404(_('%(name)s object with primary key %(key)r does not exist.') % {'name': force_str(self.model._meta.verbose_name), 'key': escape(object_id)}) # # content_type = 'application/json' # return HttpResponse(serialize([ obj ], sort_keys=True, indent=3)[2:-2], content_type=content_type) @@ -264,6 +268,6 @@ def json_view(self, request, filter=None, expand=None): qd = dict( ( k, json.loads(v)[0] ) for k,v in items ) except (FieldError, ValueError) as e: return HttpResponse(json.dumps({"error": str(e)}, sort_keys=True, indent=3), content_type=content_type) - text = json.dumps({smart_text(self.model._meta): qd}, sort_keys=True, indent=3) + text = json.dumps({smart_str(self.model._meta): qd}, sort_keys=True, indent=3) return HttpResponse(text, content_type=content_type) diff --git a/ietf/api/serializers_rpc.py b/ietf/api/serializers_rpc.py new file mode 100644 index 0000000000..d888de4586 --- /dev/null +++ b/ietf/api/serializers_rpc.py @@ -0,0 +1,804 @@ +# Copyright The IETF Trust 2025-2026, All Rights Reserved +import datetime +from pathlib import Path +from typing import Literal, Optional + +from django.db import transaction +from django.urls import reverse as urlreverse +from django.utils import timezone +from drf_spectacular.types import OpenApiTypes +from drf_spectacular.utils import extend_schema_field +from rest_framework import serializers + +from ietf.doc.expire import move_draft_files_to_archive +from ietf.doc.models import ( + DocumentAuthor, + Document, + RelatedDocument, + State, + DocEvent, + RfcAuthor, +) +from ietf.doc.serializers import RfcAuthorSerializer +from ietf.doc.tasks import trigger_red_precomputer_task, update_rfc_searchindex_task +from ietf.doc.utils import ( + default_consensus, + prettify_std_name, + update_action_holders, + update_rfcauthors, +) +from ietf.group.models import Group, Role +from ietf.group.serializers import AreaSerializer +from ietf.name.models import StreamName, StdLevelName +from ietf.person.models import Person +from ietf.utils import log + + +class PersonSerializer(serializers.ModelSerializer): + email = serializers.EmailField(read_only=True) + picture = serializers.URLField(source="cdn_photo_url", read_only=True) + url = serializers.SerializerMethodField( + help_text="relative URL for datatracker person page" + ) + + class Meta: + model = Person + fields = ["id", "plain_name", "email", "picture", "url"] + read_only_fields = ["id", "plain_name", "email", "picture", "url"] + + @extend_schema_field(OpenApiTypes.URI) + def get_url(self, object: Person): + return urlreverse( + "ietf.person.views.profile", + kwargs={"email_or_name": object.email_address() or object.name}, + ) + + +class EmailPersonSerializer(serializers.Serializer): + email = serializers.EmailField(source="address") + person_pk = serializers.IntegerField(source="person.pk") + name = serializers.CharField(source="person.name") + last_name = serializers.CharField(source="person.last_name") + initials = serializers.CharField(source="person.initials") + + +class LowerCaseEmailField(serializers.EmailField): + def to_representation(self, value): + return super().to_representation(value).lower() + + +class AuthorPersonSerializer(serializers.ModelSerializer): + person_pk = serializers.IntegerField(source="pk", read_only=True) + last_name = serializers.CharField() + initials = serializers.CharField() + email_addresses = serializers.ListField( + source="email_set.all", child=LowerCaseEmailField() + ) + + class Meta: + model = Person + fields = ["person_pk", "name", "last_name", "initials", "email_addresses"] + + +class RfcWithAuthorsSerializer(serializers.ModelSerializer): + authors = AuthorPersonSerializer(many=True, source="author_persons") + + class Meta: + model = Document + fields = ["rfc_number", "authors"] + + +class DraftWithAuthorsSerializer(serializers.ModelSerializer): + draft_name = serializers.CharField(source="name") + authors = AuthorPersonSerializer(many=True, source="author_persons") + + class Meta: + model = Document + fields = ["draft_name", "authors"] + + +class WgChairSerializer(serializers.Serializer): + """Serialize a WG chair's name and email from a Role""" + + name = serializers.SerializerMethodField() + email = serializers.SerializerMethodField() + + @extend_schema_field(serializers.CharField) + def get_name(self, role: Role) -> str: + return role.person.plain_name() + + @extend_schema_field(serializers.EmailField) + def get_email(self, role: Role) -> str: + return role.email.email_address() + + +class DocumentAuthorSerializer(serializers.ModelSerializer): + """Serializer for a Person in a response""" + + plain_name = serializers.SerializerMethodField() + + class Meta: + model = DocumentAuthor + fields = ["person", "plain_name", "affiliation"] + + def get_plain_name(self, document_author: DocumentAuthor) -> str: + return document_author.person.plain_name() + + +class FullDraftSerializer(serializers.ModelSerializer): + # Redefine these fields so they don't pick up the regex validator patterns. + # There seem to be some non-compliant drafts in the system! If this serializer + # is used for a writeable view, the validation will need to be added back. + name = serializers.CharField(max_length=255) + title = serializers.CharField(max_length=255) + group = serializers.SlugRelatedField(slug_field="acronym", read_only=True) + area = AreaSerializer(read_only=True) + + # Other fields we need to add / adjust + source_format = serializers.SerializerMethodField() + authors = DocumentAuthorSerializer(many=True, source="documentauthor_set") + shepherd = serializers.PrimaryKeyRelatedField( + source="shepherd.person", read_only=True + ) + consensus = serializers.SerializerMethodField() + wg_chairs = serializers.SerializerMethodField() + + class Meta: + model = Document + fields = [ + "id", + "name", + "rev", + "stream", + "title", + "group", + "area", + "abstract", + "pages", + "source_format", + "authors", + "intended_std_level", + "consensus", + "shepherd", + "ad", + "wg_chairs", + ] + + def get_consensus(self, doc: Document) -> Optional[bool]: + return default_consensus(doc) + + @extend_schema_field(WgChairSerializer(many=True)) + def get_wg_chairs(self, doc: Document): + if doc.group is None: + return [] + chairs = doc.group.role_set.filter(name_id="chair").select_related( + "person", "email" + ) + return WgChairSerializer(chairs, many=True).data + + def get_source_format( + self, doc: Document + ) -> Literal["unknown", "xml-v2", "xml-v3", "txt"]: + submission = doc.submission() + if submission is None: + return "unknown" + if ".xml" in submission.file_types: + if submission.xml_version == "3": + return "xml-v3" + else: + return "xml-v2" + elif ".txt" in submission.file_types: + return "txt" + return "unknown" + + +class DraftSerializer(FullDraftSerializer): + class Meta: + model = Document + fields = [ + "id", + "name", + "rev", + "stream", + "title", + "group", + "pages", + "source_format", + "authors", + "consensus", + ] + + +class SubmittedToQueueSerializer(FullDraftSerializer): + submitted = serializers.SerializerMethodField() + consensus = serializers.SerializerMethodField() + + class Meta: + model = Document + fields = [ + "id", + "name", + "stream", + "submitted", + "consensus", + ] + + def get_submitted(self, doc) -> Optional[datetime.datetime]: + event = doc.sent_to_rfc_editor_event() + return None if event is None else event.time + + def get_consensus(self, doc) -> Optional[bool]: + return default_consensus(doc) + + +class OriginalStreamSerializer(serializers.ModelSerializer): + stream = serializers.CharField(read_only=True, source="orig_stream_id") + + class Meta: + model = Document + fields = ["rfc_number", "stream"] + + +class ReferenceSerializer(serializers.ModelSerializer): + class Meta: + model = Document + fields = ["id", "name"] + read_only_fields = ["id", "name"] + + +def _update_authors(rfc, authors_data): + # Construct unsaved instances from validated author data + new_authors = [RfcAuthor(**authdata) for authdata in authors_data] + # Update the RFC with the new author set + with transaction.atomic(): + change_events = update_rfcauthors(rfc, new_authors) + for event in change_events: + event.save() + return change_events + + +class SubseriesNameField(serializers.RegexField): + + def __init__(self, **kwargs): + # pattern: no leading 0, finite length (arbitrarily set to 5 digits) + regex = r"^(bcp|std|fyi)[1-9][0-9]{0,4}$" + super().__init__(regex, **kwargs) + + + +class RfcPubSerializer(serializers.ModelSerializer): + """Write-only serializer for RFC publication""" + # publication-related fields + published = serializers.DateTimeField(default_timezone=datetime.timezone.utc) + draft_name = serializers.RegexField( + required=False, regex=r"^draft-[a-zA-Z0-9-]+$" + ) + draft_rev = serializers.RegexField( + required=False, regex=r"^[0-9][0-9]$" + ) + + # fields on the RFC Document that need tweaking from ModelSerializer defaults + rfc_number = serializers.IntegerField(min_value=1, required=True) + group = serializers.SlugRelatedField( + slug_field="acronym", queryset=Group.objects.all(), required=False + ) + stream = serializers.PrimaryKeyRelatedField( + queryset=StreamName.objects.filter(used=True) + ) + std_level = serializers.PrimaryKeyRelatedField( + queryset=StdLevelName.objects.filter(used=True), + ) + ad = serializers.PrimaryKeyRelatedField( + queryset=Person.objects.all(), + allow_null=True, + required=False, + ) + obsoletes = serializers.SlugRelatedField( + many=True, + required=False, + slug_field="rfc_number", + queryset=Document.objects.filter(type_id="rfc"), + ) + updates = serializers.SlugRelatedField( + many=True, + required=False, + slug_field="rfc_number", + queryset=Document.objects.filter(type_id="rfc"), + ) + subseries = serializers.ListField(child=SubseriesNameField(required=False)) + # N.b., authors is _not_ a field on Document! + authors = RfcAuthorSerializer(many=True) + + class Meta: + model = Document + fields = [ + "published", + "draft_name", + "draft_rev", + "rfc_number", + "title", + "authors", + "group", + "stream", + "abstract", + "pages", + "std_level", + "ad", + "obsoletes", + "updates", + "subseries", + "keywords", + ] + + def validate(self, data): + if "draft_name" in data or "draft_rev" in data: + if "draft_name" not in data: + raise serializers.ValidationError( + {"draft_name": "Missing draft_name"}, + code="invalid-draft-spec", + ) + if "draft_rev" not in data: + raise serializers.ValidationError( + {"draft_rev": "Missing draft_rev"}, + code="invalid-draft-spec", + ) + return data + + def update(self, instance, validated_data): + raise RuntimeError("Cannot update with this serializer") + + def create(self, validated_data): + """Publish an RFC""" + published = validated_data.pop("published") + draft_name = validated_data.pop("draft_name", None) + draft_rev = validated_data.pop("draft_rev", None) + obsoletes = validated_data.pop("obsoletes", []) + updates = validated_data.pop("updates", []) + subseries = validated_data.pop("subseries", []) + + system_person = Person.objects.get(name="(System)") + + # If specified, retrieve draft and extract RFC default values from it + if draft_name is None: + draft = None + else: + # validation enforces that draft_name and draft_rev are both present + draft = Document.objects.filter( + type_id="draft", + name=draft_name, + rev=draft_rev, + ).first() + if draft is None: + raise serializers.ValidationError( + { + "draft_name": "No such draft", + "draft_rev": "No such draft", + }, + code="invalid-draft" + ) + elif draft.get_state_slug() == "rfc": + raise serializers.ValidationError( + { + "draft_name": "Draft already published as RFC", + }, + code="already-published-draft", + ) + + # Transaction to clean up if something fails + with transaction.atomic(): + # create rfc, letting validated request data override draft defaults + rfc = self._create_rfc(validated_data) + DocEvent.objects.create( + doc=rfc, + rev=rfc.rev, + type="published_rfc", + time=published, + by=system_person, + desc="RFC published", + ) + rfc.set_state(State.objects.get(used=True, type_id="rfc", slug="published")) + + # create updates / obsoletes relations + for obsoleted_rfc_pk in obsoletes: + RelatedDocument.objects.get_or_create( + source=rfc, target=obsoleted_rfc_pk, relationship_id="obs" + ) + for updated_rfc_pk in updates: + RelatedDocument.objects.get_or_create( + source=rfc, target=updated_rfc_pk, relationship_id="updates" + ) + + # create subseries relations + for subseries_doc_name in subseries: + ss_slug = subseries_doc_name[:3] + subseries_doc, ss_doc_created = Document.objects.get_or_create( + type_id=ss_slug, name=subseries_doc_name + ) + if ss_doc_created: + subseries_doc.docevent_set.create( + type=f"{ss_slug}_doc_created", + by=system_person, + desc=f"Created {subseries_doc_name} via publication of {rfc.name}", + ) + _, ss_rel_created = subseries_doc.relateddocument_set.get_or_create( + relationship_id="contains", target=rfc + ) + if ss_rel_created: + subseries_doc.docevent_set.create( + type="sync_from_rfc_editor", + by=system_person, + desc=f"Added {rfc.name} to {subseries_doc.name}", + ) + rfc.docevent_set.create( + type="sync_from_rfc_editor", + by=system_person, + desc=f"Added {rfc.name} to {subseries_doc.name}", + ) + + + # create relation with draft and update draft state + if draft is not None: + draft_changes = [] + draft_events = [] + if draft.get_state_slug() != "rfc": + draft.set_state( + State.objects.get(used=True, type="draft", slug="rfc") + ) + move_draft_files_to_archive(draft, draft.rev) + draft_changes.append(f"changed state to {draft.get_state()}") + + r, created_relateddoc = RelatedDocument.objects.get_or_create( + source=draft, target=rfc, relationship_id="became_rfc", + ) + if created_relateddoc: + change = "created {rel_name} relationship between {pretty_draft_name} and {pretty_rfc_name}".format( + rel_name=r.relationship.name.lower(), + pretty_draft_name=prettify_std_name(draft_name), + pretty_rfc_name=prettify_std_name(rfc.name), + ) + draft_changes.append(change) + + # Always set the "draft-iesg" state. This state should be set for all drafts, so + # log a warning if it is not set. What should happen here is that ietf stream + # RFCs come in as "rfcqueue" and are set to "pub" when they appear in the RFC index. + # Other stream documents should normally be "idexists" and be left that way. The + # code here *actually* leaves "draft-iesg" state alone if it is "idexists" or "pub", + # and changes any other state to "pub". If unset, it changes it to "idexists". + # This reflects historical behavior and should probably be updated, but a migration + # of existing drafts (and validation of the change) is needed before we change the + # handling. + prev_iesg_state = draft.get_state("draft-iesg") + if prev_iesg_state is None: + log.log(f'Warning while processing {rfc.name}: {draft.name} has no "draft-iesg" state') + new_iesg_state = State.objects.get(type_id="draft-iesg", slug="idexists") + elif prev_iesg_state.slug not in ("pub", "idexists"): + if prev_iesg_state.slug != "rfcqueue": + log.log( + 'Warning while processing {}: {} is in "draft-iesg" state {} (expected "rfcqueue")'.format( + rfc.name, draft.name, prev_iesg_state.slug + ) + ) + new_iesg_state = State.objects.get(type_id="draft-iesg", slug="pub") + else: + new_iesg_state = prev_iesg_state + + if new_iesg_state != prev_iesg_state: + draft.set_state(new_iesg_state) + draft_changes.append(f"changed {new_iesg_state.type.label} to {new_iesg_state}") + e = update_action_holders(draft, prev_iesg_state, new_iesg_state) + if e: + draft_events.append(e) + + # If the draft and RFC streams agree, move draft to "pub" stream state. If not, complain. + if draft.stream != rfc.stream: + log.log("Warning while processing {}: draft {} stream is {} but RFC stream is {}".format( + rfc.name, draft.name, draft.stream, rfc.stream + )) + elif draft.stream.slug in ["iab", "irtf", "ise", "editorial"]: + stream_slug = f"draft-stream-{draft.stream.slug}" + prev_state = draft.get_state(stream_slug) + if prev_state is not None and prev_state.slug != "pub": + new_state = State.objects.select_related("type").get(used=True, type__slug=stream_slug, slug="pub") + draft.set_state(new_state) + draft_changes.append( + f"changed {new_state.type.label} to {new_state}" + ) + e = update_action_holders(draft, prev_state, new_state) + if e: + draft_events.append(e) + if draft_changes: + draft_events.append( + DocEvent.objects.create( + doc=draft, + rev=draft.rev, + by=system_person, + type="sync_from_rfc_editor", + desc=f"Updated while publishing {rfc.name} ({', '.join(draft_changes)})", + ) + ) + draft.save_with_history(draft_events) + + return rfc + + def _create_rfc(self, validated_data): + authors_data = validated_data.pop("authors") + rfc = Document.objects.create( + type_id="rfc", + name=f"rfc{validated_data['rfc_number']}", + **validated_data, + ) + for order, author_data in enumerate(authors_data): + rfc.rfcauthor_set.create( + order=order, + **author_data, + ) + return rfc + + +class EditableRfcSerializer(serializers.ModelSerializer): + # Would be nice to reconcile this with ietf.doc.serializers.RfcSerializer. + # The purposes of that serializer (representing data for Red) and this one + # (accepting updates from Purple) are different enough that separate formats + # may be needed, but if not it'd be nice to have a single RfcSerializer that + # can serve both. + # + # Should also consider whether this and RfcPubSerializer should merge. + # + # Treats published and subseries fields as write-only. This isn't quite correct, + # but makes it easier and we don't currently use the serialized value except for + # debugging. + published = serializers.DateTimeField( + default_timezone=datetime.timezone.utc, + write_only=True, + ) + authors = RfcAuthorSerializer(many=True, min_length=1, source="rfcauthor_set") + subseries = serializers.ListField( + child=SubseriesNameField(required=False), + write_only=True, + ) + + class Meta: + model = Document + fields = [ + "published", + "title", + "authors", + "stream", + "abstract", + "pages", + "std_level", + "subseries", + "keywords", + ] + + def create(self, validated_data): + raise RuntimeError("Cannot create with this serializer") + + def update(self, instance, validated_data): + assert isinstance(instance, Document) + assert instance.type_id == "rfc" + rfc = instance # get better name + + system_person = Person.objects.get(name="(System)") + + # Remove data that needs special handling. Use a singleton object to detect + # missing values in case we ever support a value that needs None as an option. + omitted = object() + published = validated_data.pop("published", omitted) + subseries = validated_data.pop("subseries", omitted) + authors_data = validated_data.pop("rfcauthor_set", omitted) + + # Transaction to clean up if something fails + with transaction.atomic(): + # update the rfc Document itself + rfc_changes = [] + rfc_events = [] + + for attr, new_value in validated_data.items(): + old_value = getattr(rfc, attr) + if new_value != old_value: + rfc_changes.append( + f"changed {attr} to '{new_value}' from '{old_value}'" + ) + setattr(rfc, attr, new_value) + if len(rfc_changes) > 0: + rfc_change_summary = f"{', '.join(rfc_changes)}" + rfc_events.append( + DocEvent.objects.create( + doc=rfc, + rev=rfc.rev, + by=system_person, + type="sync_from_rfc_editor", + desc=f"Changed metadata: {rfc_change_summary}", + ) + ) + if authors_data is not omitted: + rfc_events.extend(_update_authors(instance, authors_data)) + + if published is not omitted: + published_event = rfc.latest_event(type="published_rfc") + if published_event is None: + # unexpected, but possible in theory + rfc_events.append( + DocEvent.objects.create( + doc=rfc, + rev=rfc.rev, + type="published_rfc", + time=published, + by=system_person, + desc="RFC published", + ) + ) + rfc_events.append( + DocEvent.objects.create( + doc=rfc, + rev=rfc.rev, + type="sync_from_rfc_editor", + by=system_person, + desc=( + f"Set publication timestamp to {published.isoformat()}" + ), + ) + ) + else: + original_pub_time = published_event.time + if published != original_pub_time: + published_event.time = published + published_event.save() + rfc_events.append( + DocEvent.objects.create( + doc=rfc, + rev=rfc.rev, + type="sync_from_rfc_editor", + by=system_person, + desc=( + f"Changed publication time to " + f"{published.isoformat()} from " + f"{original_pub_time.isoformat()}" + ) + ) + ) + + # update subseries relations + if subseries is not omitted: + for subseries_doc_name in subseries: + ss_slug = subseries_doc_name[:3] + subseries_doc, ss_doc_created = Document.objects.get_or_create( + type_id=ss_slug, name=subseries_doc_name + ) + if ss_doc_created: + subseries_doc.docevent_set.create( + type=f"{ss_slug}_doc_created", + by=system_person, + desc=f"Created {subseries_doc_name} via update of {rfc.name}", + ) + _, ss_rel_created = subseries_doc.relateddocument_set.get_or_create( + relationship_id="contains", target=rfc + ) + if ss_rel_created: + subseries_doc.docevent_set.create( + type="sync_from_rfc_editor", + by=system_person, + desc=f"Added {rfc.name} to {subseries_doc.name}", + ) + rfc_events.append( + rfc.docevent_set.create( + type="sync_from_rfc_editor", + by=system_person, + desc=f"Added {rfc.name} to {subseries_doc.name}", + ) + ) + # Delete subseries relations that are no longer current + stale_subseries_relations = rfc.relations_that("contains").exclude( + source__name__in=subseries + ) + for stale_relation in stale_subseries_relations: + stale_subseries_doc = stale_relation.source + rfc_events.append( + rfc.docevent_set.create( + type="sync_from_rfc_editor", + by=system_person, + desc=f"Removed {rfc.name} from {stale_subseries_doc.name}", + ) + ) + stale_subseries_doc.docevent_set.create( + type="sync_from_rfc_editor", + by=system_person, + desc=f"Removed {rfc.name} from {stale_subseries_doc.name}", + ) + stale_subseries_relations.delete() + if len(rfc_events) > 0: + rfc.save_with_history(rfc_events) + # Gather obs and updates in both directions as a title/author change to + # this doc affects the info rendering of all of the other RFCs + needs_updating = sorted( + [ + d.rfc_number + for d in [rfc] + + rfc.related_that_doc(("obs", "updates")) + + rfc.related_that(("obs", "updates")) + ] + ) + trigger_red_precomputer_task.delay(rfc_number_list=needs_updating) + # Update the search index also + update_rfc_searchindex_task.delay(rfc.rfc_number) + return rfc + + +class RfcFileSerializer(serializers.Serializer): + # The structure of this serializer is constrained by what openapi-generator-cli's + # python generator can correctly serialize as multipart/form-data. It does not + # handle nested serializers well (or perhaps at all). ListFields with child + # ChoiceField or RegexField do not serialize correctly. DictFields don't seem + # to work. + # + # It does seem to correctly send filenames along with FileFields, even as a child + # in a ListField, so we use that to convey the file format of each item. There + # are other options we could consider (e.g., a structured CharField) but this + # works. + allowed_extensions = ( + ".html", + ".json", + ".notprepped.xml", + ".pdf", + ".txt", + ".xml", + ) + + rfc = serializers.SlugRelatedField( + slug_field="rfc_number", + queryset=Document.objects.filter(type_id="rfc"), + help_text="RFC number to which the contents belong", + ) + contents = serializers.ListField( + child=serializers.FileField( + allow_empty_file=False, + use_url=False, + ), + help_text=( + "List of content files. Filename extensions are used to identify " + "file types, but filenames are otherwise ignored." + ), + ) + mtime = serializers.DateTimeField( + required=False, + default=timezone.now, + default_timezone=datetime.UTC, + help_text="Modification timestamp to apply to uploaded files", + ) + replace = serializers.BooleanField( + required=False, + default=False, + help_text=( + "Replace existing files for this RFC. Defaults to false. When false, " + "if _any_ files already exist for the specified RFC the upload will be " + "rejected regardless of which files are being uploaded. When true," + "existing files will be removed and new ones will be put in place. BE" + "VERY CAREFUL WITH THIS OPTION IN PRODUCTION." + ), + ) + + def validate_contents(self, data): + found_extensions = [] + for uploaded_file in data: + if not hasattr(uploaded_file, "name"): + raise serializers.ValidationError( + "filename not specified for uploaded file", + code="missing-filename", + ) + ext = "".join(Path(uploaded_file.name).suffixes) + if ext not in self.allowed_extensions: + raise serializers.ValidationError( + f"File uploaded with invalid extension '{ext}'", + code="invalid-filename-ext", + ) + if ext in found_extensions: + raise serializers.ValidationError( + f"More than one file uploaded with extension '{ext}'", + code="duplicate-filename-ext", + ) + return data + + +class NotificationAckSerializer(serializers.Serializer): + message = serializers.CharField(default="ack") diff --git a/ietf/api/tests.py b/ietf/api/tests.py index 1be220f0f7..2a44791a5c 100644 --- a/ietf/api/tests.py +++ b/ietf/api/tests.py @@ -1,19 +1,23 @@ -# Copyright The IETF Trust 2015-2020, All Rights Reserved +# Copyright The IETF Trust 2015-2024, All Rights Reserved # -*- coding: utf-8 -*- - +import base64 +import copy import datetime import json import html +from unittest import mock import os import sys from importlib import import_module -from mock import patch from pathlib import Path +from random import randrange from django.apps import apps from django.conf import settings -from django.test import Client +from django.http import HttpResponseForbidden +from django.test import Client, RequestFactory +from django.test.utils import override_settings from django.urls import reverse as urlreverse from django.utils import timezone @@ -22,42 +26,36 @@ import debug # pyflakes:ignore import ietf +from ietf.doc.storage_utils import retrieve_str from ietf.doc.utils import get_unicode_document_content +from ietf.doc.models import RelatedDocument, State +from ietf.doc.factories import IndividualDraftFactory, WgDraftFactory, WgRfcFactory from ietf.group.factories import RoleFactory from ietf.meeting.factories import MeetingFactory, SessionFactory -from ietf.meeting.test_data import make_meeting_test_data -from ietf.meeting.models import Session -from ietf.person.factories import PersonFactory, random_faker -from ietf.person.models import User -from ietf.person.models import PersonalApiKey -from ietf.stats.models import MeetingRegistration -from ietf.utils.mail import outbox, get_payload_text +from ietf.meeting.models import Session, Registration +from ietf.nomcom.models import Volunteer +from ietf.nomcom.factories import NomComFactory, nomcom_kwargs_for_year +from ietf.person.factories import PersonFactory, random_faker, EmailFactory, PersonalApiKeyFactory +from ietf.person.models import Email, User +from ietf.utils.mail import empty_outbox, outbox, get_payload_text from ietf.utils.models import DumpInfo -from ietf.utils.test_utils import TestCase, login_testing_unauthorized +from ietf.utils.test_utils import TestCase, login_testing_unauthorized, reload_db_objects + +from . import Serializer +from .ietf_utils import is_valid_token, requires_api_token +from .views import EmailIngestionError OMITTED_APPS = ( 'ietf.secr.meetings', 'ietf.secr.proceedings', 'ietf.ipr', + 'ietf.status', + 'ietf.blobdb', ) class CustomApiTests(TestCase): settings_temp_path_overrides = TestCase.settings_temp_path_overrides + ['AGENDA_PATH'] - # Using mock to patch the import functions in ietf.meeting.views, where - # api_import_recordings() are using them: - @patch('ietf.meeting.views.import_audio_files') - def test_notify_meeting_import_audio_files(self, mock_import_audio): - meeting = make_meeting_test_data() - client = Client(Accept='application/json') - # try invalid method GET - url = urlreverse('ietf.meeting.views.api_import_recordings', kwargs={'number':meeting.number}) - r = client.get(url) - self.assertEqual(r.status_code, 405) - # try valid method POST - r = client.post(url) - self.assertEqual(r.status_code, 201) - def test_api_help_page(self): url = urlreverse('ietf.api.views.api_help') r = self.client.get(url) @@ -68,14 +66,14 @@ def test_api_openid_issuer(self): r = self.client.get(url) self.assertContains(r, 'OpenID Connect Issuer', status_code=200) - def test_api_set_session_video_url(self): + def test_deprecated_api_set_session_video_url(self): url = urlreverse('ietf.meeting.views.api_set_session_video_url') recmanrole = RoleFactory(group__type_id='ietf', name_id='recman') recman = recmanrole.person meeting = MeetingFactory(type_id='ietf') session = SessionFactory(group__type_id='wg', meeting=meeting) group = session.group - apikey = PersonalApiKey.objects.create(endpoint=url, person=recman) + apikey = PersonalApiKeyFactory(endpoint=url, person=recman) video = 'https://foo.example.com/bar/beer/' # error cases @@ -83,7 +81,7 @@ def test_api_set_session_video_url(self): self.assertContains(r, "Missing apikey parameter", status_code=400) badrole = RoleFactory(group__type_id='ietf', name_id='ad') - badapikey = PersonalApiKey.objects.create(endpoint=url, person=badrole.person) + badapikey = PersonalApiKeyFactory(endpoint=url, person=badrole.person) badrole.person.user.last_login = timezone.now() badrole.person.user.save() r = self.client.post(url, {'apikey': badapikey.hash()} ) @@ -97,7 +95,7 @@ def test_api_set_session_video_url(self): r = self.client.get(url, {'apikey': apikey.hash()} ) self.assertContains(r, "Method not allowed", status_code=405) - r = self.client.post(url, {'apikey': apikey.hash()} ) + r = self.client.post(url, {'apikey': apikey.hash(), 'group': group.acronym} ) self.assertContains(r, "Missing meeting parameter", status_code=400) @@ -149,17 +147,160 @@ def test_api_set_session_video_url(self): event = doc.latest_event() self.assertEqual(event.by, recman) - def test_api_add_session_attendees(self): + def test_api_set_session_video_url(self): + url = urlreverse("ietf.meeting.views.api_set_session_video_url") + recmanrole = RoleFactory(group__type_id="ietf", name_id="recman") + recman = recmanrole.person + meeting = MeetingFactory(type_id="ietf") + session = SessionFactory(group__type_id="wg", meeting=meeting) + apikey = PersonalApiKeyFactory(endpoint=url, person=recman) + video = "https://foo.example.com/bar/beer/" + + # error cases + r = self.client.post(url, {}) + self.assertContains(r, "Missing apikey parameter", status_code=400) + + badrole = RoleFactory(group__type_id="ietf", name_id="ad") + badapikey = PersonalApiKeyFactory(endpoint=url, person=badrole.person) + badrole.person.user.last_login = timezone.now() + badrole.person.user.save() + r = self.client.post(url, {"apikey": badapikey.hash()}) + self.assertContains(r, "Restricted to role: Recording Manager", status_code=403) + + r = self.client.post(url, {"apikey": apikey.hash()}) + self.assertContains(r, "Too long since last regular login", status_code=400) + recman.user.last_login = timezone.now() + recman.user.save() + + r = self.client.get(url, {"apikey": apikey.hash()}) + self.assertContains(r, "Method not allowed", status_code=405) + + r = self.client.post(url, {"apikey": apikey.hash()}) + self.assertContains(r, "Missing session_id parameter", status_code=400) + + r = self.client.post(url, {"apikey": apikey.hash(), "session_id": session.pk}) + self.assertContains(r, "Missing url parameter", status_code=400) + + bad_pk = int(Session.objects.order_by("-pk").first().pk) + 1 + r = self.client.post( + url, + { + "apikey": apikey.hash(), + "session_id": bad_pk, + "url": video, + }, + ) + self.assertContains(r, "Session not found", status_code=400) + + r = self.client.post( + url, + { + "apikey": apikey.hash(), + "session_id": "foo", + "url": video, + }, + ) + self.assertContains(r, "Invalid session_id", status_code=400) + + r = self.client.post( + url, + { + "apikey": apikey.hash(), + "session_id": session.pk, + "url": "foobar", + }, + ) + self.assertContains(r, "Invalid url value: 'foobar'", status_code=400) + + r = self.client.post( + url, {"apikey": apikey.hash(), "session_id": session.pk, "url": video} + ) + self.assertContains(r, "Done", status_code=200) + + recordings = session.recordings() + self.assertEqual(len(recordings), 1) + doc = recordings[0] + self.assertEqual(doc.external_url, video) + event = doc.latest_event() + self.assertEqual(event.by, recman) + + def test_api_set_meetecho_recording_name(self): + url = urlreverse("ietf.meeting.views.api_set_meetecho_recording_name") + recmanrole = RoleFactory(group__type_id="ietf", name_id="recman") + recman = recmanrole.person + meeting = MeetingFactory(type_id="ietf") + session = SessionFactory(group__type_id="wg", meeting=meeting) + apikey = PersonalApiKeyFactory(endpoint=url, person=recman) + name = "testname" + + # error cases + r = self.client.post(url, {}) + self.assertContains(r, "Missing apikey parameter", status_code=400) + + badrole = RoleFactory(group__type_id="ietf", name_id="ad") + badapikey = PersonalApiKeyFactory(endpoint=url, person=badrole.person) + badrole.person.user.last_login = timezone.now() + badrole.person.user.save() + r = self.client.post(url, {"apikey": badapikey.hash()}) + self.assertContains(r, "Restricted to role: Recording Manager", status_code=403) + + r = self.client.post(url, {"apikey": apikey.hash()}) + self.assertContains(r, "Too long since last regular login", status_code=400) + recman.user.last_login = timezone.now() + recman.user.save() + + r = self.client.get(url, {"apikey": apikey.hash()}) + self.assertContains(r, "Method not allowed", status_code=405) + + r = self.client.post(url, {"apikey": apikey.hash()}) + self.assertContains(r, "Missing session_id parameter", status_code=400) + + r = self.client.post(url, {"apikey": apikey.hash(), "session_id": session.pk}) + self.assertContains(r, "Missing name parameter", status_code=400) + + bad_pk = int(Session.objects.order_by("-pk").first().pk) + 1 + r = self.client.post( + url, + { + "apikey": apikey.hash(), + "session_id": bad_pk, + "name": name, + }, + ) + self.assertContains(r, "Session not found", status_code=400) + + r = self.client.post( + url, + { + "apikey": apikey.hash(), + "session_id": "foo", + "name": name, + }, + ) + self.assertContains(r, "Invalid session_id", status_code=400) + + r = self.client.post( + url, {"apikey": apikey.hash(), "session_id": session.pk, "name": name} + ) + self.assertContains(r, "Done", status_code=200) + + session.refresh_from_db() + self.assertEqual(session.meetecho_recording_name, name) + + + def test_api_add_session_attendees_deprecated(self): + # Deprecated test - should be removed when we stop accepting a simple list of user PKs in + # the add_session_attendees() view url = urlreverse('ietf.meeting.views.api_add_session_attendees') otherperson = PersonFactory() recmanrole = RoleFactory(group__type_id='ietf', name_id='recman') recman = recmanrole.person meeting = MeetingFactory(type_id='ietf') session = SessionFactory(group__type_id='wg', meeting=meeting) - apikey = PersonalApiKey.objects.create(endpoint=url, person=recman) + apikey = PersonalApiKeyFactory(endpoint=url, person=recman) badrole = RoleFactory(group__type_id='ietf', name_id='ad') - badapikey = PersonalApiKey.objects.create(endpoint=url, person=badrole.person) + badapikey = PersonalApiKeyFactory(endpoint=url, person=badrole.person) badrole.person.user.last_login = timezone.now() badrole.person.user.save() @@ -215,6 +356,120 @@ def test_api_add_session_attendees(self): self.assertTrue(session.attended_set.filter(person=recman).exists()) self.assertTrue(session.attended_set.filter(person=otherperson).exists()) + def test_api_add_session_attendees(self): + url = urlreverse("ietf.meeting.views.api_add_session_attendees") + otherperson = PersonFactory() + recmanrole = RoleFactory(group__type_id="ietf", name_id="recman") + recman = recmanrole.person + meeting = MeetingFactory(type_id="ietf") + session = SessionFactory(group__type_id="wg", meeting=meeting) + apikey = PersonalApiKeyFactory(endpoint=url, person=recman) + + badrole = RoleFactory(group__type_id="ietf", name_id="ad") + badapikey = PersonalApiKeyFactory(endpoint=url, person=badrole.person) + badrole.person.user.last_login = timezone.now() + badrole.person.user.save() + + # Improper credentials, or method + r = self.client.post(url, {}) + self.assertContains(r, "Missing apikey parameter", status_code=400) + + r = self.client.post(url, {"apikey": badapikey.hash()}) + self.assertContains(r, "Restricted to role: Recording Manager", status_code=403) + + r = self.client.post(url, {"apikey": apikey.hash()}) + self.assertContains(r, "Too long since last regular login", status_code=400) + + recman.user.last_login = timezone.now() - datetime.timedelta(days=365) + recman.user.save() + r = self.client.post(url, {"apikey": apikey.hash()}) + self.assertContains(r, "Too long since last regular login", status_code=400) + + recman.user.last_login = timezone.now() + recman.user.save() + r = self.client.get(url, {"apikey": apikey.hash()}) + self.assertContains(r, "Method not allowed", status_code=405) + + recman.user.last_login = timezone.now() + recman.user.save() + + # Malformed requests + r = self.client.post(url, {"apikey": apikey.hash()}) + self.assertContains(r, "Missing attended parameter", status_code=400) + + for baddict in ( + "{}", + '{"bogons;drop table":"bogons;drop table"}', + '{"session_id":"Not an integer;drop table"}', + f'{{"session_id":{session.pk},"attendees":"not a list;drop table"}}', + f'{{"session_id":{session.pk},"attendees":"not a list;drop table"}}', + f'{{"session_id":{session.pk},"attendees":[1,2,"not an int;drop table",4]}}', + f'{{"session_id":{session.pk},"attendees":["user_id":{recman.user.pk}]}}', # no join_time + f'{{"session_id":{session.pk},"attendees":["user_id":{recman.user.pk},"join_time;drop table":"2024-01-01T00:00:00Z]}}', + f'{{"session_id":{session.pk},"attendees":["user_id":{recman.user.pk},"join_time":"not a time;drop table"]}}', + # next has no time zone indicator + f'{{"session_id":{session.pk},"attendees":["user_id":{recman.user.pk},"join_time":"2024-01-01T00:00:00"]}}', + f'{{"session_id":{session.pk},"attendees":["user_id":"not an int; drop table","join_time":"2024-01-01T00:00:00Z"]}}', + # Uncomment the next one when the _deprecated version of this test is retired + # f'{{"session_id":{session.pk},"attendees":[{recman.user.pk}, {otherperson.user.pk}]}}', + ): + r = self.client.post(url, {"apikey": apikey.hash(), "attended": baddict}) + self.assertContains(r, "Malformed post", status_code=400) + + bad_session_id = Session.objects.order_by("-pk").first().pk + 1 + r = self.client.post( + url, + { + "apikey": apikey.hash(), + "attended": f'{{"session_id":{bad_session_id},"attendees":[]}}', + }, + ) + self.assertContains(r, "Invalid session", status_code=400) + bad_user_id = User.objects.order_by("-pk").first().pk + 1 + r = self.client.post( + url, + { + "apikey": apikey.hash(), + "attended": f'{{"session_id":{session.pk},"attendees":[{{"user_id":{bad_user_id}, "join_time":"2024-01-01T00:00:00Z"}}]}}', + }, + ) + self.assertContains(r, "Invalid attendee", status_code=400) + + # Reasonable request + r = self.client.post( + url, + { + "apikey": apikey.hash(), + "attended": json.dumps( + { + "session_id": session.pk, + "attendees": [ + { + "user_id": recman.user.pk, + "join_time": "2023-09-03T12:34:56Z", + }, + { + "user_id": otherperson.user.pk, + "join_time": "2023-09-03T03:00:19Z", + }, + ], + } + ), + }, + ) + + self.assertEqual(session.attended_set.count(), 2) + self.assertTrue(session.attended_set.filter(person=recman).exists()) + self.assertEqual( + session.attended_set.get(person=recman).time, + datetime.datetime(2023, 9, 3, 12, 34, 56, tzinfo=datetime.UTC), + ) + self.assertTrue(session.attended_set.filter(person=otherperson).exists()) + self.assertEqual( + session.attended_set.get(person=otherperson).time, + datetime.datetime(2023, 9, 3, 3, 0, 19, tzinfo=datetime.UTC), + ) + def test_api_upload_polls_and_chatlog(self): recmanrole = RoleFactory(group__type_id='ietf', name_id='recman') recmanrole.person.user.last_login = timezone.now() @@ -264,8 +519,8 @@ def test_api_upload_polls_and_chatlog(self): ), ): url = urlreverse(f"ietf.meeting.views.api_upload_{type_id}") - apikey = PersonalApiKey.objects.create(endpoint=url, person=recmanrole.person) - badapikey = PersonalApiKey.objects.create(endpoint=url, person=badrole.person) + apikey = PersonalApiKeyFactory(endpoint=url, person=recmanrole.person) + badapikey = PersonalApiKeyFactory(endpoint=url, person=badrole.person) r = self.client.post(url, {}) self.assertContains(r, "Missing apikey parameter", status_code=400) @@ -298,24 +553,27 @@ def test_api_upload_polls_and_chatlog(self): r = self.client.post(url,{'apikey':apikey.hash(),'apidata': f'{{"session_id":{session.pk}, "{type_id}":{content}}}'}) self.assertEqual(r.status_code, 200) - newdoc = session.sessionpresentation_set.get(document__type_id=type_id).document + newdoc = session.presentations.get(document__type_id=type_id).document newdoccontent = get_unicode_document_content(newdoc.name, Path(session.meeting.get_materials_path()) / type_id / newdoc.uploaded_filename) self.assertEqual(json.loads(content), json.loads(newdoccontent)) + self.assertEqual( + json.loads(retrieve_str(type_id, newdoc.uploaded_filename)), + json.loads(content) + ) def test_api_upload_bluesheet(self): - url = urlreverse('ietf.meeting.views.api_upload_bluesheet') - recmanrole = RoleFactory(group__type_id='ietf', name_id='recman') + url = urlreverse("ietf.meeting.views.api_upload_bluesheet") + recmanrole = RoleFactory(group__type_id="ietf", name_id="recman") recman = recmanrole.person - meeting = MeetingFactory(type_id='ietf') - session = SessionFactory(group__type_id='wg', meeting=meeting) - group = session.group - apikey = PersonalApiKey.objects.create(endpoint=url, person=recman) - + meeting = MeetingFactory(type_id="ietf") + session = SessionFactory(group__type_id="wg", meeting=meeting) + apikey = PersonalApiKeyFactory(endpoint=url, person=recman) + people = [ - {"name":"Andrea Andreotti", "affiliation": "Azienda"}, - {"name":"Bosse Bernadotte", "affiliation": "Bolag"}, - {"name":"Charles Charlemagne", "affiliation": "Compagnie"}, - ] + {"name": "Andrea Andreotti", "affiliation": "Azienda"}, + {"name": "Bosse Bernadotte", "affiliation": "Bolag"}, + {"name": "Charles Charlemagne", "affiliation": "Compagnie"}, + ] for i in range(3): faker = random_faker() people.append(dict(name=faker.name(), affiliation=faker.company())) @@ -325,80 +583,93 @@ def test_api_upload_bluesheet(self): r = self.client.post(url, {}) self.assertContains(r, "Missing apikey parameter", status_code=400) - badrole = RoleFactory(group__type_id='ietf', name_id='ad') - badapikey = PersonalApiKey.objects.create(endpoint=url, person=badrole.person) + badrole = RoleFactory(group__type_id="ietf", name_id="ad") + badapikey = PersonalApiKeyFactory(endpoint=url, person=badrole.person) badrole.person.user.last_login = timezone.now() badrole.person.user.save() - r = self.client.post(url, {'apikey': badapikey.hash()} ) - self.assertContains(r, "Restricted to roles: Recording Manager, Secretariat", status_code=403) + r = self.client.post(url, {"apikey": badapikey.hash()}) + self.assertContains( + r, "Restricted to roles: Recording Manager, Secretariat", status_code=403 + ) - r = self.client.post(url, {'apikey': apikey.hash()} ) + r = self.client.post(url, {"apikey": apikey.hash()}) self.assertContains(r, "Too long since last regular login", status_code=400) recman.user.last_login = timezone.now() recman.user.save() - r = self.client.get(url, {'apikey': apikey.hash()} ) + r = self.client.get(url, {"apikey": apikey.hash()}) self.assertContains(r, "Method not allowed", status_code=405) - r = self.client.post(url, {'apikey': apikey.hash()} ) - self.assertContains(r, "Missing meeting parameter", status_code=400) - + r = self.client.post(url, {"apikey": apikey.hash()}) + self.assertContains(r, "Missing session_id parameter", status_code=400) - r = self.client.post(url, {'apikey': apikey.hash(), 'meeting': meeting.number, } ) - self.assertContains(r, "Missing group parameter", status_code=400) - - r = self.client.post(url, {'apikey': apikey.hash(), 'meeting': meeting.number, 'group': group.acronym} ) - self.assertContains(r, "Missing item parameter", status_code=400) - - r = self.client.post(url, {'apikey': apikey.hash(), 'meeting': meeting.number, 'group': group.acronym, 'item': '1'} ) + r = self.client.post(url, {"apikey": apikey.hash(), "session_id": session.pk}) self.assertContains(r, "Missing bluesheet parameter", status_code=400) - r = self.client.post(url, {'apikey': apikey.hash(), 'meeting': '1', 'group': group.acronym, - 'item': '1', 'bluesheet': bluesheet, }) - self.assertContains(r, "No sessions found for meeting", status_code=400) - - r = self.client.post(url, {'apikey': apikey.hash(), 'meeting': meeting.number, 'group': 'bogous', - 'item': '1', 'bluesheet': bluesheet, }) - self.assertContains(r, "No sessions found in meeting '%s' for group 'bogous'"%meeting.number, status_code=400) - - r = self.client.post(url, {'apikey': apikey.hash(), 'meeting': meeting.number, 'group': group.acronym, - 'item': '1', 'bluesheet': "foobar", }) - self.assertContains(r, "Invalid json value: 'foobar'", status_code=400) - - r = self.client.post(url, {'apikey': apikey.hash(), 'meeting': meeting.number, 'group': group.acronym, - 'item': '5', 'bluesheet': bluesheet, }) - self.assertContains(r, "No item '5' found in list of sessions for group", status_code=400) - - r = self.client.post(url, {'apikey': apikey.hash(), 'meeting': meeting.number, 'group': group.acronym, - 'item': 'foo', 'bluesheet': bluesheet, }) - self.assertContains(r, "Expected a numeric value for 'item', found 'foo'", status_code=400) - - r = self.client.post(url, {'apikey': apikey.hash(), 'meeting': meeting.number, 'group': group.acronym, - 'item': '1', 'bluesheet': bluesheet, }) + bad_session_pk = int(Session.objects.order_by("-pk").first().pk) + 1 + r = self.client.post( + url, + { + "apikey": apikey.hash(), + "session_id": bad_session_pk, + "bluesheet": bluesheet, + }, + ) + self.assertContains(r, "Session not found", status_code=400) + + r = self.client.post( + url, + { + "apikey": apikey.hash(), + "session_id": "foo", + "bluesheet": bluesheet, + }, + ) + self.assertContains(r, "Invalid session_id", status_code=400) + + r = self.client.post( + url, + { + "apikey": apikey.hash(), + "session_id": session.pk, + "bluesheet": bluesheet, + }, + ) self.assertContains(r, "Done", status_code=200) # Submit again, with slightly different content, as an updated version - people[1]['affiliation'] = 'Bolaget AB' + people[1]["affiliation"] = "Bolaget AB" bluesheet = json.dumps(people) - r = self.client.post(url, {'apikey': apikey.hash(), 'meeting': meeting.number, 'group': group.acronym, - 'item': '1', 'bluesheet': bluesheet, }) + r = self.client.post( + url, + { + "apikey": apikey.hash(), + "session_id": session.pk, + "bluesheet": bluesheet, + }, + ) self.assertContains(r, "Done", status_code=200) - bluesheet = session.sessionpresentation_set.filter(document__type__slug='bluesheets').first().document + bluesheet = ( + session.presentations.filter(document__type__slug="bluesheets") + .first() + .document + ) # We've submitted an update; check that the rev is right - self.assertEqual(bluesheet.rev, '01') + self.assertEqual(bluesheet.rev, "01") # Check the content with open(bluesheet.get_file_name()) as file: text = file.read() for p in people: - self.assertIn(p['name'], html.unescape(text)) - self.assertIn(p['affiliation'], html.unescape(text)) + self.assertIn(p["name"], html.unescape(text)) + self.assertIn(p["affiliation"], html.unescape(text)) def test_person_export(self): person = PersonFactory() url = urlreverse('ietf.api.views.PersonalInformationExportView') login_testing_unauthorized(self, person.user.username, url) r = self.client.get(url) + self.assertEqual(r.status_code, 200) jsondata = r.json() data = jsondata['person.person'][str(person.id)] self.assertEqual(data['name'], person.name) @@ -409,14 +680,14 @@ def test_api_v2_person_export_view(self): url = urlreverse('ietf.api.views.ApiV2PersonExportView') robot = PersonFactory(user__is_staff=True) RoleFactory(name_id='robot', person=robot, email=robot.email(), group__acronym='secretariat') - apikey = PersonalApiKey.objects.create(endpoint=url, person=robot) + apikey = PersonalApiKeyFactory(endpoint=url, person=robot) # error cases r = self.client.post(url, {}) self.assertContains(r, "Missing apikey parameter", status_code=400) badrole = RoleFactory(group__type_id='ietf', name_id='ad') - badapikey = PersonalApiKey.objects.create(endpoint=url, person=badrole.person) + badapikey = PersonalApiKeyFactory(endpoint=url, person=badrole.person) badrole.person.user.last_login = timezone.now() badrole.person.user.save() r = self.client.post(url, {'apikey': badapikey.hash()}) @@ -434,86 +705,179 @@ def test_api_v2_person_export_view(self): self.assertEqual(data['ascii'], robot.ascii) self.assertEqual(data['user']['email'], robot.user.email) - def test_api_new_meeting_registration(self): + @override_settings(APP_API_TOKENS={"ietf.api.views.api_new_meeting_registration_v2": ["valid-token"]}) + def test_api_new_meeting_registration_v2(self): meeting = MeetingFactory(type_id='ietf') - reg = { - 'apikey': 'invalid', + person = PersonFactory() + reg_detail = { + 'email': person.email().address, + 'first_name': person.first_name(), + 'last_name': person.last_name(), + 'meeting': meeting.number, 'affiliation': "Alguma Corporação", 'country_code': 'PT', - 'email': 'foo@example.pt', - 'first_name': 'Foo', - 'last_name': 'Bar', - 'meeting': meeting.number, - 'reg_type': 'hackathon', - 'ticket_type': '', - 'checkedin': 'False', + 'checkedin': False, + 'is_nomcom_volunteer': False, + 'cancelled': False, + 'tickets': [{'attendance_type': 'onsite', 'ticket_type': 'week_pass'}], } - url = urlreverse('ietf.api.views.api_new_meeting_registration') - r = self.client.post(url, reg) - self.assertContains(r, 'Invalid apikey', status_code=403) - oidcp = PersonFactory(user__is_staff=True) - # Make sure 'oidcp' has an acceptable role - RoleFactory(name_id='robot', person=oidcp, email=oidcp.email(), group__acronym='secretariat') - key = PersonalApiKey.objects.create(person=oidcp, endpoint=url) - reg['apikey'] = key.hash() - # - # Test valid POST - # FIXME: sometimes, there seems to be something in the outbox? - old_len = len(outbox) - r = self.client.post(url, reg) - self.assertContains(r, "Accepted, New registration, Email sent", status_code=202) - # - # Check outgoing mail - self.assertEqual(len(outbox), old_len + 1) - body = get_payload_text(outbox[-1]) - self.assertIn(reg['email'], outbox[-1]['To'] ) - self.assertIn(reg['email'], body) - self.assertIn('account creation request', body) + reg_data = {'objects': {reg_detail['email']: reg_detail}} + url = urlreverse('ietf.api.views.api_new_meeting_registration_v2') # - # Check record - obj = MeetingRegistration.objects.get(email=reg['email'], meeting__number=reg['meeting']) - for key in ['affiliation', 'country_code', 'first_name', 'last_name', 'person', 'reg_type', 'ticket_type', 'checkedin']: - self.assertEqual(getattr(obj, key), False if key=='checkedin' else reg.get(key) , "Bad data for field '%s'" % key) + # Test invalid key + r = self.client.post(url, data=json.dumps(reg_data), content_type='application/json', headers={"X-Api-Key": "invalid-token"}) + self.assertEqual(r.status_code, 403) # - # Test with existing user - person = PersonFactory() - reg['email'] = person.email().address - reg['first_name'] = person.first_name() - reg['last_name'] = person.last_name() + # Test invalid data + bad_reg_data = copy.deepcopy(reg_data) + del bad_reg_data['objects'][reg_detail['email']]['email'] + r = self.client.post(url, data=json.dumps(bad_reg_data), content_type='application/json', headers={"X-Api-Key": "valid-token"}) + self.assertEqual(r.status_code, 400) # - r = self.client.post(url, reg) - self.assertContains(r, "Accepted, New registration", status_code=202) + # Test valid POST + r = self.client.post(url, data=json.dumps(reg_data), content_type='application/json', headers={"X-Api-Key": "valid-token"}) + self.assertContains(r, "Success", status_code=202) # - # There should be no new outgoing mail - self.assertEqual(len(outbox), old_len + 1) + # Check record + objects = Registration.objects.filter(email=reg_detail['email'], meeting__number=reg_detail['meeting']) + self.assertEqual(objects.count(), 1) + obj = objects[0] + for key in ['affiliation', 'country_code', 'first_name', 'last_name', 'checkedin']: + self.assertEqual(getattr(obj, key), False if key == 'checkedin' else reg_detail.get(key), f"Bad data for field {key}") + self.assertEqual(obj.tickets.count(), 1) + ticket = obj.tickets.first() + self.assertEqual(ticket.ticket_type.slug, reg_detail['tickets'][0]['ticket_type']) + self.assertEqual(ticket.attendance_type.slug, reg_detail['tickets'][0]['attendance_type']) + self.assertEqual(obj.person, person) # - # Test multiple reg types - reg['reg_type'] = 'remote' - reg['ticket_type'] = 'full_week_pass' - r = self.client.post(url, reg) - self.assertContains(r, "Accepted, New registration", status_code=202) - objs = MeetingRegistration.objects.filter(email=reg['email'], meeting__number=reg['meeting']) - self.assertEqual(len(objs), 2) - self.assertEqual(objs.filter(reg_type='hackathon').count(), 1) - self.assertEqual(objs.filter(reg_type='remote', ticket_type='full_week_pass').count(), 1) - self.assertEqual(len(outbox), old_len + 1) + # Test update (switch to remote) + reg_detail = { + 'affiliation': "Alguma Corporação", + 'country_code': 'PT', + 'email': person.email().address, + 'first_name': person.first_name(), + 'last_name': person.last_name(), + 'meeting': meeting.number, + 'checkedin': False, + 'is_nomcom_volunteer': False, + 'cancelled': False, + 'tickets': [{'attendance_type': 'remote', 'ticket_type': 'week_pass'}], + } + reg_data = {'objects': {reg_detail['email']: reg_detail}} + r = self.client.post(url, data=json.dumps(reg_data), content_type='application/json', headers={"X-Api-Key": "valid-token"}) + self.assertContains(r, "Success", status_code=202) + objects = Registration.objects.filter(email=reg_detail['email'], meeting__number=reg_detail['meeting']) + self.assertEqual(objects.count(), 1) + obj = objects[0] + self.assertEqual(obj.tickets.count(), 1) + ticket = obj.tickets.first() + self.assertEqual(ticket.ticket_type.slug, reg_detail['tickets'][0]['ticket_type']) + self.assertEqual(ticket.attendance_type.slug, reg_detail['tickets'][0]['attendance_type']) # - # Test incomplete POST - drop_fields = ['affiliation', 'first_name', 'reg_type'] - for field in drop_fields: - del reg[field] - r = self.client.post(url, reg) - self.assertContains(r, 'Missing parameters:', status_code=400) - err, fields = r.content.decode().split(':', 1) - missing_fields = [f.strip() for f in fields.split(',')] - self.assertEqual(set(missing_fields), set(drop_fields)) + # Test multiple + reg_detail = { + 'affiliation': "Alguma Corporação", + 'country_code': 'PT', + 'email': person.email().address, + 'first_name': person.first_name(), + 'last_name': person.last_name(), + 'meeting': meeting.number, + 'checkedin': False, + 'is_nomcom_volunteer': False, + 'cancelled': False, + 'tickets': [ + {'attendance_type': 'onsite', 'ticket_type': 'one_day'}, + {'attendance_type': 'remote', 'ticket_type': 'week_pass'}, + ], + } + reg_data = {'objects': {reg_detail['email']: reg_detail}} + r = self.client.post(url, data=json.dumps(reg_data), content_type='application/json', headers={"X-Api-Key": "valid-token"}) + self.assertContains(r, "Success", status_code=202) + objects = Registration.objects.filter(email=reg_detail['email'], meeting__number=reg_detail['meeting']) + self.assertEqual(objects.count(), 1) + obj = objects[0] + self.assertEqual(obj.tickets.count(), 2) + self.assertEqual(obj.tickets.filter(attendance_type__slug='onsite').count(), 1) + self.assertEqual(obj.tickets.filter(attendance_type__slug='remote').count(), 1) + + @override_settings(APP_API_TOKENS={"ietf.api.views.api_new_meeting_registration_v2": ["valid-token"]}) + def test_api_new_meeting_registration_v2_cancelled(self): + meeting = MeetingFactory(type_id='ietf') + person = PersonFactory() + reg_detail = { + 'affiliation': "Acme", + 'country_code': 'US', + 'email': person.email().address, + 'first_name': person.first_name(), + 'last_name': person.last_name(), + 'meeting': meeting.number, + 'checkedin': False, + 'is_nomcom_volunteer': False, + 'cancelled': False, + 'tickets': [{'attendance_type': 'onsite', 'ticket_type': 'week_pass'}], + } + reg_data = {'objects': {reg_detail['email']: reg_detail}} + url = urlreverse('ietf.api.views.api_new_meeting_registration_v2') + self.assertEqual(Registration.objects.count(), 0) + r = self.client.post(url, data=json.dumps(reg_data), content_type='application/json', headers={"X-Api-Key": "valid-token"}) + self.assertContains(r, "Success", status_code=202) + self.assertEqual(Registration.objects.count(), 1) + reg_detail['cancelled'] = True + r = self.client.post(url, data=json.dumps(reg_data), content_type='application/json', headers={"X-Api-Key": "valid-token"}) + self.assertContains(r, "Success", status_code=202) + self.assertEqual(Registration.objects.count(), 0) + + @override_settings(APP_API_TOKENS={"ietf.api.views.api_new_meeting_registration_v2": ["valid-token"]}) + def test_api_new_meeting_registration_v2_nomcom(self): + meeting = MeetingFactory(type_id='ietf') + person = PersonFactory() + reg_detail = { + 'affiliation': "Acme", + 'country_code': 'US', + 'email': person.email().address, + 'first_name': person.first_name(), + 'last_name': person.last_name(), + 'meeting': meeting.number, + 'checkedin': False, + 'is_nomcom_volunteer': False, + 'cancelled': False, + 'tickets': [{'attendance_type': 'onsite', 'ticket_type': 'week_pass'}], + } + reg_data = {'objects': {reg_detail['email']: reg_detail}} + url = urlreverse('ietf.api.views.api_new_meeting_registration_v2') + now = datetime.datetime.now() + if now.month > 10: + year = now.year + 1 + else: + year = now.year + # create appropriate group and nomcom objects + nomcom = NomComFactory.create(is_accepting_volunteers=True, **nomcom_kwargs_for_year(year)) + + # first test is_nomcom_volunteer False + r = self.client.post(url, data=json.dumps(reg_data), content_type='application/json', headers={"X-Api-Key": "valid-token"}) + self.assertContains(r, "Success", status_code=202) + # assert no Volunteers exists + self.assertEqual(Volunteer.objects.count(), 0) + + # test is_nomcom_volunteer True + reg_detail['is_nomcom_volunteer'] = True + r = self.client.post(url, data=json.dumps(reg_data), content_type='application/json', headers={"X-Api-Key": "valid-token"}) + self.assertContains(r, "Success", status_code=202) + # assert Volunteer exists + self.assertEqual(Volunteer.objects.count(), 1) + volunteer = Volunteer.objects.last() + self.assertEqual(volunteer.person, person) + self.assertEqual(volunteer.nomcom, nomcom) + self.assertEqual(volunteer.origin, 'registration') def test_api_version(self): - DumpInfo.objects.create(date=timezone.datetime(2022,8,31,7,10,1,tzinfo=timezone.utc), host='testapi.example.com',tz='UTC') + DumpInfo.objects.create(date=timezone.datetime(2022,8,31,7,10,1,tzinfo=datetime.UTC), host='testapi.example.com',tz='UTC') url = urlreverse('ietf.api.views.version') r = self.client.get(url) data = r.json() self.assertEqual(data['version'], ietf.__version__+ietf.__patch__) + for lib in settings.ADVERTISE_VERSIONS: + self.assertIn(lib, data['other']) self.assertEqual(data['dumptime'], "2022-08-31 07:10:01 +0000") DumpInfo.objects.update(tz='PST8PDT') r = self.client.get(url) @@ -522,30 +886,618 @@ def test_api_version(self): def test_api_appauth(self): - url = urlreverse('ietf.api.views.app_auth') - person = PersonFactory() - apikey = PersonalApiKey.objects.create(endpoint=url, person=person) + for app in ["authortools", "bibxml"]: + url = urlreverse('ietf.api.views.app_auth', kwargs={"app": app}) + person = PersonFactory() + apikey = PersonalApiKeyFactory(endpoint=url, person=person) + + self.client.login(username=person.user.username,password=f'{person.user.username}+password') + self.client.logout() + + # error cases + # missing apikey + r = self.client.post(url, {}) + self.assertContains(r, 'Missing apikey parameter', status_code=400) + + # invalid apikey + r = self.client.post(url, {'apikey': 'foobar'}) + self.assertContains(r, 'Invalid apikey', status_code=403) + + # working case + r = self.client.post(url, {'apikey': apikey.hash()}) + self.assertEqual(r.status_code, 200) + jsondata = r.json() + self.assertEqual(jsondata['success'], True) + self.client.logout() - self.client.login(username=person.user.username,password=f'{person.user.username}+password') - self.client.logout() + @override_settings(APP_API_TOKENS={"ietf.api.views.nfs_metrics": ["valid-token"]}) + def test_api_nfs_metrics(self): + url = urlreverse("ietf.api.views.nfs_metrics") + r = self.client.get(url) + self.assertEqual(r.status_code, 403) + r = self.client.get(url, headers={"X-Api-Key": "valid-token"}) + self.assertContains(r, 'nfs_latency_seconds{operation="write"}') - # error cases - # missing apikey - r = self.client.post(url, {}) - self.assertContains(r, 'Missing apikey parameter', status_code=400) + def test_api_get_session_matherials_no_agenda_meeting_url(self): + meeting = MeetingFactory(type_id='ietf') + session = SessionFactory(meeting=meeting) + url = urlreverse('ietf.meeting.views.api_get_session_materials', kwargs={'session_id': session.pk}) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + + @override_settings(APP_API_TOKENS={"ietf.api.views.draft_aliases": ["valid-token"]}) + @mock.patch("ietf.api.views.DraftAliasGenerator") + def test_draft_aliases(self, mock): + mock.return_value = (("alias1", ("a1", "a2")), ("alias2", ("a3", "a4"))) + url = urlreverse("ietf.api.views.draft_aliases") + r = self.client.get(url, headers={"X-Api-Key": "valid-token"}) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-type"], "application/json") + self.assertEqual( + json.loads(r.content), + { + "aliases": [ + {"alias": "alias1", "domains": ["ietf"], "addresses": ["a1", "a2"]}, + {"alias": "alias2", "domains": ["ietf"], "addresses": ["a3", "a4"]}, + ]} + ) + # some invalid cases + self.assertEqual( + self.client.get(url, headers={}).status_code, + 403, + ) + self.assertEqual( + self.client.get(url, headers={"X-Api-Key": "something-else"}).status_code, + 403, + ) + self.assertEqual( + self.client.post(url, headers={"X-Api-Key": "something-else"}).status_code, + 403, + ) + self.assertEqual( + self.client.post(url, headers={"X-Api-Key": "valid-token"}).status_code, + 405, + ) + + @override_settings(APP_API_TOKENS={"ietf.api.views.group_aliases": ["valid-token"]}) + @mock.patch("ietf.api.views.GroupAliasGenerator") + def test_group_aliases(self, mock): + mock.return_value = (("alias1", ("ietf",), ("a1", "a2")), ("alias2", ("ietf", "iab"), ("a3", "a4"))) + url = urlreverse("ietf.api.views.group_aliases") + r = self.client.get(url, headers={"X-Api-Key": "valid-token"}) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-type"], "application/json") + self.assertEqual( + json.loads(r.content), + { + "aliases": [ + {"alias": "alias1", "domains": ["ietf"], "addresses": ["a1", "a2"]}, + {"alias": "alias2", "domains": ["ietf", "iab"], "addresses": ["a3", "a4"]}, + ]} + ) + # some invalid cases + self.assertEqual( + self.client.get(url, headers={}).status_code, + 403, + ) + self.assertEqual( + self.client.get(url, headers={"X-Api-Key": "something-else"}).status_code, + 403, + ) + self.assertEqual( + self.client.post(url, headers={"X-Api-Key": "something-else"}).status_code, + 403, + ) + self.assertEqual( + self.client.post(url, headers={"X-Api-Key": "valid-token"}).status_code, + 405, + ) + + @override_settings(APP_API_TOKENS={"ietf.api.views.active_email_list": ["valid-token"]}) + def test_active_email_list(self): + EmailFactory(active=True) # make sure there's at least one active email... + EmailFactory(active=False) # ... and at least one non-active emai + url = urlreverse("ietf.api.views.active_email_list") + r = self.client.post(url, headers={}) + self.assertEqual(r.status_code, 403) + r = self.client.get(url, headers={}) + self.assertEqual(r.status_code, 403) + r = self.client.get(url, headers={"X-Api-Key": "not-the-valid-token"}) + self.assertEqual(r.status_code, 403) + r = self.client.post(url, headers={"X-Api-Key": "not-the-valid-token"}) + self.assertEqual(r.status_code, 403) + r = self.client.post(url, headers={"X-Api-Key": "valid-token"}) + self.assertEqual(r.status_code, 405) + r = self.client.get(url, headers={"X-Api-Key": "valid-token"}) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-Type"], "application/json") + result = json.loads(r.content) + self.assertCountEqual(result.keys(), ["addresses"]) + self.assertCountEqual(result["addresses"], Email.objects.filter(active=True).values_list("address", flat=True)) + + @override_settings(APP_API_TOKENS={"ietf.api.views.related_email_list": ["valid-token"]}) + def test_related_email_list(self): + joe = EmailFactory(address='joe@work.com') + EmailFactory(address='joe@home.com', person=joe.person) + EmailFactory(address='jòe@spain.com', person=joe.person) + url = urlreverse("ietf.api.views.related_email_list", kwargs={'email': 'joe@home.com'}) + # no api key + r = self.client.get(url, headers={}) + self.assertEqual(r.status_code, 403) + # invalid api key + r = self.client.get(url, headers={"X-Api-Key": "not-the-valid-token"}) + self.assertEqual(r.status_code, 403) + # wrong method + r = self.client.post(url, headers={"X-Api-Key": "valid-token"}) + self.assertEqual(r.status_code, 405) + # valid + r = self.client.get(url, headers={"X-Api-Key": "valid-token"}) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-Type"], "application/json") + result = json.loads(r.content) + self.assertCountEqual(result.keys(), ["addresses"]) + self.assertCountEqual(result["addresses"], joe.person.email_set.values_list("address", flat=True)) + # non-ascii + non_ascii_url = urlreverse("ietf.api.views.related_email_list", kwargs={'email': 'jòe@spain.com'}) + r = self.client.get(non_ascii_url, headers={"X-Api-Key": "valid-token"}) + self.assertEqual(r.status_code, 200) + result = json.loads(r.content) + self.assertTrue('joe@home.com' in result["addresses"]) + # email not found + not_found_url = urlreverse("ietf.api.views.related_email_list", kwargs={'email': 'nobody@nowhere.com'}) + r = self.client.get(not_found_url, headers={"X-Api-Key": "valid-token"}) + self.assertEqual(r.status_code, 404) + + @override_settings(APP_API_TOKENS={"ietf.api.views.role_holder_addresses": ["valid-token"]}) + def test_role_holder_addresses(self): + url = urlreverse("ietf.api.views.role_holder_addresses") + r = self.client.get(url, headers={}) + self.assertEqual(r.status_code, 403, "No api token, no access") + r = self.client.get(url, headers={"X-Api-Key": "not-valid-token"}) + self.assertEqual(r.status_code, 403, "Bad api token, no access") + r = self.client.post(url, headers={"X-Api-Key": "valid-token"}) + self.assertEqual(r.status_code, 405, "Bad method, no access") + + emails = EmailFactory.create_batch(5) + email_queryset = Email.objects.filter(pk__in=[e.pk for e in emails]) + with mock.patch("ietf.api.views.role_holder_emails", return_value=email_queryset): + r = self.client.get(url, headers={"X-Api-Key": "valid-token"}) + self.assertEqual(r.status_code, 200, "Good api token and method, access") + content_dict = json.loads(r.content) + self.assertCountEqual(content_dict.keys(), ["addresses"]) + self.assertEqual( + content_dict["addresses"], + sorted(e.address for e in emails), + ) + + @override_settings( + APP_API_TOKENS={"ietf.api.views.ingest_email": "valid-token", "ietf.api.views.ingest_email_test": "test-token"} + ) + @mock.patch("ietf.api.views.iana_ingest_review_email") + @mock.patch("ietf.api.views.ipr_ingest_response_email") + @mock.patch("ietf.api.views.nomcom_ingest_feedback_email") + def test_ingest_email( + self, mock_nomcom_ingest, mock_ipr_ingest, mock_iana_ingest + ): + mocks = {mock_nomcom_ingest, mock_ipr_ingest, mock_iana_ingest} + empty_outbox() + url = urlreverse("ietf.api.views.ingest_email") + test_mode_url = urlreverse("ietf.api.views.ingest_email_test") + + # test various bad calls + r = self.client.get(url) + self.assertEqual(r.status_code, 403) + self.assertFalse(any(m.called for m in mocks)) + r = self.client.get(test_mode_url) + self.assertEqual(r.status_code, 403) + self.assertFalse(any(m.called for m in mocks)) + + r = self.client.post(url) + self.assertEqual(r.status_code, 403) + self.assertFalse(any(m.called for m in mocks)) + r = self.client.post(test_mode_url) + self.assertEqual(r.status_code, 403) + self.assertFalse(any(m.called for m in mocks)) + + r = self.client.get(url, headers={"X-Api-Key": "valid-token"}) + self.assertEqual(r.status_code, 405) + self.assertFalse(any(m.called for m in mocks)) + r = self.client.get(test_mode_url, headers={"X-Api-Key": "test-token"}) + self.assertEqual(r.status_code, 405) + self.assertFalse(any(m.called for m in mocks)) + + r = self.client.post(url, headers={"X-Api-Key": "valid-token"}) + self.assertEqual(r.status_code, 415) + self.assertFalse(any(m.called for m in mocks)) + r = self.client.post(test_mode_url, headers={"X-Api-Key": "test-token"}) + self.assertEqual(r.status_code, 415) + self.assertFalse(any(m.called for m in mocks)) + + r = self.client.post( + url, content_type="application/json", headers={"X-Api-Key": "valid-token"} + ) + self.assertEqual(r.status_code, 400) + self.assertFalse(any(m.called for m in mocks)) + r = self.client.post( + test_mode_url, content_type="application/json", headers={"X-Api-Key": "test-token"} + ) + self.assertEqual(r.status_code, 400) + self.assertFalse(any(m.called for m in mocks)) + + r = self.client.post( + url, + "this is not JSON!", + content_type="application/json", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 400) + self.assertFalse(any(m.called for m in mocks)) + r = self.client.post( + test_mode_url, + "this is not JSON!", + content_type="application/json", + headers={"X-Api-Key": "test-token"}, + ) + self.assertEqual(r.status_code, 400) + self.assertFalse(any(m.called for m in mocks)) + + r = self.client.post( + url, + {"json": "yes", "valid_schema": False}, + content_type="application/json", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 400) + self.assertFalse(any(m.called for m in mocks)) + r = self.client.post( + test_mode_url, + {"json": "yes", "valid_schema": False}, + content_type="application/json", + headers={"X-Api-Key": "test-token"}, + ) + self.assertEqual(r.status_code, 400) + self.assertFalse(any(m.called for m in mocks)) + + # bad destination + message_b64 = base64.b64encode(b"This is a message").decode() + r = self.client.post( + url, + {"dest": "not-a-destination", "message": message_b64}, + content_type="application/json", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-Type"], "application/json") + self.assertEqual(json.loads(r.content), {"result": "bad_dest"}) + self.assertFalse(any(m.called for m in mocks)) + r = self.client.post( + test_mode_url, + {"dest": "not-a-destination", "message": message_b64}, + content_type="application/json", + headers={"X-Api-Key": "test-token"}, + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-Type"], "application/json") + self.assertEqual(json.loads(r.content), {"result": "bad_dest"}) + self.assertFalse(any(m.called for m in mocks)) + + # test that valid requests call handlers appropriately + r = self.client.post( + url, + {"dest": "iana-review", "message": message_b64}, + content_type="application/json", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-Type"], "application/json") + self.assertEqual(json.loads(r.content), {"result": "ok"}) + self.assertTrue(mock_iana_ingest.called) + self.assertEqual(mock_iana_ingest.call_args, mock.call(b"This is a message")) + self.assertFalse(any(m.called for m in (mocks - {mock_iana_ingest}))) + mock_iana_ingest.reset_mock() + + # the test mode endpoint should _not_ call the handler + r = self.client.post( + test_mode_url, + {"dest": "iana-review", "message": message_b64}, + content_type="application/json", + headers={"X-Api-Key": "test-token"}, + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-Type"], "application/json") + self.assertEqual(json.loads(r.content), {"result": "ok"}) + self.assertFalse(any(m.called for m in mocks)) + mock_iana_ingest.reset_mock() + + r = self.client.post( + url, + {"dest": "ipr-response", "message": message_b64}, + content_type="application/json", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-Type"], "application/json") + self.assertEqual(json.loads(r.content), {"result": "ok"}) + self.assertTrue(mock_ipr_ingest.called) + self.assertEqual(mock_ipr_ingest.call_args, mock.call(b"This is a message")) + self.assertFalse(any(m.called for m in (mocks - {mock_ipr_ingest}))) + mock_ipr_ingest.reset_mock() + + # the test mode endpoint should _not_ call the handler + r = self.client.post( + test_mode_url, + {"dest": "ipr-response", "message": message_b64}, + content_type="application/json", + headers={"X-Api-Key": "test-token"}, + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-Type"], "application/json") + self.assertEqual(json.loads(r.content), {"result": "ok"}) + self.assertFalse(any(m.called for m in mocks)) + mock_ipr_ingest.reset_mock() + + # bad nomcom-feedback dest + for bad_nomcom_dest in [ + "nomcom-feedback", # no suffix + "nomcom-feedback-", # no year + "nomcom-feedback-squid", # not a year, + "nomcom-feedback-2024-2025", # also not a year + ]: + r = self.client.post( + url, + {"dest": bad_nomcom_dest, "message": message_b64}, + content_type="application/json", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-Type"], "application/json") + self.assertEqual(json.loads(r.content), {"result": "bad_dest"}) + self.assertFalse(any(m.called for m in mocks)) + r = self.client.post( + test_mode_url, + {"dest": bad_nomcom_dest, "message": message_b64}, + content_type="application/json", + headers={"X-Api-Key": "test-token"}, + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-Type"], "application/json") + self.assertEqual(json.loads(r.content), {"result": "bad_dest"}) + self.assertFalse(any(m.called for m in mocks)) + + # good nomcom-feedback dest + random_year = randrange(100000) + r = self.client.post( + url, + {"dest": f"nomcom-feedback-{random_year}", "message": message_b64}, + content_type="application/json", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-Type"], "application/json") + self.assertEqual(json.loads(r.content), {"result": "ok"}) + self.assertTrue(mock_nomcom_ingest.called) + self.assertEqual(mock_nomcom_ingest.call_args, mock.call(b"This is a message", random_year)) + self.assertFalse(any(m.called for m in (mocks - {mock_nomcom_ingest}))) + mock_nomcom_ingest.reset_mock() + + # the test mode endpoint should _not_ call the handler + r = self.client.post( + test_mode_url, + {"dest": f"nomcom-feedback-{random_year}", "message": message_b64}, + content_type="application/json", + headers={"X-Api-Key": "test-token"}, + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-Type"], "application/json") + self.assertEqual(json.loads(r.content), {"result": "ok"}) + self.assertFalse(any(m.called for m in mocks)) + mock_nomcom_ingest.reset_mock() + + # test that exceptions lead to email being sent - assumes that iana-review handling is representative + mock_iana_ingest.side_effect = EmailIngestionError("Error: don't send email") + r = self.client.post( + url, + {"dest": "iana-review", "message": message_b64}, + content_type="application/json", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-Type"], "application/json") + self.assertEqual(json.loads(r.content), {"result": "bad_msg"}) + self.assertTrue(mock_iana_ingest.called) + self.assertEqual(mock_iana_ingest.call_args, mock.call(b"This is a message")) + self.assertFalse(any(m.called for m in (mocks - {mock_iana_ingest}))) + self.assertEqual(len(outbox), 0) # implicitly tests that _none_ of the earlier tests sent email + mock_iana_ingest.reset_mock() + + # test default recipients and attached original message + mock_iana_ingest.side_effect = EmailIngestionError( + "Error: do send email", + email_body="This is my email\n", + email_original_message=b"This is the original message" + ) + with override_settings(ADMINS=[("Some Admin", "admin@example.com")]): + r = self.client.post( + url, + {"dest": "iana-review", "message": message_b64}, + content_type="application/json", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-Type"], "application/json") + self.assertEqual(json.loads(r.content), {"result": "bad_msg"}) + self.assertTrue(mock_iana_ingest.called) + self.assertEqual(mock_iana_ingest.call_args, mock.call(b"This is a message")) + self.assertFalse(any(m.called for m in (mocks - {mock_iana_ingest}))) + self.assertEqual(len(outbox), 1) + self.assertIn("admin@example.com", outbox[0]["To"]) + self.assertEqual("Error: do send email", outbox[0]["Subject"]) + self.assertEqual("This is my email\n", get_payload_text(outbox[0].get_body())) + attachments = list(a for a in outbox[0].iter_attachments()) + self.assertEqual(len(attachments), 1) + self.assertEqual(attachments[0].get_filename(), "original-message") + self.assertEqual(attachments[0].get_content_type(), "application/octet-stream") + self.assertEqual(attachments[0].get_content(), b"This is the original message") + mock_iana_ingest.reset_mock() + empty_outbox() + + # test overridden recipients and no attached original message + mock_iana_ingest.side_effect = EmailIngestionError( + "Error: do send email", + email_body="This is my email\n", + email_recipients=("thatguy@example.com") + ) + with override_settings(ADMINS=[("Some Admin", "admin@example.com")]): + r = self.client.post( + url, + {"dest": "iana-review", "message": message_b64}, + content_type="application/json", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-Type"], "application/json") + self.assertEqual(json.loads(r.content), {"result": "bad_msg"}) + self.assertTrue(mock_iana_ingest.called) + self.assertEqual(mock_iana_ingest.call_args, mock.call(b"This is a message")) + self.assertFalse(any(m.called for m in (mocks - {mock_iana_ingest}))) + self.assertEqual(len(outbox), 1) + self.assertNotIn("admin@example.com", outbox[0]["To"]) + self.assertIn("thatguy@example.com", outbox[0]["To"]) + self.assertEqual("Error: do send email", outbox[0]["Subject"]) + self.assertEqual("This is my email\n", get_payload_text(outbox[0])) + mock_iana_ingest.reset_mock() + empty_outbox() + + # test attached traceback + mock_iana_ingest.side_effect = EmailIngestionError( + "Error: do send email", + email_body="This is my email\n", + email_attach_traceback=True, + ) + with override_settings(ADMINS=[("Some Admin", "admin@example.com")]): + r = self.client.post( + url, + {"dest": "iana-review", "message": message_b64}, + content_type="application/json", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-Type"], "application/json") + self.assertEqual(json.loads(r.content), {"result": "bad_msg"}) + self.assertTrue(mock_iana_ingest.called) + self.assertEqual(mock_iana_ingest.call_args, mock.call(b"This is a message")) + self.assertFalse(any(m.called for m in (mocks - {mock_iana_ingest}))) + self.assertEqual(len(outbox), 1) + self.assertIn("admin@example.com", outbox[0]["To"]) + self.assertEqual("Error: do send email", outbox[0]["Subject"]) + self.assertEqual("This is my email\n", get_payload_text(outbox[0].get_body())) + attachments = list(a for a in outbox[0].iter_attachments()) + self.assertEqual(len(attachments), 1) + self.assertEqual(attachments[0].get_filename(), "traceback.txt") + self.assertEqual(attachments[0].get_content_type(), "text/plain") + self.assertIn("ietf.api.views.EmailIngestionError: Error: do send email", attachments[0].get_content()) + mock_iana_ingest.reset_mock() + empty_outbox() + + +class DirectAuthApiTests(TestCase): + + def setUp(self): + super().setUp() + self.valid_token = "nSZJDerbau6WZwbEAYuQ" + self.invalid_token = self.valid_token + while self.invalid_token == self.valid_token: + self.invalid_token = User.objects.make_random_password(20) + self.url = urlreverse("ietf.api.views.directauth") + self.valid_person = PersonFactory() + self.valid_password = self.valid_person.user.username+"+password" + self.invalid_password = self.valid_password + while self.invalid_password == self.valid_password: + self.invalid_password = User.objects.make_random_password(20) + + self.valid_body_with_good_password = self.post_dict(authtoken=self.valid_token, username=self.valid_person.user.username, password=self.valid_password) + self.valid_body_with_bad_password = self.post_dict(authtoken=self.valid_token, username=self.valid_person.user.username, password=self.invalid_password) + self.valid_body_with_unknown_user = self.post_dict(authtoken=self.valid_token, username="notauser@nowhere.nada", password=self.valid_password) + + def post_dict(self, authtoken, username, password): + data = dict() + if authtoken is not None: + data["authtoken"] = authtoken + if username is not None: + data["username"] = username + if password is not None: + data["password"] = password + return dict(data = json.dumps(data)) + + def response_data(self, response): + try: + data = json.loads(response.content) + except json.decoder.JSONDecodeError: + data = None + self.assertIsNotNone(data) + return data + + def test_bad_methods(self): + for method in (self.client.get, self.client.put, self.client.head, self.client.delete, self.client.patch): + r = method(self.url) + self.assertEqual(r.status_code, 405) + + def test_bad_post(self): + for bad in [ + self.post_dict(authtoken=None, username=self.valid_person.user.username, password=self.valid_password), + self.post_dict(authtoken=self.valid_token, username=None, password=self.valid_password), + self.post_dict(authtoken=self.valid_token, username=self.valid_person.user.username, password=None), + self.post_dict(authtoken=None, username=None, password=self.valid_password), + self.post_dict(authtoken=self.valid_token, username=None, password=None), + self.post_dict(authtoken=None, username=self.valid_person.user.username, password=None), + self.post_dict(authtoken=None, username=None, password=None), + ]: + r = self.client.post(self.url, bad) + self.assertEqual(r.status_code, 200) + data = self.response_data(r) + self.assertEqual(data["result"], "failure") + self.assertEqual(data["reason"], "invalid post") - # invalid apikey - r = self.client.post(url, {'apikey': 'foobar'}) - self.assertContains(r, 'Invalid apikey', status_code=403) + bad = dict(authtoken=self.valid_token, username=self.valid_person.user.username, password=self.valid_password) + r = self.client.post(self.url, bad) + self.assertEqual(r.status_code, 200) + data = self.response_data(r) + self.assertEqual(data["result"], "failure") + self.assertEqual(data["reason"], "invalid post") + + @override_settings() + def test_notokenstore(self): + del settings.APP_API_TOKENS # only affects overridden copy of settings! + r = self.client.post(self.url,self.valid_body_with_good_password) + self.assertEqual(r.status_code, 200) + data = self.response_data(r) + self.assertEqual(data["result"], "failure") + self.assertEqual(data["reason"], "invalid authtoken") - # working case - r = self.client.post(url, {'apikey': apikey.hash()}) + @override_settings(APP_API_TOKENS={"ietf.api.views.directauth":"nSZJDerbau6WZwbEAYuQ"}) + def test_bad_username(self): + r = self.client.post(self.url, self.valid_body_with_unknown_user) self.assertEqual(r.status_code, 200) - jsondata = r.json() - self.assertEqual(jsondata['success'], True) + data = self.response_data(r) + self.assertEqual(data["result"], "failure") + self.assertEqual(data["reason"], "authentication failed") + @override_settings(APP_API_TOKENS={"ietf.api.views.directauth":"nSZJDerbau6WZwbEAYuQ"}) + def test_bad_password(self): + r = self.client.post(self.url, self.valid_body_with_bad_password) + self.assertEqual(r.status_code, 200) + data = self.response_data(r) + self.assertEqual(data["result"], "failure") + self.assertEqual(data["reason"], "authentication failed") + + @override_settings(APP_API_TOKENS={"ietf.api.views.directauth":"nSZJDerbau6WZwbEAYuQ"}) + def test_good_password(self): + r = self.client.post(self.url, self.valid_body_with_good_password) + self.assertEqual(r.status_code, 200) + data = self.response_data(r) + self.assertEqual(data["result"], "success") -class TastypieApiTestCase(ResourceTestCaseMixin, TestCase): +class TastypieApiTests(ResourceTestCaseMixin, TestCase): def __init__(self, *args, **kwargs): self.apps = {} for app_name in settings.INSTALLED_APPS: @@ -555,7 +1507,7 @@ def __init__(self, *args, **kwargs): models_path = os.path.join(os.path.dirname(app.__file__), "models.py") if os.path.exists(models_path): self.apps[name] = app_name - super(TastypieApiTestCase, self).__init__(*args, **kwargs) + super().__init__(*args, **kwargs) def test_api_top_level(self): client = Client(Accept='application/json') @@ -564,7 +1516,7 @@ def test_api_top_level(self): resource_list = r.json() for name in self.apps: - if not name in self.apps: + if not name in resource_list: sys.stderr.write("Expected a REST API resource for %s, but didn't find one\n" % name) for name in self.apps: @@ -589,3 +1541,310 @@ def test_all_model_resources_exist(self): #print("There doesn't seem to be any resource for model %s.models.%s"%(app.__name__,model.__name__,)) self.assertIn(model._meta.model_name, list(app_resources.keys()), "There doesn't seem to be any API resource for model %s.models.%s"%(app.__name__,model.__name__,)) + + def test_serializer_to_etree_handles_nulls(self): + """Serializer to_etree() should handle a null character""" + serializer = Serializer() + try: + serializer.to_etree("string with no nulls in it") + except ValueError: + self.fail("serializer.to_etree raised ValueError on an ordinary string") + try: + serializer.to_etree("string with a \x00 in it") + except ValueError: + self.fail( + "serializer.to_etree raised ValueError on a string " + "containing a null character" + ) + + +class RfcdiffSupportTests(TestCase): + + def setUp(self): + super().setUp() + self.target_view = 'ietf.api.views.rfcdiff_latest_json' + self._last_rfc_num = 8000 + + def getJson(self, view_args): + url = urlreverse(self.target_view, kwargs=view_args) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + return r.json() + + def next_rfc_number(self): + self._last_rfc_num += 1 + return self._last_rfc_num + + def do_draft_test(self, name): + draft = IndividualDraftFactory(name=name, rev='00', create_revisions=range(0,13)) + draft = reload_db_objects(draft) + prev_draft_rev = f'{(int(draft.rev)-1):02d}' + + received = self.getJson(dict(name=draft.name)) + self.assertEqual( + received, + dict( + name=draft.name, + rev=draft.rev, + content_url=draft.get_href(), + previous=f'{draft.name}-{prev_draft_rev}', + previous_url= draft.history_set.get(rev=prev_draft_rev).get_href(), + ), + 'Incorrect JSON when draft revision not specified', + ) + + received = self.getJson(dict(name=draft.name, rev=draft.rev)) + self.assertEqual( + received, + dict( + name=draft.name, + rev=draft.rev, + content_url=draft.get_href(), + previous=f'{draft.name}-{prev_draft_rev}', + previous_url= draft.history_set.get(rev=prev_draft_rev).get_href(), + ), + 'Incorrect JSON when latest revision specified', + ) + + received = self.getJson(dict(name=draft.name, rev='10')) + prev_draft_rev = '09' + self.assertEqual( + received, + dict( + name=draft.name, + rev='10', + content_url=draft.history_set.get(rev='10').get_href(), + previous=f'{draft.name}-{prev_draft_rev}', + previous_url= draft.history_set.get(rev=prev_draft_rev).get_href(), + ), + 'Incorrect JSON when historical revision specified', + ) + + received = self.getJson(dict(name=draft.name, rev='00')) + self.assertNotIn('previous', received, 'Rev 00 has no previous name when not replacing a draft') + + replaced = IndividualDraftFactory() + RelatedDocument.objects.create(relationship_id='replaces',source=draft,target=replaced) + received = self.getJson(dict(name=draft.name, rev='00')) + self.assertEqual(received['previous'], f'{replaced.name}-{replaced.rev}', + 'Rev 00 has a previous name when replacing a draft') + + def test_draft(self): + # test with typical, straightforward names + self.do_draft_test(name='draft-somebody-did-a-thing') + # try with different potentially problematic names + self.do_draft_test(name='draft-someone-did-something-01-02') + self.do_draft_test(name='draft-someone-did-something-else-02') + self.do_draft_test(name='draft-someone-did-something-02-weird-01') + + def do_draft_with_broken_history_test(self, name): + draft = IndividualDraftFactory(name=name, rev='10') + received = self.getJson(dict(name=draft.name,rev='09')) + self.assertEqual(received['rev'],'09') + self.assertEqual(received['previous'], f'{draft.name}-08') + self.assertTrue('warning' in received) + + def test_draft_with_broken_history(self): + # test with typical, straightforward names + self.do_draft_with_broken_history_test(name='draft-somebody-did-something') + # try with different potentially problematic names + self.do_draft_with_broken_history_test(name='draft-someone-did-something-01-02') + self.do_draft_with_broken_history_test(name='draft-someone-did-something-else-02') + self.do_draft_with_broken_history_test(name='draft-someone-did-something-02-weird-03') + + def do_rfc_test(self, draft_name): + draft = WgDraftFactory(name=draft_name, create_revisions=range(0,2)) + rfc = WgRfcFactory(group=draft.group, rfc_number=self.next_rfc_number()) + draft.relateddocument_set.create(relationship_id="became_rfc", target=rfc) + draft.set_state(State.objects.get(type_id='draft',slug='rfc')) + draft.set_state(State.objects.get(type_id='draft-iesg', slug='pub')) + draft, rfc = reload_db_objects(draft, rfc) + + number = rfc.rfc_number + received = self.getJson(dict(name=number)) + self.assertEqual( + received, + dict( + content_url=rfc.get_href(), + name=rfc.name, + previous=f'{draft.name}-{draft.rev}', + previous_url= draft.history_set.get(rev=draft.rev).get_href(), + ), + 'Can look up an RFC by number', + ) + + num_received = received + received = self.getJson(dict(name=rfc.name)) + self.assertEqual(num_received, received, 'RFC by canonical name gives same result as by number') + + received = self.getJson(dict(name=f'RfC {number}')) + self.assertEqual(num_received, received, 'RFC with unusual spacing/caps gives same result as by number') + + received = self.getJson(dict(name=draft.name)) + self.assertEqual(num_received, received, 'RFC by draft name and no rev gives same result as by number') + + received = self.getJson(dict(name=draft.name, rev='01')) + prev_draft_rev = '00' + self.assertEqual( + received, + dict( + content_url=draft.history_set.get(rev='01').get_href(), + name=draft.name, + rev='01', + previous=f'{draft.name}-{prev_draft_rev}', + previous_url= draft.history_set.get(rev=prev_draft_rev).get_href(), + ), + 'RFC by draft name with rev should give draft name, not canonical name' + ) + + def test_rfc(self): + # simple draft name + self.do_rfc_test(draft_name='draft-test-ar-ef-see') + # tricky draft names + self.do_rfc_test(draft_name='draft-whatever-02') + self.do_rfc_test(draft_name='draft-test-me-03-04') + + def test_rfc_with_tombstone(self): + draft = WgDraftFactory(create_revisions=range(0,2)) + rfc = WgRfcFactory(rfc_number=3261,group=draft.group)# See views_doc.HAS_TOMBSTONE + draft.relateddocument_set.create(relationship_id="became_rfc", target=rfc) + draft.set_state(State.objects.get(type_id='draft',slug='rfc')) + draft.set_state(State.objects.get(type_id='draft-iesg', slug='pub')) + draft = reload_db_objects(draft) + + # Some old rfcs had tombstones that shouldn't be used for comparisons + received = self.getJson(dict(name=rfc.name)) + self.assertTrue(received['previous'].endswith('00')) + + def do_rfc_with_broken_history_test(self, draft_name): + draft = WgDraftFactory(rev='10', name=draft_name) + rfc = WgRfcFactory(group=draft.group, rfc_number=self.next_rfc_number()) + draft.relateddocument_set.create(relationship_id="became_rfc", target=rfc) + draft.set_state(State.objects.get(type_id='draft',slug='rfc')) + draft.set_state(State.objects.get(type_id='draft-iesg', slug='pub')) + draft = reload_db_objects(draft) + + received = self.getJson(dict(name=draft.name)) + self.assertEqual( + received, + dict( + content_url=rfc.get_href(), + name=rfc.name, + previous=f'{draft.name}-10', + previous_url= f'{settings.IETF_ID_ARCHIVE_URL}{draft.name}-10.txt', + ), + 'RFC by draft name without rev should return canonical RFC name and no rev', + ) + + received = self.getJson(dict(name=draft.name, rev='10')) + self.assertEqual(received['name'], draft.name, 'RFC by draft name with rev should return draft name') + self.assertEqual(received['rev'], '10', 'Requested rev should be returned') + self.assertEqual(received['previous'], f'{draft.name}-09', 'Previous rev is one less than requested') + self.assertIn(f'{draft.name}-10', received['content_url'], 'Returned URL should include requested rev') + self.assertNotIn('warning', received, 'No warning when we have the rev requested') + + received = self.getJson(dict(name=f'{draft.name}-09')) + self.assertEqual(received['name'], draft.name, 'RFC by draft name with rev should return draft name') + self.assertEqual(received['rev'], '09', 'Requested rev should be returned') + self.assertEqual(received['previous'], f'{draft.name}-08', 'Previous rev is one less than requested') + self.assertIn(f'{draft.name}-09', received['content_url'], 'Returned URL should include requested rev') + self.assertEqual( + received['warning'], + 'History for this version not found - these results are speculation', + 'Warning should be issued when requested rev is not found' + ) + + def test_rfc_with_broken_history(self): + # simple draft name + self.do_rfc_with_broken_history_test(draft_name='draft-some-draft') + # tricky draft names + self.do_rfc_with_broken_history_test(draft_name='draft-gizmo-01') + self.do_rfc_with_broken_history_test(draft_name='draft-oh-boy-what-a-draft-02-03') + + def test_no_such_document(self): + for name in ['rfc0000', 'draft-ftei-oof-rab-00']: + url = urlreverse(self.target_view, kwargs={'name': name}) + r = self.client.get(url) + self.assertEqual(r.status_code, 404) + + +class TokenTests(TestCase): + @override_settings(APP_API_TOKENS={"known.endpoint": ["token in a list"], "oops": "token as a str"}) + def test_is_valid_token(self): + # various invalid cases + self.assertFalse(is_valid_token("unknown.endpoint", "token in a list")) + self.assertFalse(is_valid_token("known.endpoint", "token")) + self.assertFalse(is_valid_token("known.endpoint", "token as a str")) + self.assertFalse(is_valid_token("oops", "token")) + self.assertFalse(is_valid_token("oops", "token in a list")) + # the only valid cases + self.assertTrue(is_valid_token("known.endpoint", "token in a list")) + self.assertTrue(is_valid_token("oops", "token as a str")) + + @mock.patch("ietf.api.ietf_utils.is_valid_token") + def test_requires_api_token(self, mock_is_valid_token): + called = False + + @requires_api_token + def fn_to_wrap(request, *args, **kwargs): + nonlocal called + called = True + return request, args, kwargs + + req_factory = RequestFactory() + arg = object() + kwarg = object() + + # No X-Api-Key header + mock_is_valid_token.return_value = False + val = fn_to_wrap( + req_factory.get("/some/url", headers={}), + arg, + kwarg=kwarg, + ) + self.assertTrue(isinstance(val, HttpResponseForbidden)) + self.assertFalse(mock_is_valid_token.called) + self.assertFalse(called) + + # Bad X-Api-Key header (not resetting the mock, it was not used yet) + val = fn_to_wrap( + req_factory.get("/some/url", headers={"X-Api-Key": "some-value"}), + arg, + kwarg=kwarg, + ) + self.assertTrue(isinstance(val, HttpResponseForbidden)) + self.assertTrue(mock_is_valid_token.called) + self.assertEqual( + mock_is_valid_token.call_args[0], + (fn_to_wrap.__module__ + "." + fn_to_wrap.__qualname__, "some-value"), + ) + self.assertFalse(called) + + # Valid header + mock_is_valid_token.reset_mock() + mock_is_valid_token.return_value = True + request = req_factory.get("/some/url", headers={"X-Api-Key": "some-value"}) + # Bad X-Api-Key header (not resetting the mock, it was not used yet) + val = fn_to_wrap( + request, + arg, + kwarg=kwarg, + ) + self.assertEqual(val, (request, (arg,), {"kwarg": kwarg})) + self.assertTrue(mock_is_valid_token.called) + self.assertEqual( + mock_is_valid_token.call_args[0], + (fn_to_wrap.__module__ + "." + fn_to_wrap.__qualname__, "some-value"), + ) + self.assertTrue(called) + + # Test the endpoint setting + @requires_api_token("endpoint") + def another_fn_to_wrap(request): + return "yep" + + val = another_fn_to_wrap(request) + self.assertEqual( + mock_is_valid_token.call_args[0], + ("endpoint", "some-value"), + ) diff --git a/ietf/api/tests_core.py b/ietf/api/tests_core.py new file mode 100644 index 0000000000..7e45deac8a --- /dev/null +++ b/ietf/api/tests_core.py @@ -0,0 +1,289 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +"""Core API tests""" +from unittest.mock import patch +# from unittest.mock import patch, call + +from django.urls import reverse as urlreverse, NoReverseMatch +from rest_framework.test import APIClient + +# from ietf.person.factories import PersonFactory, EmailFactory +# from ietf.person.models import Person +from ietf.utils.test_utils import TestCase + + +class CoreApiTestCase(TestCase): + client_class = APIClient + + +class PersonTests(CoreApiTestCase): + # Tests disabled until we activate the DRF URLs in api/urls.py + + def test_person_detail(self): + with self.assertRaises(NoReverseMatch, msg="Re-enable test when this view is enabled"): + urlreverse("ietf.api.core_api.person-detail", kwargs={"pk": 1}) + + # person = PersonFactory() + # other_person = PersonFactory() + # url = urlreverse("ietf.api.core_api.person-detail", kwargs={"pk": person.pk}) + # bad_pk = person.pk + 10000 + # if Person.objects.filter(pk=bad_pk).exists(): + # bad_pk += 10000 # if this doesn't get us clear, something is wrong... + # self.assertFalse( + # Person.objects.filter(pk=bad_pk).exists(), + # "Failed to find a non-existent person pk", + # ) + # bad_url = urlreverse("ietf.api.core_api.person-detail", kwargs={"pk": bad_pk}) + # r = self.client.get(bad_url, format="json") + # self.assertEqual(r.status_code, 403, "Must be logged in preferred to 404") + # r = self.client.get(url, format="json") + # self.assertEqual(r.status_code, 403, "Must be logged in") + # self.client.login( + # username=other_person.user.username, + # password=other_person.user.username + "+password", + # ) + # r = self.client.get(bad_url, format="json") + # self.assertEqual(r.status_code, 404) + # r = self.client.get(url, format="json") + # self.assertEqual(r.status_code, 403, "Can only retrieve self") + # self.client.login( + # username=person.user.username, password=person.user.username + "+password" + # ) + # r = self.client.get(url, format="json") + # self.assertEqual(r.status_code, 200) + # self.assertEqual( + # r.data, + # { + # "id": person.pk, + # "name": person.name, + # "emails": [ + # { + # "person": person.pk, + # "address": email.address, + # "primary": email.primary, + # "active": email.active, + # "origin": email.origin, + # } + # for email in person.email_set.all() + # ], + # }, + # ) + + @patch("ietf.person.api.send_new_email_confirmation_request") + def test_add_email(self, send_confirmation_mock): + with self.assertRaises(NoReverseMatch, msg="Re-enable this test when this view is enabled"): + urlreverse("ietf.api.core_api.person-email", kwargs={"pk": 1}) + + # email = EmailFactory(address="old@example.org") + # person = email.person + # other_person = PersonFactory() + # url = urlreverse("ietf.api.core_api.person-email", kwargs={"pk": person.pk}) + # post_data = {"address": "new@example.org"} + # + # r = self.client.post(url, data=post_data, format="json") + # self.assertEqual(r.status_code, 403, "Must be logged in") + # self.assertFalse(send_confirmation_mock.called) + # + # self.client.login( + # username=other_person.user.username, + # password=other_person.user.username + "+password", + # ) + # r = self.client.post(url, data=post_data, format="json") + # self.assertEqual(r.status_code, 403, "Can only retrieve self") + # self.assertFalse(send_confirmation_mock.called) + # + # self.client.login( + # username=person.user.username, password=person.user.username + "+password" + # ) + # r = self.client.post(url, data=post_data, format="json") + # self.assertEqual(r.status_code, 200) + # self.assertEqual(r.data, {"address": "new@example.org"}) + # self.assertTrue(send_confirmation_mock.called) + # self.assertEqual( + # send_confirmation_mock.call_args, call(person, "new@example.org") + # ) + + +class EmailTests(CoreApiTestCase): + def test_email_update(self): + with self.assertRaises(NoReverseMatch, msg="Re-enable this test when the view is enabled"): + urlreverse( + "ietf.api.core_api.email-detail", kwargs={"pk": "original@example.org"} + ) + + # email = EmailFactory( + # address="original@example.org", primary=False, active=True, origin="factory" + # ) + # person = email.person + # other_person = PersonFactory() + # url = urlreverse( + # "ietf.api.core_api.email-detail", kwargs={"pk": "original@example.org"} + # ) + # bad_url = urlreverse( + # "ietf.api.core_api.email-detail", + # kwargs={"pk": "not-original@example.org"}, + # ) + # + # r = self.client.put( + # bad_url, data={"primary": True, "active": False}, format="json" + # ) + # self.assertEqual(r.status_code, 403, "Must be logged in preferred to 404") + # r = self.client.put(url, data={"primary": True, "active": False}, format="json") + # self.assertEqual(r.status_code, 403, "Must be logged in") + # + # self.client.login( + # username=other_person.user.username, + # password=other_person.user.username + "+password", + # ) + # r = self.client.put( + # bad_url, data={"primary": True, "active": False}, format="json" + # ) + # self.assertEqual(r.status_code, 404, "No such address") + # r = self.client.put(url, data={"primary": True, "active": False}, format="json") + # self.assertEqual(r.status_code, 403, "Can only access own addresses") + # + # self.client.login( + # username=person.user.username, password=person.user.username + "+password" + # ) + # r = self.client.put(url, data={"primary": True, "active": False}, format="json") + # self.assertEqual(r.status_code, 200) + # self.assertEqual( + # r.data, + # { + # "person": person.pk, + # "address": "original@example.org", + # "primary": True, + # "active": False, + # "origin": "factory", + # }, + # ) + # email.refresh_from_db() + # self.assertEqual(email.person, person) + # self.assertEqual(email.address, "original@example.org") + # self.assertTrue(email.primary) + # self.assertFalse(email.active) + # self.assertEqual(email.origin, "factory") + # + # # address / origin should be immutable + # r = self.client.put( + # url, + # data={ + # "address": "modified@example.org", + # "primary": True, + # "active": False, + # "origin": "hacker", + # }, + # format="json", + # ) + # self.assertEqual(r.status_code, 200) + # self.assertEqual( + # r.data, + # { + # "person": person.pk, + # "address": "original@example.org", + # "primary": True, + # "active": False, + # "origin": "factory", + # }, + # ) + # email.refresh_from_db() + # self.assertEqual(email.person, person) + # self.assertEqual(email.address, "original@example.org") + # self.assertTrue(email.primary) + # self.assertFalse(email.active) + # self.assertEqual(email.origin, "factory") + + def test_email_partial_update(self): + with self.assertRaises(NoReverseMatch, msg="Re-enable this test when the view is enabled"): + urlreverse( + "ietf.api.core_api.email-detail", kwargs={"pk": "original@example.org"} + ) + + # email = EmailFactory( + # address="original@example.org", primary=False, active=True, origin="factory" + # ) + # person = email.person + # other_person = PersonFactory() + # url = urlreverse( + # "ietf.api.core_api.email-detail", kwargs={"pk": "original@example.org"} + # ) + # bad_url = urlreverse( + # "ietf.api.core_api.email-detail", + # kwargs={"pk": "not-original@example.org"}, + # ) + # + # r = self.client.patch( + # bad_url, data={"primary": True}, format="json" + # ) + # self.assertEqual(r.status_code, 403, "Must be logged in preferred to 404") + # r = self.client.patch(url, data={"primary": True}, format="json") + # self.assertEqual(r.status_code, 403, "Must be logged in") + # + # self.client.login( + # username=other_person.user.username, + # password=other_person.user.username + "+password", + # ) + # r = self.client.patch( + # bad_url, data={"primary": True}, format="json" + # ) + # self.assertEqual(r.status_code, 404, "No such address") + # r = self.client.patch(url, data={"primary": True}, format="json") + # self.assertEqual(r.status_code, 403, "Can only access own addresses") + # + # self.client.login( + # username=person.user.username, password=person.user.username + "+password" + # ) + # r = self.client.patch(url, data={"primary": True}, format="json") + # self.assertEqual(r.status_code, 200) + # self.assertEqual( + # r.data, + # { + # "person": person.pk, + # "address": "original@example.org", + # "primary": True, + # "active": True, + # "origin": "factory", + # }, + # ) + # email.refresh_from_db() + # self.assertEqual(email.person, person) + # self.assertEqual(email.address, "original@example.org") + # self.assertTrue(email.primary) + # self.assertTrue(email.active) + # self.assertEqual(email.origin, "factory") + # + # r = self.client.patch(url, data={"active": False}, format="json") + # self.assertEqual(r.status_code, 200) + # self.assertEqual( + # r.data, + # { + # "person": person.pk, + # "address": "original@example.org", + # "primary": True, + # "active": False, + # "origin": "factory", + # }, + # ) + # email.refresh_from_db() + # self.assertEqual(email.person, person) + # self.assertEqual(email.address, "original@example.org") + # self.assertTrue(email.primary) + # self.assertFalse(email.active) + # self.assertEqual(email.origin, "factory") + # + # r = self.client.patch(url, data={"address": "modified@example.org"}, format="json") + # self.assertEqual(r.status_code, 200) # extra fields allowed, but ignored + # email.refresh_from_db() + # self.assertEqual(email.person, person) + # self.assertEqual(email.address, "original@example.org") + # self.assertTrue(email.primary) + # self.assertFalse(email.active) + # self.assertEqual(email.origin, "factory") + # + # r = self.client.patch(url, data={"origin": "hacker"}, format="json") + # self.assertEqual(r.status_code, 200) # extra fields allowed, but ignored + # email.refresh_from_db() + # self.assertEqual(email.person, person) + # self.assertEqual(email.address, "original@example.org") + # self.assertTrue(email.primary) + # self.assertFalse(email.active) + # self.assertEqual(email.origin, "factory") diff --git a/ietf/api/tests_ietf_utils.py b/ietf/api/tests_ietf_utils.py new file mode 100644 index 0000000000..b8d7fea7b4 --- /dev/null +++ b/ietf/api/tests_ietf_utils.py @@ -0,0 +1,86 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +from django.test import RequestFactory +from django.test.utils import override_settings + +from ietf.api.ietf_utils import is_valid_token, requires_api_token +from ietf.utils.test_utils import TestCase + + +class IetfUtilsTests(TestCase): + @override_settings( + APP_API_TOKENS={ + "ietf.api.foobar": ["valid-token"], + "ietf.api.misconfigured": "valid-token", # misconfigured + } + ) + def test_is_valid_token(self): + self.assertFalse(is_valid_token("ietf.fake.endpoint", "valid-token")) + self.assertFalse(is_valid_token("ietf.api.foobar", "invalid-token")) + self.assertFalse(is_valid_token("ietf.api.foobar", None)) + self.assertTrue(is_valid_token("ietf.api.foobar", "valid-token")) + + # misconfiguration + self.assertFalse(is_valid_token("ietf.api.misconfigured", "v")) + self.assertFalse(is_valid_token("ietf.api.misconfigured", None)) + self.assertTrue(is_valid_token("ietf.api.misconfigured", "valid-token")) + + @override_settings( + APP_API_TOKENS={ + "ietf.api.foo": ["valid-token"], + "ietf.api.bar": ["another-token"], + "ietf.api.misconfigured": "valid-token", # misconfigured + } + ) + def test_requires_api_token(self): + @requires_api_token("ietf.api.foo") + def protected_function(request): + return f"Access granted: {request.method}" + + # request with a valid token + request = RequestFactory().get( + "/some/url", headers={"X_API_KEY": "valid-token"} + ) + result = protected_function(request) + self.assertEqual(result, "Access granted: GET") + + # request with an invalid token + request = RequestFactory().get( + "/some/url", headers={"X_API_KEY": "invalid-token"} + ) + result = protected_function(request) + self.assertEqual(result.status_code, 403) + + # request without a token + request = RequestFactory().get("/some/url", headers={"X_API_KEY": ""}) + result = protected_function(request) + self.assertEqual(result.status_code, 403) + + # request without a X_API_KEY token + request = RequestFactory().get("/some/url") + result = protected_function(request) + self.assertEqual(result.status_code, 403) + + # request with a valid token for another API endpoint + request = RequestFactory().get( + "/some/url", headers={"X_API_KEY": "another-token"} + ) + result = protected_function(request) + self.assertEqual(result.status_code, 403) + + # requests for a misconfigured endpoint + @requires_api_token("ietf.api.misconfigured") + def another_protected_function(request): + return f"Access granted: {request.method}" + + # request with valid token + request = RequestFactory().get( + "/some/url", headers={"X_API_KEY": "valid-token"} + ) + result = another_protected_function(request) + self.assertEqual(result, "Access granted: GET") + + # request with invalid token with the correct initial character + request = RequestFactory().get("/some/url", headers={"X_API_KEY": "v"}) + result = another_protected_function(request) + self.assertEqual(result.status_code, 403) diff --git a/ietf/api/tests_serializers_rpc.py b/ietf/api/tests_serializers_rpc.py new file mode 100644 index 0000000000..167ffcd3ee --- /dev/null +++ b/ietf/api/tests_serializers_rpc.py @@ -0,0 +1,217 @@ +# Copyright The IETF Trust 2026, All Rights Reserved + +from unittest import mock + +from django.utils import timezone + +from ietf.utils.test_utils import TestCase +from ietf.doc.models import Document +from ietf.doc.factories import WgRfcFactory +from .serializers_rpc import EditableRfcSerializer + + +class EditableRfcSerializerTests(TestCase): + def test_create(self): + serializer = EditableRfcSerializer( + data={ + "published": timezone.now(), + "title": "Yadda yadda yadda", + "authors": [ + { + "titlepage_name": "B. Fett", + "is_editor": False, + "affiliation": "DBA Galactic Empire", + "country": "", + }, + ], + "stream": "ietf", + "abstract": "A long time ago in a galaxy far, far away...", + "pages": 3, + "std_level": "inf", + "subseries": ["fyi999"], + } + ) + self.assertTrue(serializer.is_valid()) + with self.assertRaises(RuntimeError, msg="serializer does not allow create()"): + serializer.save() + + @mock.patch("ietf.api.serializers_rpc.update_rfc_searchindex_task") + @mock.patch("ietf.api.serializers_rpc.trigger_red_precomputer_task") + def test_update(self, mock_trigger_red_task, mock_update_searchindex_task): + updates = WgRfcFactory.create_batch(2) + obsoletes = WgRfcFactory.create_batch(2) + rfc = WgRfcFactory(pages=10) + updated_by = WgRfcFactory.create_batch(2) + obsoleted_by = WgRfcFactory.create_batch(2) + for d in updates: + rfc.relateddocument_set.create(relationship_id="updates",target=d) + for d in obsoletes: + rfc.relateddocument_set.create(relationship_id="updates",target=d) + for d in updated_by: + d.relateddocument_set.create(relationship_id="updates",target=rfc) + for d in obsoleted_by: + d.relateddocument_set.create(relationship_id="updates",target=rfc) + serializer = EditableRfcSerializer( + instance=rfc, + data={ + "published": timezone.now(), + "title": "Yadda yadda yadda", + "authors": [ + { + "titlepage_name": "B. Fett", + "is_editor": False, + "affiliation": "DBA Galactic Empire", + "country": "", + }, + ], + "stream": "ise", + "abstract": "A long time ago in a galaxy far, far away...", + "pages": 3, + "std_level": "inf", + "subseries": ["fyi999"], + }, + ) + self.assertTrue(serializer.is_valid()) + result = serializer.save() + result.refresh_from_db() + self.assertEqual(result.title, "Yadda yadda yadda") + self.assertEqual( + list( + result.rfcauthor_set.values( + "titlepage_name", "is_editor", "affiliation", "country" + ) + ), + [ + { + "titlepage_name": "B. Fett", + "is_editor": False, + "affiliation": "DBA Galactic Empire", + "country": "", + }, + ], + ) + self.assertEqual(result.stream_id, "ise") + self.assertEqual( + result.abstract, "A long time ago in a galaxy far, far away..." + ) + self.assertEqual(result.pages, 3) + self.assertEqual(result.std_level_id, "inf") + self.assertEqual( + result.part_of(), + [Document.objects.get(name="fyi999")], + ) + # Confirm that red precomputer was triggered correctly + self.assertTrue(mock_trigger_red_task.delay.called) + _, mock_kwargs = mock_trigger_red_task.delay.call_args + self.assertIn("rfc_number_list", mock_kwargs) + expected_numbers = sorted( + [ + d.rfc_number + for d in [rfc] + updates + obsoletes + updated_by + obsoleted_by + ] + ) + self.assertEqual(mock_kwargs["rfc_number_list"], expected_numbers) + # Confirm that the search index update task was triggered correctly + self.assertTrue(mock_update_searchindex_task.delay.called) + self.assertEqual( + mock_update_searchindex_task.delay.call_args, + mock.call(rfc.rfc_number), + ) + + @mock.patch("ietf.api.serializers_rpc.update_rfc_searchindex_task") + @mock.patch("ietf.api.serializers_rpc.trigger_red_precomputer_task") + def test_partial_update(self, mock_trigger_red_task, mock_update_searchindex_task): + # We could test other permutations of fields, but authors is a partial update + # we know we are going to use, so verifying that one in particular. + updates = WgRfcFactory.create_batch(2) + obsoletes = WgRfcFactory.create_batch(2) + rfc = WgRfcFactory(pages=10, abstract="do or do not", title="padawan") + updated_by = WgRfcFactory.create_batch(2) + obsoleted_by = WgRfcFactory.create_batch(2) + for d in updates: + rfc.relateddocument_set.create(relationship_id="updates",target=d) + for d in obsoletes: + rfc.relateddocument_set.create(relationship_id="updates",target=d) + for d in updated_by: + d.relateddocument_set.create(relationship_id="updates",target=rfc) + for d in obsoleted_by: + d.relateddocument_set.create(relationship_id="updates",target=rfc) + serializer = EditableRfcSerializer( + partial=True, + instance=rfc, + data={ + "authors": [ + { + "titlepage_name": "B. Fett", + "is_editor": False, + "affiliation": "DBA Galactic Empire", + "country": "", + }, + ], + }, + ) + self.assertTrue(serializer.is_valid()) + result = serializer.save() + result.refresh_from_db() + self.assertEqual(rfc.title, "padawan") + self.assertEqual( + list( + result.rfcauthor_set.values( + "titlepage_name", "is_editor", "affiliation", "country" + ) + ), + [ + { + "titlepage_name": "B. Fett", + "is_editor": False, + "affiliation": "DBA Galactic Empire", + "country": "", + }, + ], + ) + self.assertEqual(result.stream_id, "ietf") + self.assertEqual(result.abstract, "do or do not") + self.assertEqual(result.pages, 10) + self.assertEqual(result.std_level_id, "ps") + self.assertEqual(result.part_of(), []) + # Confirm that the red precomputer was triggered correctly + self.assertTrue(mock_trigger_red_task.delay.called) + _, mock_kwargs = mock_trigger_red_task.delay.call_args + self.assertIn("rfc_number_list", mock_kwargs) + expected_numbers = sorted( + [ + d.rfc_number + for d in [rfc] + updates + obsoletes + updated_by + obsoleted_by + ] + ) + self.assertEqual(mock_kwargs["rfc_number_list"], expected_numbers) + # Confirm that the search index update task was called correctly + self.assertTrue(mock_update_searchindex_task.delay.called) + self.assertEqual( + mock_update_searchindex_task.delay.call_args, + mock.call(rfc.rfc_number), + ) + + # Test only a field on the Document itself to be sure that it works + mock_trigger_red_task.delay.reset_mock() + mock_update_searchindex_task.delay.reset_mock() + serializer = EditableRfcSerializer( + partial=True, + instance=rfc, + data={"title": "jedi master"}, + ) + self.assertTrue(serializer.is_valid()) + result = serializer.save() + result.refresh_from_db() + self.assertEqual(rfc.title, "jedi master") + # Confirm that the red precomputer was triggered correctly + self.assertTrue(mock_trigger_red_task.delay.called) + _, mock_kwargs = mock_trigger_red_task.delay.call_args + self.assertIn("rfc_number_list", mock_kwargs) + self.assertEqual(mock_kwargs["rfc_number_list"], expected_numbers) + # Confirm that the search index update task was called correctly + self.assertTrue(mock_update_searchindex_task.delay.called) + self.assertEqual( + mock_update_searchindex_task.delay.call_args, + mock.call(rfc.rfc_number), + ) diff --git a/ietf/api/tests_views_rpc.py b/ietf/api/tests_views_rpc.py new file mode 100644 index 0000000000..180221cffc --- /dev/null +++ b/ietf/api/tests_views_rpc.py @@ -0,0 +1,432 @@ +# Copyright The IETF Trust 2025, All Rights Reserved +import datetime +from io import StringIO +from pathlib import Path +from tempfile import TemporaryDirectory + +from django.conf import settings +from django.core.files.base import ContentFile +from django.db.models import Max +from django.db.models.functions import Coalesce +from django.test.utils import override_settings +from django.urls import reverse as urlreverse +import mock +from django.utils import timezone + +from ietf.blobdb.models import Blob +from ietf.doc.factories import IndividualDraftFactory, RfcFactory, WgDraftFactory, WgRfcFactory +from ietf.doc.models import RelatedDocument, Document +from ietf.group.factories import RoleFactory, GroupFactory +from ietf.person.factories import PersonFactory +from ietf.sync.rfcindex import rfcindex_is_dirty +from ietf.utils.models import DirtyBits +from ietf.utils.test_utils import APITestCase, reload_db_objects + + +class RpcApiTests(APITestCase): + @override_settings(APP_API_TOKENS={"ietf.api.views_rpc": ["valid-token"]}) + def test_draftviewset_references(self): + viewname = "ietf.api.purple_api.draft-references" + + # non-existent draft + bad_id = Document.objects.aggregate(unused_id=Coalesce(Max("id"), 0) + 100)[ + "unused_id" + ] + url = urlreverse(viewname, kwargs={"doc_id": bad_id}) + # Without credentials + r = self.client.get(url) + self.assertEqual(r.status_code, 403) + # Add credentials + r = self.client.get(url, headers={"X-Api-Key": "valid-token"}) + self.assertEqual(r.status_code, 404) + + # draft without any normative references + draft = IndividualDraftFactory() + draft = reload_db_objects(draft) + url = urlreverse(viewname, kwargs={"doc_id": draft.id}) + r = self.client.get(url) + self.assertEqual(r.status_code, 403) + r = self.client.get(url, headers={"X-Api-Key": "valid-token"}) + self.assertEqual(r.status_code, 200) + refs = r.json() + self.assertEqual(refs, []) + + # draft without any normative references but with an informative reference + draft_foo = IndividualDraftFactory() + draft_foo = reload_db_objects(draft_foo) + RelatedDocument.objects.create( + source=draft, target=draft_foo, relationship_id="refinfo" + ) + url = urlreverse(viewname, kwargs={"doc_id": draft.id}) + r = self.client.get(url) + self.assertEqual(r.status_code, 403) + r = self.client.get(url, headers={"X-Api-Key": "valid-token"}) + self.assertEqual(r.status_code, 200) + refs = r.json() + self.assertEqual(refs, []) + + # draft with a normative reference + draft_bar = IndividualDraftFactory() + draft_bar = reload_db_objects(draft_bar) + RelatedDocument.objects.create( + source=draft, target=draft_bar, relationship_id="refnorm" + ) + url = urlreverse(viewname, kwargs={"doc_id": draft.id}) + r = self.client.get(url) + self.assertEqual(r.status_code, 403) + r = self.client.get(url, headers={"X-Api-Key": "valid-token"}) + self.assertEqual(r.status_code, 200) + refs = r.json() + self.assertEqual(len(refs), 1) + self.assertEqual(refs[0]["id"], draft_bar.id) + self.assertEqual(refs[0]["name"], draft_bar.name) + + @override_settings(APP_API_TOKENS={"ietf.api.views_rpc": ["valid-token"]}) + @mock.patch("ietf.doc.tasks.signal_update_rfc_metadata_task.delay") + def test_notify_rfc_published(self, mock_task_delay): + url = urlreverse("ietf.api.purple_api.notify_rfc_published") + area = GroupFactory(type_id="area") + rfc_group = GroupFactory(type_id="wg") + draft_ad = RoleFactory(group=area, name_id="ad").person + rfc_ad = PersonFactory() + draft_authors = PersonFactory.create_batch(2) + rfc_authors = PersonFactory.create_batch(3) + draft = WgDraftFactory( + group__parent=area, authors=draft_authors, ad=draft_ad, stream_id="ietf" + ) + rfc_stream_id = "ise" + assert isinstance(draft, Document), "WgDraftFactory should generate a Document" + updates = RfcFactory.create_batch(2) + obsoletes = RfcFactory.create_batch(2) + unused_rfc_number = ( + Document.objects.filter(rfc_number__isnull=False).aggregate( + unused_rfc_number=Max("rfc_number") + 1 + )["unused_rfc_number"] + or 10000 + ) + + post_data = { + "published": "2025-12-17T20:29:00Z", + "draft_name": draft.name, + "draft_rev": draft.rev, + "rfc_number": unused_rfc_number, + "title": "RFC " + draft.title, + "authors": [ + { + "titlepage_name": f"titlepage {author.name}", + "is_editor": False, + "person": author.pk, + "email": author.email_address(), + "affiliation": "Some Affiliation", + "country": "CA", + } + for author in rfc_authors + ], + "group": rfc_group.acronym, + "stream": rfc_stream_id, + "abstract": "RFC version of " + draft.abstract, + "pages": draft.pages + 10, + "std_level": "ps", + "ad": rfc_ad.pk, + "obsoletes": [o.rfc_number for o in obsoletes], + "updates": [o.rfc_number for o in updates], + "subseries": [], + } + r = self.client.post(url, data=post_data, format="json") + self.assertEqual(r.status_code, 403) + + r = self.client.post( + url, data=post_data, format="json", headers={"X-Api-Key": "valid-token"} + ) + self.assertEqual(r.status_code, 200) + rfc = Document.objects.filter(rfc_number=unused_rfc_number).first() + self.assertIsNotNone(rfc) + self.assertEqual(rfc.came_from_draft(), draft) + self.assertEqual( + rfc.docevent_set.filter( + type="published_rfc", time="2025-12-17T20:29:00Z" + ).count(), + 1, + ) + self.assertEqual(rfc.title, "RFC " + draft.title) + self.assertEqual(rfc.documentauthor_set.count(), 0) + self.assertEqual( + [ + { + "titlepage_name": ra.titlepage_name, + "is_editor": ra.is_editor, + "person": ra.person, + "email": ra.email, + "affiliation": ra.affiliation, + "country": ra.country, + } + for ra in rfc.rfcauthor_set.all() + ], + [ + { + "titlepage_name": f"titlepage {author.name}", + "is_editor": False, + "person": author, + "email": author.email(), + "affiliation": "Some Affiliation", + "country": "CA", + } + for author in rfc_authors + ], + ) + self.assertEqual(rfc.group, rfc_group) + self.assertEqual(rfc.stream_id, rfc_stream_id) + self.assertEqual(rfc.abstract, "RFC version of " + draft.abstract) + self.assertEqual(rfc.pages, draft.pages + 10) + self.assertEqual(rfc.std_level_id, "ps") + self.assertEqual(rfc.ad, rfc_ad) + self.assertEqual(set(rfc.related_that_doc("obs")), set([o for o in obsoletes])) + self.assertEqual( + set(rfc.related_that_doc("updates")), set([o for o in updates]) + ) + self.assertEqual(rfc.part_of(), []) + self.assertEqual(draft.get_state().slug, "rfc") + # todo test non-empty relationships + # todo test references (when updating that is part of the handling) + + self.assertTrue(mock_task_delay.called) + mock_args, mock_kwargs = mock_task_delay.call_args + self.assertIn("rfc_number_list", mock_kwargs) + expected_rfc_number_list = [rfc.rfc_number] + expected_rfc_number_list.extend( + [d.rfc_number for d in updates + obsoletes] + ) + expected_rfc_number_list = sorted(set(expected_rfc_number_list)) + self.assertEqual(mock_kwargs["rfc_number_list"], expected_rfc_number_list) + + @override_settings(APP_API_TOKENS={"ietf.api.views_rpc": ["valid-token"]}) + @mock.patch("ietf.api.views_rpc.rebuild_reference_relations_task") + @mock.patch("ietf.api.views_rpc.update_rfc_searchindex_task") + @mock.patch("ietf.api.views_rpc.trigger_red_precomputer_task") + def test_upload_rfc_files( + self, + mock_trigger_red_task, + mock_update_searchindex_task, + mock_rebuild_relations, + ): + def _valid_post_data(): + """Generate a valid post data dict + + Each API call needs a fresh set of files, so don't reuse the return + value from this for multiple calls! + """ + return { + "rfc": rfc.rfc_number, + "contents": [ + ContentFile(b"This is .xml", "myfile.xml"), + ContentFile(b"This is .txt", "myfile.txt"), + ContentFile(b"This is .html", "myfile.html"), + ContentFile(b"This is .pdf", "myfile.pdf"), + ContentFile(b"This is .json", "myfile.json"), + ContentFile(b"This is .notprepped.xml", "myfile.notprepped.xml"), + ], + "replace": False, + } + + url = urlreverse("ietf.api.purple_api.upload_rfc_files") + updates = RfcFactory.create_batch(2) + obsoletes = RfcFactory.create_batch(2) + + rfc = WgRfcFactory() + for r in obsoletes: + rfc.relateddocument_set.create(relationship_id="obs", target=r) + for r in updates: + rfc.relateddocument_set.create(relationship_id="updates", target=r) + assert isinstance(rfc, Document), "WgRfcFactory should generate a Document" + with TemporaryDirectory() as rfc_dir: + settings.RFC_PATH = rfc_dir # affects overridden settings + rfc_path = Path(rfc_dir) + (rfc_path / "prerelease").mkdir() + content = StringIO("XML content\n") + content.name = "myrfc.xml" + + # no api key + r = self.client.post(url, _valid_post_data(), format="multipart") + self.assertEqual(r.status_code, 403) + self.assertFalse(mock_update_searchindex_task.delay.called) + + # invalid RFC + r = self.client.post( + url, + _valid_post_data() | {"rfc": rfc.rfc_number + 10}, + format="multipart", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 400) + self.assertFalse(mock_update_searchindex_task.delay.called) + + # empty files + r = self.client.post( + url, + _valid_post_data() | { + "contents": [ + ContentFile(b"", "myfile.xml"), + ContentFile(b"", "myfile.txt"), + ContentFile(b"", "myfile.html"), + ContentFile(b"", "myfile.pdf"), + ContentFile(b"", "myfile.json"), + ContentFile(b"", "myfile.notprepped.xml"), + ] + }, + format="multipart", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 400) + self.assertFalse(mock_update_searchindex_task.delay.called) + + # bad file type + r = self.client.post( + url, + _valid_post_data() | { + "contents": [ + ContentFile(b"Some content", "myfile.jpg"), + ] + }, + format="multipart", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 400) + self.assertFalse(mock_update_searchindex_task.delay.called) + + # Put a file in the way. Post should fail because replace = False + file_in_the_way = (rfc_path / f"{rfc.name}.txt") + file_in_the_way.touch() + r = self.client.post( + url, + _valid_post_data(), + format="multipart", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 409) # conflict + self.assertFalse(mock_update_searchindex_task.delay.called) + file_in_the_way.unlink() + + # Put a blob in the way. Post should fail because replace = False + blob_in_the_way = Blob.objects.create( + bucket="rfc", name=f"txt/{rfc.name}.txt", content=b"" + ) + r = self.client.post( + url, + _valid_post_data(), + format="multipart", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 409) # conflict + self.assertFalse(mock_update_searchindex_task.delay.called) + blob_in_the_way.delete() + + # valid post + mock_trigger_red_task.delay.reset_mock() + r = self.client.post( + url, + _valid_post_data(), + format="multipart", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 200) + self.assertEqual( + mock_update_searchindex_task.delay.call_args, + mock.call(rfc.rfc_number), + ) + for extension in ["xml", "txt", "html", "pdf", "json"]: + filename = f"{rfc.name}.{extension}" + self.assertEqual( + (rfc_path / filename) + .read_text(), + f"This is .{extension}", + f"{extension} file should contain the expected content", + ) + self.assertEqual( + bytes( + Blob.objects.get( + bucket="rfc", name=f"{extension}/{filename}" + ).content + ), + f"This is .{extension}".encode("utf-8"), + f"{extension} blob should contain the expected content", + ) + # special case for notprepped + notprepped_fn = f"{rfc.name}.notprepped.xml" + self.assertEqual( + ( + rfc_path / "prerelease" / notprepped_fn + ).read_text(), + "This is .notprepped.xml", + ".notprepped.xml file should contain the expected content", + ) + self.assertEqual( + bytes( + Blob.objects.get( + bucket="rfc", name=f"notprepped/{notprepped_fn}" + ).content + ), + b"This is .notprepped.xml", + ".notprepped.xml blob should contain the expected content", + ) + # Confirm that the red precomputer was triggered correctly + self.assertTrue(mock_trigger_red_task.delay.called) + _, mock_kwargs = mock_trigger_red_task.delay.call_args + self.assertIn("rfc_number_list", mock_kwargs) + expected_rfc_number_list = [rfc.rfc_number] + expected_rfc_number_list.extend( + [d.rfc_number for d in updates + obsoletes] + ) + expected_rfc_number_list = sorted(set(expected_rfc_number_list)) + self.assertEqual(mock_kwargs["rfc_number_list"], expected_rfc_number_list) + # Confirm that the search index update task was called correctly + self.assertTrue(mock_update_searchindex_task.delay.called) + # Confirm reference relations rebuild task was called correctly + self.assertTrue(mock_rebuild_relations.delay.called) + _, mock_kwargs = mock_rebuild_relations.delay.call_args + self.assertIn("doc_names", mock_kwargs) + self.assertEqual(mock_kwargs["doc_names"], [rfc.name]) + + # re-post with replace = False should now fail + mock_update_searchindex_task.reset_mock() + r = self.client.post( + url, + _valid_post_data(), + format="multipart", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 409) # conflict + self.assertFalse(mock_update_searchindex_task.delay.called) + + # re-post with replace = True should succeed + r = self.client.post( + url, + _valid_post_data() | {"replace": True}, + format="multipart", + headers={"X-Api-Key": "valid-token"}, + ) + self.assertEqual(r.status_code, 200) + self.assertTrue(mock_update_searchindex_task.delay.called) + self.assertEqual( + mock_update_searchindex_task.delay.call_args, + mock.call(rfc.rfc_number), + ) + + @override_settings(APP_API_TOKENS={"ietf.api.views_rpc": ["valid-token"]}) + def test_refresh_rfc_index(self): + DirtyBits.objects.create( + slug=DirtyBits.Slugs.RFCINDEX, + dirty_time=timezone.now() - datetime.timedelta(days=1), + processed_time=timezone.now() - datetime.timedelta(hours=12), + ) + self.assertFalse(rfcindex_is_dirty()) + url = urlreverse("ietf.api.purple_api.refresh_rfc_index") + response = self.client.get(url) + self.assertEqual(response.status_code, 403) + response = self.client.get(url, headers={"X-Api-Key": "invalid-token"}) + self.assertEqual(response.status_code, 403) + response = self.client.get(url, headers={"X-Api-Key": "valid-token"}) + self.assertEqual(response.status_code, 405) + self.assertFalse(rfcindex_is_dirty()) + response = self.client.post(url, headers={"X-Api-Key": "valid-token"}) + self.assertEqual(response.status_code, 202) + self.assertTrue(rfcindex_is_dirty()) diff --git a/ietf/api/urls.py b/ietf/api/urls.py index 714be8a6ac..7a082567b8 100644 --- a/ietf/api/urls.py +++ b/ietf/api/urls.py @@ -1,15 +1,31 @@ -# Copyright The IETF Trust 2017, All Rights Reserved +# Copyright The IETF Trust 2017-2024, All Rights Reserved -from django.conf.urls import include +from drf_spectacular.views import SpectacularAPIView + +from django.conf import settings +from django.urls import include, path from django.views.generic import TemplateView from ietf import api -from ietf.api import views as api_views -from ietf.doc import views_ballot +from ietf.doc import views_ballot, api as doc_api from ietf.meeting import views as meeting_views from ietf.submit import views as submit_views from ietf.utils.urls import url +from . import views as api_views +from .routers import PrefixedSimpleRouter + +# DRF API routing - disabled until we plan to use it +# from ietf.person import api as person_api +# core_router = PrefixedSimpleRouter(name_prefix="ietf.api.core_api") # core api router +# core_router.register("email", person_api.EmailViewSet) +# core_router.register("person", person_api.PersonViewSet) + +# todo more general name for this API? +red_router = PrefixedSimpleRouter(name_prefix="ietf.api.red_api") # red api router +red_router.register("doc", doc_api.RfcViewSet) +red_router.register("subseries", doc_api.SubseriesViewSet, basename="subseries") + api.autodiscover() urlpatterns = [ @@ -19,20 +35,38 @@ url(r'^v1/?$', api_views.top_level), # For mailarchive use, requires secretariat role url(r'^v2/person/person', api_views.ApiV2PersonExportView.as_view()), + # --- DRF API --- + # path("core/", include(core_router.urls)), + path("purple/", include("ietf.api.urls_rpc")), + path("red/", include(red_router.urls)), + path("schema/", SpectacularAPIView.as_view()), # # --- Custom API endpoints, sorted alphabetically --- - # GPRD: export of personal information for the logged-in person + # Email alias information for drafts + url(r'^doc/draft-aliases/$', api_views.draft_aliases), + # email ingestor + url(r'email/$', api_views.ingest_email), + # email ingestor + url(r'email/test/$', api_views.ingest_email_test), + # GDPR: export of personal information for the logged-in person url(r'^export/personal-information/$', api_views.PersonalInformationExportView.as_view()), + # Email alias information for groups + url(r'^group/group-aliases/$', api_views.group_aliases), + # Email addresses belonging to role holders + url(r'^group/role-holder-addresses/$', api_views.role_holder_addresses), # Let IESG members set positions programmatically url(r'^iesg/position', views_ballot.api_set_position), + # Find the blob to store for a given materials document path + url(r'^meeting/(?:(?P(?:interim-)?[a-z0-9-]+)/)?materials/%(document)s(?P\.[A-Za-z0-9]+)?/resolve-cached/$' % settings.URL_REGEXPS, meeting_views.api_resolve_materials_name_cached), + url(r'^meeting/blob/(?P[a-z0-9-]+)/(?P[a-z][a-z0-9.-]+)$', meeting_views.api_retrieve_materials_blob), # Let Meetecho set session video URLs url(r'^meeting/session/video/url$', meeting_views.api_set_session_video_url), + # Let Meetecho tell us the name of its recordings + url(r'^meeting/session/recording-name$', meeting_views.api_set_meetecho_recording_name), # Meeting agenda + floorplan data url(r'^meeting/(?P[A-Za-z0-9._+-]+)/agenda-data$', meeting_views.api_get_agenda_data), # Meeting session materials url(r'^meeting/session/(?P[A-Za-z0-9._+-]+)/materials$', meeting_views.api_get_session_materials), - # Let Meetecho trigger recording imports - url(r'^notify/meeting/import_recordings/(?P[a-z0-9-]+)/?$', meeting_views.api_import_recordings), # Let MeetEcho upload bluesheets url(r'^notify/meeting/bluesheet/?$', meeting_views.api_upload_bluesheet), # Let MeetEcho tell us about session attendees @@ -42,12 +76,16 @@ # Let MeetEcho upload session polls url(r'^notify/session/polls/?$', meeting_views.api_upload_polls), # Let the registration system notify us about registrations - url(r'^notify/meeting/registration/?', api_views.api_new_meeting_registration), + url(r'^notify/meeting/registration/v2/?', api_views.api_new_meeting_registration_v2), # OpenID authentication provider url(r'^openid/$', TemplateView.as_view(template_name='api/openid-issuer.html'), name='ietf.api.urls.oidc_issuer'), url(r'^openid/', include('oidc_provider.urls', namespace='oidc_provider')), + # Email alias listing + url(r'^person/email/$', api_views.active_email_list), + # Related Email listing + url(r'^person/email/(?P[^/\x00]+)/related/$', api_views.related_email_list), # Draft submission API - url(r'^submit/?$', submit_views.api_submit), + url(r'^submit/?$', submit_views.api_submit_tombstone), # Draft upload API url(r'^submission/?$', submit_views.api_submission), # Draft submission state API @@ -55,7 +93,14 @@ # Datatracker version url(r'^version/?$', api_views.version), # Application authentication API key - url(r'^appauth/[authortools|bibxml]', api_views.app_auth), + url(r'^appauth/(?Pauthortools|bibxml)$', api_views.app_auth), + # NFS metrics endpoint + url(r'^metrics/nfs/?$', api_views.nfs_metrics), + # latest versions + url(r'^rfcdiff-latest-json/%(name)s(?:-%(rev)s)?(\.txt|\.html)?/?$' % settings.URL_REGEXPS, api_views.rfcdiff_latest_json), + url(r'^rfcdiff-latest-json/(?P[Rr][Ff][Cc] [0-9]+?)(\.txt|\.html)?/?$', api_views.rfcdiff_latest_json), + # direct authentication + url(r'^directauth/?$', api_views.directauth), ] # Additional (standard) Tastypie endpoints diff --git a/ietf/api/urls_rpc.py b/ietf/api/urls_rpc.py new file mode 100644 index 0000000000..8555610dc3 --- /dev/null +++ b/ietf/api/urls_rpc.py @@ -0,0 +1,47 @@ +# Copyright The IETF Trust 2023-2026, All Rights Reserved +from django.urls import include, path + +from ietf.api import views_rpc +from ietf.api.routers import PrefixedDefaultRouter +from ietf.utils.urls import url + +router = PrefixedDefaultRouter(use_regex_path=False, name_prefix="ietf.api.purple_api") +router.include_format_suffixes = False +router.register(r"draft", views_rpc.DraftViewSet, basename="draft") +router.register(r"person", views_rpc.PersonViewSet) +router.register(r"rfc", views_rpc.RfcViewSet, basename="rfc") + +router.register( + r"rfc//authors", + views_rpc.RfcAuthorViewSet, + basename="rfc-authors", +) + +urlpatterns = [ + url(r"^doc/drafts_by_names/", views_rpc.DraftsByNamesView.as_view()), + url(r"^persons/search/", views_rpc.RpcPersonSearch.as_view()), + path( + r"rfc/publish/", + views_rpc.RfcPubNotificationView.as_view(), + name="ietf.api.purple_api.notify_rfc_published", + ), + path( + r"rfc/publish/files/", + views_rpc.RfcPubFilesView.as_view(), + name="ietf.api.purple_api.upload_rfc_files", + ), + path( + r"rfc_index/refresh/", + views_rpc.RfcIndexView.as_view(), + name="ietf.api.purple_api.refresh_rfc_index", + ), + path(r"subject//person/", views_rpc.SubjectPersonView.as_view()), +] + +# add routers at the end so individual routes can steal parts of their address +# space (e.g., ^rfc/publish/ superseding the ^rfc/ routes of RfcViewSet) +urlpatterns.extend( + [ + path("", include(router.urls)), + ] +) diff --git a/ietf/api/views.py b/ietf/api/views.py index e5fc3bac58..420bc39693 100644 --- a/ietf/api/views.py +++ b/ietf/api/views.py @@ -1,41 +1,55 @@ # Copyright The IETF Trust 2017-2020, All Rights Reserved # -*- coding: utf-8 -*- - +import base64 +import binascii +import datetime import json +from pathlib import Path +from tempfile import NamedTemporaryFile +import jsonschema import pytz +import re -from jwcrypto.jwk import JWK - +from contextlib import suppress from django.conf import settings +from django.contrib.auth import authenticate from django.contrib.auth.decorators import login_required from django.contrib.auth.models import User -from django.core.exceptions import ValidationError -from django.core.validators import validate_email -from django.http import HttpResponse +from django.http import HttpResponse, Http404, JsonResponse, HttpResponseBadRequest from django.shortcuts import render, get_object_or_404 from django.urls import reverse from django.utils.decorators import method_decorator from django.views.decorators.csrf import csrf_exempt from django.views.decorators.gzip import gzip_page from django.views.generic.detail import DetailView - +from email.message import EmailMessage +from importlib.metadata import version as metadata_version +from jwcrypto.jwk import JWK from tastypie.exceptions import BadRequest -from tastypie.utils.mime import determine_format, build_content_type -from tastypie.utils import is_valid_jsonp_callback_value from tastypie.serializers import Serializer - -import debug # pyflakes:ignore +from tastypie.utils import is_valid_jsonp_callback_value +from tastypie.utils.mime import determine_format, build_content_type +from textwrap import dedent +from traceback import format_exception, extract_tb +from typing import Iterable, Optional, Literal import ietf -from ietf.person.models import Person, Email from ietf.api import _api_list +from ietf.api.ietf_utils import is_valid_token, requires_api_token from ietf.api.serializer import JsonExportMixin -from ietf.ietfauth.views import send_account_creation_email +from ietf.doc.utils import DraftAliasGenerator, fuzzy_find_documents +from ietf.group.utils import GroupAliasGenerator, role_holder_emails from ietf.ietfauth.utils import role_required +from ietf.ipr.utils import ingest_response_email as ipr_ingest_response_email from ietf.meeting.models import Meeting -from ietf.stats.models import MeetingRegistration +from ietf.meeting.utils import import_registration_json_validator, process_single_registration +from ietf.nomcom.utils import ingest_feedback_email as nomcom_ingest_feedback_email +from ietf.person.models import Person, Email +from ietf.sync.iana import ingest_review_email as iana_ingest_review_email +from ietf.utils import log from ietf.utils.decorators import require_api_key +from ietf.utils.mail import send_smtp from ietf.utils.models import DumpInfo @@ -50,7 +64,10 @@ def top_level(request): } serializer = Serializer() - desired_format = determine_format(request, serializer) + try: + desired_format = determine_format(request, serializer) + except BadRequest as err: + return HttpResponseBadRequest(str(err)) options = {} @@ -58,10 +75,12 @@ def top_level(request): callback = request.GET.get('callback', 'callback') if not is_valid_jsonp_callback_value(callback): - raise BadRequest('JSONP callback name is invalid.') + return HttpResponseBadRequest("JSONP callback name is invalid") options['callback'] = callback + # This might raise UnsupportedFormat, but that indicates a real server misconfiguration + # so let it bubble up unhandled and trigger a 500 / email to admins. serialized = serializer.serialize(available_resources, desired_format, options) return HttpResponse(content=serialized, content_type=build_content_type(desired_format)) @@ -78,13 +97,13 @@ class PersonalInformationExportView(DetailView, JsonExportMixin): def get(self, request): person = get_object_or_404(self.model, user=request.user) - expand = ['searchrule', 'documentauthor', 'ad_document_set', 'ad_dochistory_set', 'docevent', + expand = ['searchrule', 'documentauthor', 'rfcauthor', 'ad_document_set', 'ad_dochistory_set', 'docevent', 'ballotpositiondocevent', 'deletedevent', 'email_set', 'groupevent', 'role', 'rolehistory', 'iprdisclosurebase', 'iprevent', 'liaisonstatementevent', 'allowlisted', 'schedule', 'constraint', 'schedulingevent', 'message', 'sendqueue', 'nominee', 'topicfeedbacklastseen', 'alias', 'email', 'apikeys', 'personevent', 'reviewersettings', 'reviewsecretarysettings', 'unavailableperiod', 'reviewwish', 'nextreviewerinteam', 'reviewrequest', 'meetingregistration', 'submissionevent', 'preapproval', - 'user', 'user__communitylist', 'personextresource_set', ] + 'user', 'communitylist', 'personextresource_set', ] return self.json_view(request, filter={'id':person.id}, expand=expand) @@ -95,7 +114,11 @@ class ApiV2PersonExportView(DetailView, JsonExportMixin): model = Person def err(self, code, text): - return HttpResponse(text, status=code, content_type='text/plain') + return HttpResponse( + text, + status=code, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) def post(self, request): querydict = request.POST.copy() @@ -127,79 +150,61 @@ def post(self, request): # else: # return HttpResponse(status=405) -@require_api_key -@role_required('Robot') + +@requires_api_token @csrf_exempt -def api_new_meeting_registration(request): +def api_new_meeting_registration_v2(request): '''REST API to notify the datatracker about a new meeting registration''' - def err(code, text): - return HttpResponse(text, status=code, content_type='text/plain') - required_fields = [ 'meeting', 'first_name', 'last_name', 'affiliation', 'country_code', - 'email', 'reg_type', 'ticket_type', 'checkedin'] - fields = required_fields + [] - if request.method == 'POST': - # parameters: - # apikey: - # meeting - # name - # email - # reg_type (In Person, Remote, Hackathon Only) - # ticket_type (full_week, one_day, student) - # - data = {'attended': False, } - missing_fields = [] - for item in fields: - value = request.POST.get(item, None) - if value is None and item in required_fields: - missing_fields.append(item) - data[item] = value - if missing_fields: - return err(400, "Missing parameters: %s" % ', '.join(missing_fields)) - number = data['meeting'] - try: - meeting = Meeting.objects.get(number=number) - except Meeting.DoesNotExist: - return err(400, "Invalid meeting value: '%s'" % (number, )) - reg_type = data['reg_type'] - email = data['email'] - try: - validate_email(email) - except ValidationError: - return err(400, "Invalid email value: '%s'" % (email, )) - if request.POST.get('cancelled', 'false') == 'true': - MeetingRegistration.objects.filter( - meeting_id=meeting.pk, - email=email, - reg_type=reg_type).delete() - return HttpResponse('OK', status=200, content_type='text/plain') - else: - object, created = MeetingRegistration.objects.get_or_create( - meeting_id=meeting.pk, - email=email, - reg_type=reg_type) - try: - # Update attributes - for key in set(data.keys())-set(['attended', 'apikey', 'meeting', 'email']): - if key == 'checkedin': - new = bool(data.get(key).lower() == 'true') - else: - new = data.get(key) - setattr(object, key, new) - person = Person.objects.filter(email__address=email) - if person.exists(): - object.person = person.first() - object.save() - except ValueError as e: - return err(400, "Unexpected POST data: %s" % e) - response = "Accepted, New registration" if created else "Accepted, Updated registration" - if User.objects.filter(username=email).exists() or Email.objects.filter(address=email).exists(): - pass - else: - send_account_creation_email(request, email) - response += ", Email sent" - return HttpResponse(response, status=202, content_type='text/plain') - else: - return HttpResponse(status=405) + def _http_err(code, text): + return HttpResponse( + text, + status=code, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) + + def _api_response(result): + return JsonResponse(data={"result": result}) + + if request.method != "POST": + return _http_err(405, "Method not allowed") + + if request.content_type != "application/json": + return _http_err(415, "Content-Type must be application/json") + + # Validate + try: + payload = json.loads(request.body) + import_registration_json_validator.validate(payload) + except json.decoder.JSONDecodeError as err: + return _http_err(400, f"JSON parse error at line {err.lineno} col {err.colno}: {err.msg}") + except jsonschema.exceptions.ValidationError as err: + return _http_err(400, f"JSON schema error at {err.json_path}: {err.message}") + except Exception: + return _http_err(400, "Invalid request format") + + # Get the meeting ID from the first registration, the API only deals with one meeting at a time + first_email = next(iter(payload['objects'])) + meeting_number = payload['objects'][first_email]['meeting'] + try: + meeting = Meeting.objects.get(number=meeting_number) + except Meeting.DoesNotExist: + return _http_err(400, f"Invalid meeting value: {meeting_number}") + + # confirm email exists + try: + Email.objects.get(address=first_email) + except Email.DoesNotExist: + return _http_err(400, f"Unknown email: {first_email}") + + reg_data = payload['objects'][first_email] + + process_single_registration(reg_data, meeting) + + return HttpResponse( + 'Success', + status=202, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) def version(request): @@ -210,9 +215,16 @@ def version(request): if dumpinfo.tz != "UTC": dumpdate = pytz.timezone(dumpinfo.tz).localize(dumpinfo.date.replace(tzinfo=None)) dumptime = dumpdate.strftime('%Y-%m-%d %H:%M:%S %z') if dumpinfo else None + + # important libraries + __version_extra__ = {} + for lib in settings.ADVERTISE_VERSIONS: + __version_extra__[lib] = metadata_version(lib) + return HttpResponse( json.dumps({ 'version': ietf.__version__+ietf.__patch__, + 'other': __version_extra__, 'dumptime': dumptime, }), content_type='application/json', @@ -221,7 +233,504 @@ def version(request): @require_api_key @csrf_exempt -def app_auth(request): +def app_auth(request, app: Literal["authortools", "bibxml"]): return HttpResponse( json.dumps({'success': True}), content_type='application/json') + +@requires_api_token +@csrf_exempt +def nfs_metrics(request): + with NamedTemporaryFile(dir=settings.NFS_METRICS_TMP_DIR,delete=False) as fp: + fp.close() + mark = datetime.datetime.now() + with open(fp.name, mode="w") as f: + f.write("whyioughta"*1024) + write_latency = (datetime.datetime.now() - mark).total_seconds() + mark = datetime.datetime.now() + with open(fp.name, "r") as f: + _=f.read() + read_latency = (datetime.datetime.now() - mark).total_seconds() + Path(f.name).unlink() + response=f'nfs_latency_seconds{{operation="write"}} {write_latency}\nnfs_latency_seconds{{operation="read"}} {read_latency}\n' + return HttpResponse(response) + +def find_doc_for_rfcdiff(name, rev): + """rfcdiff lookup heuristics + + Returns a tuple with: + [0] - condition string + [1] - document found (or None) + [2] - historic version + [3] - revision actually found (may differ from :rev: input) + """ + found = fuzzy_find_documents(name, rev) + condition = 'no such document' + if found.documents.count() != 1: + return (condition, None, None, rev) + doc = found.documents.get() + if found.matched_rev is None or doc.rev == found.matched_rev: + condition = 'current version' + return (condition, doc, None, found.matched_rev) + else: + candidate = doc.history_set.filter(rev=found.matched_rev).order_by("-time").first() + if candidate: + condition = 'historic version' + return (condition, doc, candidate, found.matched_rev) + else: + condition = 'version dochistory not found' + return (condition, doc, None, found.matched_rev) + +# This is a proof of concept of a service that would redirect to the current revision +# def rfcdiff_latest(request, name, rev=None): +# condition, doc, history = find_doc_for_rfcdiff(name, rev) +# if not doc: +# raise Http404 +# if history: +# return redirect(history.get_href()) +# else: +# return redirect(doc.get_href()) + +HAS_TOMBSTONE = [ + 2821, 2822, 2873, 2919, 2961, 3023, 3029, 3031, 3032, 3033, 3034, 3035, 3036, + 3037, 3038, 3042, 3044, 3050, 3052, 3054, 3055, 3056, 3057, 3059, 3060, 3061, + 3062, 3063, 3064, 3067, 3068, 3069, 3070, 3071, 3072, 3073, 3074, 3075, 3076, + 3077, 3078, 3080, 3081, 3082, 3084, 3085, 3086, 3087, 3088, 3089, 3090, 3094, + 3095, 3096, 3097, 3098, 3101, 3102, 3103, 3104, 3105, 3106, 3107, 3108, 3109, + 3110, 3111, 3112, 3113, 3114, 3115, 3116, 3117, 3118, 3119, 3120, 3121, 3123, + 3124, 3126, 3127, 3128, 3130, 3131, 3132, 3133, 3134, 3135, 3136, 3137, 3138, + 3139, 3140, 3141, 3142, 3143, 3144, 3145, 3147, 3149, 3150, 3151, 3152, 3153, + 3154, 3155, 3156, 3157, 3158, 3159, 3160, 3161, 3162, 3163, 3164, 3165, 3166, + 3167, 3168, 3169, 3170, 3171, 3172, 3173, 3174, 3176, 3179, 3180, 3181, 3182, + 3183, 3184, 3185, 3186, 3187, 3188, 3189, 3190, 3191, 3192, 3193, 3194, 3197, + 3198, 3201, 3202, 3203, 3204, 3205, 3206, 3207, 3208, 3209, 3210, 3211, 3212, + 3213, 3214, 3215, 3216, 3217, 3218, 3220, 3221, 3222, 3224, 3225, 3226, 3227, + 3228, 3229, 3230, 3231, 3232, 3233, 3234, 3235, 3236, 3237, 3238, 3240, 3241, + 3242, 3243, 3244, 3245, 3246, 3247, 3248, 3249, 3250, 3253, 3254, 3255, 3256, + 3257, 3258, 3259, 3260, 3261, 3262, 3263, 3264, 3265, 3266, 3267, 3268, 3269, + 3270, 3271, 3272, 3273, 3274, 3275, 3276, 3278, 3279, 3280, 3281, 3282, 3283, + 3284, 3285, 3286, 3287, 3288, 3289, 3290, 3291, 3292, 3293, 3294, 3295, 3296, + 3297, 3298, 3301, 3302, 3303, 3304, 3305, 3307, 3308, 3309, 3310, 3311, 3312, + 3313, 3315, 3317, 3318, 3319, 3320, 3321, 3322, 3323, 3324, 3325, 3326, 3327, + 3329, 3330, 3331, 3332, 3334, 3335, 3336, 3338, 3340, 3341, 3342, 3343, 3346, + 3348, 3349, 3351, 3352, 3353, 3354, 3355, 3356, 3360, 3361, 3362, 3363, 3364, + 3366, 3367, 3368, 3369, 3370, 3371, 3372, 3374, 3375, 3377, 3378, 3379, 3383, + 3384, 3385, 3386, 3387, 3388, 3389, 3390, 3391, 3394, 3395, 3396, 3397, 3398, + 3401, 3402, 3403, 3404, 3405, 3406, 3407, 3408, 3409, 3410, 3411, 3412, 3413, + 3414, 3415, 3416, 3417, 3418, 3419, 3420, 3421, 3422, 3423, 3424, 3425, 3426, + 3427, 3428, 3429, 3430, 3431, 3433, 3434, 3435, 3436, 3437, 3438, 3439, 3440, + 3441, 3443, 3444, 3445, 3446, 3447, 3448, 3449, 3450, 3451, 3452, 3453, 3454, + 3455, 3458, 3459, 3460, 3461, 3462, 3463, 3464, 3465, 3466, 3467, 3468, 3469, + 3470, 3471, 3472, 3473, 3474, 3475, 3476, 3477, 3480, 3481, 3483, 3485, 3488, + 3494, 3495, 3496, 3497, 3498, 3501, 3502, 3503, 3504, 3505, 3506, 3507, 3508, + 3509, 3511, 3512, 3515, 3516, 3517, 3518, 3520, 3521, 3522, 3523, 3524, 3525, + 3527, 3529, 3530, 3532, 3533, 3534, 3536, 3537, 3538, 3539, 3541, 3543, 3544, + 3545, 3546, 3547, 3548, 3549, 3550, 3551, 3552, 3555, 3556, 3557, 3558, 3559, + 3560, 3562, 3563, 3564, 3565, 3568, 3569, 3570, 3571, 3572, 3573, 3574, 3575, + 3576, 3577, 3578, 3579, 3580, 3581, 3582, 3583, 3584, 3588, 3589, 3590, 3591, + 3592, 3593, 3594, 3595, 3597, 3598, 3601, 3607, 3609, 3610, 3612, 3614, 3615, + 3616, 3625, 3627, 3630, 3635, 3636, 3637, 3638 +] + + +def get_previous_url(name, rev=None): + '''Return previous url''' + condition, document, history, found_rev = find_doc_for_rfcdiff(name, rev) + previous_url = '' + if condition in ('historic version', 'current version'): + doc = history if history else document + previous_url = doc.get_href() + elif condition == 'version dochistory not found': + document.rev = found_rev + previous_url = document.get_href() + return previous_url + + +def rfcdiff_latest_json(request, name, rev=None): + response = dict() + condition, document, history, found_rev = find_doc_for_rfcdiff(name, rev) + if document and document.type_id == "rfc": + draft = document.came_from_draft() + if condition == 'no such document': + raise Http404 + elif condition in ('historic version', 'current version'): + doc = history if history else document + if doc.type_id == "rfc": + response['content_url'] = doc.get_href() + response['name']=doc.name + if draft: + prev_rev = draft.rev + if doc.rfc_number in HAS_TOMBSTONE and prev_rev != '00': + prev_rev = f'{(int(draft.rev)-1):02d}' + response['previous'] = f'{draft.name}-{prev_rev}' + response['previous_url'] = get_previous_url(draft.name, prev_rev) + elif doc.type_id == "draft" and not found_rev and doc.relateddocument_set.filter(relationship_id="became_rfc").exists(): + rfc = doc.related_that_doc("became_rfc")[0] + response['content_url'] = rfc.get_href() + response['name']=rfc.name + prev_rev = doc.rev + if rfc.rfc_number in HAS_TOMBSTONE and prev_rev != '00': + prev_rev = f'{(int(doc.rev)-1):02d}' + response['previous'] = f'{doc.name}-{prev_rev}' + response['previous_url'] = get_previous_url(doc.name, prev_rev) + else: + response['content_url'] = doc.get_href() + response['rev'] = doc.rev + response['name'] = doc.name + if doc.rev == '00': + replaces_docs = (history.doc if condition=='historic version' else doc).related_that_doc('replaces') + if replaces_docs: + replaces = replaces_docs[0] + response['previous'] = f'{replaces.name}-{replaces.rev}' + response['previous_url'] = get_previous_url(replaces.name, replaces.rev) + else: + match = re.search("-(rfc)?([0-9][0-9][0-9]+)bis(-.*)?$", name) + if match and match.group(2): + response['previous'] = f'rfc{match.group(2)}' + response['previous_url'] = get_previous_url(f'rfc{match.group(2)}') + else: + # not sure what to do if non-numeric values come back, so at least log it + log.assertion('doc.rev.isdigit()') + prev_rev = f'{(int(doc.rev)-1):02d}' + response['previous'] = f'{doc.name}-{prev_rev}' + response['previous_url'] = get_previous_url(doc.name, prev_rev) + elif condition == 'version dochistory not found': + response['warning'] = 'History for this version not found - these results are speculation' + response['name'] = document.name + response['rev'] = found_rev + document.rev = found_rev + response['content_url'] = document.get_href() + # not sure what to do if non-numeric values come back, so at least log it + log.assertion('found_rev.isdigit()') + if int(found_rev) > 0: + prev_rev = f'{(int(found_rev)-1):02d}' + response['previous'] = f'{document.name}-{prev_rev}' + response['previous_url'] = get_previous_url(document.name, prev_rev) + else: + match = re.search("-(rfc)?([0-9][0-9][0-9]+)bis(-.*)?$", name) + if match and match.group(2): + response['previous'] = f'rfc{match.group(2)}' + response['previous_url'] = get_previous_url(f'rfc{match.group(2)}') + if not response: + raise Http404 + return HttpResponse(json.dumps(response), content_type='application/json') + +@csrf_exempt +def directauth(request): + if request.method == "POST": + raw_data = request.POST.get("data", None) + if raw_data: + try: + data = json.loads(raw_data) + except json.decoder.JSONDecodeError: + data = None + + if raw_data is None or data is None: + log.log("Request body is either missing or invalid") + return HttpResponse(json.dumps(dict(result="failure",reason="invalid post")), content_type='application/json') + + authtoken = data.get('authtoken', None) + username = data.get('username', None) + password = data.get('password', None) + + if any([item is None for item in (authtoken, username, password)]): + log.log("One or more mandatory fields are missing: authtoken, username, password") + return HttpResponse(json.dumps(dict(result="failure",reason="invalid post")), content_type='application/json') + + if not is_valid_token("ietf.api.views.directauth", authtoken): + log.log("Auth token provided is invalid") + return HttpResponse(json.dumps(dict(result="failure",reason="invalid authtoken")), content_type='application/json') + + user_query = User.objects.filter(username__iexact=username) + + # Matching email would be consistent with auth everywhere else in the app, but until we can map users well + # in the imap server, people's annotations are associated with a very specific login. + # If we get a second user of this API, add an "allow_any_email" argument. + + + # Note well that we are using user.username, not what was passed to the API. + user_count = user_query.count() + if user_count == 1 and authenticate(username = user_query.first().username, password = password): + user = user_query.get() + if user_query.filter(person__isnull=True).count() == 1: # Can't inspect user.person direclty here + log.log(f"Direct auth success (personless user): {user.pk}:{user.username}") + else: + log.log(f"Direct auth success: {user.pk}:{user.person.plain_name()}") + return HttpResponse(json.dumps(dict(result="success")), content_type='application/json') + + log.log(f"Direct auth failure: {username} ({user_count} user(s) found)") + return HttpResponse(json.dumps(dict(result="failure", reason="authentication failed")), content_type='application/json') + + else: + log.log(f"Request must be POST: {request.method} received") + return HttpResponse(status=405) + + +@requires_api_token +@csrf_exempt +def draft_aliases(request): + if request.method == "GET": + return JsonResponse( + { + "aliases": [ + { + "alias": alias, + "domains": ["ietf"], + "addresses": address_list, + } + for alias, address_list in DraftAliasGenerator() + ] + } + ) + return HttpResponse(status=405) + + +@requires_api_token +@csrf_exempt +def group_aliases(request): + if request.method == "GET": + return JsonResponse( + { + "aliases": [ + { + "alias": alias, + "domains": domains, + "addresses": address_list, + } + for alias, domains, address_list in GroupAliasGenerator() + ] + } + ) + return HttpResponse(status=405) + + +@requires_api_token +@csrf_exempt +def active_email_list(request): + if request.method == "GET": + return JsonResponse( + { + "addresses": list(Email.objects.filter(active=True).values_list("address", flat=True)), + } + ) + return HttpResponse(status=405) + + +@requires_api_token +@csrf_exempt +def related_email_list(request, email): + """Given an email address, returns all other email addresses known + to Datatracker, via Person object + """ + def _http_err(code, text): + return HttpResponse( + text, + status=code, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) + + if request.method == "GET": + try: + email_obj = Email.objects.get(address=email) + except Email.DoesNotExist: + return _http_err(404, "Email not found") + person = email_obj.person + if not person: + return JsonResponse({"addresses": []}) + return JsonResponse( + { + "addresses": list(person.email_set.values_list("address", flat=True)), + } + ) + return HttpResponse(status=405) + + +@requires_api_token +def role_holder_addresses(request): + if request.method == "GET": + return JsonResponse( + { + "addresses": list( + role_holder_emails() + .order_by("address") + .values_list("address", flat=True) + ) + } + ) + return HttpResponse(status=405) + + +_response_email_json_validator = jsonschema.Draft202012Validator( + schema={ + "type": "object", + "properties": { + "dest": { + "type": "string", + }, + "message": { + "type": "string", # base64-encoded mail message + }, + }, + "required": ["dest", "message"], + } +) + + +class EmailIngestionError(Exception): + """Exception indicating ingestion failed""" + def __init__( + self, + msg="Message rejected", + *, + email_body: Optional[str] = None, + email_recipients: Optional[Iterable[str]] = None, + email_attach_traceback=False, + email_original_message: Optional[bytes]=None, + ): + self.msg = msg + self.email_body = email_body + self.email_subject = msg + self.email_recipients = email_recipients + self.email_attach_traceback = email_attach_traceback + self.email_original_message = email_original_message + self.email_from = settings.SERVER_EMAIL + + @staticmethod + def _summarize_error(error): + frame = extract_tb(error.__traceback__)[-1] + return dedent(f"""\ + Error details: + Exception type: {type(error).__module__}.{type(error).__name__} + File: {frame.filename} + Line: {frame.lineno}""") + + def as_emailmessage(self) -> Optional[EmailMessage]: + """Generate an EmailMessage to report an error""" + if self.email_body is None: + return None + error = self if self.__cause__ is None else self.__cause__ + format_values = dict( + error=error, + error_summary=self._summarize_error(error), + ) + msg = EmailMessage() + if self.email_recipients is None: + msg["To"] = tuple(adm[1] for adm in settings.ADMINS) + else: + msg["To"] = self.email_recipients + msg["From"] = self.email_from + msg["Subject"] = self.msg + msg.set_content( + self.email_body.format(**format_values) + ) + if self.email_attach_traceback: + msg.add_attachment( + "".join(format_exception(None, error, error.__traceback__)), + filename="traceback.txt", + ) + if self.email_original_message is not None: + # Attach incoming message if it was provided. Send as a generic media + # type because we don't know for sure that it was actually a valid + # message. + msg.add_attachment( + self.email_original_message, + 'application', 'octet-stream', # media type + filename='original-message', + ) + return msg + + +def ingest_email_handler(request, test_mode=False): + """Ingest incoming email - handler + + Returns a 4xx or 5xx status code if the HTTP request was invalid or something went + wrong while processing it. If the request was valid, returns a 200. This may or may + not indicate that the message was accepted. + + If test_mode is true, actual processing of a valid message will be skipped. In this + mode, a valid request with a valid destination will be treated as accepted. The + "bad_dest" error may still be returned. + """ + + def _http_err(code, text): + return HttpResponse( + text, + status=code, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) + + def _api_response(result): + return JsonResponse(data={"result": result}) + + if request.method != "POST": + return _http_err(405, "Method not allowed") + + if request.content_type != "application/json": + return _http_err(415, "Content-Type must be application/json") + + # Validate + try: + payload = json.loads(request.body) + _response_email_json_validator.validate(payload) + except json.decoder.JSONDecodeError as err: + return _http_err(400, f"JSON parse error at line {err.lineno} col {err.colno}: {err.msg}") + except jsonschema.exceptions.ValidationError as err: + return _http_err(400, f"JSON schema error at {err.json_path}: {err.message}") + except Exception: + return _http_err(400, "Invalid request format") + + try: + message = base64.b64decode(payload["message"], validate=True) + except binascii.Error: + return _http_err(400, "Invalid message: bad base64 encoding") + + dest = payload["dest"] + valid_dest = False + try: + if dest == "iana-review": + valid_dest = True + if not test_mode: + iana_ingest_review_email(message) + elif dest == "ipr-response": + valid_dest = True + if not test_mode: + ipr_ingest_response_email(message) + elif dest.startswith("nomcom-feedback-"): + maybe_year = dest[len("nomcom-feedback-"):] + if maybe_year.isdecimal(): + valid_dest = True + if not test_mode: + nomcom_ingest_feedback_email(message, int(maybe_year)) + except EmailIngestionError as err: + error_email = err.as_emailmessage() + if error_email is not None: + with suppress(Exception): # send_smtp logs its own exceptions, ignore them here + send_smtp(error_email) + return _api_response("bad_msg") + + if not valid_dest: + return _api_response("bad_dest") + + return _api_response("ok") + + +@requires_api_token +@csrf_exempt +def ingest_email(request): + """Ingest incoming email + + Hands off to ingest_email_handler() with test_mode=False. This allows @requires_api_token to + give the test endpoint a distinct token from the real one. + """ + return ingest_email_handler(request, test_mode=False) + + +@requires_api_token +@csrf_exempt +def ingest_email_test(request): + """Ingest incoming email test endpoint + + Hands off to ingest_email_handler() with test_mode=True. This allows @requires_api_token to + give the test endpoint a distinct token from the real one. + """ + return ingest_email_handler(request, test_mode=True) diff --git a/ietf/api/views_rpc.py b/ietf/api/views_rpc.py new file mode 100644 index 0000000000..6bc45fe3da --- /dev/null +++ b/ietf/api/views_rpc.py @@ -0,0 +1,552 @@ +# Copyright The IETF Trust 2023-2026, All Rights Reserved +import os +import shutil +from pathlib import Path +from tempfile import TemporaryDirectory + +from django.conf import settings +from django.db import IntegrityError +from drf_spectacular.utils import OpenApiParameter +from rest_framework import mixins, parsers, serializers, viewsets, status +from rest_framework.decorators import action +from rest_framework.exceptions import APIException +from rest_framework.views import APIView +from rest_framework.response import Response + +from django.db.models import CharField as ModelCharField, OuterRef, Subquery, Q +from django.db.models.functions import Coalesce +from django.http import Http404 +from drf_spectacular.utils import extend_schema_view, extend_schema +from rest_framework import generics +from rest_framework.fields import CharField as DrfCharField +from rest_framework.filters import SearchFilter +from rest_framework.pagination import LimitOffsetPagination + +from ietf.api.serializers_rpc import ( + PersonSerializer, + FullDraftSerializer, + DraftSerializer, + SubmittedToQueueSerializer, + OriginalStreamSerializer, + ReferenceSerializer, + EmailPersonSerializer, + RfcWithAuthorsSerializer, + DraftWithAuthorsSerializer, + NotificationAckSerializer, + RfcPubSerializer, + RfcFileSerializer, + EditableRfcSerializer, +) +from ietf.doc.models import Document, DocHistory, RfcAuthor, DocEvent +from ietf.doc.serializers import RfcAuthorSerializer +from ietf.doc.storage_utils import remove_from_storage, store_file, exists_in_storage +from ietf.doc.tasks import ( + signal_update_rfc_metadata_task, + rebuild_reference_relations_task, + trigger_red_precomputer_task, + update_rfc_searchindex_task, +) +from ietf.person.models import Email, Person +from ietf.sync.rfcindex import mark_rfcindex_as_dirty + + +class Conflict(APIException): + status_code = status.HTTP_409_CONFLICT + default_detail = "Conflict." + default_code = "conflict" + + +@extend_schema_view( + retrieve=extend_schema( + operation_id="get_person_by_id", + summary="Find person by ID", + description="Returns a single person", + parameters=[ + OpenApiParameter( + name="person_id", + type=int, + location="path", + description="Person ID identifying this person.", + ), + ], + ), +) +class PersonViewSet(mixins.RetrieveModelMixin, viewsets.GenericViewSet): + queryset = Person.objects.all() + serializer_class = PersonSerializer + api_key_endpoint = "ietf.api.views_rpc" + lookup_url_kwarg = "person_id" + + @extend_schema( + operation_id="get_persons", + summary="Get a batch of persons", + description="Returns a list of persons matching requested ids. Omits any that are missing.", + request=list[int], + responses=PersonSerializer(many=True), + ) + @action(detail=False, methods=["post"]) + def batch(self, request): + """Get a batch of rpc person names""" + pks = request.data + return Response( + self.get_serializer(Person.objects.filter(pk__in=pks), many=True).data + ) + + @extend_schema( + operation_id="persons_by_email", + summary="Get a batch of persons by email addresses", + description=( + "Returns a list of persons matching requested ids. " + "Omits any that are missing." + ), + request=list[str], + responses=EmailPersonSerializer(many=True), + ) + @action(detail=False, methods=["post"], serializer_class=EmailPersonSerializer) + def batch_by_email(self, request): + emails = Email.objects.filter(address__in=request.data, person__isnull=False) + serializer = self.get_serializer(emails, many=True) + return Response(serializer.data) + + +class SubjectPersonView(APIView): + api_key_endpoint = "ietf.api.views_rpc" + + @extend_schema( + operation_id="get_subject_person_by_id", + summary="Find person for OIDC subject by ID", + description="Returns a single person", + responses=PersonSerializer, + parameters=[ + OpenApiParameter( + name="subject_id", + type=str, + description="subject ID of person to return", + location="path", + ), + ], + ) + def get(self, request, subject_id: str): + try: + user_id = int(subject_id) + except ValueError: + raise serializers.ValidationError( + {"subject_id": "This field must be an integer value."} + ) + person = Person.objects.filter(user__pk=user_id).first() + if person: + return Response(PersonSerializer(person).data) + raise Http404 + + +class RpcLimitOffsetPagination(LimitOffsetPagination): + default_limit = 10 + max_limit = 100 + + +class SingleTermSearchFilter(SearchFilter): + """SearchFilter backend that does not split terms + + The default SearchFilter treats comma or whitespace-separated terms as individual + search terms. This backend instead searches for the exact term. + """ + + def get_search_terms(self, request): + value = request.query_params.get(self.search_param, "") + field = DrfCharField(trim_whitespace=False, allow_blank=True) + cleaned_value = field.run_validation(value) + return [cleaned_value] + + +@extend_schema_view( + get=extend_schema( + operation_id="search_person", + description="Get a list of persons, matching by partial name or email", + ), +) +class RpcPersonSearch(generics.ListAPIView): + # n.b. the OpenAPI schema for this can be generated by running + # ietf/manage.py spectacular --file spectacular.yaml + # and extracting / touching up the rpc_person_search_list operation + api_key_endpoint = "ietf.api.views_rpc" + queryset = Person.objects.all() + serializer_class = PersonSerializer + pagination_class = RpcLimitOffsetPagination + + # Searchable on all name-like fields or email addresses + filter_backends = [SingleTermSearchFilter] + search_fields = ["name", "plain", "email__address"] + + +@extend_schema_view( + retrieve=extend_schema( + operation_id="get_draft_by_id", + summary="Get a draft", + description="Returns the draft for the requested ID", + parameters=[ + OpenApiParameter( + name="doc_id", + type=int, + location="path", + description="Doc ID identifying this draft.", + ), + ], + ), + submitted_to_rpc=extend_schema( + operation_id="submitted_to_rpc", + summary="List documents ready to enter the RFC Editor Queue", + description="List documents ready to enter the RFC Editor Queue", + responses=SubmittedToQueueSerializer(many=True), + ), +) +class DraftViewSet(mixins.RetrieveModelMixin, viewsets.GenericViewSet): + queryset = Document.objects.filter(type_id="draft") + serializer_class = FullDraftSerializer + api_key_endpoint = "ietf.api.views_rpc" + lookup_url_kwarg = "doc_id" + + @action(detail=False, serializer_class=SubmittedToQueueSerializer) + def submitted_to_rpc(self, request): + """Return documents in datatracker that have been submitted to the RPC but are not yet in the queue + + Those queries overreturn - there may be things, particularly not from the IETF stream that are already in the queue. + """ + ietf_docs = Q(states__type_id="draft-iesg", states__slug__in=["ann"]) + irtf_iab_ise_editorial_docs = Q( + states__type_id__in=[ + "draft-stream-iab", + "draft-stream-irtf", + "draft-stream-ise", + "draft-stream-editorial", + ], + states__slug__in=["rfc-edit"], + ) + docs = ( + self.get_queryset() + .filter(type_id="draft") + .filter(ietf_docs | irtf_iab_ise_editorial_docs) + ) + serializer = self.get_serializer(docs, many=True) + return Response(serializer.data) + + @extend_schema( + operation_id="get_draft_references", + summary="Get normative references to I-Ds", + description=( + "Returns the id and name of each normatively " + "referenced Internet-Draft for the given docId" + ), + parameters=[ + OpenApiParameter( + name="doc_id", + type=int, + location="path", + description="Doc ID identifying this draft.", + ), + ], + responses=ReferenceSerializer(many=True), + ) + @action(detail=True, serializer_class=ReferenceSerializer) + def references(self, request, doc_id=None): + doc = self.get_object() + serializer = self.get_serializer( + [ + reference + for reference in doc.related_that_doc("refnorm") + if reference.type_id == "draft" + ], + many=True, + ) + return Response(serializer.data) + + @extend_schema( + operation_id="get_draft_authors", + summary="Gather authors of the drafts with the given names", + description="returns a list mapping draft names to objects describing authors", + request=list[str], + responses=DraftWithAuthorsSerializer(many=True), + ) + @action(detail=False, methods=["post"], serializer_class=DraftWithAuthorsSerializer) + def bulk_authors(self, request): + drafts = self.get_queryset().filter(name__in=request.data) + serializer = self.get_serializer(drafts, many=True) + return Response(serializer.data) + + +@extend_schema_view( + rfc_original_stream=extend_schema( + operation_id="get_rfc_original_streams", + summary="Get the streams RFCs were originally published into", + description="returns a list of dicts associating an RFC with its originally published stream", + responses=OriginalStreamSerializer(many=True), + ) +) +class RfcViewSet(mixins.UpdateModelMixin, viewsets.GenericViewSet): + queryset = Document.objects.filter(type_id="rfc") + api_key_endpoint = "ietf.api.views_rpc" + lookup_field = "rfc_number" + serializer_class = EditableRfcSerializer + + def perform_update(self, serializer): + DocEvent.objects.create( + doc=serializer.instance, + rev=serializer.instance.rev, + by=Person.objects.get(name="(System)"), + type="sync_from_rfc_editor", + desc="Metadata update from RFC Editor", + ) + super().perform_update(serializer) + + @action(detail=False, serializer_class=OriginalStreamSerializer) + def rfc_original_stream(self, request): + rfcs = self.get_queryset().annotate( + orig_stream_id=Coalesce( + Subquery( + DocHistory.objects.filter(doc=OuterRef("pk")) + .exclude(stream__isnull=True) + .order_by("time") + .values_list("stream_id", flat=True)[:1] + ), + "stream_id", + output_field=ModelCharField(), + ), + ) + serializer = self.get_serializer(rfcs, many=True) + return Response(serializer.data) + + @extend_schema( + operation_id="get_rfc_authors", + summary="Gather authors of the RFCs with the given numbers", + description="returns a list mapping rfc numbers to objects describing authors", + request=list[int], + responses=RfcWithAuthorsSerializer(many=True), + ) + @action(detail=False, methods=["post"], serializer_class=RfcWithAuthorsSerializer) + def bulk_authors(self, request): + rfcs = self.get_queryset().filter(rfc_number__in=request.data) + serializer = self.get_serializer(rfcs, many=True) + return Response(serializer.data) + + +class DraftsByNamesView(APIView): + api_key_endpoint = "ietf.api.views_rpc" + + @extend_schema( + operation_id="get_drafts_by_names", + summary="Get a batch of drafts by draft names", + description="returns a list of drafts with matching names", + request=list[str], + responses=DraftSerializer(many=True), + ) + def post(self, request): + names = request.data + docs = Document.objects.filter(type_id="draft", name__in=names) + return Response(DraftSerializer(docs, many=True).data) + + +class RfcAuthorViewSet(viewsets.ReadOnlyModelViewSet): + """ViewSet for RfcAuthor model + + Router needs to provide rfc_number as a kwarg + """ + + api_key_endpoint = "ietf.api.views_rpc" + + queryset = RfcAuthor.objects.all() + serializer_class = RfcAuthorSerializer + lookup_url_kwarg = "author_id" + rfc_number_param = "rfc_number" + + def get_queryset(self): + return ( + super() + .get_queryset() + .filter( + document__type_id="rfc", + document__rfc_number=self.kwargs[self.rfc_number_param], + ) + ) + + +class RfcPubNotificationView(APIView): + api_key_endpoint = "ietf.api.views_rpc" + + @extend_schema( + operation_id="notify_rfc_published", + summary="Notify datatracker of RFC publication", + request=RfcPubSerializer, + responses=NotificationAckSerializer, + ) + def post(self, request): + serializer = RfcPubSerializer(data=request.data) + serializer.is_valid(raise_exception=True) + # Create RFC + try: + rfc = serializer.save() + except IntegrityError as err: + if Document.objects.filter( + rfc_number=serializer.validated_data["rfc_number"] + ): + raise serializers.ValidationError( + "RFC with that number already exists", + code="rfc-number-in-use", + ) + raise serializers.ValidationError( + f"Unable to publish: {err}", + code="unknown-integrity-error", + ) + rfc_number_list = [rfc.rfc_number] + rfc_number_list.extend( + [d.rfc_number for d in rfc.related_that_doc(("updates", "obs"))] + ) + rfc_number_list = sorted(set(rfc_number_list)) + signal_update_rfc_metadata_task.delay(rfc_number_list=rfc_number_list) + return Response(NotificationAckSerializer().data) + + +class RfcPubFilesView(APIView): + api_key_endpoint = "ietf.api.views_rpc" + parser_classes = [parsers.MultiPartParser] + + def _fs_destination(self, filename: str | Path) -> Path: + """Destination for an uploaded RFC file in the filesystem + + Strips any path components in filename and returns an absolute Path. + """ + rfc_path = Path(settings.RFC_PATH) + filename = Path(filename) # could potentially have directory components + extension = "".join(filename.suffixes) + if extension == ".notprepped.xml": + return rfc_path / "prerelease" / filename.name + return rfc_path / filename.name + + def _blob_destination(self, filename: str | Path) -> str: + """Destination name for an uploaded RFC file in the blob store + + Strips any path components in filename and returns an absolute Path. + """ + filename = Path(filename) # could potentially have directory components + extension = "".join(filename.suffixes) + if extension == ".notprepped.xml": + file_type = "notprepped" + elif extension[0] == ".": + file_type = extension[1:] + else: + raise serializers.ValidationError( + f"Extension does not begin with '.'!? ({filename})", + ) + return f"{file_type}/{filename.name}" + + @extend_schema( + operation_id="upload_rfc_files", + summary="Upload files for a published RFC", + request=RfcFileSerializer, + responses=NotificationAckSerializer, + ) + def post(self, request): + serializer = RfcFileSerializer( + # many=True, + data=request.data, + ) + serializer.is_valid(raise_exception=True) + rfc = serializer.validated_data["rfc"] + uploaded_files = serializer.validated_data["contents"] # list[UploadedFile] + replace = serializer.validated_data["replace"] + dest_stem = f"rfc{rfc.rfc_number}" + mtime = serializer.validated_data["mtime"] + mtimestamp = mtime.timestamp() + blob_kind = "rfc" + + # List of files that might exist for an RFC + possible_rfc_files = [ + self._fs_destination(dest_stem + ext) + for ext in serializer.allowed_extensions + ] + possible_rfc_blobs = [ + self._blob_destination(dest_stem + ext) + for ext in serializer.allowed_extensions + ] + if not replace: + # this is the default: refuse to overwrite anything if not replacing + for possible_existing_file in possible_rfc_files: + if possible_existing_file.exists(): + raise Conflict( + "File(s) already exist for this RFC", + code="files-exist", + ) + for possible_existing_blob in possible_rfc_blobs: + if exists_in_storage(kind=blob_kind, name=possible_existing_blob): + raise Conflict( + "Blob(s) already exist for this RFC", + code="blobs-exist", + ) + + with TemporaryDirectory() as tempdir: + # Save files in a temporary directory. Use the uploaded filename + # extensions to identify files, but ignore the stems and generate our own. + files_to_move = [] # list[Path] + tmpfile_stem = Path(tempdir) / dest_stem + for upfile in uploaded_files: + uploaded_filename = Path(upfile.name) # name supplied by request + uploaded_ext = "".join(uploaded_filename.suffixes) + tempfile_path = tmpfile_stem.with_suffix(uploaded_ext) + with tempfile_path.open("wb") as dest: + for chunk in upfile.chunks(): + dest.write(chunk) + os.utime(tempfile_path, (mtimestamp, mtimestamp)) + files_to_move.append(tempfile_path) + # copy files to final location, removing any existing ones first if the + # remove flag was set + if replace: + for possible_existing_file in possible_rfc_files: + possible_existing_file.unlink(missing_ok=True) + for possible_existing_blob in possible_rfc_blobs: + remove_from_storage( + blob_kind, possible_existing_blob, warn_if_missing=False + ) + for ftm in files_to_move: + with ftm.open("rb") as f: + store_file( + kind=blob_kind, + name=self._blob_destination(ftm), + file=f, + doc_name=rfc.name, + doc_rev=rfc.rev, # expect blank, but match whatever it is + mtime=mtime, + ) + destination = self._fs_destination(ftm) + if ( + settings.SERVER_MODE != "production" + and not destination.parent.exists() + ): + destination.parent.mkdir() + shutil.move(ftm, destination) + + # Trigger red precomputer + needs_updating = [rfc.rfc_number] + for rel in rfc.relateddocument_set.filter( + relationship_id__in=["obs", "updates"] + ): + needs_updating.append(rel.target.rfc_number) + trigger_red_precomputer_task.delay(rfc_number_list=sorted(needs_updating)) + # Trigger search index update + update_rfc_searchindex_task.delay(rfc.rfc_number) + # Trigger reference relation srebuild + rebuild_reference_relations_task.delay(doc_names=[rfc.name]) + + return Response(NotificationAckSerializer().data) + + +class RfcIndexView(APIView): + api_key_endpoint = "ietf.api.views_rpc" + + @extend_schema( + operation_id="refresh_rfc_index", + summary="Refresh rfc-index files", + description="Requests creation of various index files.", + responses={202: None}, + request=None, + ) + def post(self, request): + mark_rfcindex_as_dirty() + return Response(status=202) diff --git a/ietf/bin/.gitignore b/ietf/bin/.gitignore deleted file mode 100644 index c7013ced9d..0000000000 --- a/ietf/bin/.gitignore +++ /dev/null @@ -1,2 +0,0 @@ -/*.pyc -/settings_local.py diff --git a/ietf/bin/2016-05-25-collect-photos b/ietf/bin/2016-05-25-collect-photos deleted file mode 100755 index dedda767a8..0000000000 --- a/ietf/bin/2016-05-25-collect-photos +++ /dev/null @@ -1,296 +0,0 @@ -#!/usr/bin/env python - -import os, re, sys, shutil, pathlib -from collections import namedtuple -from PIL import Image - -# boilerplate -basedir = os.path.abspath(os.path.join(os.path.dirname(__file__), "../..")) -sys.path = [ basedir ] + sys.path -os.environ["DJANGO_SETTINGS_MODULE"] = "ietf.settings" - -virtualenv_activation = os.path.join(basedir, "env", "bin", "activate_this.py") -if os.path.exists(virtualenv_activation): - execfile(virtualenv_activation, dict(__file__=virtualenv_activation)) - -import django -django.setup() - -from django.conf import settings -from django.utils.text import slugify - -import debug - -from ietf.group.models import Role, Person -from ietf.person.name import name_parts - -old_images_dir = '' -new_images_dir = settings.PHOTOS_DIR - -if not os.path.exists(new_images_dir): - print("New images directory does not exist: %s" % new_images_dir) - sys.exit(1) - -old_image_files = [] -for dir in settings.OLD_PHOTO_DIRS: - if not os.path.exists(dir): - print("Old images directory does not exist: %s" % dir) - sys.exit(1) - old_image_files += [ f for f in pathlib.Path(dir).iterdir() if f.is_file() and f.suffix.lower() in ['.jpg', '.jpeg', '.png'] ] - -photo = namedtuple('photo', ['path', 'name', 'ext', 'width', 'height', 'time', 'file']) - -old_images = [] -for f in old_image_files: - path = str(f) - img = Image.open(path) - old_images.append(photo(path, f.stem.decode('utf8'), f.suffix, img.size[0], img.size[1], f.stat().st_mtime, f)) - -# Fix up some names: - -def fix_missing_surnames(images): - replacement = { - "alissa": "alissa-cooper", - "alissa1": "alissa-cooper", - "andrei": "andrei-robachevsky", - "bernard": "bernard-aboba", - "danny": "danny-mcpherson", - "danny1": "danny-mcpherson", - "dthaler": "dave-thaler", - "eliot-mug": "eliot-lear", - "erik.nordmark-300": "erik-nordmark", - "hannes": "hannes-tschofenig", - "hildebrand": "joe-hildebrand", - "housley": "russ-housley", - "jariarkko": "jari-arkko", - "joel": "joel-jaeggli", - "joel1": "joel-jaeggli", - "joel2": "joel-jaeggli", - "jon": "jon-peterson", - "kessens": "david-kessens", - "klensin": "john-klensin", - "lars": "lars-eggert", - "lars1": "lars-eggert", - "marc_blanchet": "marc-blanchet", - "marcelo": "marcelo-bagnulo", - "olaf": "olaf-kolkman", - "olaf1": "olaf-kolkman", - "ross": "ross-callon", - "spencer": "spencer-dawkins", - "spencer1": "spencer-dawkins", - "vijay": "vijay-gurbani", - "xing": "xing-li", - } - - for i in range(len(images)): - img = images[i] - name = re.sub('-[0-9]+x[0-9]+', '', img.name) - if '/iab/' in img.path and name in replacement: - name = replacement[name] - images[i] = photo(img.path, name, img.ext, img.width, img.height, img.time, img.file) - - -fix_missing_surnames(old_images) - -interesting_persons = set(Person.objects.all()) - -name_alias = { - u"andy": [u"andrew", ], - u"ben": [u"benjamin", ], - u"bill": [u"william", ], - u"bob": [u"robert", ], - u"chris": [u"christopher", u"christian"], - u"dan": [u"daniel", ], - u"dave": [u"david", ], - u"dick": [u"richard", ], - u"fred": [u"alfred", ], - u"geoff": [u"geoffrey", ], - u"jake": [u"jacob", ], - u"jerry": [u"gerald", ], - u"jim": [u"james", ], - u"joe": [u"joseph", ], - u"jon": [u"jonathan", ], - u"mike": [u"michael", ], - u"ned": [u"edward", ], - u"pete": [u"peter", ], - u"ron": [u"ronald", ], - u"russ": [u"russel", ], - u"steve": [u"stephen", ], - u"ted": [u"edward", ], - u"terry": [u"terence", ], - u"tom": [u"thomas", ], - u"wes": [u"wesley", ], - u"will": [u"william", ], - - u"beth": [u"elizabeth", ], - u"liz": [u"elizabeth", ], - u"lynn": [u"carolyn", ], - u"pat": [u"patricia", u"patrick", ], - u"sue": [u"susan", ], -} -# Add lookups from long to short, from the initial set -for key,value in name_alias.items(): - for item in value: - if item in name_alias: - name_alias[item] += [ key ]; - else: - name_alias[item] = [ key ]; - -exceptions = { -'Aboba' : 'aboba-bernard', -'Bernardos' : 'cano-carlos', -'Bormann' : 'bormann-carsten', -'Hinden' : 'hinden-bob', -'Hutton' : 'hutton-andy', -'Narten' : 'narten-thomas', # but there's no picture of him -'O\'Donoghue' : 'odonoghue-karen', -'Przygienda' : 'przygienda-antoni', -'Salowey' : 'salowey-joe', -'Gunter Van de Velde' : 'vandevelde-gunter', -'Eric Vyncke' : 'vynke-eric', -'Zuniga' : 'zuniga-carlos-juan', -'Zhen Cao' : 'zhen-cao', -'Jamal Hadi Salim': 'hadi-salim-jamal', -} - -# Manually copied Bo Burman and Thubert Pascal from wg/photos/ -# Manually copied Victor Pascual (main image, not thumb) from wg/ -# Manually copied Eric Vync?ke (main image, not thumb) from wg/photos/ -# Manually copied Danial King (main image, not thumb) from wg/photos/ -# Manually copied the thumb (not labelled as such) for Tianran Zhou as both the main and thumb image from wg/photos/ - -processed_files = [] - -for person in sorted(list(interesting_persons),key=lambda x:x.last_name()+x.ascii): - substr_pattern = None - for exception in exceptions: - if exception in person.ascii: - substr_pattern = exceptions[exception] - break - if not person.ascii.strip(): - print(" Setting person.ascii for %s" % person.name) - person.ascii = person.name.encode('ascii', errors='replace').decode('ascii') - - _, first, _, last, _ = person.ascii_parts() - first = first.lower() - last = last. lower() - if not substr_pattern: - substr_pattern = slugify("%s %s" % (last, first)) - - if first in ['', '<>'] or last in ['', '<>']: - continue - - #debug.show('1, substr_pattern') - - candidates = [x for x in old_images if x.name.lower().startswith(substr_pattern)] - # Also check the reverse the name order (necessary for Deng Hui, for instance) - substr_pattern = slugify("%s %s" % (first, last)) - #debug.show('2, substr_pattern') - prev_len = len(candidates) - candidates += [x for x in old_images if x.name.lower().startswith(substr_pattern)] - if prev_len < len(candidates) : - print(" Found match with '%s %s' for '%s %s'" % (last, first, first, last, )) - # If no joy, try a short name - if first in name_alias: - prev_len = len(candidates) - for alias in name_alias[first]: - substr_pattern = slugify("%s %s" % (last, alias)) - #debug.show('3, substr_pattern') - candidates += [x for x in old_images if x.name.lower().startswith(substr_pattern)] - if prev_len < len(candidates): - print(" Found match with '%s %s' for '%s %s'" % (alias, last, first, last, )) - - -# # If still no joy, try with Person.plain_name() (necessary for Donald Eastlake) -# if not candidates: -# prefix, first, middle, last, suffix = person.name_parts() -# name_parts = person.plain_name().lower().split() -# -# substr_pattern = u'-'.join(name_parts[-1:]+name_parts[0:1]) -# candidates = [x for x in old_images if x.name.lower().startswith(substr_pattern)] -# # If no joy, try a short name -# if not candidates and first in name_alias: -# prev_len = len(candidates) -# for alias in name_alias[first]: -# substr_pattern = u'-'.join(name_parts[-1:]+[alias]) -# candidates += [x for x in old_images if x.name.lower().startswith(substr_pattern)] -# if prev_len < len(candidates) : -# print(" Used '%s %s' instead of '%s %s'" % (alias, last, first, last, )) - -# # Fixup for other exceptional cases -# if person.ascii=="David Oran": -# candidates = ['oran-dave-th.jpg','oran-david.jpg'] -# -# if person.ascii=="Susan Hares": -# candidates = ['hares-sue-th.jpg','hares-susan.JPG'] -# -# if person.ascii=="Mahesh Jethanandani": -# candidates = ['Mahesh-Jethanandani-th.jpg','Jethanandani-Mahesh.jpg'] - - processed_files += [ c.path for c in candidates ] - - # We now have a list of candidate photos. - # * Consider anything less than 200x200 a thumbnail - # * For the full photo, sort by size (width) and time - # * For the thumbnail: - # - first look for a square photo less than 200x200 - # - if none found, then for the first in the sorted list less than 200x200 - # - if none found, then the smallest photo - if candidates: - candidates.sort(key=lambda x: "%04d-%d" % (x.width, x.time)) - iesg_cand = [ c for c in candidates if '/iesg/' in c.path ] - iab_cand = [ c for c in candidates if '/iab/' in c.path ] - if iesg_cand: - full = iesg_cand[-1] - thumb = iesg_cand[-1] - elif iab_cand: - full = iab_cand[-1] - thumb = iab_cand[0] - else: - full = candidates[-1] - thumbs = [ c for c in candidates if c.width==c.height and c.width <= 200 ] - if not thumbs: - thumbs = [ c for c in candidates if c.width==c.height ] - if not thumbs: - thumbs = [ c for c in candidates if c.width <= 200 ] - if not thumbs: - thumbs = candidates[:1] - thumb = thumbs[-1] - candidates = [ thumb, full ] - - # At this point we either have no candidates or two. If two, the first will be the thumb - - def copy(old, new): - if not os.path.exists(new): - print("Copying "+old+" to "+new) - shutil.copy(old, new) - shutil.copystat(old, new) - - assert(len(candidates) in [0,2]) - if len(candidates)==2: - thumb, full = candidates - - new_name = person.photo_name(thumb=False)+full.ext.lower() - new_thumb_name = person.photo_name(thumb=True)+thumb.ext.lower() - - copy( full.path, os.path.join(new_images_dir,new_name) ) - - # - copy( thumb.path, os.path.join(new_images_dir,new_thumb_name) ) - - -print("") -not_processed = 0 -for file in old_image_files: - if ( file.is_file() - and not file.suffix.lower() in ['.txt', '.lck', '.html',] - and not file.name.startswith('index.') - and not file.name.startswith('milestoneupdate') - and not file.name.startswith('nopicture') - and not file.name.startswith('robots.txt') - ): - if not str(file).decode('utf8') in processed_files: - not_processed += 1 - print(u"Not processed: "+str(file).decode('utf8')) -print("") -print("Not processed: %s files" % not_processed) diff --git a/ietf/bin/aliases-from-json.py b/ietf/bin/aliases-from-json.py new file mode 100644 index 0000000000..0da5d1f8b9 --- /dev/null +++ b/ietf/bin/aliases-from-json.py @@ -0,0 +1,104 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +# +# Uses only Python standard lib +# + +import argparse +import datetime +import json +import shutil +import stat +import sys + +from pathlib import Path +from tempfile import TemporaryDirectory + +# Default options +POSTCONFIRM_PATH = "/a/postconfirm/wrapper" +VDOMAIN = "virtual.ietf.org" + +# Map from domain label to dns domain +ADOMAINS = { + "ietf": "ietf.org", + "irtf": "irtf.org", + "iab": "iab.org", +} + + +def generate_files(records, adest, vdest, postconfirm, vdomain): + """Generate files from an iterable of records + + If adest or vdest exists as a file, it will be overwritten. If it is a directory, files + with the default names (draft-aliases and draft-virtual) will be created, but existing + files _will not_ be overwritten! + """ + with TemporaryDirectory() as tmpdir: + tmppath = Path(tmpdir) + apath = tmppath / "aliases" + vpath = tmppath / "virtual" + + with apath.open("w") as afile, vpath.open("w") as vfile: + date = datetime.datetime.now(datetime.UTC) + signature = f"# Generated by {Path(__file__).absolute()} at {date}\n" + afile.write(signature) + vfile.write(signature) + vfile.write(f"{vdomain} anything\n") + + for item in records: + alias = item["alias"] + domains = item["domains"] + address_list = item["addresses"] + filtername = f"xfilter-{alias}" + afile.write(f'{filtername + ":":64s} "|{postconfirm} filter expand-{alias} {vdomain}"\n') + for dom in domains: + vfile.write(f"{f'{alias}@{ADOMAINS[dom]}':64s} {filtername}\n") + vfile.write(f"{f'expand-{alias}@{vdomain}':64s} {', '.join(sorted(address_list))}\n") + + perms = stat.S_IWUSR | stat.S_IRUSR | stat.S_IRGRP | stat.S_IROTH + apath.chmod(perms) + vpath.chmod(perms) + shutil.move(apath, adest) + shutil.move(vpath, vdest) + + +def directory_path(val): + p = Path(val) + if p.is_dir(): + return p + else: + raise argparse.ArgumentTypeError(f"{p} is not a directory") + +if __name__ == "__main__": + parser = argparse.ArgumentParser( + description="Convert a JSON stream of draft alias definitions into alias / virtual alias files." + ) + parser.add_argument( + "--prefix", + required=True, + help="Prefix for output files. Files will be named -aliases and -virtual." + ) + parser.add_argument( + "--output-dir", + default="./", + type=directory_path, + help="Destination for output files.", + ) + parser.add_argument( + "--postconfirm", + default=POSTCONFIRM_PATH, + help=f"Full path to postconfirm executable (defaults to {POSTCONFIRM_PATH}", + ) + parser.add_argument( + "--vdomain", + default=VDOMAIN, + help=f"Virtual domain (defaults to {VDOMAIN}_", + ) + args = parser.parse_args() + data = json.load(sys.stdin) + generate_files( + data["aliases"], + adest=args.output_dir / f"{args.prefix}-aliases", + vdest=args.output_dir / f"{args.prefix}-virtual", + postconfirm=args.postconfirm, + vdomain=args.vdomain, + ) diff --git a/ietf/bin/announce-header-change b/ietf/bin/announce-header-change deleted file mode 100755 index 256324e31a..0000000000 --- a/ietf/bin/announce-header-change +++ /dev/null @@ -1,48 +0,0 @@ -#!/usr/bin/env python - -import sys, os, sys -import datetime - -# boilerplate -basedir = os.path.abspath(os.path.join(os.path.dirname(__file__), "../..")) -sys.path = [ basedir ] + sys.path -os.environ["DJANGO_SETTINGS_MODULE"] = "ietf.settings" - -virtualenv_activation = os.path.join(basedir, "env", "bin", "activate_this.py") -if os.path.exists(virtualenv_activation): - execfile(virtualenv_activation, dict(__file__=virtualenv_activation)) - -import django -django.setup() - -from django.core import management -from django.template.loader import render_to_string - -from ietf import settings -from ietf.utils.mail import send_mail_preformatted -from ietf.utils.mail import send_mail - -target_date=datetime.date(year=2014,month=1,day=24) - -send_mail(request = None, - to = "IETF-Announce ", - frm = "The IESG ", - subject = "Upcoming change to announcement email header fields (using old header)", - template = "utils/header_change_content.txt", - context = dict(oldornew='old', target_date=target_date), - extra = {'Reply-To' : 'ietf@ietf.org', - 'Sender' : '', - } - ) - -send_mail(request = None, - to = "IETF-Announce:;", - frm = "The IESG ", - subject = "Upcoming change to announcement email header fields (using new header)", - template = "utils/header_change_content.txt", - context = dict(oldornew='new', target_date=target_date), - extra = {'Reply-To' : 'IETF Discussion List ', - 'Sender' : '', - }, - bcc = '', - ) diff --git a/ietf/bin/create-break-sessions b/ietf/bin/create-break-sessions deleted file mode 100755 index 52ce044d8c..0000000000 --- a/ietf/bin/create-break-sessions +++ /dev/null @@ -1,43 +0,0 @@ -#!/usr/bin/env python -# -*- coding: utf-8 -*- -# -*- Python -*- -# - -import os, sys - -basedir = os.path.abspath(os.path.join(os.path.dirname(__file__), "../..")) -sys.path = [ basedir ] + sys.path -os.environ["DJANGO_SETTINGS_MODULE"] = "ietf.settings" - -virtualenv_activation = os.path.join(basedir, "env", "bin", "activate_this.py") -if os.path.exists(virtualenv_activation): - execfile(virtualenv_activation, dict(__file__=virtualenv_activation)) - -import django -django.setup() - -from ietf.group.models import Group -from ietf.person.models import Person -from ietf.name.models import SessionStatusName -from ietf.meeting.models import Meeting, Session, ScheduledSession as ScheduleTimeslotSSessionAssignment - -secretariat = Group.objects.get(acronym='secretariat') -system = Person.objects.get(id=1, name='(System)') -scheduled = SessionStatusName.objects.get(slug='sched') - -for meeting in Meeting.objects.filter(type="ietf").order_by("date"): - print "Checking %s schedules ..." % meeting - brk, __ = Session.objects.get_or_create(meeting=meeting, group=secretariat, requested_by=system, status=scheduled, name='Break', type_id='break',) - reg, __ = Session.objects.get_or_create(meeting=meeting, group=secretariat, requested_by=system, status=scheduled, name='Registration', type_id='reg',) - - for schedule in meeting.schedule_set.all(): - print " Checking for missing Break and Reg sessions in %s" % schedule - for timeslot in meeting.timeslot_set.all(): - if timeslot.type_id == 'break' and not (schedule.base and SchedTimeSessAssignment.objects.filter(timeslot=timeslot, session=brk, schedule=schedule.base).exists()): - assignment, created = SchedTimeSessAssignment.objects.get_or_create(timeslot=timeslot, session=brk, schedule=schedule) - if created: - print " Added %s break assignment" % timeslot - if timeslot.type_id == 'reg' and not (schedule.base and SchedTimeSessAssignment.objects.filter(timeslot=timeslot, session=reg, schedule=schedule.base).exists()): - assignment, created = SchedTimeSessAssignment.objects.get_or_create(timeslot=timeslot, session=reg, schedule=schedule) - if created: - print " Added %s registration assignment" % timeslot diff --git a/ietf/bin/create-charter-newrevisiondocevents b/ietf/bin/create-charter-newrevisiondocevents deleted file mode 100755 index d91c0b5b73..0000000000 --- a/ietf/bin/create-charter-newrevisiondocevents +++ /dev/null @@ -1,72 +0,0 @@ -#!/usr/bin/env python - -import os -import sys - -version = "0.10" -program = os.path.basename(sys.argv[0]) -progdir = os.path.dirname(sys.argv[0]) - -# boilerplate -basedir = os.path.abspath(os.path.join(os.path.dirname(__file__), "../..")) -sys.path = [ basedir ] + sys.path -os.environ["DJANGO_SETTINGS_MODULE"] = "ietf.settings" - -virtualenv_activation = os.path.join(basedir, "env", "bin", "activate_this.py") -if os.path.exists(virtualenv_activation): - execfile(virtualenv_activation, dict(__file__=virtualenv_activation)) - - -# ---------------------------------------------------------------------- -def note(string): - sys.stdout.write("%s\n" % (string)) - -# ---------------------------------------------------------------------- -def warn(string): - sys.stderr.write(" * %s\n" % (string)) - -# ------------------------------------------------------------------------------ - -import re -import datetime - -import django -django.setup() - -from django.conf import settings - -from ietf.utils.path import path as Path -from ietf.doc.models import Document, NewRevisionDocEvent -from ietf.person.models import Person - -system_entity = Person.objects.get(name="(System)") - -charterdir = Path(settings.CHARTER_PATH) -for file in charterdir.files("charter-ietf-*.txt"): - fname = file.name - ftime = datetime.datetime.fromtimestamp(file.mtime, datetime.timezone.utc) - match = re.search("^(?P[a-z0-9-]+)-(?P\d\d-\d\d)\.txt$", fname) - if match: - name = match.group("name") - rev = match.group("rev") - else: - match = re.search("^(?P[a-z0-9-]+)-(?P\d\d)\.txt$", fname) - if match: - name = match.group("name") - rev = match.group("rev") - else: - warn("Failed extracting revision from filename: '%s'" % fname) - try: - doc = Document.objects.get(type="charter", name=name) - try: - event = NewRevisionDocEvent.objects.get(doc=doc, type='new_revision', rev=rev) - note(".") - except NewRevisionDocEvent.MultipleObjectsReturned, e: - warn("Multiple NewRevisionDocEvent exists for '%s'" % fname) - except NewRevisionDocEvent.DoesNotExist: - event = NewRevisionDocEvent(doc=doc, type='new_revision', rev=rev, by=system_entity, time=ftime, desc="") - event.save() - note("Created new NewRevisionDocEvent for %s-%s" % (name, rev)) - except Document.DoesNotExist: - warn("Document not found: '%s'; no NewRevisionDocEvent created for '%s'" % (name, fname)) - diff --git a/ietf/bin/dump-draft-info b/ietf/bin/dump-draft-info deleted file mode 100755 index 3ac2e4a58a..0000000000 --- a/ietf/bin/dump-draft-info +++ /dev/null @@ -1,41 +0,0 @@ -#!/usr/bin/env python - -import os -import sys - -version = "0.10" -program = os.path.basename(sys.argv[0]) -progdir = os.path.dirname(sys.argv[0]) - -# boilerplate -basedir = os.path.abspath(os.path.join(os.path.dirname(__file__), "../..")) -sys.path = [ basedir ] + sys.path -os.environ["DJANGO_SETTINGS_MODULE"] = "ietf.settings" - -virtualenv_activation = os.path.join(basedir, "env", "bin", "activate_this.py") -if os.path.exists(virtualenv_activation): - execfile(virtualenv_activation, dict(__file__=virtualenv_activation)) - -import django -django.setup() - -from django.template import Template, Context - -from ietf.doc.models import Document -from ietf.person.models import Person - -drafts = Document.objects.filter(type="draft") - -ads = {} -for p in Person.objects.filter(ad_document_set__type="draft").distinct(): - ads[p.id] = p.role_email("ad") - -for d in drafts: - d.ad_email = ads.get(d.ad_id) - -templ_text = """{% for draft in drafts %}{% if draft.notify or draft.ad_email %}{{ draft.name }}{% if draft.notify %} docnotify='{{ draft.notify|cut:"<"|cut:">" }}'{% endif %}{% if draft.ad_email %} docsponsor='{{ draft.ad_email }}'{% endif %} -{% endif %}{% endfor %}""" -template = Template(templ_text) -context = Context({ 'drafts':drafts }) - -print template.render(context).encode('utf-8') diff --git a/ietf/bin/email-sync-discrepancies b/ietf/bin/email-sync-discrepancies deleted file mode 100755 index 3593fd126f..0000000000 --- a/ietf/bin/email-sync-discrepancies +++ /dev/null @@ -1,37 +0,0 @@ -#!/usr/bin/env python - -import sys, os, sys -import syslog - -# boilerplate -basedir = os.path.abspath(os.path.join(os.path.dirname(__file__), "../..")) -sys.path = [ basedir ] + sys.path -os.environ["DJANGO_SETTINGS_MODULE"] = "ietf.settings" - -virtualenv_activation = os.path.join(basedir, "env", "bin", "activate_this.py") -if os.path.exists(virtualenv_activation): - execfile(virtualenv_activation, dict(__file__=virtualenv_activation)) - -from optparse import OptionParser - -parser = OptionParser() -parser.add_option("-t", "--to", dest="to", - help="Email address to send report to", metavar="EMAIL") - -options, args = parser.parse_args() - -syslog.openlog(os.path.basename(__file__), syslog.LOG_PID, syslog.LOG_USER) - -import django -django.setup() - -from ietf.sync.mails import email_discrepancies - -receivers = ["iesg-secretary@ietf.org"] - -if options.to: - receivers = [options.to] - -email_discrepancies(receivers) - -syslog.syslog("Emailed sync discrepancies to %s" % receivers) diff --git a/ietf/bin/expire-ids b/ietf/bin/expire-ids index 98ee8d75fe..bb0b94ee61 100755 --- a/ietf/bin/expire-ids +++ b/ietf/bin/expire-ids @@ -13,10 +13,6 @@ basedir = os.path.abspath(os.path.join(os.path.dirname(__file__), "../..")) sys.path = [ basedir ] + sys.path os.environ["DJANGO_SETTINGS_MODULE"] = "ietf.settings" -virtualenv_activation = os.path.join(basedir, "env", "bin", "activate_this.py") -if os.path.exists(virtualenv_activation): - execfile(virtualenv_activation, dict(__file__=virtualenv_activation)) - syslog.openlog(os.path.basename(__file__), syslog.LOG_PID, syslog.LOG_USER) import django diff --git a/ietf/bin/expire-last-calls b/ietf/bin/expire-last-calls deleted file mode 100755 index 83b565e192..0000000000 --- a/ietf/bin/expire-last-calls +++ /dev/null @@ -1,34 +0,0 @@ -#!/usr/bin/env python - -# This script requires that the proper virtual python environment has been -# invoked before start - -import os -import sys -import syslog - -# boilerplate -basedir = os.path.abspath(os.path.join(os.path.dirname(__file__), "../..")) -sys.path = [ basedir ] + sys.path -os.environ["DJANGO_SETTINGS_MODULE"] = "ietf.settings" - -virtualenv_activation = os.path.join(basedir, "env", "bin", "activate_this.py") -if os.path.exists(virtualenv_activation): - execfile(virtualenv_activation, dict(__file__=virtualenv_activation)) - -syslog.openlog(os.path.basename(__file__), syslog.LOG_PID, syslog.LOG_USER) - -import django -django.setup() - -# ---------------------------------------------------------------------- - -from ietf.doc.lastcall import get_expired_last_calls, expire_last_call - -drafts = get_expired_last_calls() -for doc in drafts: - try: - expire_last_call(doc) - syslog.syslog("Expired last call for %s (id=%s)" % (doc.file_tag(), doc.pk)) - except Exception as e: - syslog.syslog(syslog.LOG_ERR, "ERROR: Failed to expire last call for %s (id=%s)" % (doc.file_tag(), doc.pk)) diff --git a/ietf/bin/expire-submissions b/ietf/bin/expire-submissions index 22db38322d..113a53ddfa 100755 --- a/ietf/bin/expire-submissions +++ b/ietf/bin/expire-submissions @@ -8,10 +8,6 @@ basedir = os.path.abspath(os.path.join(os.path.dirname(__file__), "../..")) sys.path = [ basedir ] + sys.path os.environ["DJANGO_SETTINGS_MODULE"] = "ietf.settings" -virtualenv_activation = os.path.join(basedir, "env", "bin", "activate_this.py") -if os.path.exists(virtualenv_activation): - execfile(virtualenv_activation, dict(__file__=virtualenv_activation)) - syslog.openlog(os.path.basename(__file__), syslog.LOG_PID, syslog.LOG_USER) import django diff --git a/ietf/bin/find-submission-confirmation-email-in-postfix-log b/ietf/bin/find-submission-confirmation-email-in-postfix-log deleted file mode 100755 index 6bf41574a1..0000000000 --- a/ietf/bin/find-submission-confirmation-email-in-postfix-log +++ /dev/null @@ -1,106 +0,0 @@ -#!/usr/bin/env python - -import io -import os -import sys - -version = "0.10" -program = os.path.basename(sys.argv[0]) -progdir = os.path.dirname(sys.argv[0]) - -# boilerplate -basedir = os.path.abspath(os.path.join(os.path.dirname(__file__), "../..")) -sys.path = [ basedir ] + sys.path -os.environ["DJANGO_SETTINGS_MODULE"] = "ietf.settings" - -virtualenv_activation = os.path.join(basedir, "env", "bin", "activate_this.py") -if os.path.exists(virtualenv_activation): - execfile(virtualenv_activation, dict(__file__=virtualenv_activation)) - -# ---------------------------------------------------------------------- -def note(string): - sys.stdout.write("%s\n" % (string)) - -# ---------------------------------------------------------------------- -def warn(string): - sys.stderr.write(" * %s\n" % (string)) - -# ------------------------------------------------------------------------------ - -import re -from datetime import datetime as Datetime -import time -import warnings -warnings.filterwarnings('ignore', message='the sets module is deprecated', append=True) - -import django -django.setup() - -from django.conf import settings - -from ietf.utils.path import path as Path - -from ietf.submit.models import Submission -from ietf.doc.models import Document - - - -args = sys.argv[1:] -if len(args) < 3: - warn("Expected '$ %s DRAFTNAME USER.LOG POSTFIX.LOG', but found no arguments -- exiting" % program) - sys.exit(1) - -draft = args[0] -if re.search("\.txt$", draft): - draft = draft[:-4] -if re.search("-\d\d$", draft): - draft = draft[:-3] - -if len(args) == 1: - logfiles = [ arg[1] ] -else: - logfiles = args[1:] - -from_email = settings.IDSUBMIT_FROM_EMAIL -if "<" in from_email: - from_email = from_email.split("<")[1].split(">")[0] - -submission = Submission.objects.filter(name=draft).latest('submission_date') -document = Document.objects.get(name=draft) -emails = [ author.email.address for author in document.documentauthor_set.all() if author.email ] - -timestrings = [] -for file in [ Path(settings.INTERNET_DRAFT_PATH) / ("%s-%s.txt"%(draft, submission.rev)), - Path(settings.IDSUBMIT_STAGING_PATH) / ("%s-%s.txt"%(draft, submission.rev)) ]: - if os.path.exists(file): - upload_time = time.localtime(file.mtime) - ts = time.strftime("%b %d %H:%M", upload_time) - timestrings += [ ts ] - timestrings += [ ts[:-1] + chr(((ord(ts[-1])-ord('0')+1)%10)+ord('0')) ] - print "Looking for mail log lines timestamped %s, also checking %s ..." % (timestrings[0], timestrings[1]) - -for log in logfiles: - print "\n Checking %s ...\n" % log - if log.endswith('.gz'): - import gzip - logfile = gzip.open(log) - else: - logfile = io.open(log) - queue_ids = [] - for line in logfile: - if from_email in line and "Confirmation for Auto-Post of I-D "+draft in line: - ts = line[:12] - timestrings += [ ts ] - print "Found a mention of %s, adding timestamp %s: \n %s" % (draft, ts, line) - for ts in timestrings: - if line.startswith(ts): - if from_email in line: - for to_email in emails: - if to_email in line: - sys.stdout.write(line) - if "queued_as:" in line: - queue_ids += [ line.split("queued_as:")[1].split(",")[0] ] - elif queue_ids: - for qi in queue_ids: - if qi in line: - sys.stdout.write(line) diff --git a/ietf/bin/iana-review-email b/ietf/bin/iana-review-email index 5c7a7183b9..27aee4015e 100755 --- a/ietf/bin/iana-review-email +++ b/ietf/bin/iana-review-email @@ -8,10 +8,6 @@ basedir = os.path.abspath(os.path.join(os.path.dirname(__file__), "../..")) sys.path = [ basedir ] + sys.path os.environ["DJANGO_SETTINGS_MODULE"] = "ietf.settings" -virtualenv_activation = os.path.join(basedir, "env", "bin", "activate_this.py") -if os.path.exists(virtualenv_activation): - execfile(virtualenv_activation, dict(__file__=virtualenv_activation)) - syslog.openlog(os.path.basename(__file__), syslog.LOG_PID, syslog.LOG_USER) import django diff --git a/ietf/bin/interim_minutes_reminder b/ietf/bin/interim_minutes_reminder deleted file mode 100755 index 7f2f84f739..0000000000 --- a/ietf/bin/interim_minutes_reminder +++ /dev/null @@ -1,27 +0,0 @@ -#!/usr/bin/env python -# -*- coding: utf-8 -*- -# -*- Python -*- -# -''' -This script calls ietf.meeting.helpers.check_interim_minutes() which sends -a reminder email for interim meetings that occurred 10 days ago but still -don't have minutes. -''' - -# Set PYTHONPATH and load environment variables for standalone script ----------------- -import os, sys -basedir = os.path.abspath(os.path.join(os.path.dirname(__file__), "../..")) -sys.path = [ basedir ] + sys.path -os.environ["DJANGO_SETTINGS_MODULE"] = "ietf.settings" - -virtualenv_activation = os.path.join(basedir, "env", "bin", "activate_this.py") -if os.path.exists(virtualenv_activation): - execfile(virtualenv_activation, dict(__file__=virtualenv_activation)) - -import django -django.setup() -# ------------------------------------------------------------------------------------- - -from ietf.meeting.helpers import check_interim_minutes - -check_interim_minutes() diff --git a/ietf/bin/list-role-holder-emails b/ietf/bin/list-role-holder-emails deleted file mode 100755 index 6d6c160464..0000000000 --- a/ietf/bin/list-role-holder-emails +++ /dev/null @@ -1,36 +0,0 @@ -#!/usr/bin/env python - - -import os, sys -import syslog - -# boilerplate -basedir = os.path.abspath(os.path.join(os.path.dirname(__file__), "../..")) -sys.path = [ basedir ] + sys.path -os.environ["DJANGO_SETTINGS_MODULE"] = "ietf.settings" - -virtualenv_activation = os.path.join(basedir, "env", "bin", "activate_this.py") -if os.path.exists(virtualenv_activation): - execfile(virtualenv_activation, dict(__file__=virtualenv_activation)) - -syslog.openlog(os.path.basename(__file__), syslog.LOG_PID, syslog.LOG_USER) - -import django -django.setup() - -from django.utils.encoding import force_str -from ietf.group.models import Role - -addresses = set() -for role in Role.objects.filter( - group__state__slug='active', - group__type__in=['ag','area','dir','iab','ietf','irtf','nomcom','rg','team','wg','rag']): - #sys.stderr.write(str(role)+'\n') - for e in role.person.email_set.all(): - if e.active and not e.address.startswith('unknown-email-'): - addresses.add(e.address) - -addresses = list(addresses) -addresses.sort() -for a in addresses: - print(force_str(a)) diff --git a/ietf/bin/mailman_listinfo.py b/ietf/bin/mailman_listinfo.py deleted file mode 100755 index f7e4cfe4c1..0000000000 --- a/ietf/bin/mailman_listinfo.py +++ /dev/null @@ -1,50 +0,0 @@ -#!/usr/bin/python2.7 -# Copyright The IETF Trust 2022, All Rights Reserved -# Note the shebang. This specifically targets deployment on IETFA and intends to use its system python2.7. - -# This is an adaptor to pull information out of Mailman2 using its python libraries (which are only available for python2). -# It is NOT django code, and does not have access to django.conf.settings. - -import json -import sys - -from collections import defaultdict - -def main(): - - sys.path.append('/usr/lib/mailman') - - have_mailman = False - try: - from Mailman import Utils - from Mailman import MailList - from Mailman import MemberAdaptor - have_mailman = True - except ImportError: - pass - - - if not have_mailman: - sys.stderr.write("Could not import mailman modules -- skipping import of mailman list info") - sys.exit() - - names = list(Utils.list_names()) - - # need to emit dict of names, each name has an mlist, and each mlist has description, advertised, and members (calculated as below) - result = defaultdict(dict) - for name in names: - mlist = MailList.MailList(name, lock=False) - result[name] = dict() - result[name]['internal_name'] = mlist.internal_name() - result[name]['real_name'] = mlist.real_name - result[name]['description'] = mlist.description # Not attempting to change encoding - result[name]['advertised'] = mlist.advertised - result[name]['members'] = list() - if mlist.advertised: - members = mlist.getRegularMemberKeys() + mlist.getDigestMemberKeys() - members = set([ m for m in members if mlist.getDeliveryStatus(m) == MemberAdaptor.ENABLED ]) - result[name]['members'] = list(members) - json.dump(result, sys.stdout) - -if __name__ == "__main__": - main() diff --git a/ietf/bin/merge-person-records b/ietf/bin/merge-person-records deleted file mode 100755 index 155e5755f6..0000000000 --- a/ietf/bin/merge-person-records +++ /dev/null @@ -1,65 +0,0 @@ -#!/usr/bin/env python -# -*- coding: utf-8 -*- -# -*- Python -*- -# -''' -This script merges two Person records into one. It determines which record is the target -based on most current User record (last_login) unless -f (force) option is used to -force SOURCE TARGET as specified on the command line. The order of operations is -important. We must complete all source.save() operations before moving the aliases to -the target, this is to avoid extra "Possible duplicate Person" emails going out, if the -Person is saved without an alias the Person.save() creates another one, which then -conflicts with the moved one. -''' - -# Set PYTHONPATH and load environment variables for standalone script ----------------- -import os, sys -basedir = os.path.abspath(os.path.join(os.path.dirname(__file__), "../..")) -sys.path = [ basedir ] + sys.path -os.environ["DJANGO_SETTINGS_MODULE"] = "ietf.settings" - -virtualenv_activation = os.path.join(basedir, "env", "bin", "activate_this.py") -if os.path.exists(virtualenv_activation): - execfile(virtualenv_activation, dict(__file__=virtualenv_activation)) - -import django -django.setup() -# ------------------------------------------------------------------------------------- - -import argparse -from django.contrib import admin -from ietf.person.models import Person -from ietf.person.utils import (merge_persons, send_merge_notification, handle_users, - determine_merge_order) - - -def main(): - parser = argparse.ArgumentParser() - parser.add_argument("source_id",type=int) - parser.add_argument("target_id",type=int) - parser.add_argument('-f','--force', help='force merge order',action='store_true') - parser.add_argument('-v','--verbose', help='verbose output',action='store_true') - args = parser.parse_args() - - source = Person.objects.get(pk=args.source_id) - target = Person.objects.get(pk=args.target_id) - - # set merge order - if not args.force: - source,target = determine_merge_order(source,target) - - # confirm - print "Merging person {}({}) to {}({})".format(source.ascii,source.pk,target.ascii,target.pk) - print handle_users(source,target,check_only=True) - response = raw_input('Ok to continue y/n? ') - if response.lower() != 'y': - sys.exit() - - # perform merge - success, changes = merge_persons(source, target, verbose=args.verbose) - - # send email notification - send_merge_notification(target,changes) - -if __name__ == "__main__": - main() diff --git a/ietf/bin/notify-expirations b/ietf/bin/notify-expirations index 0270c13765..fc2fd86a31 100755 --- a/ietf/bin/notify-expirations +++ b/ietf/bin/notify-expirations @@ -7,10 +7,6 @@ basedir = os.path.abspath(os.path.join(os.path.dirname(__file__), "../..")) sys.path = [ basedir ] + sys.path os.environ["DJANGO_SETTINGS_MODULE"] = "ietf.settings" -virtualenv_activation = os.path.join(basedir, "env", "bin", "activate_this.py") -if os.path.exists(virtualenv_activation): - execfile(virtualenv_activation, dict(__file__=virtualenv_activation)) - import django django.setup() diff --git a/ietf/bin/pretty-xml-dump b/ietf/bin/pretty-xml-dump deleted file mode 100755 index 22abc08a64..0000000000 --- a/ietf/bin/pretty-xml-dump +++ /dev/null @@ -1,5 +0,0 @@ -#!/bin/sh -python manage.py dumpdata --format=xml "$@" | sed -e 's/<\/*object/\ - &/g' -e 's/')][1:])) - else: - retval.append(ParsedAuthor(a.strip(),None)) - - return retval - -def calculate_changes(tracker_persons,tracker_emails,names,emails): - adds = set() - deletes = set() - for email in emails: - if email and email!='none' and email not in ignore_addresses: - p = Person.objects.filter(email__address=email).first() - if p: - if not set(map(unicode.lower,p.email_set.values_list('address',flat=True))).intersection(tracker_emails): - adds.add(email) - else: - #person_name = names[emails.index(email)] - adds.add(email) - for person in tracker_persons: - if not set(map(unicode.lower,person.email_set.values_list('address',flat=True))).intersection(emails): - match = False - for index in [i for i,j in enumerate(emails) if j=='none' or not j]: - if names[index].split()[-1].lower()==person.last_name().lower(): - match = True - if not match: - deletes.add(person) - return adds, deletes - -def _main(): - - parser = argparse.ArgumentParser(description="Recalculate RFC documentauthor_set"+'\n\n'+__doc__, - formatter_class=argparse.RawDescriptionHelpFormatter,) - parser.add_argument('-v','--verbose',help="Show the action taken for each RFC",action='store_true') - parser.add_argument('--rfc',type=int, nargs='*',help="Only recalculate the given rfc numbers",dest='rfcnumberlist') - args = parser.parse_args() - - probable_email_match = set() - probable_duplicates = [] - - all_the_email = get_all_the_email() - author_names, author_emails = get_rfc_data() - - stats = { 'rfc not in tracker' :0, - 'same addresses' :0, - 'different addresses belonging to same people' :0, - 'same names, rfced emails do not match' :0, - 'rfced data is unusable' :0, - "data doesn't match but no changes found" :0, - 'changed authors' :0, } - - for rfc_num in args.rfcnumberlist or sorted(author_names.keys()): - - rfc = Document.objects.filter(docalias__name='rfc%s'%rfc_num).first() - - if not rfc: - if args.verbose: - show_verbose(rfc_num,'rfc not in tracker') - stats['rfc not in tracker'] += 1 - continue - - rfced_emails = set(author_emails[rfc_num]) - tracker_emails = set(map(unicode.lower,rfc.authors.values_list('address',flat=True))) - tracker_persons = set([x.person for x in rfc.authors.all()]) - matching_emails = get_matching_emails(all_the_email,rfced_emails) - rfced_persons = set([x.person for x in matching_emails]) - known_emails = set([e.l_address for e in matching_emails]) - unknown_emails = rfced_emails - known_emails - unknown_persons = tracker_persons-rfced_persons - - rfced_lastnames = sorted([n.split()[-1].lower() for n in author_names[rfc_num]]) - tracker_lastnames = sorted([p.last_name().lower() for p in tracker_persons]) - - if rfced_emails == tracker_emails: - if args.verbose: - show_verbose(rfc_num,'tracker and rfc editor have the same addresses') - stats['same addresses'] += 1 - continue - - if len(rfced_emails)==len(tracker_emails) and not 'none' in author_emails[rfc_num]: - if tracker_persons == rfced_persons: - if args.verbose: - show_verbose(rfc_num,'tracker and rfc editor have the different addresses belonging to same people') - stats['different addresses belonging to same people'] += 1 - continue - else: - if len(unknown_emails)==1 and len(tracker_persons-rfced_persons)==1: - p = list(tracker_persons-rfced_persons)[0] - probable_email_match.add(u"%s is probably %s (%s) : %s "%(list(unknown_emails)[0], p, p.pk, rfc_num)) - elif len(unknown_emails)==len(unknown_persons): - probable_email_match.add(u"%s are probably %s : %s"%(unknown_emails,[(p.ascii,p.pk) for p in unknown_persons],rfc_num)) - else: - probable_duplicates.append((tracker_persons^rfced_persons,rfc_num)) - - if tracker_lastnames == rfced_lastnames: - if args.verbose: - show_verbose(rfc_num,"emails don't match up, but person names appear to be the same") - stats[ 'same names, rfced emails do not match'] += 1 - continue - - use_rfc_data = bool(len(author_emails[rfc_num])==len(author_names[rfc_num])) - if not use_rfc_data: - if args.verbose: - print 'Ignoring rfc database for rfc%d'%rfc_num - stats[ 'rfced data is unusable'] += 1 - - if use_rfc_data: - adds, deletes = calculate_changes(tracker_persons,tracker_emails,author_names[rfc_num],author_emails[rfc_num]) - parsed_authors=get_parsed_authors(rfc_num) - parsed_adds, parsed_deletes = calculate_changes(tracker_persons,tracker_emails,[x.name for x in parsed_authors],[x.address for x in parsed_authors]) - - for e in adds.union(parsed_adds) if use_rfc_data else parsed_adds: - if not e or e in ignore_addresses: - continue - if not Person.objects.filter(email__address=e).exists(): - if e not in parsed_adds: - #print rfc_num,"Would add",e,"as",author_names[rfc_num][author_emails[rfc_num].index(e)],"(rfced database)" - print "(address='%s',name='%s'),"%(e,author_names[rfc_num][author_emails[rfc_num].index(e)]),"# (rfced %d)"%rfc_num - for p in Person.objects.filter(name__iendswith=author_names[rfc_num][author_emails[rfc_num].index(e)].split(' ')[-1]): - print "\t", p.pk, p.ascii - else: - name = [x.name for x in parsed_authors if x.address==e][0] - p = Person.objects.filter(name=name).first() - if p: - #print e,"is probably",p.pk,p - print "'%s': %d, # %s (%d)"%(e,p.pk,p.ascii,rfc_num) - - else: - p = Person.objects.filter(ascii=name).first() - if p: - print e,"is probably",p.pk,p - print "'%s': %d, # %s (%d)"%(e,p.pk,p.ascii,rfc_num) - else: - p = Person.objects.filter(ascii_short=name).first() - if p: - print e,"is probably",p.pk,p - print "'%s': %d, # %s (%d)"%(e,p.pk,p.ascii,rfc_num) - #print rfc_num,"Would add",e,"as",name,"(parsed)" - print "(address='%s',name='%s'),"%(e,name),"# (parsed %d)"%rfc_num - for p in Person.objects.filter(name__iendswith=name.split(' ')[-1]): - print "\t", p.pk, p.ascii - - if False: # This was a little useful, but the noise in the rfc_ed file keeps it from being completely useful - for p in deletes: - for n in author_names[rfc_num]: - if p.last_name().lower()==n.split()[-1].lower(): - email_candidate = author_emails[rfc_num][author_names[rfc_num].index(n)] - email_found = Email.objects.filter(address=email_candidate).first() - if email_found: - probable_duplicates.append((set([p,email_found.person]),rfc_num)) - else: - probable_email_match.add(u"%s is probably %s (%s) : %s"%(email_candidate, p, p.pk, rfc_num)) - - if args.verbose: - if use_rfc_data: - working_adds = parsed_adds - seen_people = set(Email.objects.get(address=e).person for e in parsed_adds) - for addr in adds: - person = Email.objects.get(address=addr).person - if person not in seen_people: - working_adds.add(addr) - seen_people.add(person) - working_deletes = deletes.union(parsed_deletes) - else: - working_adds = parsed_adds - working_deletes = parsed_deletes - # unique_adds = set() # TODO don't add different addresses for the same person from the two sources - if working_adds or working_deletes: - show_verbose(rfc_num,"Changing original list",tracker_persons,"by adding",working_adds," and deleting",working_deletes) - print "(",rfc_num,",",[e for e in working_adds],",",[p.pk for p in working_deletes],"), #",[p.ascii for p in working_deletes] - else: - stats["data doesn't match but no changes found"] += 1 - show_verbose(rfc_num,"Couldn't figure out what to change") - - if False: - #if tracker_persons: - #if any(['iab@' in e for e in adds]) or any(['iesg@' in e for e in adds]) or any(['IESG'==p.name for p in deletes]) or any(['IAB'==p.name for p in deletes]): - print rfc_num - print "tracker_persons",tracker_persons - print "author_names",author_names[rfc_num] - print "author_emails",author_emails[rfc_num] - print "Adds:", adds - print "Deletes:", deletes - - stats['changed authors'] += 1 - - if False: - debug.show('rfc_num') - debug.show('rfced_emails') - debug.show('tracker_emails') - debug.show('known_emails') - debug.show('unknown_emails') - debug.show('tracker_persons') - debug.show('rfced_persons') - debug.show('tracker_persons==rfced_persons') - debug.show('[p.id for p in tracker_persons]') - debug.show('[p.id for p in rfced_persons]') - exit() - - if True: - for p in sorted(list(probable_email_match)): - print p - if True: - print "Probable duplicate persons" - for d,r in sorted(probable_duplicates): - print [(p,p.pk) for p in d], r - else: - print len(probable_duplicates)," probable duplicate persons" - - print stats - -if __name__ == "__main__": - _main() - diff --git a/ietf/bin/redirect-dump b/ietf/bin/redirect-dump deleted file mode 100755 index ef35bbf0de..0000000000 --- a/ietf/bin/redirect-dump +++ /dev/null @@ -1,9 +0,0 @@ -#!/bin/sh -# -# Copyright The IETF Trust 2007, All Rights Reserved -# -#python manage.py dumpdata --format=xml redirects | xmllint --format - -python manage.py dumpdata --format=xml redirects | sed -e 's/<\/*object/\ - &/g' -e 's/ {self.bucket}:{self.blob}" diff --git a/ietf/blobdb/replication.py b/ietf/blobdb/replication.py new file mode 100644 index 0000000000..d251d3b95c --- /dev/null +++ b/ietf/blobdb/replication.py @@ -0,0 +1,178 @@ +# Copyright The IETF Trust 2025, All Rights Reserved +import datetime +from dataclasses import dataclass +from io import BytesIO +from typing import Optional + +from django.conf import settings +from django.core.files import File +from django.core.files.storage import storages, InvalidStorageError +from django.db import connections + +from ietf.utils import log + +DEFAULT_SETTINGS = { + "ENABLED": False, + "DEST_STORAGE_PATTERN": "r2-{bucket}", + "INCLUDE_BUCKETS": (), # empty means include all + "EXCLUDE_BUCKETS": (), # empty means exclude none + "VERBOSE_LOGGING": False, +} + + +class SimpleMetadataFile(File): + def __init__(self, file, name=None): + super().__init__(file, name) + self.custom_metadata = {} + self.content_type = "" + + +def get_replication_settings(): + return DEFAULT_SETTINGS | getattr(settings, "BLOBDB_REPLICATION", {}) + + +def validate_replication_settings(): + replicator_settings = get_replication_settings() + # No extra settings allowed + unknown_settings = set(DEFAULT_SETTINGS.keys()) - set(replicator_settings.keys()) + if len(unknown_settings) > 0: + raise RuntimeError( + f"Unrecognized BLOBDB_REPLICATOR settings: {', '.join(str(unknown_settings))}" + ) + # destination storage pattern must be a string that includes {bucket} + pattern = replicator_settings["DEST_STORAGE_PATTERN"] + if not isinstance(pattern, str): + raise RuntimeError( + f"DEST_STORAGE_PATTERN must be a str, not {type(pattern).__name__}" + ) + if "{bucket}" not in pattern: + raise RuntimeError( + f"DEST_STORAGE_PATTERN must contain the substring '{{bucket}}' (found '{pattern}')" + ) + # include/exclude buckets must be list-like + include_buckets = replicator_settings["INCLUDE_BUCKETS"] + if not isinstance(include_buckets, (list, tuple, set)): + raise RuntimeError("INCLUDE_BUCKETS must be a list, tuple, or set") + exclude_buckets = replicator_settings["EXCLUDE_BUCKETS"] + if not isinstance(exclude_buckets, (list, tuple, set)): + raise RuntimeError("EXCLUDE_BUCKETS must be a list, tuple, or set") + # if we have explicit include_buckets, make sure the necessary storages exist + if len(include_buckets) > 0: + include_storages = {destination_storage_name_for(b) for b in include_buckets} + exclude_storages = {destination_storage_name_for(b) for b in exclude_buckets} + configured_storages = set(settings.STORAGES.keys()) + missing_storages = include_storages - exclude_storages - configured_storages + if len(missing_storages) > 0: + raise RuntimeError( + f"Replication requires unknown storage(s): {', '.join(missing_storages)}" + ) + + +def destination_storage_name_for(bucket: str): + pattern = get_replication_settings()["DEST_STORAGE_PATTERN"] + return pattern.format(bucket=bucket) + + +def destination_storage_for(bucket: str): + storage_name = destination_storage_name_for(bucket) + return storages[storage_name] + + +def replication_enabled(bucket: str): + replication_settings = get_replication_settings() + if not replication_settings["ENABLED"]: + return False + # Default is all buckets are included + included = ( + len(replication_settings["INCLUDE_BUCKETS"]) == 0 + or bucket in replication_settings["INCLUDE_BUCKETS"] + ) + # Default is no buckets are excluded + excluded = ( + len(replication_settings["EXCLUDE_BUCKETS"]) > 0 + and bucket in replication_settings["EXCLUDE_BUCKETS"] + ) + return included and not excluded + + +def verbose_logging_enabled(): + return bool(get_replication_settings()["VERBOSE_LOGGING"]) + + +@dataclass +class SqlBlob: + content: bytes + checksum: str + modified: datetime.datetime + mtime: Optional[datetime.datetime] + content_type: str + + +def fetch_blob_via_sql(bucket: str, name: str) -> Optional[SqlBlob]: + blobdb_connection = connections["blobdb"] + cursor = blobdb_connection.cursor() + cursor.execute( + """ + SELECT content, checksum, modified, mtime, content_type FROM blobdb_blob + WHERE bucket=%s AND name=%s LIMIT 1 + """, + [bucket, name], + ) + row = cursor.fetchone() + col_names = [col[0] for col in cursor.description] + return None if row is None else SqlBlob(**{ + col_name: row_val + for col_name, row_val in zip(col_names, row) + }) + + +def replicate_blob(bucket, name): + """Replicate a Blobdb blob to a Storage""" + if not replication_enabled(bucket): + if verbose_logging_enabled(): + log.log( + f"Not replicating {bucket}:{name} because replication is not enabled for {bucket}" + ) + return + + try: + destination_storage = destination_storage_for(bucket) + except InvalidStorageError as e: + log.log( + f"Failed to replicate {bucket}:{name} because destination storage for {bucket} is not configured" + ) + raise ReplicationError from e + + blob = fetch_blob_via_sql(bucket, name) + if blob is None: + if verbose_logging_enabled(): + log.log(f"Deleting {bucket}:{name} from replica") + try: + destination_storage.delete(name) + except Exception as e: + log.log(f"Failed to delete {bucket}:{name} from replica: {e}") + raise ReplicationError from e + else: + # Add metadata expected by the MetadataS3Storage + file_with_metadata = SimpleMetadataFile(file=BytesIO(blob.content)) + file_with_metadata.content_type = blob.content_type + file_with_metadata.custom_metadata = { + "sha384": blob.checksum, + "mtime": (blob.mtime or blob.modified).isoformat(), + } + if verbose_logging_enabled(): + log.log( + f"Saving {bucket}:{name} to replica (" + f"sha384: '{file_with_metadata.custom_metadata['sha384'][:16]}...', " + f"content_type: '{file_with_metadata.content_type}', " + f"mtime: '{file_with_metadata.custom_metadata['mtime']})" + ) + try: + destination_storage.save(name, file_with_metadata) + except Exception as e: + log.log(f"Failed to save {bucket}:{name} to replica: {e}") + raise ReplicationError from e + + +class ReplicationError(Exception): + pass diff --git a/ietf/blobdb/routers.py b/ietf/blobdb/routers.py new file mode 100644 index 0000000000..319c0fbc71 --- /dev/null +++ b/ietf/blobdb/routers.py @@ -0,0 +1,58 @@ +# Copyright The IETF Trust 2025, All Rights Reserved +from django.apps import apps + +from .apps import BlobdbConfig, get_blobdb + + +class BlobdbStorageRouter: + """Database router for the Blobdb""" + + _app_label = None + + @property + def app_label(self): + if self._app_label is None: + for app in apps.get_app_configs(): + if isinstance(app, BlobdbConfig): + self._app_label = app.label + break + if self._app_label is None: + raise RuntimeError( + f"{BlobdbConfig.name} is not present in the Django app registry" + ) + return self._app_label + + @property + def db(self): + return get_blobdb() + + def db_for_read(self, model, **hints): + """Suggest the database that should be used for read operations for objects of type model + + Returns None if there is no suggestion. + """ + if model._meta.app_label == self.app_label: + return self.db + return None # no suggestion + + def db_for_write(self, model, **hints): + """Suggest the database that should be used for write of objects of type model + + Returns None if there is no suggestion. + """ + if model._meta.app_label == self.app_label: + return self.db + return None # no suggestion + + def allow_migrate(self, db, app_label, model_name=None, **hints): + """Determine if the migration operation is allowed to run on the database with alias db + + Return True if the operation should run, False if it shouldn’t run, or + None if the router has no opinion. + """ + if self.db is None: + return None # no opinion, use the default db + is_our_app = app_label == self.app_label + is_our_db = db == self.db + if is_our_app or is_our_db: + return is_our_app and is_our_db diff --git a/ietf/blobdb/storage.py b/ietf/blobdb/storage.py new file mode 100644 index 0000000000..4213ec801d --- /dev/null +++ b/ietf/blobdb/storage.py @@ -0,0 +1,96 @@ +# Copyright The IETF Trust 2025, All Rights Reserved +from typing import Optional + +from django.core.exceptions import SuspiciousFileOperation +from django.core.files.base import ContentFile +from django.core.files.storage import Storage +from django.db.models.functions import Length +from django.utils.deconstruct import deconstructible +from django.utils import timezone + +from ietf.utils.storage import MetadataFile +from .models import Blob + + +class BlobFile(MetadataFile): + + def __init__(self, content, name=None, mtime=None, content_type=""): + super().__init__( + file=ContentFile(content), + name=name, + mtime=mtime, + content_type=content_type, + ) + + +@deconstructible +class BlobdbStorage(Storage): + + def __init__(self, bucket_name: Optional[str]=None): + if bucket_name is None: + raise ValueError("BlobdbStorage bucket_name must be specified") + self.bucket_name = bucket_name + + def get_queryset(self): + return Blob.objects.filter(bucket=self.bucket_name) + + def delete(self, name): + blob = self.get_queryset().filter(name=name).first() + if blob is not None: + blob.delete() + + def exists(self, name): + return self.get_queryset().filter(name=name).exists() + + def size(self, name): + sizes = ( + self.get_queryset() + .filter(name=name) + .annotate(object_size=Length("content")) + .values_list("object_size", flat=True) + ) + if len(sizes) == 0: + raise FileNotFoundError( + f"No object '{name}' exists in bucket '{self.bucket_name}'" + ) + return sizes[0] # unique constraint guarantees 0 or 1 entry + + def _open(self, name, mode="rb"): + try: + blob = self.get_queryset().get(name=name) + except Blob.DoesNotExist: + raise FileNotFoundError( + f"No object '{name}' exists in bucket '{self.bucket_name}'" + ) + return BlobFile( + content=blob.content, + name=blob.name, + mtime=blob.mtime or blob.modified, # fall back to modified time + content_type=blob.content_type, + ) + + def _save(self, name, content): + """Perform the save operation + + The storage API allows _save() to save to a different name than was requested. This method will + never do that, instead overwriting the existing blob. + """ + Blob.objects.update_or_create( + name=name, + bucket=self.bucket_name, + defaults={ + "content": content.read(), + "modified": timezone.now(), + "mtime": getattr(content, "mtime", None), + "content_type": getattr(content, "content_type", ""), + }, + ) + return name + + def get_available_name(self, name, max_length=None): + if max_length is not None and len(name) > max_length: + raise SuspiciousFileOperation( + f"BlobdbStorage only allows names up to {max_length}, but was" + f"asked to store the name '{name[:5]}...{name[-5:]} of length {len(name)}" + ) + return name # overwrite is permitted diff --git a/ietf/blobdb/tasks.py b/ietf/blobdb/tasks.py new file mode 100644 index 0000000000..538d415830 --- /dev/null +++ b/ietf/blobdb/tasks.py @@ -0,0 +1,17 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +import json + +from celery import shared_task + +from .replication import replicate_blob, ReplicationError + + +@shared_task( + autoretry_for=(ReplicationError,), retry_backoff=10, retry_kwargs={"max_retries": 5} +) +def pybob_the_blob_replicator_task(body: str): + request = json.loads(body) + bucket = request["bucket"] + name = request["name"] + replicate_blob(bucket, name) diff --git a/ietf/blobdb/tests.py b/ietf/blobdb/tests.py new file mode 100644 index 0000000000..0eadad0a1f --- /dev/null +++ b/ietf/blobdb/tests.py @@ -0,0 +1,80 @@ +# Copyright The IETF Trust 2025, All Rights Reserved +import datetime + +from django.core.files.base import ContentFile + +from ietf.utils.test_utils import TestCase +from .factories import BlobFactory +from .models import Blob +from .storage import BlobFile, BlobdbStorage + + +class StorageTests(TestCase): + def test_save(self): + storage = BlobdbStorage(bucket_name="my-bucket") + timestamp = datetime.datetime( + 2025, + 3, + 17, + 1, + 2, + 3, + tzinfo=datetime.timezone.utc, + ) + # Create file to save + my_file = BlobFile( + content=b"These are my bytes.", + mtime=timestamp, + content_type="application/x-my-content-type", + ) + # save the file + saved_name = storage.save("myfile.txt", my_file) + # validate the outcome + self.assertEqual(saved_name, "myfile.txt") + blob = Blob.objects.filter(bucket="my-bucket", name="myfile.txt").first() + self.assertIsNotNone(blob) # validates bucket and name + self.assertEqual(bytes(blob.content), b"These are my bytes.") + self.assertEqual(blob.mtime, timestamp) + self.assertEqual(blob.content_type, "application/x-my-content-type") + + def test_save_naive_file(self): + storage = BlobdbStorage(bucket_name="my-bucket") + my_naive_file = ContentFile(content=b"These are my naive bytes.") + # save the file + saved_name = storage.save("myfile.txt", my_naive_file) + # validate the outcome + self.assertEqual(saved_name, "myfile.txt") + blob = Blob.objects.filter(bucket="my-bucket", name="myfile.txt").first() + self.assertIsNotNone(blob) # validates bucket and name + self.assertEqual(bytes(blob.content), b"These are my naive bytes.") + self.assertIsNone(blob.mtime) + self.assertEqual(blob.content_type, "") + + def test_open(self): + """BlobdbStorage open yields a BlobFile with specific mtime and content_type""" + mtime = datetime.datetime(2021, 1, 2, 3, 45, tzinfo=datetime.timezone.utc) + blob = BlobFactory(mtime=mtime, content_type="application/x-oh-no-you-didnt") + storage = BlobdbStorage(bucket_name=blob.bucket) + with storage.open(blob.name, "rb") as f: + self.assertTrue(isinstance(f, BlobFile)) + assert isinstance(f, BlobFile) # redundant, narrows type for linter + self.assertEqual(f.read(), bytes(blob.content)) + self.assertEqual(f.mtime, mtime) + self.assertEqual(f.content_type, "application/x-oh-no-you-didnt") + + def test_open_null_mtime(self): + """BlobdbStorage open yields a BlobFile with default mtime and content_type""" + blob = BlobFactory(content_type="application/x-oh-no-you-didnt") # does not set mtime + storage = BlobdbStorage(bucket_name=blob.bucket) + with storage.open(blob.name, "rb") as f: + self.assertTrue(isinstance(f, BlobFile)) + assert isinstance(f, BlobFile) # redundant, narrows type for linter + self.assertEqual(f.read(), bytes(blob.content)) + self.assertIsNotNone(f.mtime) + self.assertEqual(f.mtime, blob.modified) + self.assertEqual(f.content_type, "application/x-oh-no-you-didnt") + + def test_open_file_not_found(self): + storage = BlobdbStorage(bucket_name="not-a-bucket") + with self.assertRaises(FileNotFoundError): + storage.open("definitely/not-a-file.txt") diff --git a/ietf/celeryapp.py b/ietf/celeryapp.py index cefde3a8d3..fda89c30be 100644 --- a/ietf/celeryapp.py +++ b/ietf/celeryapp.py @@ -1,11 +1,20 @@ import os +import scout_apm.celery + +import celery +from scout_apm.api import Config + + +# Disable celery's internal logging configuration, we set it up via Django +@celery.signals.setup_logging.connect +def on_setup_logging(**kwargs): + pass -from celery import Celery # Set the default Django settings module for the 'celery' program os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'ietf.settings') -app = Celery('ietf') +app = celery.Celery('ietf') # Using a string here means the worker doesn't have to serialize # the configuration object to child processes. @@ -13,6 +22,28 @@ # should have a `CELERY_` prefix. app.config_from_object('django.conf:settings', namespace='CELERY') +# Turn on Scout APM celery instrumentation if configured in the environment +scout_key = os.environ.get("DATATRACKER_SCOUT_KEY", None) +if scout_key is not None: + scout_name = os.environ.get("DATATRACKER_SCOUT_NAME", "Datatracker") + scout_core_agent_socket_path = "tcp://{host}:{port}".format( + host=os.environ.get("DATATRACKER_SCOUT_CORE_AGENT_HOST", "localhost"), + port=os.environ.get("DATATRACKER_SCOUT_CORE_AGENT_PORT", "6590"), + ) + Config.set( + key=scout_key, + name=scout_name, + monitor=True, + core_agent_download=False, + core_agent_launch=False, + core_agent_path=scout_core_agent_socket_path, + ) + # Note: Passing the Celery app to install() method as recommended in the + # Scout documentation causes failure at startup, likely because Scout + # ingests the config greedily before Django is ready. Have not found a + # workaround for this other than explicitly configuring Scout. + scout_apm.celery.install() + # Load task modules from all registered Django apps. app.autodiscover_tasks() diff --git a/ietf/checks.py b/ietf/checks.py index 53e695a769..f911d081f0 100644 --- a/ietf/checks.py +++ b/ietf/checks.py @@ -28,81 +28,6 @@ def already_ran(): checks_run.append(name) return False -@checks.register('directories') -def check_cdn_directory_exists(app_configs, **kwargs): - """This checks that the path from which the CDN will serve static files for - this version of the datatracker actually exists. In development and test - mode STATIC_ROOT will normally be just static/, but in production it will be - set to a different part of the file system which is served via CDN, and the - path will contain the datatracker release version. - """ - if already_ran(): - return [] - # - errors = [] - if settings.SERVER_MODE == 'production' and not os.path.exists(settings.STATIC_ROOT): - errors.append(checks.Error( - "The static files directory has not been set up.", - hint="Please run 'ietf/manage.py collectstatic'.", - obj=None, - id='datatracker.E001', - )) - return errors - -@checks.register('files') -def check_group_email_aliases_exists(app_configs, **kwargs): - from ietf.group.views import check_group_email_aliases - # - if already_ran(): - return [] - # - errors = [] - try: - ok = check_group_email_aliases() - if not ok: - errors.append(checks.Error( - "Found no aliases in the group email aliases file\n'%s'."%settings.GROUP_ALIASES_PATH, - hint="Please run the generate_group_aliases management command to generate them.", - obj=None, - id="datatracker.E0002", - )) - except IOError as e: - errors.append(checks.Error( - "Could not read group email aliases:\n %s" % e, - hint="Please run the generate_group_aliases management command to generate them.", - obj=None, - id="datatracker.E0003", - )) - - return errors - -@checks.register('files') -def check_doc_email_aliases_exists(app_configs, **kwargs): - from ietf.doc.views_doc import check_doc_email_aliases - # - if already_ran(): - return [] - # - errors = [] - try: - ok = check_doc_email_aliases() - if not ok: - errors.append(checks.Error( - "Found no aliases in the document email aliases file\n'%s'."%settings.DRAFT_VIRTUAL_PATH, - hint="Please run the generate_draft_aliases management command to generate them.", - obj=None, - id="datatracker.E0004", - )) - except IOError as e: - errors.append(checks.Error( - "Could not read document email aliases:\n %s" % e, - hint="Please run the generate_draft_aliases management command to generate them.", - obj=None, - id="datatracker.E0005", - )) - - return errors - @checks.register('directories') def check_id_submission_directories(app_configs, **kwargs): # @@ -114,7 +39,7 @@ def check_id_submission_directories(app_configs, **kwargs): p = getattr(settings, s) if not os.path.exists(p): errors.append(checks.Critical( - "A directory used by the ID submission tool does not\n" + "A directory used by the I-D submission tool does not\n" "exist at the path given in the settings file. The setting is:\n" " %s = %s" % (s, p), hint = ("Please either update the local settings to point at the correct\n" @@ -134,7 +59,7 @@ def check_id_submission_files(app_configs, **kwargs): p = getattr(settings, s) if not os.path.exists(p): errors.append(checks.Critical( - "A file used by the ID submission tool does not exist\n" + "A file used by the I-D submission tool does not exist\n" "at the path given in the settings file. The setting is:\n" " %s = %s" % (s, p), hint = ("Please either update the local settings to point at the correct\n" @@ -179,7 +104,7 @@ def check_id_submission_checkers(app_configs, **kwargs): except Exception as e: errors.append(checks.Critical( "An exception was raised when trying to import the\n" - "draft submission checker class '%s':\n %s" % (checker_path, e), + "Internet-Draft submission checker class '%s':\n %s" % (checker_path, e), hint = "Please check that the class exists and can be imported.\n", id = "datatracker.E0008", )) @@ -188,7 +113,7 @@ def check_id_submission_checkers(app_configs, **kwargs): except Exception as e: errors.append(checks.Critical( "An exception was raised when trying to instantiate\n" - "the draft submission checker class '%s':\n %s" % (checker_path, e), + "the Internet-Draft submission checker class '%s':\n %s" % (checker_path, e), hint = "Please check that the class can be instantiated.\n", id = "datatracker.E0009", )) @@ -196,7 +121,7 @@ def check_id_submission_checkers(app_configs, **kwargs): for attr in ('name',): if not hasattr(checker, attr): errors.append(checks.Critical( - "The draft submission checker\n '%s'\n" + "The Internet-Draft submission checker\n '%s'\n" "has no attribute '%s', which is required" % (checker_path, attr), hint = "Please update the class.\n", id = "datatracker.E0010", @@ -207,7 +132,7 @@ def check_id_submission_checkers(app_configs, **kwargs): break else: errors.append(checks.Critical( - "The draft submission checker\n '%s'\n" + "The Internet-Draft submission checker\n '%s'\n" " has no recognised checker method; " "should be one or more of %s." % (checker_path, checker_methods), hint = "Please update the class.\n", diff --git a/ietf/community/admin.py b/ietf/community/admin.py index 890819d9d9..4c947ad3f7 100644 --- a/ietf/community/admin.py +++ b/ietf/community/admin.py @@ -7,8 +7,8 @@ from ietf.community.models import CommunityList, SearchRule, EmailSubscription class CommunityListAdmin(admin.ModelAdmin): - list_display = ['id', 'user', 'group'] - raw_id_fields = ['user', 'group', 'added_docs'] + list_display = ['id', 'person', 'group'] + raw_id_fields = ['person', 'group', 'added_docs'] admin.site.register(CommunityList, CommunityListAdmin) class SearchRuleAdmin(admin.ModelAdmin): diff --git a/ietf/community/apps.py b/ietf/community/apps.py new file mode 100644 index 0000000000..ab0a6d6054 --- /dev/null +++ b/ietf/community/apps.py @@ -0,0 +1,12 @@ +# Copyright The IETF Trust 2024, All Rights Reserved + +from django.apps import AppConfig + + +class CommunityConfig(AppConfig): + name = "ietf.community" + + def ready(self): + """Initialize the app after the registry is populated""" + # implicitly connects @receiver-decorated signals + from . import signals # pyflakes: ignore diff --git a/ietf/community/forms.py b/ietf/community/forms.py index a8709c787a..d3fa01dd19 100644 --- a/ietf/community/forms.py +++ b/ietf/community/forms.py @@ -13,7 +13,7 @@ from ietf.person.fields import SearchablePersonField class AddDocumentsForm(forms.Form): - documents = SearchableDocumentsField(label="Add documents to track", doc_type="draft") + documents = SearchableDocumentsField(label="Add Internet-Drafts to track", doc_type="draft") class SearchRuleTypeForm(forms.Form): rule_type = forms.ChoiceField(choices=[('', '--------------')] + SearchRule.RULE_TYPES) @@ -30,6 +30,8 @@ def __init__(self, clist, rule_type, *args, **kwargs): super(SearchRuleForm, self).__init__(*args, **kwargs) def restrict_state(state_type, slug=None): + if "state" not in self.fields: + raise RuntimeError(f"Rule type {rule_type} cannot include state filtering") f = self.fields['state'] f.queryset = f.queryset.filter(used=True).filter(type=state_type) if slug: @@ -38,11 +40,15 @@ def restrict_state(state_type, slug=None): f.initial = f.queryset[0].pk f.widget = forms.HiddenInput() + if rule_type.endswith("_rfc"): + del self.fields["state"] # rfc rules must not look at document states + if rule_type in ["group", "group_rfc", "area", "area_rfc", "group_exp"]: if rule_type == "group_exp": restrict_state("draft", "expired") else: - restrict_state("draft", "rfc" if rule_type.endswith("rfc") else "active") + if not rule_type.endswith("_rfc"): + restrict_state("draft", "active") if rule_type.startswith("area"): self.fields["group"].label = "Area" @@ -70,7 +76,8 @@ def restrict_state(state_type, slug=None): del self.fields["text"] elif rule_type in ["author", "author_rfc", "shepherd", "ad"]: - restrict_state("draft", "rfc" if rule_type.endswith("rfc") else "active") + if not rule_type.endswith("_rfc"): + restrict_state("draft", "active") if rule_type.startswith("author"): self.fields["person"].label = "Author" @@ -84,7 +91,8 @@ def restrict_state(state_type, slug=None): del self.fields["text"] elif rule_type == "name_contains": - restrict_state("draft", "rfc" if rule_type.endswith("rfc") else "active") + if not rule_type.endswith("_rfc"): + restrict_state("draft", "active") del self.fields["person"] del self.fields["group"] @@ -106,14 +114,13 @@ def clean_text(self): class SubscriptionForm(forms.ModelForm): - def __init__(self, user, clist, *args, **kwargs): + def __init__(self, person, clist, *args, **kwargs): self.clist = clist - self.user = user super(SubscriptionForm, self).__init__(*args, **kwargs) self.fields["notify_on"].widget = forms.RadioSelect(choices=self.fields["notify_on"].choices) - self.fields["email"].queryset = self.fields["email"].queryset.filter(person__user=user, active=True).order_by("-primary") + self.fields["email"].queryset = self.fields["email"].queryset.filter(person=person, active=True).order_by("-primary") self.fields["email"].widget = forms.RadioSelect(choices=[t for t in self.fields["email"].choices if t[0]]) if self.fields["email"].queryset: diff --git a/ietf/community/migrations/0001_initial.py b/ietf/community/migrations/0001_initial.py index dc1ae75a3f..44154687f3 100644 --- a/ietf/community/migrations/0001_initial.py +++ b/ietf/community/migrations/0001_initial.py @@ -1,10 +1,6 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-20 10:52 - - -from typing import List, Tuple # pyflakes:ignore +# Generated by Django 2.2.28 on 2023-03-20 19:22 +from typing import List, Tuple from django.db import migrations, models import django.db.models.deletion import ietf.utils.models @@ -14,8 +10,8 @@ class Migration(migrations.Migration): initial = True - dependencies = [ - ] # type: List[Tuple[str]] + dependencies: List[Tuple[str, str]] = [ + ] operations = [ migrations.CreateModel( @@ -35,7 +31,7 @@ class Migration(migrations.Migration): name='SearchRule', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('rule_type', models.CharField(choices=[('group', 'All I-Ds associated with a particular group'), ('area', 'All I-Ds associated with all groups in a particular Area'), ('group_rfc', 'All RFCs associated with a particular group'), ('area_rfc', 'All RFCs associated with all groups in a particular Area'), ('state_iab', 'All I-Ds that are in a particular IAB state'), ('state_iana', 'All I-Ds that are in a particular IANA state'), ('state_iesg', 'All I-Ds that are in a particular IESG state'), ('state_irtf', 'All I-Ds that are in a particular IRTF state'), ('state_ise', 'All I-Ds that are in a particular ISE state'), ('state_rfceditor', 'All I-Ds that are in a particular RFC Editor state'), ('state_ietf', 'All I-Ds that are in a particular Working Group state'), ('author', 'All I-Ds with a particular author'), ('author_rfc', 'All RFCs with a particular author'), ('ad', 'All I-Ds with a particular responsible AD'), ('shepherd', 'All I-Ds with a particular document shepherd'), ('name_contains', 'All I-Ds with particular text/regular expression in the name')], max_length=30)), + ('rule_type', models.CharField(choices=[('group', 'All I-Ds associated with a particular group'), ('area', 'All I-Ds associated with all groups in a particular Area'), ('group_rfc', 'All RFCs associated with a particular group'), ('area_rfc', 'All RFCs associated with all groups in a particular Area'), ('group_exp', 'All expired I-Ds of a particular group'), ('state_iab', 'All I-Ds that are in a particular IAB state'), ('state_iana', 'All I-Ds that are in a particular IANA state'), ('state_iesg', 'All I-Ds that are in a particular IESG state'), ('state_irtf', 'All I-Ds that are in a particular IRTF state'), ('state_ise', 'All I-Ds that are in a particular ISE state'), ('state_rfceditor', 'All I-Ds that are in a particular RFC Editor state'), ('state_ietf', 'All I-Ds that are in a particular Working Group state'), ('author', 'All I-Ds with a particular author'), ('author_rfc', 'All RFCs with a particular author'), ('ad', 'All I-Ds with a particular responsible AD'), ('shepherd', 'All I-Ds with a particular document shepherd'), ('name_contains', 'All I-Ds with particular text/regular expression in the name')], max_length=30)), ('text', models.CharField(blank=True, default='', max_length=255, verbose_name='Text/RegExp')), ('community_list', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='community.CommunityList')), ], diff --git a/ietf/community/migrations/0002_auto_20180220_1052.py b/ietf/community/migrations/0002_auto_20180220_1052.py deleted file mode 100644 index cb5658ac19..0000000000 --- a/ietf/community/migrations/0002_auto_20180220_1052.py +++ /dev/null @@ -1,70 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-20 10:52 - - -from django.conf import settings -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - initial = True - - dependencies = [ - ('doc', '0001_initial'), - ('group', '0001_initial'), - migrations.swappable_dependency(settings.AUTH_USER_MODEL), - ('community', '0001_initial'), - ('person', '0001_initial'), - ] - - operations = [ - migrations.AddField( - model_name='searchrule', - name='group', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='group.Group'), - ), - migrations.AddField( - model_name='searchrule', - name='name_contains_index', - field=models.ManyToManyField(to='doc.Document'), - ), - migrations.AddField( - model_name='searchrule', - name='person', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='person.Person'), - ), - migrations.AddField( - model_name='searchrule', - name='state', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.State'), - ), - migrations.AddField( - model_name='emailsubscription', - name='community_list', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='community.CommunityList'), - ), - migrations.AddField( - model_name='emailsubscription', - name='email', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Email'), - ), - migrations.AddField( - model_name='communitylist', - name='added_docs', - field=models.ManyToManyField(to='doc.Document'), - ), - migrations.AddField( - model_name='communitylist', - name='group', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='group.Group'), - ), - migrations.AddField( - model_name='communitylist', - name='user', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL), - ), - ] diff --git a/ietf/community/migrations/0002_auto_20230320_1222.py b/ietf/community/migrations/0002_auto_20230320_1222.py new file mode 100644 index 0000000000..f552cc06e1 --- /dev/null +++ b/ietf/community/migrations/0002_auto_20230320_1222.py @@ -0,0 +1,67 @@ +# Generated by Django 2.2.28 on 2023-03-20 19:22 + +from django.conf import settings +from django.db import migrations, models +import django.db.models.deletion +import ietf.utils.models + + +class Migration(migrations.Migration): + + initial = True + + dependencies = [ + ('person', '0001_initial'), + migrations.swappable_dependency(settings.AUTH_USER_MODEL), + ('community', '0001_initial'), + ('group', '0001_initial'), + ('doc', '0001_initial'), + ] + + operations = [ + migrations.AddField( + model_name='searchrule', + name='group', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='group.Group'), + ), + migrations.AddField( + model_name='searchrule', + name='name_contains_index', + field=models.ManyToManyField(to='doc.Document'), + ), + migrations.AddField( + model_name='searchrule', + name='person', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='person.Person'), + ), + migrations.AddField( + model_name='searchrule', + name='state', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.State'), + ), + migrations.AddField( + model_name='emailsubscription', + name='community_list', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='community.CommunityList'), + ), + migrations.AddField( + model_name='emailsubscription', + name='email', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Email'), + ), + migrations.AddField( + model_name='communitylist', + name='added_docs', + field=models.ManyToManyField(to='doc.Document'), + ), + migrations.AddField( + model_name='communitylist', + name='group', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='group.Group'), + ), + migrations.AddField( + model_name='communitylist', + name='user', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL), + ), + ] diff --git a/ietf/community/migrations/0003_add_communitylist_docs2_m2m.py b/ietf/community/migrations/0003_add_communitylist_docs2_m2m.py deleted file mode 100644 index 3bfaee9aa9..0000000000 --- a/ietf/community/migrations/0003_add_communitylist_docs2_m2m.py +++ /dev/null @@ -1,44 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-21 14:23 - - -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('community', '0002_auto_20180220_1052'), - ] - - operations = [ - migrations.CreateModel( - name='CommunityListDocs', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('communitylist', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='community.CommunityList')), - ('document', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field=b'id')), - ], - ), - migrations.AddField( - model_name='communitylist', - name='added_docs2', - field=models.ManyToManyField(related_name='communitylists', through='community.CommunityListDocs', to='doc.Document'), - ), - migrations.CreateModel( - name='SearchRuleDocs', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('document', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field=b'id')), - ('searchrule', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='community.SearchRule')), - ], - ), - migrations.AddField( - model_name='searchrule', - name='name_contains_index2', - field=models.ManyToManyField(related_name='searchrules', through='community.SearchRuleDocs', to='doc.Document'), - ), - ] diff --git a/ietf/community/migrations/0003_track_rfcs.py b/ietf/community/migrations/0003_track_rfcs.py new file mode 100644 index 0000000000..3c2d04097d --- /dev/null +++ b/ietf/community/migrations/0003_track_rfcs.py @@ -0,0 +1,50 @@ +# Generated by Django 4.2.3 on 2023-07-07 18:33 + +from django.db import migrations + + +def forward(apps, schema_editor): + """Track any RFCs that were created from tracked drafts""" + CommunityList = apps.get_model("community", "CommunityList") + RelatedDocument = apps.get_model("doc", "RelatedDocument") + + # Handle individually tracked documents + for cl in CommunityList.objects.all(): + for rfc in set( + RelatedDocument.objects.filter( + source__in=cl.added_docs.all(), + relationship__slug="became_rfc", + ).values_list("target__docs", flat=True) + ): + cl.added_docs.add(rfc) + + # Handle rules - rules ending with _rfc should no longer filter by state. + # There are 9 CommunityLists with invalid author_rfc rules that are filtering + # by (draft, active) instead of (draft, rfc) state before migration. All but one + # also includes an author rule for (draft, active), so these will start following + # RFCs as well. The one exception will start tracking RFCs instead of I-Ds, which + # is probably what was intended, but will be a change in their user experience. + SearchRule = apps.get_model("community", "SearchRule") + rfc_rules = SearchRule.objects.filter(rule_type__endswith="_rfc") + rfc_rules.update(state=None) + +def reverse(apps, schema_editor): + Document = apps.get_model("doc", "Document") + for rfc in Document.objects.filter(type__slug="rfc"): + rfc.communitylist_set.clear() + + # See the comment above regarding author_rfc + SearchRule = apps.get_model("community", "SearchRule") + State = apps.get_model("doc", "State") + SearchRule.objects.filter(rule_type__endswith="_rfc").update( + state=State.objects.get(type_id="draft", slug="rfc") + ) + + +class Migration(migrations.Migration): + dependencies = [ + ("community", "0002_auto_20230320_1222"), + ("doc", "0014_move_rfc_docaliases"), + ] + + operations = [migrations.RunPython(forward, reverse)] diff --git a/ietf/community/migrations/0004_delete_useless_community_lists.py b/ietf/community/migrations/0004_delete_useless_community_lists.py new file mode 100644 index 0000000000..9f657a3c34 --- /dev/null +++ b/ietf/community/migrations/0004_delete_useless_community_lists.py @@ -0,0 +1,26 @@ +# Generated by Django 4.2.9 on 2024-01-05 21:28 + +from django.db import migrations + + +def forward(apps, schema_editor): + CommunityList = apps.get_model("community", "CommunityList") + # As of 2024-01-05, there are 570 personal CommunityLists with a user + # who has no associated Person. None of these has an EmailSubscription, + # so the lists are doing nothing and can be safely deleted. + personal_lists_no_person = CommunityList.objects.exclude( + user__isnull=True + ).filter( + user__person__isnull=True + ) + # Confirm the assumption that none of the lists to be deleted has an EmailSubscription + assert not personal_lists_no_person.filter(emailsubscription__isnull=False).exists() + personal_lists_no_person.delete() + + +class Migration(migrations.Migration): + dependencies = [ + ("community", "0003_track_rfcs"), + ] + + operations = [migrations.RunPython(forward)] diff --git a/ietf/community/migrations/0004_set_document_m2m_keys.py b/ietf/community/migrations/0004_set_document_m2m_keys.py deleted file mode 100644 index 60e51fc368..0000000000 --- a/ietf/community/migrations/0004_set_document_m2m_keys.py +++ /dev/null @@ -1,56 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-21 14:27 - - -import sys - -from tqdm import tqdm - -from django.db import migrations - - -def forward(apps, schema_editor): - - Document = apps.get_model('doc','Document') - CommunityList = apps.get_model('community', 'CommunityList') - CommunityListDocs = apps.get_model('community', 'CommunityListDocs') - SearchRule = apps.get_model('community', 'SearchRule') - SearchRuleDocs = apps.get_model('community', 'SearchRuleDocs') - - # Document id fixup ------------------------------------------------------------ - - objs = Document.objects.in_bulk() - nameid = { o.name: o.id for id, o in objs.items() } - - sys.stderr.write('\n') - - sys.stderr.write(' %s.%s:\n' % (CommunityList.__name__, 'added_docs')) - count = 0 - for l in tqdm(CommunityList.objects.all()): - for d in l.added_docs.all(): - count += 1 - CommunityListDocs.objects.get_or_create(communitylist=l, document_id=nameid[d.name]) - sys.stderr.write(' %s CommunityListDocs objects created\n' % (count, )) - - sys.stderr.write(' %s.%s:\n' % (SearchRule.__name__, 'name_contains_index')) - count = 0 - for r in tqdm(SearchRule.objects.all()): - for d in r.name_contains_index.all(): - count += 1 - SearchRuleDocs.objects.get_or_create(searchrule=r, document_id=nameid[d.name]) - sys.stderr.write(' %s SearchRuleDocs objects created\n' % (count, )) - -def reverse(apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('community', '0003_add_communitylist_docs2_m2m'), - ('doc', '0014_set_document_docalias_id'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/community/migrations/0005_1_del_docs_m2m_table.py b/ietf/community/migrations/0005_1_del_docs_m2m_table.py deleted file mode 100644 index 9a7b7f9453..0000000000 --- a/ietf/community/migrations/0005_1_del_docs_m2m_table.py +++ /dev/null @@ -1,32 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-22 08:15 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('community', '0004_set_document_m2m_keys'), - ] - - # The implementation of AlterField in Django 1.11 applies - # 'ALTER TABLE MODIFY ...;' in order to fix foregn keys - # to the altered field, but as it seems does _not_ fix up m2m - # intermediary tables in an equivalent manner, so here we remove and - # then recreate the m2m tables so they will have the appropriate field - # types. - - operations = [ - # Remove fields - migrations.RemoveField( - model_name='communitylist', - name='added_docs', - ), - migrations.RemoveField( - model_name='searchrule', - name='name_contains_index', - ), - ] diff --git a/ietf/community/migrations/0005_2_add_docs_m2m_table.py b/ietf/community/migrations/0005_2_add_docs_m2m_table.py deleted file mode 100644 index ef1f9bc4f5..0000000000 --- a/ietf/community/migrations/0005_2_add_docs_m2m_table.py +++ /dev/null @@ -1,34 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-22 08:15 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('community', '0005_1_del_docs_m2m_table'), - ] - - # The implementation of AlterField in Django 1.11 applies - # 'ALTER TABLE
MODIFY ...;' in order to fix foregn keys - # to the altered field, but as it seems does _not_ fix up m2m - # intermediary tables in an equivalent manner, so here we remove and - # then recreate the m2m tables so they will have the appropriate field - # types. - - operations = [ - # Add fields back (will create the m2m tables with the right field types) - migrations.AddField( - model_name='communitylist', - name='added_docs', - field=models.ManyToManyField(to='doc.Document'), - ), - migrations.AddField( - model_name='searchrule', - name='name_contains_index', - field=models.ManyToManyField(to='doc.Document'), - ), - ] diff --git a/ietf/community/migrations/0005_user_to_person.py b/ietf/community/migrations/0005_user_to_person.py new file mode 100644 index 0000000000..01d8950edb --- /dev/null +++ b/ietf/community/migrations/0005_user_to_person.py @@ -0,0 +1,54 @@ +# Generated by Django 4.2.2 on 2023-06-12 19:35 + +from django.conf import settings +from django.db import migrations +import django.db.models.deletion +import ietf.utils.models + + +def forward(apps, schema_editor): + CommunityList = apps.get_model('community', 'CommunityList') + for clist in CommunityList.objects.all(): + try: + clist.person = clist.user.person + except: + clist.person = None + clist.save() + +def reverse(apps, schema_editor): + CommunityList = apps.get_model('community', 'CommunityList') + for clist in CommunityList.objects.all(): + try: + clist.user = clist.person.user + except: + clist.user = None + clist.save() + +class Migration(migrations.Migration): + dependencies = [ + ("community", "0004_delete_useless_community_lists"), + ("person", "0001_initial"), + ] + + operations = [ + migrations.AddField( + model_name="communitylist", + name="person", + field=ietf.utils.models.ForeignKey( + blank=True, + null=True, + on_delete=django.db.models.deletion.CASCADE, + to="person.Person", + ), + ), + migrations.RunPython(forward, reverse), + migrations.RemoveField( + model_name="communitylist", + name="user", + field=ietf.utils.models.ForeignKey( + blank=True, + null=True, + on_delete=django.db.models.deletion.CASCADE, + to=settings.AUTH_USER_MODEL), + ), + ] diff --git a/ietf/community/migrations/0006_copy_docs_m2m_table.py b/ietf/community/migrations/0006_copy_docs_m2m_table.py deleted file mode 100644 index 6d1a5e3a88..0000000000 --- a/ietf/community/migrations/0006_copy_docs_m2m_table.py +++ /dev/null @@ -1,57 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-27 05:56 - - -from django.db import migrations - -import sys, time - -from tqdm import tqdm - - -def forward(apps, schema_editor): - - CommunityList = apps.get_model('community', 'CommunityList') - CommunityListDocs = apps.get_model('community', 'CommunityListDocs') - SearchRule = apps.get_model('community', 'SearchRule') - SearchRuleDocs = apps.get_model('community', 'SearchRuleDocs') - - # Document id fixup ------------------------------------------------------------ - - - sys.stderr.write('\n') - - sys.stderr.write(' %s.%s:\n' % (CommunityList.__name__, 'added_docs')) - for l in tqdm(CommunityList.objects.all()): - l.added_docs.set([ d.document for d in CommunityListDocs.objects.filter(communitylist=l) ]) - - sys.stderr.write(' %s.%s:\n' % (SearchRule.__name__, 'name_contains_index')) - for r in tqdm(SearchRule.objects.all()): - r.name_contains_index.set([ d.document for d in SearchRuleDocs.objects.filter(searchrule=r) ]) - -def reverse(apps, schema_editor): - pass - -def timestamp(apps, schema_editor): - sys.stderr.write('\n %s' % time.strftime('%Y-%m-%d %H:%M:%S')) - -class Migration(migrations.Migration): - - dependencies = [ - ('community', '0005_2_add_docs_m2m_table'), - ] - - operations = [ - #migrations.RunPython(forward, reverse), - # Alternative: - migrations.RunPython(timestamp, timestamp), - migrations.RunSQL( - "INSERT INTO community_communitylist_added_docs SELECT * FROM community_communitylistdocs;", - ""), - migrations.RunPython(timestamp, timestamp), - migrations.RunSQL( - "INSERT INTO community_searchrule_name_contains_index SELECT * FROM community_searchruledocs;", - ""), - migrations.RunPython(timestamp, timestamp), - ] diff --git a/ietf/community/migrations/0007_remove_docs2_m2m.py b/ietf/community/migrations/0007_remove_docs2_m2m.py deleted file mode 100644 index a454fbf072..0000000000 --- a/ietf/community/migrations/0007_remove_docs2_m2m.py +++ /dev/null @@ -1,46 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-30 03:06 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('community', '0006_copy_docs_m2m_table'), - ] - - operations = [ - migrations.RemoveField( - model_name='communitylistdocs', - name='communitylist', - ), - migrations.RemoveField( - model_name='communitylistdocs', - name='document', - ), - migrations.RemoveField( - model_name='searchruledocs', - name='document', - ), - migrations.RemoveField( - model_name='searchruledocs', - name='searchrule', - ), - migrations.RemoveField( - model_name='communitylist', - name='added_docs2', - ), - migrations.RemoveField( - model_name='searchrule', - name='name_contains_index2', - ), - migrations.DeleteModel( - name='CommunityListDocs', - ), - migrations.DeleteModel( - name='SearchRuleDocs', - ), - ] diff --git a/ietf/community/migrations/0008_add_group_exp_rule.py b/ietf/community/migrations/0008_add_group_exp_rule.py deleted file mode 100644 index b5ddebe638..0000000000 --- a/ietf/community/migrations/0008_add_group_exp_rule.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.2.28 on 2022-06-30 05:59 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('community', '0007_remove_docs2_m2m'), - ] - - operations = [ - migrations.AlterField( - model_name='searchrule', - name='rule_type', - field=models.CharField(choices=[('group', 'All I-Ds associated with a particular group'), ('area', 'All I-Ds associated with all groups in a particular Area'), ('group_rfc', 'All RFCs associated with a particular group'), ('area_rfc', 'All RFCs associated with all groups in a particular Area'), ('group_exp', 'All expired I-Ds of a particular group'), ('state_iab', 'All I-Ds that are in a particular IAB state'), ('state_iana', 'All I-Ds that are in a particular IANA state'), ('state_iesg', 'All I-Ds that are in a particular IESG state'), ('state_irtf', 'All I-Ds that are in a particular IRTF state'), ('state_ise', 'All I-Ds that are in a particular ISE state'), ('state_rfceditor', 'All I-Ds that are in a particular RFC Editor state'), ('state_ietf', 'All I-Ds that are in a particular Working Group state'), ('author', 'All I-Ds with a particular author'), ('author_rfc', 'All RFCs with a particular author'), ('ad', 'All I-Ds with a particular responsible AD'), ('shepherd', 'All I-Ds with a particular document shepherd'), ('name_contains', 'All I-Ds with particular text/regular expression in the name')], max_length=30), - ), - ] diff --git a/ietf/community/migrations/0009_add_group_exp_rule_to_groups.py b/ietf/community/migrations/0009_add_group_exp_rule_to_groups.py deleted file mode 100644 index 1d3222dc6e..0000000000 --- a/ietf/community/migrations/0009_add_group_exp_rule_to_groups.py +++ /dev/null @@ -1,30 +0,0 @@ -# Generated by Django 2.2.28 on 2022-06-30 23:15 - - -from django.db import migrations - - -def forward(apps, schema_editor): - SearchRule = apps.get_model('community', 'SearchRule') - CommunityList = apps.get_model('community', 'CommunityList') - Group = apps.get_model('group', 'Group') - State = apps.get_model('doc', 'State') - for group in Group.objects.filter(type_id__in=['wg','rg'], state_id='active'): - com_list = CommunityList.objects.filter(group=group).first() - if com_list is not None: - SearchRule.objects.create(community_list=com_list, rule_type="group_exp", group=group, state=State.objects.get(slug="expired", type="draft"),) - - -def reverse(apps, schema_editor): - SearchRule = apps.get_model('community', 'SearchRule') - SearchRule.objects.filter(rule_type='group_exp').delete() - - -class Migration(migrations.Migration): - dependencies = [ - ('community', '0008_add_group_exp_rule'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/community/models.py b/ietf/community/models.py index 8938b1c097..6945918f9a 100644 --- a/ietf/community/models.py +++ b/ietf/community/models.py @@ -1,37 +1,35 @@ # Copyright The IETF Trust 2012-2020, All Rights Reserved # -*- coding: utf-8 -*- - -from django.contrib.auth.models import User from django.db import models -from django.db.models import signals from django.urls import reverse as urlreverse -from ietf.doc.models import Document, DocEvent, State +from ietf.doc.models import Document, State from ietf.group.models import Group from ietf.person.models import Person, Email from ietf.utils.models import ForeignKey + class CommunityList(models.Model): - user = ForeignKey(User, blank=True, null=True) + person = ForeignKey(Person, blank=True, null=True) group = ForeignKey(Group, blank=True, null=True) added_docs = models.ManyToManyField(Document) def long_name(self): - if self.user: - return 'Personal ID list of %s' % self.user.username + if self.person: + return 'Personal I-D list of %s' % self.person.plain_name() elif self.group: - return 'ID list for %s' % self.group.name + return 'I-D list for %s' % self.group.name else: - return 'ID list' + return 'I-D list' def __str__(self): return self.long_name() def get_absolute_url(self): import ietf.community.views - if self.user: - return urlreverse(ietf.community.views.view_list, kwargs={ 'username': self.user.username }) + if self.person: + return urlreverse(ietf.community.views.view_list, kwargs={ 'email_or_name': self.person.email() }) elif self.group: return urlreverse("ietf.group.views.group_documents", kwargs={ 'acronym': self.group.acronym }) return "" @@ -95,20 +93,3 @@ class EmailSubscription(models.Model): def __str__(self): return "%s to %s (%s changes)" % (self.email, self.community_list, self.notify_on) - - -def notify_events(sender, instance, **kwargs): - if not isinstance(instance, DocEvent): - return - - if instance.doc.type_id != 'draft': - return - - if getattr(instance, "skip_community_list_notification", False): - return - - from ietf.community.utils import notify_event_to_subscribers - notify_event_to_subscribers(instance) - - -signals.post_save.connect(notify_events) diff --git a/ietf/community/signals.py b/ietf/community/signals.py new file mode 100644 index 0000000000..20ee761129 --- /dev/null +++ b/ietf/community/signals.py @@ -0,0 +1,44 @@ +# Copyright The IETF Trust 2024, All Rights Reserved + +from django.conf import settings +from django.db import transaction +from django.db.models.signals import post_save +from django.dispatch import receiver + +from ietf.doc.models import DocEvent +from .tasks import notify_event_to_subscribers_task + + +def notify_of_event(event: DocEvent): + """Send subscriber notification emails for a 'draft'-related DocEvent + + If the event is attached to a draft of type 'doc', queues a task to send notification emails to + community list subscribers. No emails will be sent when SERVER_MODE is 'test'. + """ + if event.doc.type_id != "draft": + return + + if getattr(event, "skip_community_list_notification", False): + return + + # kludge alert: queuing a celery task in response to a signal can cause unexpected attempts to + # start a Celery task during tests. To prevent this, don't queue a celery task if we're running + # tests. + if settings.SERVER_MODE != "test": + # Wrap in on_commit in case a transaction is open + transaction.on_commit( + lambda: notify_event_to_subscribers_task.delay(event_id=event.pk) + ) + + +# dispatch_uid ensures only a single signal receiver binding is made +@receiver(post_save, dispatch_uid="notify_of_events_receiver_uid") +def notify_of_events_receiver(sender, instance, **kwargs): + """Call notify_of_event after saving a new DocEvent""" + if not isinstance(instance, DocEvent): + return + + if not kwargs.get("created", False): + return # only notify on creation + + notify_of_event(instance) diff --git a/ietf/community/tasks.py b/ietf/community/tasks.py new file mode 100644 index 0000000000..763a596495 --- /dev/null +++ b/ietf/community/tasks.py @@ -0,0 +1,15 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +from celery import shared_task + +from ietf.doc.models import DocEvent +from ietf.utils.log import log + + +@shared_task +def notify_event_to_subscribers_task(event_id): + from .utils import notify_event_to_subscribers + event = DocEvent.objects.filter(pk=event_id).first() + if event is None: + log(f"Unable to send subscriber notifications because DocEvent {event_id} was not found") + else: + notify_event_to_subscribers(event) diff --git a/ietf/community/tests.py b/ietf/community/tests.py index 3dd86f70e3..04f1433d61 100644 --- a/ietf/community/tests.py +++ b/ietf/community/tests.py @@ -1,60 +1,112 @@ -# Copyright The IETF Trust 2016-2020, All Rights Reserved +# Copyright The IETF Trust 2016-2023, All Rights Reserved # -*- coding: utf-8 -*- - +from unittest import mock from pyquery import PyQuery +from django.test.utils import override_settings from django.urls import reverse as urlreverse -from django.contrib.auth.models import User - -from django_webtest import WebTest +from lxml import etree -import debug # pyflakes:ignore +import debug # pyflakes:ignore from ietf.community.models import CommunityList, SearchRule, EmailSubscription -from ietf.community.utils import docs_matching_community_list_rule, community_list_rules_matching_doc -from ietf.community.utils import reset_name_contains_index_for_rule +from ietf.community.signals import notify_of_event +from ietf.community.utils import ( + docs_matching_community_list_rule, + community_list_rules_matching_doc, +) +from ietf.community.utils import ( + reset_name_contains_index_for_rule, + notify_event_to_subscribers, +) +from ietf.community.tasks import notify_event_to_subscribers_task import ietf.community.views from ietf.group.models import Group from ietf.group.utils import setup_default_community_list_for_group +from ietf.doc.factories import DocumentFactory from ietf.doc.models import State from ietf.doc.utils import add_state_change_event -from ietf.person.models import Person, Email -from ietf.utils.test_utils import login_testing_unauthorized -from ietf.utils.mail import outbox -from ietf.doc.factories import WgDraftFactory +from ietf.person.models import Person, Email, Alias +from ietf.utils.test_utils import TestCase, login_testing_unauthorized +from ietf.doc.factories import DocEventFactory, WgDraftFactory from ietf.group.factories import GroupFactory, RoleFactory -from ietf.person.factories import PersonFactory +from ietf.person.factories import PersonFactory, EmailFactory, AliasFactory + -class CommunityListTests(WebTest): +class CommunityListTests(TestCase): def test_rule_matching(self): - plain = PersonFactory(user__username='plain') - ad = Person.objects.get(user__username='ad') + plain = PersonFactory(user__username="plain") + ad = Person.objects.get(user__username="ad") draft = WgDraftFactory( - group__parent=Group.objects.get(acronym='farfut' ), + group__parent=Group.objects.get(acronym="farfut"), authors=[ad], ad=ad, shepherd=plain.email(), - states=[('draft-iesg','lc'),('draft','active')], + states=[("draft-iesg", "lc"), ("draft", "active")], ) - clist = CommunityList.objects.create(user=User.objects.get(username="plain")) + clist = CommunityList.objects.create(person=plain) - rule_group = SearchRule.objects.create(rule_type="group", group=draft.group, state=State.objects.get(type="draft", slug="active"), community_list=clist) - rule_group_rfc = SearchRule.objects.create(rule_type="group_rfc", group=draft.group, state=State.objects.get(type="draft", slug="rfc"), community_list=clist) - rule_area = SearchRule.objects.create(rule_type="area", group=draft.group.parent, state=State.objects.get(type="draft", slug="active"), community_list=clist) + rule_group = SearchRule.objects.create( + rule_type="group", + group=draft.group, + state=State.objects.get(type="draft", slug="active"), + community_list=clist, + ) + rule_group_rfc = SearchRule.objects.create( + rule_type="group_rfc", + group=draft.group, + state=State.objects.get(type="rfc", slug="published"), + community_list=clist, + ) + rule_area = SearchRule.objects.create( + rule_type="area", + group=draft.group.parent, + state=State.objects.get(type="draft", slug="active"), + community_list=clist, + ) - rule_state_iesg = SearchRule.objects.create(rule_type="state_iesg", state=State.objects.get(type="draft-iesg", slug="lc"), community_list=clist) + rule_state_iesg = SearchRule.objects.create( + rule_type="state_iesg", + state=State.objects.get(type="draft-iesg", slug="lc"), + community_list=clist, + ) - rule_author = SearchRule.objects.create(rule_type="author", state=State.objects.get(type="draft", slug="active"), person=Person.objects.filter(documentauthor__document=draft).first(), community_list=clist) + rule_author = SearchRule.objects.create( + rule_type="author", + state=State.objects.get(type="draft", slug="active"), + person=Person.objects.filter(documentauthor__document=draft).first(), + community_list=clist, + ) - rule_ad = SearchRule.objects.create(rule_type="ad", state=State.objects.get(type="draft", slug="active"), person=draft.ad, community_list=clist) + rule_ad = SearchRule.objects.create( + rule_type="ad", + state=State.objects.get(type="draft", slug="active"), + person=draft.ad, + community_list=clist, + ) - rule_shepherd = SearchRule.objects.create(rule_type="shepherd", state=State.objects.get(type="draft", slug="active"), person=draft.shepherd.person, community_list=clist) + rule_shepherd = SearchRule.objects.create( + rule_type="shepherd", + state=State.objects.get(type="draft", slug="active"), + person=draft.shepherd.person, + community_list=clist, + ) - rule_group_exp = SearchRule.objects.create(rule_type="group_exp", group=draft.group, state=State.objects.get(type="draft", slug="expired"), community_list=clist) + rule_group_exp = SearchRule.objects.create( + rule_type="group_exp", + group=draft.group, + state=State.objects.get(type="draft", slug="expired"), + community_list=clist, + ) - rule_name_contains = SearchRule.objects.create(rule_type="name_contains", state=State.objects.get(type="draft", slug="active"), text="draft-.*" + "-".join(draft.name.split("-")[2:]), community_list=clist) + rule_name_contains = SearchRule.objects.create( + rule_type="name_contains", + state=State.objects.get(type="draft", slug="active"), + text="draft-.*" + "-".join(draft.name.split("-")[2:]), + community_list=clist, + ) reset_name_contains_index_for_rule(rule_name_contains) # doc -> rules @@ -71,37 +123,71 @@ def test_rule_matching(self): # rule -> docs self.assertTrue(draft in list(docs_matching_community_list_rule(rule_group))) - self.assertTrue(draft not in list(docs_matching_community_list_rule(rule_group_rfc))) + self.assertTrue( + draft not in list(docs_matching_community_list_rule(rule_group_rfc)) + ) self.assertTrue(draft in list(docs_matching_community_list_rule(rule_area))) - self.assertTrue(draft in list(docs_matching_community_list_rule(rule_state_iesg))) + self.assertTrue( + draft in list(docs_matching_community_list_rule(rule_state_iesg)) + ) self.assertTrue(draft in list(docs_matching_community_list_rule(rule_author))) self.assertTrue(draft in list(docs_matching_community_list_rule(rule_ad))) self.assertTrue(draft in list(docs_matching_community_list_rule(rule_shepherd))) - self.assertTrue(draft in list(docs_matching_community_list_rule(rule_name_contains))) - self.assertTrue(draft not in list(docs_matching_community_list_rule(rule_group_exp))) + self.assertTrue( + draft in list(docs_matching_community_list_rule(rule_name_contains)) + ) + self.assertTrue( + draft not in list(docs_matching_community_list_rule(rule_group_exp)) + ) - draft.set_state(State.objects.get(type='draft', slug='expired')) + draft.set_state(State.objects.get(type="draft", slug="expired")) # doc -> rules matching_rules = list(community_list_rules_matching_doc(draft)) self.assertTrue(rule_group_exp in matching_rules) # rule -> docs - self.assertTrue(draft in list(docs_matching_community_list_rule(rule_group_exp))) + self.assertTrue( + draft in list(docs_matching_community_list_rule(rule_group_exp)) + ) - def test_view_list(self): - PersonFactory(user__username='plain') - draft = WgDraftFactory() + def test_view_list_duplicates(self): + person = PersonFactory( + name="John Q. Public", user__username="bazquux@example.com" + ) + PersonFactory(name="John Q. Public", user__username="foobar@example.com") - url = urlreverse(ietf.community.views.view_list, kwargs={ "username": "plain" }) + url = urlreverse( + ietf.community.views.view_list, + kwargs={"email_or_name": person.plain_name()}, + ) + r = self.client.get(url) + self.assertEqual(r.status_code, 404) + def complex_person(self, *args, **kwargs): + person = PersonFactory(*args, **kwargs) + EmailFactory(person=person) + AliasFactory(person=person) + return person + + def email_or_name_set(self, person): + return [e for e in Email.objects.filter(person=person)] + [ + a for a in Alias.objects.filter(person=person) + ] + + def do_view_list_test(self, person): + draft = WgDraftFactory() # without list - r = self.client.get(url) - self.assertEqual(r.status_code, 200) + for id in self.email_or_name_set(person): + url = urlreverse( + ietf.community.views.view_list, kwargs={"email_or_name": id} + ) + r = self.client.get(url) + self.assertEqual(r.status_code, 200, msg=f"id='{id}', url='{url}'") # with list - clist = CommunityList.objects.create(user=User.objects.get(username="plain")) - if not draft in clist.added_docs.all(): + clist = CommunityList.objects.create(person=person) + if draft not in clist.added_docs.all(): clist.added_docs.add(draft) SearchRule.objects.create( community_list=clist, @@ -109,86 +195,135 @@ def test_view_list(self): state=State.objects.get(type="draft", slug="active"), text="test", ) - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - self.assertContains(r, draft.name) + for id in self.email_or_name_set(person): + url = urlreverse( + ietf.community.views.view_list, kwargs={"email_or_name": id} + ) + r = self.client.get(url) + self.assertEqual(r.status_code, 200, msg=f"id='{id}', url='{url}'") + self.assertContains(r, draft.name) - def test_manage_personal_list(self): + def test_view_list(self): + person = self.complex_person(user__username="plain") + self.do_view_list_test(person) + + def test_view_list_without_active_email(self): + person = self.complex_person(user__username="plain") + person.email_set.update(active=False) + self.do_view_list_test(person) - PersonFactory(user__username='plain') - ad = Person.objects.get(user__username='ad') + def test_manage_personal_list(self): + person = self.complex_person(user__username="plain") + ad = Person.objects.get(user__username="ad") draft = WgDraftFactory(authors=[ad]) - url = urlreverse(ietf.community.views.manage_list, kwargs={ "username": "plain" }) + url = urlreverse( + ietf.community.views.manage_list, kwargs={"email_or_name": person.email()} + ) login_testing_unauthorized(self, "plain", url) - page = self.app.get(url, user='plain') - self.assertEqual(page.status_int, 200) - - # add document - self.assertIn('add_document', page.forms) - form = page.forms['add_document'] - form['documents'].options=[(draft.pk, True, draft.name)] - page = form.submit('action',value='add_documents') - self.assertEqual(page.status_int, 302) - clist = CommunityList.objects.get(user__username="plain") - self.assertTrue(clist.added_docs.filter(pk=draft.pk)) - page = page.follow() - - self.assertContains(page, draft.name) - - # remove document - self.assertIn('remove_document_%s' % draft.pk, page.forms) - form = page.forms['remove_document_%s' % draft.pk] - page = form.submit('action',value='remove_document') - self.assertEqual(page.status_int, 302) - clist = CommunityList.objects.get(user__username="plain") - self.assertTrue(not clist.added_docs.filter(pk=draft.pk)) - page = page.follow() - - # add rule - r = self.client.post(url, { - "action": "add_rule", - "rule_type": "author_rfc", - "author_rfc-person": Person.objects.filter(documentauthor__document=draft).first().pk, - "author_rfc-state": State.objects.get(type="draft", slug="rfc").pk, - }) - self.assertEqual(r.status_code, 302) - clist = CommunityList.objects.get(user__username="plain") - self.assertTrue(clist.searchrule_set.filter(rule_type="author_rfc")) - - # add name_contains rule - r = self.client.post(url, { - "action": "add_rule", - "rule_type": "name_contains", - "name_contains-text": "draft.*mars", - "name_contains-state": State.objects.get(type="draft", slug="active").pk, - }) - self.assertEqual(r.status_code, 302) - clist = CommunityList.objects.get(user__username="plain") - self.assertTrue(clist.searchrule_set.filter(rule_type="name_contains")) - - # rule shows up on GET - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - rule = clist.searchrule_set.filter(rule_type="author_rfc").first() - q = PyQuery(r.content) - self.assertEqual(len(q('#r%s' % rule.pk)), 1) - - # remove rule - r = self.client.post(url, { - "action": "remove_rule", - "rule": rule.pk, - }) - - clist = CommunityList.objects.get(user__username="plain") - self.assertTrue(not clist.searchrule_set.filter(rule_type="author_rfc")) + for id in self.email_or_name_set(person): + url = urlreverse( + ietf.community.views.manage_list, kwargs={"email_or_name": id} + ) + r = self.client.get(url, user="plain") + self.assertEqual(r.status_code, 200, msg=f"id='{id}', url='{url}'") + + # We can't call post() with follow=True because that 404's if + # the url contains unicode, because the django test client + # apparently re-encodes the already-encoded url. + def follow(r): + redirect_url = r.url or url + return self.client.get(redirect_url, user="plain") + + # add document + self.assertContains(r, "add_document") + r = self.client.post( + url, {"action": "add_documents", "documents": draft.pk} + ) + self.assertEqual(r.status_code, 302, msg=f"id='{id}', url='{url}'") + clist = CommunityList.objects.get(person__user__username="plain") + self.assertTrue(clist.added_docs.filter(pk=draft.pk)) + r = follow(r) + self.assertContains(r, draft.name, status_code=200) + + # remove document + self.assertContains(r, "remove_document_%s" % draft.pk) + r = self.client.post( + url, {"action": "remove_document", "document": draft.pk} + ) + self.assertEqual(r.status_code, 302, msg=f"id='{id}', url='{url}'") + clist = CommunityList.objects.get(person__user__username="plain") + self.assertTrue(not clist.added_docs.filter(pk=draft.pk)) + r = follow(r) + self.assertNotContains(r, draft.name, status_code=200) + + # add rule + r = self.client.post( + url, + { + "action": "add_rule", + "rule_type": "author_rfc", + "author_rfc-person": Person.objects.filter( + documentauthor__document=draft + ) + .first() + .pk, + "author_rfc-state": State.objects.get( + type="rfc", slug="published" + ).pk, + }, + ) + self.assertEqual(r.status_code, 302, msg=f"id='{id}', url='{url}'") + clist = CommunityList.objects.get(person__user__username="plain") + self.assertTrue(clist.searchrule_set.filter(rule_type="author_rfc")) + + # add name_contains rule + r = self.client.post( + url, + { + "action": "add_rule", + "rule_type": "name_contains", + "name_contains-text": "draft.*mars", + "name_contains-state": State.objects.get( + type="draft", slug="active" + ).pk, + }, + ) + self.assertEqual(r.status_code, 302, msg=f"id='{id}', url='{url}'") + clist = CommunityList.objects.get(person__user__username="plain") + self.assertTrue(clist.searchrule_set.filter(rule_type="name_contains")) + + # rule shows up on GET + r = self.client.get(url) + self.assertEqual(r.status_code, 200, msg=f"id='{id}', url='{url}'") + rule = clist.searchrule_set.filter(rule_type="author_rfc").first() + q = PyQuery(r.content) + self.assertEqual(len(q("#r%s" % rule.pk)), 1) + + # remove rule + r = self.client.post( + url, + { + "action": "remove_rule", + "rule": rule.pk, + }, + ) + + clist = CommunityList.objects.get(person__user__username="plain") + self.assertTrue(not clist.searchrule_set.filter(rule_type="author_rfc")) def test_manage_group_list(self): - draft = WgDraftFactory(group__acronym='mars') - RoleFactory(group__acronym='mars',name_id='chair',person=PersonFactory(user__username='marschairman')) + draft = WgDraftFactory(group__acronym="mars") + RoleFactory( + group__acronym="mars", + name_id="chair", + person=PersonFactory(user__username="marschairman"), + ) - url = urlreverse(ietf.community.views.manage_list, kwargs={ "acronym": draft.group.acronym }) + url = urlreverse( + ietf.community.views.manage_list, kwargs={"acronym": draft.group.acronym} + ) setup_default_community_list_for_group(draft.group) login_testing_unauthorized(self, "marschairman", url) @@ -197,95 +332,132 @@ def test_manage_group_list(self): self.assertEqual(r.status_code, 200) # Verify GET also works with non-WG and RG groups - for gtype in ['area','program']: + for gtype in ["area", "program"]: g = GroupFactory.create(type_id=gtype) # make sure the group's features have been initialized to improve coverage - _ = g.features # pyflakes:ignore + _ = g.features # pyflakes:ignore p = PersonFactory() - g.role_set.create(name_id={'area':'ad','program':'lead'}[gtype],person=p, email=p.email()) - url = urlreverse(ietf.community.views.manage_list, kwargs={ "acronym": g.acronym }) + g.role_set.create( + name_id={"area": "ad", "program": "lead"}[gtype], + person=p, + email=p.email(), + ) + url = urlreverse( + ietf.community.views.manage_list, kwargs={"acronym": g.acronym} + ) setup_default_community_list_for_group(g) - self.client.login(username=p.user.username,password=p.user.username+"+password") + self.client.login( + username=p.user.username, password=p.user.username + "+password" + ) r = self.client.get(url) self.assertEqual(r.status_code, 200) def test_track_untrack_document(self): - PersonFactory(user__username='plain') + person = self.complex_person(user__username="plain") draft = WgDraftFactory() - url = urlreverse(ietf.community.views.track_document, kwargs={ "username": "plain", "name": draft.name }) + url = urlreverse( + ietf.community.views.track_document, + kwargs={"email_or_name": person.email(), "name": draft.name}, + ) login_testing_unauthorized(self, "plain", url) - # track - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - - r = self.client.post(url) - self.assertEqual(r.status_code, 302) - clist = CommunityList.objects.get(user__username="plain") - self.assertEqual(list(clist.added_docs.all()), [draft]) + for id in self.email_or_name_set(person): + url = urlreverse( + ietf.community.views.track_document, + kwargs={"email_or_name": id, "name": draft.name}, + ) - # untrack - url = urlreverse(ietf.community.views.untrack_document, kwargs={ "username": "plain", "name": draft.name }) - r = self.client.get(url) - self.assertEqual(r.status_code, 200) + # track + r = self.client.get(url) + self.assertEqual(r.status_code, 200, msg=f"id='{id}', url='{url}'") + + r = self.client.post(url) + self.assertEqual(r.status_code, 302, msg=f"id='{id}', url='{url}'") + clist = CommunityList.objects.get(person__user__username="plain") + self.assertEqual(list(clist.added_docs.all()), [draft]) + + # untrack + url = urlreverse( + ietf.community.views.untrack_document, + kwargs={"email_or_name": id, "name": draft.name}, + ) + r = self.client.get(url) + self.assertEqual(r.status_code, 200, msg=f"id='{id}', url='{url}'") - r = self.client.post(url) - self.assertEqual(r.status_code, 302) - clist = CommunityList.objects.get(user__username="plain") - self.assertEqual(list(clist.added_docs.all()), []) + r = self.client.post(url) + self.assertEqual(r.status_code, 302, msg=f"id='{id}', url='{url}'") + clist = CommunityList.objects.get(person__user__username="plain") + self.assertEqual(list(clist.added_docs.all()), []) def test_track_untrack_document_through_ajax(self): - PersonFactory(user__username='plain') + person = self.complex_person(user__username="plain") draft = WgDraftFactory() - url = urlreverse(ietf.community.views.track_document, kwargs={ "username": "plain", "name": draft.name }) + url = urlreverse( + ietf.community.views.track_document, + kwargs={"email_or_name": person.email(), "name": draft.name}, + ) login_testing_unauthorized(self, "plain", url) - # track - r = self.client.post(url, HTTP_X_REQUESTED_WITH='XMLHttpRequest') - self.assertEqual(r.status_code, 200) - self.assertEqual(r.json()["success"], True) - clist = CommunityList.objects.get(user__username="plain") - self.assertEqual(list(clist.added_docs.all()), [draft]) - - # untrack - url = urlreverse(ietf.community.views.untrack_document, kwargs={ "username": "plain", "name": draft.name }) - r = self.client.post(url, HTTP_X_REQUESTED_WITH='XMLHttpRequest') - self.assertEqual(r.status_code, 200) - self.assertEqual(r.json()["success"], True) - clist = CommunityList.objects.get(user__username="plain") - self.assertEqual(list(clist.added_docs.all()), []) + for id in self.email_or_name_set(person): + url = urlreverse( + ietf.community.views.track_document, + kwargs={"email_or_name": id, "name": draft.name}, + ) + + # track + r = self.client.post(url, HTTP_X_REQUESTED_WITH="XMLHttpRequest") + self.assertEqual(r.status_code, 200, msg=f"id='{id}', url='{url}'") + self.assertEqual(r.json()["success"], True) + clist = CommunityList.objects.get(person__user__username="plain") + self.assertEqual(list(clist.added_docs.all()), [draft]) + + # untrack + url = urlreverse( + ietf.community.views.untrack_document, + kwargs={"email_or_name": id, "name": draft.name}, + ) + r = self.client.post(url, HTTP_X_REQUESTED_WITH="XMLHttpRequest") + self.assertEqual(r.status_code, 200, msg=f"id='{id}', url='{url}'") + self.assertEqual(r.json()["success"], True) + clist = CommunityList.objects.get(person__user__username="plain") + self.assertEqual(list(clist.added_docs.all()), []) def test_csv(self): - PersonFactory(user__username='plain') + person = self.complex_person(user__username="plain") draft = WgDraftFactory() - url = urlreverse(ietf.community.views.export_to_csv, kwargs={ "username": "plain" }) + for id in self.email_or_name_set(person): + url = urlreverse( + ietf.community.views.export_to_csv, kwargs={"email_or_name": id} + ) - # without list - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - - # with list - clist = CommunityList.objects.create(user=User.objects.get(username="plain")) - if not draft in clist.added_docs.all(): - clist.added_docs.add(draft) - SearchRule.objects.create( - community_list=clist, - rule_type="name_contains", - state=State.objects.get(type="draft", slug="active"), - text="test", - ) - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - # this is a simple-minded test, we don't actually check the fields - self.assertContains(r, draft.name) + # without list + r = self.client.get(url) + self.assertEqual(r.status_code, 200, msg=f"id='{id}', url='{url}'") + + # with list + clist = CommunityList.objects.create(person=person) + if draft not in clist.added_docs.all(): + clist.added_docs.add(draft) + SearchRule.objects.create( + community_list=clist, + rule_type="name_contains", + state=State.objects.get(type="draft", slug="active"), + text="test", + ) + r = self.client.get(url) + self.assertEqual(r.status_code, 200, msg=f"id='{id}', url='{url}'") + # this is a simple-minded test, we don't actually check the fields + self.assertContains(r, draft.name) def test_csv_for_group(self): draft = WgDraftFactory() - url = urlreverse(ietf.community.views.export_to_csv, kwargs={ "acronym": draft.group.acronym }) + url = urlreverse( + ietf.community.views.export_to_csv, kwargs={"acronym": draft.group.acronym} + ) setup_default_community_list_for_group(draft.group) @@ -294,60 +466,85 @@ def test_csv_for_group(self): self.assertEqual(r.status_code, 200) def test_feed(self): - PersonFactory(user__username='plain') + person = self.complex_person(user__username="plain") draft = WgDraftFactory() - url = urlreverse(ietf.community.views.feed, kwargs={ "username": "plain" }) - - # without list - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - - # with list - clist = CommunityList.objects.create(user=User.objects.get(username="plain")) - if not draft in clist.added_docs.all(): - clist.added_docs.add(draft) - SearchRule.objects.create( - community_list=clist, - rule_type="name_contains", - state=State.objects.get(type="draft", slug="active"), - text="test", - ) - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - self.assertContains(r, draft.name) + for id in self.email_or_name_set(person): + url = urlreverse(ietf.community.views.feed, kwargs={"email_or_name": id}) - # only significant - r = self.client.get(url + "?significant=1") - self.assertEqual(r.status_code, 200) - self.assertNotContains(r, '') + # without list + r = self.client.get(url) + self.assertEqual(r.status_code, 200, msg=f"id='{id}', url='{url}'") + + # with list + clist = CommunityList.objects.create(person=person) + if draft not in clist.added_docs.all(): + clist.added_docs.add(draft) + SearchRule.objects.create( + community_list=clist, + rule_type="name_contains", + state=State.objects.get(type="draft", slug="active"), + text="test", + ) + r = self.client.get(url) + self.assertEqual(r.status_code, 200, msg=f"id='{id}', url='{url}'") + self.assertContains(r, draft.name) + + # test atom xml + xml = etree.fromstring(r.content) + ns = {"atom": "http://www.w3.org/2005/Atom"} + updated = xml.xpath("/atom:feed/atom:updated", namespaces=ns)[0].text + entries = xml.xpath("/atom:feed/atom:entry", namespaces=ns) + self.assertIn("+00:00", updated) # RFC 3339 compatible UTC TZ + for entry in entries: + updated = entry.xpath("atom:updated", namespaces=ns)[0].text + published = entry.xpath("atom:published", namespaces=ns)[0].text + entry_id = entry.xpath("atom:id", namespaces=ns)[0].text + self.assertIn("+00:00", updated) + self.assertIn("+00:00", published) + self.assertIn( + "urn:datatracker-ietf-org:event:", entry_id + ) # atom:entry:id must be a valid URN + + # only significant + r = self.client.get(url + "?significant=1") + self.assertEqual(r.status_code, 200, msg=f"id='{id}', url='{url}'") + self.assertNotContains(r, "") def test_feed_for_group(self): draft = WgDraftFactory() - url = urlreverse(ietf.community.views.feed, kwargs={ "acronym": draft.group.acronym }) + url = urlreverse( + ietf.community.views.feed, kwargs={"acronym": draft.group.acronym} + ) setup_default_community_list_for_group(draft.group) # test GET, rest is tested with personal list r = self.client.get(url) self.assertEqual(r.status_code, 200) - + def test_subscription(self): - PersonFactory(user__username='plain') + person = self.complex_person(user__username="plain") draft = WgDraftFactory() - url = urlreverse(ietf.community.views.subscription, kwargs={ "username": "plain" }) - + url = urlreverse( + ietf.community.views.subscription, kwargs={"email_or_name": person.email()} + ) login_testing_unauthorized(self, "plain", url) - # subscription without list - r = self.client.get(url) - self.assertEqual(r.status_code, 404) + for id in self.email_or_name_set(person): + url = urlreverse( + ietf.community.views.subscription, kwargs={"email_or_name": id} + ) + + # subscription without list + r = self.client.get(url) + self.assertEqual(r.status_code, 404, msg=f"id='{id}', url='{url}'") # subscription with list - clist = CommunityList.objects.create(user=User.objects.get(username="plain")) - if not draft in clist.added_docs.all(): + clist = CommunityList.objects.create(person=person) + if draft not in clist.added_docs.all(): clist.added_docs.add(draft) SearchRule.objects.create( community_list=clist, @@ -355,28 +552,51 @@ def test_subscription(self): state=State.objects.get(type="draft", slug="active"), text="test", ) - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - # subscribe - email = Email.objects.filter(person__user__username="plain").first() - r = self.client.post(url, { "email": email.pk, "notify_on": "significant", "action": "subscribe" }) - self.assertEqual(r.status_code, 302) + for email in Email.objects.filter(person=person): + url = urlreverse( + ietf.community.views.subscription, kwargs={"email_or_name": email} + ) - subscription = EmailSubscription.objects.filter(community_list=clist, email=email, notify_on="significant").first() - - self.assertTrue(subscription) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) - # delete subscription - r = self.client.post(url, { "subscription_id": subscription.pk, "action": "unsubscribe" }) - self.assertEqual(r.status_code, 302) - self.assertEqual(EmailSubscription.objects.filter(community_list=clist, email=email, notify_on="significant").count(), 0) + # subscribe + r = self.client.post( + url, + {"email": email.pk, "notify_on": "significant", "action": "subscribe"}, + ) + self.assertEqual(r.status_code, 302) + + subscription = EmailSubscription.objects.filter( + community_list=clist, email=email, notify_on="significant" + ).first() + + self.assertTrue(subscription) + + # delete subscription + r = self.client.post( + url, {"subscription_id": subscription.pk, "action": "unsubscribe"} + ) + self.assertEqual(r.status_code, 302) + self.assertEqual( + EmailSubscription.objects.filter( + community_list=clist, email=email, notify_on="significant" + ).count(), + 0, + ) def test_subscription_for_group(self): - draft = WgDraftFactory(group__acronym='mars') - RoleFactory(group__acronym='mars',name_id='chair',person=PersonFactory(user__username='marschairman')) + draft = WgDraftFactory(group__acronym="mars") + RoleFactory( + group__acronym="mars", + name_id="chair", + person=PersonFactory(user__username="marschairman"), + ) - url = urlreverse(ietf.community.views.subscription, kwargs={ "acronym": draft.group.acronym }) + url = urlreverse( + ietf.community.views.subscription, kwargs={"acronym": draft.group.acronym} + ) setup_default_community_list_for_group(draft.group) @@ -385,27 +605,136 @@ def test_subscription_for_group(self): # test GET, rest is tested with personal list r = self.client.get(url) self.assertEqual(r.status_code, 200) - - def test_notification(self): - PersonFactory(user__username='plain') + + @mock.patch("ietf.community.signals.notify_of_event") + def test_notification_signal_receiver(self, mock_notify_of_event): + """Saving a newly created DocEvent should notify subscribers + + This implicitly tests that notify_of_event_receiver is hooked up to the post_save signal. + """ + # Arbitrary model that's not a DocEvent + person = PersonFactory.build() # builds but does not save... + mock_notify_of_event.reset_mock() # clear any calls that resulted from the factories + person.save() + self.assertFalse(mock_notify_of_event.called) + + # build a DocEvent that is not yet persisted + doc = DocumentFactory() + event = DocEventFactory.build(by=person, doc=doc) # builds but does not save... + mock_notify_of_event.reset_mock() # clear any calls that resulted from the factories + event.save() + self.assertEqual( + mock_notify_of_event.call_count, + 1, + "notify_task should be run on creation of DocEvent", + ) + self.assertEqual(mock_notify_of_event.call_args, mock.call(event)) + + # save the existing DocEvent and see that no notification is sent + mock_notify_of_event.reset_mock() + event.save() + self.assertFalse( + mock_notify_of_event.called, + "notify_task should not be run save of on existing DocEvent", + ) + + # Mock out the on_commit call so we can tell whether the task was actually queued + @mock.patch("ietf.submit.views.transaction.on_commit", side_effect=lambda x: x()) + @mock.patch("ietf.community.signals.notify_event_to_subscribers_task") + def test_notify_of_event(self, mock_notify_task, mock_on_commit): + """The community notification task should be called as intended""" + person = PersonFactory() # builds but does not save... + doc = DocumentFactory() + event = DocEventFactory(by=person, doc=doc) + # be careful overriding SERVER_MODE - we do it here because the method + # under test does not make this call when in "test" mode + with override_settings(SERVER_MODE="not-test"): + notify_of_event(event) + self.assertTrue( + mock_notify_task.delay.called, + "notify_task should run for a DocEvent on a draft", + ) + mock_notify_task.reset_mock() + + event.skip_community_list_notification = True + # be careful overriding SERVER_MODE - we do it here because the method + # under test does not make this call when in "test" mode + with override_settings(SERVER_MODE="not-test"): + notify_of_event(event) + self.assertFalse( + mock_notify_task.delay.called, + "notify_task should not run when skip_community_list_notification is set", + ) + + event = DocEventFactory.build(by=person, doc=DocumentFactory(type_id="rfc")) + # be careful overriding SERVER_MODE - we do it here because the method + # under test does not make this call when in "test" mode + with override_settings(SERVER_MODE="not-test"): + notify_of_event(event) + self.assertFalse( + mock_notify_task.delay.called, + "notify_task should not run on a document with type 'rfc'", + ) + + @mock.patch("ietf.utils.mail.send_mail_text") + def test_notify_event_to_subscribers(self, mock_send_mail_text): + person = PersonFactory(user__username="plain") draft = WgDraftFactory() - clist = CommunityList.objects.create(user=User.objects.get(username="plain")) - if not draft in clist.added_docs.all(): + clist = CommunityList.objects.create(person=person) + if draft not in clist.added_docs.all(): clist.added_docs.add(draft) - EmailSubscription.objects.create(community_list=clist, email=Email.objects.filter(person__user__username="plain").first(), notify_on="significant") + sub_to_significant = EmailSubscription.objects.create( + community_list=clist, + email=Email.objects.filter(person__user__username="plain").first(), + notify_on="significant", + ) + sub_to_all = EmailSubscription.objects.create( + community_list=clist, + email=Email.objects.filter(person__user__username="plain").first(), + notify_on="all", + ) - mailbox_before = len(outbox) active_state = State.objects.get(type="draft", slug="active") system = Person.objects.get(name="(System)") - add_state_change_event(draft, system, None, active_state) - self.assertEqual(len(outbox), mailbox_before) + event = add_state_change_event(draft, system, None, active_state) + notify_event_to_subscribers(event) + self.assertEqual(mock_send_mail_text.call_count, 1) + address = mock_send_mail_text.call_args[0][1] + subject = mock_send_mail_text.call_args[0][3] + content = mock_send_mail_text.call_args[0][4] + self.assertEqual(address, sub_to_all.email.address) + self.assertIn(draft.name, subject) + self.assertIn(clist.long_name(), content) - mailbox_before = len(outbox) rfc_state = State.objects.get(type="draft", slug="rfc") - add_state_change_event(draft, system, active_state, rfc_state) - self.assertEqual(len(outbox), mailbox_before + 1) - self.assertTrue(draft.name in outbox[-1]["Subject"]) - - \ No newline at end of file + event = add_state_change_event(draft, system, active_state, rfc_state) + mock_send_mail_text.reset_mock() + notify_event_to_subscribers(event) + self.assertEqual(mock_send_mail_text.call_count, 2) + addresses = [ + call_args[0][1] for call_args in mock_send_mail_text.call_args_list + ] + subjects = {call_args[0][3] for call_args in mock_send_mail_text.call_args_list} + contents = {call_args[0][4] for call_args in mock_send_mail_text.call_args_list} + self.assertCountEqual( + addresses, + [sub_to_significant.email.address, sub_to_all.email.address], + ) + self.assertEqual(len(subjects), 1) + self.assertIn(draft.name, subjects.pop()) + self.assertEqual(len(contents), 1) + self.assertIn(clist.long_name(), contents.pop()) + + @mock.patch("ietf.community.utils.notify_event_to_subscribers") + def test_notify_event_to_subscribers_task(self, mock_notify): + d = DocEventFactory() + notify_event_to_subscribers_task(event_id=d.pk) + self.assertEqual(mock_notify.call_count, 1) + self.assertEqual(mock_notify.call_args, mock.call(d)) + mock_notify.reset_mock() + + d.delete() + notify_event_to_subscribers_task(event_id=d.pk) + self.assertFalse(mock_notify.called) diff --git a/ietf/community/urls.py b/ietf/community/urls.py index f80547ffad..3ab132f2dc 100644 --- a/ietf/community/urls.py +++ b/ietf/community/urls.py @@ -4,11 +4,11 @@ from ietf.utils.urls import url urlpatterns = [ - url(r'^personal/(?P[^/]+)/$', views.view_list), - url(r'^personal/(?P[^/]+)/manage/$', views.manage_list), - url(r'^personal/(?P[^/]+)/trackdocument/(?P[^/]+)/$', views.track_document), - url(r'^personal/(?P[^/]+)/untrackdocument/(?P[^/]+)/$', views.untrack_document), - url(r'^personal/(?P[^/]+)/csv/$', views.export_to_csv), - url(r'^personal/(?P[^/]+)/feed/$', views.feed), - url(r'^personal/(?P[^/]+)/subscription/$', views.subscription), + url(r'^personal/(?P[^/]+)/$', views.view_list), + url(r'^personal/(?P[^/]+)/manage/$', views.manage_list), + url(r'^personal/(?P[^/]+)/trackdocument/(?P[^/]+)/$', views.track_document), + url(r'^personal/(?P[^/]+)/untrackdocument/(?P[^/]+)/$', views.untrack_document), + url(r'^personal/(?P[^/]+)/csv/$', views.export_to_csv), + url(r'^personal/(?P[^/]+)/feed/$', views.feed), + url(r'^personal/(?P[^/]+)/subscription/$', views.subscription), ] diff --git a/ietf/community/utils.py b/ietf/community/utils.py index 06da50011c..b6137095ef 100644 --- a/ietf/community/utils.py +++ b/ietf/community/utils.py @@ -1,4 +1,4 @@ -# Copyright The IETF Trust 2016-2020, All Rights Reserved +# Copyright The IETF Trust 2016-2023, All Rights Reserved # -*- coding: utf-8 -*- @@ -11,11 +11,9 @@ from ietf.community.models import CommunityList, EmailSubscription, SearchRule from ietf.doc.models import Document, State -from ietf.group.models import Role, Group +from ietf.group.models import Role from ietf.person.models import Person from ietf.ietfauth.utils import has_role -from django.contrib.auth.models import User -from django.shortcuts import get_object_or_404 from ietf.utils.mail import send_mail @@ -29,24 +27,12 @@ def states_of_significant_change(): Q(type="draft", slug__in=['rfc', 'dead']) ) -def lookup_community_list(username=None, acronym=None): - assert username or acronym - - if acronym: - group = get_object_or_404(Group, acronym=acronym) - clist = CommunityList.objects.filter(group=group).first() or CommunityList(group=group) - else: - user = get_object_or_404(User, username=username) - clist = CommunityList.objects.filter(user=user).first() or CommunityList(user=user) - - return clist - def can_manage_community_list(user, clist): if not user or not user.is_authenticated: return False - if clist.user: - return user == clist.user + if clist.person: + return user == clist.person.user elif clist.group: if has_role(user, 'Secretariat'): return True @@ -60,7 +46,7 @@ def reset_name_contains_index_for_rule(rule): if not rule.rule_type == "name_contains": return - rule.name_contains_index.set(Document.objects.filter(docalias__name__regex=rule.text)) + rule.name_contains_index.set(Document.objects.filter(name__regex=rule.text)) def update_name_contains_indexes_with_new_doc(doc): for r in SearchRule.objects.filter(rule_type="name_contains"): @@ -71,71 +57,113 @@ def update_name_contains_indexes_with_new_doc(doc): if re.search(r.text, doc.name) and not doc in r.name_contains_index.all(): r.name_contains_index.add(doc) + def docs_matching_community_list_rule(rule): docs = Document.objects.all() + + if rule.rule_type.endswith("_rfc"): + docs = docs.filter(type_id="rfc") # rule.state is ignored for RFCs + else: + docs = docs.filter(type_id="draft", states=rule.state) + if rule.rule_type in ['group', 'area', 'group_rfc', 'area_rfc']: - return docs.filter(Q(group=rule.group_id) | Q(group__parent=rule.group_id), states=rule.state) + return docs.filter(Q(group=rule.group_id) | Q(group__parent=rule.group_id)) elif rule.rule_type in ['group_exp']: - return docs.filter(group=rule.group_id, states=rule.state) + return docs.filter(group=rule.group_id) elif rule.rule_type.startswith("state_"): - return docs.filter(states=rule.state) - elif rule.rule_type in ["author", "author_rfc"]: - return docs.filter(states=rule.state, documentauthor__person=rule.person) + return docs + elif rule.rule_type == "author": + return docs.filter(documentauthor__person=rule.person) + elif rule.rule_type == "author_rfc": + return docs.filter(Q(rfcauthor__person=rule.person)|Q(rfcauthor__isnull=True,documentauthor__person=rule.person)) elif rule.rule_type == "ad": - return docs.filter(states=rule.state, ad=rule.person) + return docs.filter(ad=rule.person) elif rule.rule_type == "shepherd": - return docs.filter(states=rule.state, shepherd__person=rule.person) + return docs.filter(shepherd__person=rule.person) elif rule.rule_type == "name_contains": - return docs.filter(states=rule.state, searchrule=rule) + return docs.filter(searchrule=rule) raise NotImplementedError -def community_list_rules_matching_doc(doc): - states = list(doc.states.values_list("pk", flat=True)) +def community_list_rules_matching_doc(doc): rules = SearchRule.objects.none() + if doc.type_id not in ["draft", "rfc"]: + return rules # none + states = list(doc.states.values_list("pk", flat=True)) + # group and area rules if doc.group_id: groups = [doc.group_id] if doc.group.parent_id: groups.append(doc.group.parent_id) + rules_to_add = SearchRule.objects.filter(group__in=groups) + if doc.type_id == "rfc": + rules_to_add = rules_to_add.filter(rule_type__in=["group_rfc", "area_rfc"]) + else: + rules_to_add = rules_to_add.filter( + rule_type__in=["group", "area", "group_exp"], + state__in=states, + ) + rules |= rules_to_add + + # state rules (only relevant for I-Ds) + if doc.type_id == "draft": rules |= SearchRule.objects.filter( - rule_type__in=['group', 'area', 'group_rfc', 'area_rfc', 'group_exp'], + rule_type__in=[ + "state_iab", + "state_iana", + "state_iesg", + "state_irtf", + "state_ise", + "state_rfceditor", + "state_ietf", + ], state__in=states, - group__in=groups ) - rules |= SearchRule.objects.filter( - rule_type__in=['state_iab', 'state_iana', 'state_iesg', 'state_irtf', 'state_ise', 'state_rfceditor', 'state_ietf'], - state__in=states, - ) - - rules |= SearchRule.objects.filter( - rule_type__in=["author", "author_rfc"], - state__in=states, - person__in=list(Person.objects.filter(documentauthor__document=doc)), - ) - - if doc.ad_id: + # author rules + if doc.type_id == "rfc": + has_rfcauthors = doc.rfcauthor_set.exists() + rules |= SearchRule.objects.filter( + rule_type="author_rfc", + person__in=list( + Person.objects.filter( + Q(rfcauthor__document=doc) + if has_rfcauthors + else Q(documentauthor__document=doc) + ) + ), + ) + else: rules |= SearchRule.objects.filter( - rule_type="ad", + rule_type="author", state__in=states, - person=doc.ad_id, + person__in=list(Person.objects.filter(documentauthor__document=doc)), ) - if doc.shepherd_id: + # Other draft-only rules rules + if doc.type_id == "draft": + if doc.ad_id: + rules |= SearchRule.objects.filter( + rule_type="ad", + state__in=states, + person=doc.ad_id, + ) + + if doc.shepherd_id: + rules |= SearchRule.objects.filter( + rule_type="shepherd", + state__in=states, + person__email=doc.shepherd_id, + ) + rules |= SearchRule.objects.filter( - rule_type="shepherd", + rule_type="name_contains", state__in=states, - person__email=doc.shepherd_id, + name_contains_index=doc, # search our materialized index to avoid full scan ) - rules |= SearchRule.objects.filter( - rule_type="name_contains", - state__in=states, - name_contains_index=doc, # search our materialized index to avoid full scan - ) - return rules @@ -146,7 +174,11 @@ def docs_tracked_by_community_list(clist): # in theory, we could use an OR query, but databases seem to have # trouble with OR queries and complicated joins so do the OR'ing # manually - doc_ids = set(clist.added_docs.values_list("pk", flat=True)) + doc_ids = set() + for doc in clist.added_docs.all(): + doc_ids.add(doc.pk) + doc_ids.update(rfc.pk for rfc in doc.related_that_doc("became_rfc")) + for rule in clist.searchrule_set.all(): doc_ids = doc_ids | set(docs_matching_community_list_rule(rule).values_list("pk", flat=True)) diff --git a/ietf/community/views.py b/ietf/community/views.py index b0646424a3..08b1c24fe5 100644 --- a/ietf/community/views.py +++ b/ietf/community/views.py @@ -1,4 +1,4 @@ -# Copyright The IETF Trust 2012-2020, All Rights Reserved +# Copyright The IETF Trust 2012-2023, All Rights Reserved # -*- coding: utf-8 -*- @@ -13,60 +13,119 @@ from django.utils import timezone from django.utils.html import strip_tags -import debug # pyflakes:ignore - -from ietf.community.models import SearchRule, EmailSubscription -from ietf.community.forms import SearchRuleTypeForm, SearchRuleForm, AddDocumentsForm, SubscriptionForm -from ietf.community.utils import lookup_community_list, can_manage_community_list -from ietf.community.utils import docs_tracked_by_community_list, docs_matching_community_list_rule -from ietf.community.utils import states_of_significant_change, reset_name_contains_index_for_rule +import debug # pyflakes:ignore + +from ietf.community.models import CommunityList, EmailSubscription, SearchRule +from ietf.community.forms import ( + SearchRuleTypeForm, + SearchRuleForm, + AddDocumentsForm, + SubscriptionForm, +) +from ietf.community.utils import can_manage_community_list +from ietf.community.utils import ( + docs_tracked_by_community_list, + docs_matching_community_list_rule, +) +from ietf.community.utils import ( + states_of_significant_change, + reset_name_contains_index_for_rule, +) +from ietf.group.models import Group from ietf.doc.models import DocEvent, Document from ietf.doc.utils_search import prepare_document_table +from ietf.person.utils import lookup_persons +from ietf.utils.decorators import ignore_view_kwargs +from ietf.utils.http import is_ajax from ietf.utils.response import permission_denied -def view_list(request, username=None): - clist = lookup_community_list(username) +def lookup_community_list(request, email_or_name=None, acronym=None): + """Finds a CommunityList for a person or group + + Instantiates an unsaved CommunityList if one is not found. + + If the person or group cannot be found and uniquely identified, raises an Http404 exception + """ + assert email_or_name or acronym + + if acronym: + group = get_object_or_404(Group, acronym=acronym) + clist = CommunityList.objects.filter(group=group).first() or CommunityList( + group=group + ) + else: + persons = lookup_persons(email_or_name) + if len(persons) > 1: + if hasattr(request.user, "person") and request.user.person in persons: + person = request.user.person + else: + raise Http404( + f"Unable to identify the CommunityList for {email_or_name}" + ) + else: + person = persons[0] + clist = CommunityList.objects.filter(person=person).first() or CommunityList( + person=person + ) + return clist + + +def view_list(request, email_or_name=None): + clist = lookup_community_list(request, email_or_name) # may raise Http404 docs = docs_tracked_by_community_list(clist) docs, meta = prepare_document_table(request, docs, request.GET) - subscribed = request.user.is_authenticated and EmailSubscription.objects.filter(community_list=clist, email__person__user=request.user) + subscribed = request.user.is_authenticated and ( + EmailSubscription.objects.none() + if clist.pk is None + else EmailSubscription.objects.filter( + community_list=clist, email__person__user=request.user + ) + ) + + return render( + request, + "community/view_list.html", + { + "clist": clist, + "docs": docs, + "meta": meta, + "can_manage_list": can_manage_community_list(request.user, clist), + "subscribed": subscribed, + "email_or_name": email_or_name, + }, + ) - return render(request, 'community/view_list.html', { - 'clist': clist, - 'docs': docs, - 'meta': meta, - 'can_manage_list': can_manage_community_list(request.user, clist), - 'subscribed': subscribed, - }) @login_required -def manage_list(request, username=None, acronym=None, group_type=None): +@ignore_view_kwargs("group_type") +def manage_list(request, email_or_name=None, acronym=None): # we need to be a bit careful because clist may not exist in the # database so we can't call related stuff on it yet - clist = lookup_community_list(username, acronym) + clist = lookup_community_list(request, email_or_name, acronym) # may raise Http404 if not can_manage_community_list(request.user, clist): permission_denied(request, "You do not have permission to access this view") - action = request.POST.get('action') + action = request.POST.get("action") - if request.method == 'POST' and action == 'add_documents': + if request.method == "POST" and action == "add_documents": add_doc_form = AddDocumentsForm(request.POST) if add_doc_form.is_valid(): if clist.pk is None: clist.save() - for d in add_doc_form.cleaned_data['documents']: - if not d in clist.added_docs.all(): + for d in add_doc_form.cleaned_data["documents"]: + if d not in clist.added_docs.all(): clist.added_docs.add(d) return HttpResponseRedirect("") else: add_doc_form = AddDocumentsForm() - if request.method == 'POST' and action == 'remove_document': - document_id = request.POST.get('document') + if request.method == "POST" and action == "remove_document": + document_id = request.POST.get("document") if clist.pk is not None and document_id: document = get_object_or_404(clist.added_docs, id=document_id) clist.added_docs.remove(document) @@ -74,30 +133,29 @@ def manage_list(request, username=None, acronym=None, group_type=None): return HttpResponseRedirect("") rule_form = None - if request.method == 'POST' and action == 'add_rule': + if request.method == "POST" and action == "add_rule": rule_type_form = SearchRuleTypeForm(request.POST) if rule_type_form.is_valid(): - rule_type = rule_type_form.cleaned_data['rule_type'] - - if rule_type: - rule_form = SearchRuleForm(clist, rule_type, request.POST) - if rule_form.is_valid(): - if clist.pk is None: - clist.save() - - rule = rule_form.save(commit=False) - rule.community_list = clist - rule.rule_type = rule_type - rule.save() - if rule.rule_type == "name_contains": - reset_name_contains_index_for_rule(rule) + rule_type = rule_type_form.cleaned_data["rule_type"] + if rule_type: + rule_form = SearchRuleForm(clist, rule_type, request.POST) + if rule_form.is_valid(): + if clist.pk is None: + clist.save() + + rule = rule_form.save(commit=False) + rule.community_list = clist + rule.rule_type = rule_type + rule.save() + if rule.rule_type == "name_contains": + reset_name_contains_index_for_rule(rule) return HttpResponseRedirect("") else: rule_type_form = SearchRuleTypeForm() - if request.method == 'POST' and action == 'remove_rule': - rule_pk = request.POST.get('rule') + if request.method == "POST" and action == "remove_rule": + rule_pk = request.POST.get("rule") if clist.pk is not None and rule_pk: rule = get_object_or_404(SearchRule, pk=rule_pk, community_list=clist) rule.delete() @@ -108,53 +166,74 @@ def manage_list(request, username=None, acronym=None, group_type=None): for r in rules: r.matching_documents_count = docs_matching_community_list_rule(r).count() - empty_rule_forms = { rule_type: SearchRuleForm(clist, rule_type) for rule_type, _ in SearchRule.RULE_TYPES } + empty_rule_forms = { + rule_type: SearchRuleForm(clist, rule_type) + for rule_type, _ in SearchRule.RULE_TYPES + } total_count = docs_tracked_by_community_list(clist).count() - all_forms = [f for f in [rule_type_form, rule_form, add_doc_form, *empty_rule_forms.values()] - if f is not None] - return render(request, 'community/manage_list.html', { - 'clist': clist, - 'rules': rules, - 'individually_added': clist.added_docs.all() if clist.pk is not None else [], - 'rule_type_form': rule_type_form, - 'rule_form': rule_form, - 'empty_rule_forms': empty_rule_forms, - 'total_count': total_count, - 'add_doc_form': add_doc_form, - 'all_forms': all_forms, - }) + all_forms = [ + f + for f in [rule_type_form, rule_form, add_doc_form, *empty_rule_forms.values()] + if f is not None + ] + return render( + request, + "community/manage_list.html", + { + "clist": clist, + "rules": rules, + "individually_added": ( + clist.added_docs.all() if clist.pk is not None else [] + ), + "rule_type_form": rule_type_form, + "rule_form": rule_form, + "empty_rule_forms": empty_rule_forms, + "total_count": total_count, + "add_doc_form": add_doc_form, + "all_forms": all_forms, + }, + ) @login_required -def track_document(request, name, username=None, acronym=None): - doc = get_object_or_404(Document, docalias__name=name) +def track_document(request, name, email_or_name=None, acronym=None): + doc = get_object_or_404(Document, name=name) if request.method == "POST": - clist = lookup_community_list(username, acronym) + clist = lookup_community_list( + request, email_or_name, acronym + ) # may raise Http404 if not can_manage_community_list(request.user, clist): permission_denied(request, "You do not have permission to access this view") if clist.pk is None: clist.save() - if not doc in clist.added_docs.all(): + if doc not in clist.added_docs.all(): clist.added_docs.add(doc) - if request.is_ajax(): - return HttpResponse(json.dumps({ 'success': True }), content_type='application/json') + if is_ajax(request): + return HttpResponse( + json.dumps({"success": True}), content_type="application/json" + ) else: return HttpResponseRedirect(clist.get_absolute_url()) - return render(request, "community/track_document.html", { - "name": doc.name, - }) + return render( + request, + "community/track_document.html", + { + "name": doc.name, + }, + ) + @login_required -def untrack_document(request, name, username=None, acronym=None): - doc = get_object_or_404(Document, docalias__name=name) - clist = lookup_community_list(username, acronym) +def untrack_document(request, name, email_or_name=None, acronym=None): + doc = get_object_or_404(Document, name=name) + clist = lookup_community_list(request, email_or_name, acronym) # may raise Http404 if not can_manage_community_list(request.user, clist): permission_denied(request, "You do not have permission to access this view") @@ -162,29 +241,35 @@ def untrack_document(request, name, username=None, acronym=None): if clist.pk is not None: clist.added_docs.remove(doc) - if request.is_ajax(): - return HttpResponse(json.dumps({ 'success': True }), content_type='application/json') + if is_ajax(request): + return HttpResponse( + json.dumps({"success": True}), content_type="application/json" + ) else: return HttpResponseRedirect(clist.get_absolute_url()) - return render(request, "community/untrack_document.html", { - "name": doc.name, - }) - + return render( + request, + "community/untrack_document.html", + { + "name": doc.name, + }, + ) -def export_to_csv(request, username=None, acronym=None, group_type=None): - clist = lookup_community_list(username, acronym) - response = HttpResponse(content_type='text/csv') +@ignore_view_kwargs("group_type") +def export_to_csv(request, email_or_name=None, acronym=None): + clist = lookup_community_list(request, email_or_name, acronym) # may raise Http404 + response = HttpResponse(content_type="text/csv") if clist.group: filename = "%s-draft-list.csv" % clist.group.acronym else: filename = "draft-list.csv" - response['Content-Disposition'] = 'attachment; filename=%s' % filename + response["Content-Disposition"] = "attachment; filename=%s" % filename - writer = csv.writer(response, dialect=csv.excel, delimiter=str(',')) + writer = csv.writer(response, dialect=csv.excel, delimiter=str(",")) header = [ "Name", @@ -197,12 +282,12 @@ def export_to_csv(request, username=None, acronym=None, group_type=None): ] writer.writerow(header) - docs = docs_tracked_by_community_list(clist).select_related('type', 'group', 'ad') + docs = docs_tracked_by_community_list(clist).select_related("type", "group", "ad") for doc in docs.prefetch_related("states", "tags"): row = [] row.append(doc.name) row.append(doc.title) - e = doc.latest_event(type='new_revision') + e = doc.latest_event(type="new_revision") row.append(e.time.strftime("%Y-%m-%d") if e else "") row.append(strip_tags(doc.friendly_state())) row.append(doc.group.acronym if doc.group else "") @@ -213,53 +298,73 @@ def export_to_csv(request, username=None, acronym=None, group_type=None): return response -def feed(request, username=None, acronym=None, group_type=None): - clist = lookup_community_list(username, acronym) - significant = request.GET.get('significant', '') == '1' +@ignore_view_kwargs("group_type") +def feed(request, email_or_name=None, acronym=None): + clist = lookup_community_list(request, email_or_name, acronym) # may raise Http404 + significant = request.GET.get("significant", "") == "1" - documents = docs_tracked_by_community_list(clist).values_list('pk', flat=True) - since = timezone.now() - datetime.timedelta(days=14) + documents = docs_tracked_by_community_list(clist).values_list("pk", flat=True) + updated = timezone.now() + since = updated - datetime.timedelta(days=14) - events = DocEvent.objects.filter( - doc__id__in=documents, - time__gte=since, - ).distinct().order_by('-time', '-id').select_related("doc") + events = ( + DocEvent.objects.filter( + doc__id__in=documents, + time__gte=since, + ) + .distinct() + .order_by("-time", "-id") + .select_related("doc") + ) if significant: - events = events.filter(type="changed_state", statedocevent__state__in=list(states_of_significant_change())) + events = events.filter( + type="changed_state", + statedocevent__state__in=list(states_of_significant_change()), + ) host = request.get_host() - feed_url = 'https://%s%s' % (host, request.get_full_path()) + feed_url = "https://%s%s" % (host, request.get_full_path()) feed_id = uuid.uuid5(uuid.NAMESPACE_URL, str(feed_url)) - title = '%s RSS Feed' % clist.long_name() + title = "%s RSS Feed" % clist.long_name() if significant: - subtitle = 'Significant document changes' + subtitle = "Significant document changes" else: - subtitle = 'Document changes' - - return render(request, 'community/atom.xml', { - 'clist': clist, - 'entries': events[:50], - 'title': title, - 'subtitle': subtitle, - 'id': feed_id.urn, - 'updated': timezone.now(), - }, content_type='text/xml') + subtitle = "Document changes" + + return render( + request, + "community/atom.xml", + { + "clist": clist, + "entries": events[:50], + "title": title, + "subtitle": subtitle, + "id": feed_id.urn, + "updated": updated, + }, + content_type="text/xml", + ) @login_required -def subscription(request, username=None, acronym=None, group_type=None): - clist = lookup_community_list(username, acronym) +@ignore_view_kwargs("group_type") +def subscription(request, email_or_name=None, acronym=None): + clist = lookup_community_list(request, email_or_name, acronym) # may raise Http404 if clist.pk is None: raise Http404 - existing_subscriptions = EmailSubscription.objects.filter(community_list=clist, email__person__user=request.user) + person = request.user.person + + existing_subscriptions = EmailSubscription.objects.filter( + community_list=clist, email__person=person + ) - if request.method == 'POST': + if request.method == "POST": action = request.POST.get("action") if action == "subscribe": - form = SubscriptionForm(request.user, clist, request.POST) + form = SubscriptionForm(person, clist, request.POST) if form.is_valid(): subscription = form.save(commit=False) subscription.community_list = clist @@ -268,14 +373,20 @@ def subscription(request, username=None, acronym=None, group_type=None): return HttpResponseRedirect("") elif action == "unsubscribe": - existing_subscriptions.filter(pk=request.POST.get("subscription_id")).delete() + existing_subscriptions.filter( + pk=request.POST.get("subscription_id") + ).delete() return HttpResponseRedirect("") else: - form = SubscriptionForm(request.user, clist) - - return render(request, 'community/subscription.html', { - 'clist': clist, - 'form': form, - 'existing_subscriptions': existing_subscriptions, - }) + form = SubscriptionForm(person, clist) + + return render( + request, + "community/subscription.html", + { + "clist": clist, + "form": form, + "existing_subscriptions": existing_subscriptions, + }, + ) diff --git a/ietf/context_processors.py b/ietf/context_processors.py index 618a896acd..5aaa4ab256 100644 --- a/ietf/context_processors.py +++ b/ietf/context_processors.py @@ -3,7 +3,9 @@ import sys import django from django.conf import settings +from django.utils import timezone from ietf import __version__, __patch__, __release_branch__, __release_hash__ +from opentelemetry.propagate import inject def server_mode(request): return {'server_mode': settings.SERVER_MODE} @@ -44,4 +46,14 @@ def sql_debug(request): def settings_info(request): return { 'settings': settings, - } \ No newline at end of file + } + +def timezone_now(request): + return { + 'timezone_now': timezone.now(), + } + +def traceparent_id(request): + context_extras = {} + inject(context_extras) + return { "otel": context_extras } diff --git a/ietf/cookies/.gitignore b/ietf/cookies/.gitignore deleted file mode 100644 index a74b07aee4..0000000000 --- a/ietf/cookies/.gitignore +++ /dev/null @@ -1 +0,0 @@ -/*.pyc diff --git a/ietf/database-notes/.gitignore b/ietf/database-notes/.gitignore deleted file mode 100644 index c7013ced9d..0000000000 --- a/ietf/database-notes/.gitignore +++ /dev/null @@ -1,2 +0,0 @@ -/*.pyc -/settings_local.py diff --git a/ietf/dbtemplate/fixtures/nomcom_templates.xml b/ietf/dbtemplate/fixtures/nomcom_templates.xml index abf0cb58f6..e7065b84cd 100644 --- a/ietf/dbtemplate/fixtures/nomcom_templates.xml +++ b/ietf/dbtemplate/fixtures/nomcom_templates.xml @@ -1,190 +1,190 @@ - - - - /nomcom/defaults/home.rst - Home page of group - - rst - Home page -========= - -This is the home page of the nomcom group. - - - - /nomcom/defaults/email/inexistent_person.txt - Email sent to chair of nomcom and secretariat when Email and Person are created if some of them don't exist - $email: Newly created email -$fullname: Fullname of the new person -$person_id: Id of the new Person object -$group: Name of the group - plain - Hello, - -A new person with name $fullname and email $email has been created. The new Person object has the following id: '$person_id'. - -Please, check if there is some more action nedeed. - - - - /nomcom/defaults/email/new_nominee.txt - Email sent to nominees when they are nominated - $nominee: Full name of the nominee -$position: Name of the position -$domain: Server domain -$accept_url: Url hash to accept nominations -$decline_url: Url hash to decline nominations - plain - Hi, - -You have been nominated for the position of $position. - -The NomCom would appreciate receiving an indication of whether or not you accept this nomination to stand for consideration as a candidate for this position. - -You can accept the nomination via web going to the following link https://$domain$accept_url or decline the nomination going the following link https://$domain$decline_url - -If you accept, you will need to fill out a questionnaire. You will receive the questionnaire by email. - -Best regards, - - - - - /nomcom/defaults/email/new_nomination.txt - Email sent to nominators and secretariat when the nominators make the nominations - $nominator: Full name of the nominator -$nominator_email: Email of the nominator -$nominee: Full name of the nominee -$nominee_email: Email of the nominee -$position: Nomination position - plain - A new nomination have been received. - -Nominator: $nominator ($nominator_email) -Nominee: $nominee ($nominee_email) -Position: $position - - - - /nomcom/defaults/position/questionnaire.txt - Questionnaire sent to the nomine - $position: Position - plain - Enter here the questionnaire for the position $position: - -Questionnaire - - - - /nomcom/defaults/position/requirements - Position requirements - $position: Position - rst - These are the requirements for the position $position: - -Requirements. - - - - /nomcom/defaults/position/header_questionnaire.txt - Header of the email that contains the questionnaire sent to the nomine - $nominee: Full name of the nomine -$position: Position - plain - Hi $nominee, this is the questionnaire for the position $position: - - - - - - /nomcom/defaults/email/nomination_accept_reminder.txt - Email sent to nominees asking them to accept (or decline) the nominations. - $positions: Nomination positions - plain - Hi, - -You have been nominated for the position of $position. - -The NomCom would appreciate receiving an indication of whether or not you accept this nomination to stand for consideration as a candidate for this position. - -You can accept the nomination via web going to the following link https://$domain$accept_url or decline the nomination going the following link https://$domain$decline_url - -If you accept, you will need to fill out a questionnaire. - -Best regards, - - - - /nomcom/defaults/email/nomination_receipt.txt - Email sent to nominator to get a confirmation mail containing feedback in cleartext - $nominee: Full name of the nominee -$position: Name of the position -$domain: Server domain -$accept_url: Url hash to accept nominations -$decline_url: Url hash to decline nominations - plain - Hi, - -Your nomination of $nominee for the position of -$position has been received and registered. - -The following comments have also been registered: - --------------------------------------------------------------------------- -$comments --------------------------------------------------------------------------- - -Thank you, - - - - /nomcom/defaults/email/feedback_receipt.txt - Email sent to feedback author to get a confirmation mail containing feedback in cleartext - $nominee: Full name of the nominee -$position: Nomination position -$comments: Comments on this candidate - plain - Hi, - -Your input regarding $about has been received and registered. - -The following comments have been registered: - --------------------------------------------------------------------------- -$comments --------------------------------------------------------------------------- - -Thank you, - - - - /nomcom/defaults/email/questionnaire_reminder.txt - Email sent to nominees reminding them to complete a questionnaire - $positions: Nomination positions - plain - -Thank you for accepting your nomination for the position of $position. - -Please remember to complete and return the questionnaire for this position at your earliest opportunity. -The questionnaire is repeated below for your convenience. - --------- - - - - /nomcom/defaults/topic/description - Description of Topic - $topic: Topic' - rst - This is a description of the topic "$topic" - -Describe the topic and add any information/instructions for the responder here. - - - - /nomcom/defaults/iesg_requirements - Generic IESG Requirements - rst - Generic IESG Requirements Yo! - - + + + + /nomcom/defaults/home.rst + Home page of group + + rst + Home page +========= + +This is the home page of the nomcom group. + + + + /nomcom/defaults/email/inexistent_person.txt + Email sent to chair of nomcom and secretariat when Email and Person are created if some of them don't exist + $email: Newly created email +$fullname: Fullname of the new person +$person_id: Id of the new Person object +$group: Name of the group + plain + Hello, + +A new person with name $fullname and email $email has been created. The new Person object has the following id: '$person_id'. + +Please, check if there is some more action nedeed. + + + + /nomcom/defaults/email/new_nominee.txt + Email sent to nominees when they are nominated + $nominee: Full name of the nominee +$position: Name of the position +$domain: Server domain +$accept_url: Url hash to accept nominations +$decline_url: Url hash to decline nominations + plain + Hi, + +You have been nominated for the position of $position. + +The NomCom would appreciate receiving an indication of whether or not you accept this nomination to stand for consideration as a candidate for this position. + +You can accept the nomination via web going to the following link https://$domain$accept_url or decline the nomination going the following link https://$domain$decline_url + +If you accept, you will need to fill out a questionnaire. You will receive the questionnaire by email. + +Best regards, + + + + + /nomcom/defaults/email/new_nomination.txt + Email sent to nominators and secretariat when the nominators make the nominations + $nominator: Full name of the nominator +$nominator_email: Email of the nominator +$nominee: Full name of the nominee +$nominee_email: Email of the nominee +$position: Nomination position + plain + A new nomination have been received. + +Nominator: $nominator ($nominator_email) +Nominee: $nominee ($nominee_email) +Position: $position + + + + /nomcom/defaults/position/questionnaire.txt + Questionnaire sent to the nomine + $position: Position + plain + Enter here the questionnaire for the position $position: + +Questionnaire + + + + /nomcom/defaults/position/requirements + Position requirements + $position: Position + rst + These are the requirements for the position $position: + +Requirements. + + + + /nomcom/defaults/position/header_questionnaire.txt + Header of the email that contains the questionnaire sent to the nomine + $nominee: Full name of the nomine +$position: Position + plain + Hi $nominee, this is the questionnaire for the position $position: + + + + + + /nomcom/defaults/email/nomination_accept_reminder.txt + Email sent to nominees asking them to accept (or decline) the nominations. + $positions: Nomination positions + plain + Hi, + +You have been nominated for the position of $position. + +The NomCom would appreciate receiving an indication of whether or not you accept this nomination to stand for consideration as a candidate for this position. + +You can accept the nomination via web going to the following link https://$domain$accept_url or decline the nomination going the following link https://$domain$decline_url + +If you accept, you will need to fill out a questionnaire. + +Best regards, + + + + /nomcom/defaults/email/nomination_receipt.txt + Email sent to nominator to get a confirmation mail containing feedback in cleartext + $nominee: Full name of the nominee +$position: Name of the position +$domain: Server domain +$accept_url: Url hash to accept nominations +$decline_url: Url hash to decline nominations + plain + Hi, + +Your nomination of $nominee for the position of +$position has been received and registered. + +The following comments have also been registered: + +-------------------------------------------------------------------------- +$comments +-------------------------------------------------------------------------- + +Thank you, + + + + /nomcom/defaults/email/feedback_receipt.txt + Email sent to feedback author to get a confirmation mail containing feedback in cleartext + $nominee: Full name of the nominee +$position: Nomination position +$comments: Comments on this candidate + plain + Hi, + +Your input regarding $about has been received and registered. + +The following comments have been registered: + +-------------------------------------------------------------------------- +$comments +-------------------------------------------------------------------------- + +Thank you, + + + + /nomcom/defaults/email/questionnaire_reminder.txt + Email sent to nominees reminding them to complete a questionnaire + $positions: Nomination positions + plain + +Thank you for accepting your nomination for the position of $position. + +Please remember to complete and return the questionnaire for this position at your earliest opportunity. +The questionnaire is repeated below for your convenience. + +-------- + + + + /nomcom/defaults/topic/description + Description of Topic + $topic: Topic' + rst + This is a description of the topic "$topic" + +Describe the topic and add any information/instructions for the responder here. + + + + /nomcom/defaults/iesg_requirements + Generic IESG Requirements + rst + Generic IESG Requirements Yo! + + diff --git a/ietf/dbtemplate/migrations/0001_initial.py b/ietf/dbtemplate/migrations/0001_initial.py index 53defbda2c..a3a23451f0 100644 --- a/ietf/dbtemplate/migrations/0001_initial.py +++ b/ietf/dbtemplate/migrations/0001_initial.py @@ -1,10 +1,6 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-20 10:52 - - -from typing import List, Tuple # pyflakes:ignore +# Generated by Django 2.2.28 on 2023-03-20 19:22 +from typing import List, Tuple from django.db import migrations, models @@ -12,8 +8,8 @@ class Migration(migrations.Migration): initial = True - dependencies = [ - ] # type: List[Tuple[str]] + dependencies: List[Tuple[str, str]] = [ + ] operations = [ migrations.CreateModel( diff --git a/ietf/dbtemplate/migrations/0002_auto_20180220_1052.py b/ietf/dbtemplate/migrations/0002_auto_20180220_1052.py deleted file mode 100644 index 9cff93f4af..0000000000 --- a/ietf/dbtemplate/migrations/0002_auto_20180220_1052.py +++ /dev/null @@ -1,32 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-20 10:52 - - -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - initial = True - - dependencies = [ - ('group', '0001_initial'), - ('name', '0001_initial'), - ('dbtemplate', '0001_initial'), - ] - - operations = [ - migrations.AddField( - model_name='dbtemplate', - name='group', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='group.Group'), - ), - migrations.AddField( - model_name='dbtemplate', - name='type', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.DBTemplateTypeName'), - ), - ] diff --git a/ietf/dbtemplate/migrations/0002_auto_20230320_1222.py b/ietf/dbtemplate/migrations/0002_auto_20230320_1222.py new file mode 100644 index 0000000000..5aa713635f --- /dev/null +++ b/ietf/dbtemplate/migrations/0002_auto_20230320_1222.py @@ -0,0 +1,29 @@ +# Generated by Django 2.2.28 on 2023-03-20 19:22 + +from django.db import migrations +import django.db.models.deletion +import ietf.utils.models + + +class Migration(migrations.Migration): + + initial = True + + dependencies = [ + ('dbtemplate', '0001_initial'), + ('group', '0001_initial'), + ('name', '0001_initial'), + ] + + operations = [ + migrations.AddField( + model_name='dbtemplate', + name='group', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='group.Group'), + ), + migrations.AddField( + model_name='dbtemplate', + name='type', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.DBTemplateTypeName'), + ), + ] diff --git a/ietf/dbtemplate/migrations/0003_adjust_review_templates.py b/ietf/dbtemplate/migrations/0003_adjust_review_templates.py deleted file mode 100644 index d3038c946c..0000000000 --- a/ietf/dbtemplate/migrations/0003_adjust_review_templates.py +++ /dev/null @@ -1,119 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-03-05 11:39 - - -from django.db import migrations - -def forward(apps, schema_editor): - DBTemplate = apps.get_model('dbtemplate', 'DBTemplate') - DBTemplate.objects.filter(id=186).update(content="""I am the assigned Gen-ART reviewer for this draft. The General Area -Review Team (Gen-ART) reviews all IETF documents being processed -by the IESG for the IETF Chair. Please treat these comments just -like any other last call comments. - -For more information, please see the FAQ at - -. - -Document: {{ assignment.review_request.doc.name }}-?? -Reviewer: {{ assignment.reviewer.person.plain_name }} -Review Date: {{ today }} -IETF LC End Date: {% if assignment.review_request.doc.most_recent_ietflc %}{{ assignment.review_request.doc.most_recent_ietflc.expires|date:"Y-m-d" }}{% else %}None{% endif %} -IESG Telechat date: {{ assignment.review_request.doc.telechat_date|default:'Not scheduled for a telechat' }} - -Summary: - -Major issues: - -Minor issues: - -Nits/editorial comments: -""") - - DBTemplate.objects.filter(id=187).update(content="""I am the assigned Gen-ART reviewer for this draft. The General Area -Review Team (Gen-ART) reviews all IETF documents being processed -by the IESG for the IETF Chair. Please wait for direction from your -document shepherd or AD before posting a new version of the draft. - -For more information, please see the FAQ at - -. - -Document: {{ assignment.review_request.doc.name }}-?? -Reviewer: {{ assignment.reviewer.person.plain_name }} -Review Date: {{ today }} -IETF LC End Date: {% if assignment.review_req.doc.most_recent_ietflc %}{{ assignment.review_request.doc.most_recent_ietflc.expires|date:"Y-m-d" }}{% else %}None{% endif %} -IESG Telechat date: {{ assignment.review_request.doc.telechat_date|default:'Not scheduled for a telechat' }} - -Summary: - -Major issues: - -Minor issues: - -Nits/editorial comments: - -""") - -def reverse(apps, schema_editor): - DBTemplate = apps.get_model('dbtemplate','DBTemplate') - DBTemplate.objects.filter(id=186).update(content="""I am the assigned Gen-ART reviewer for this draft. The General Area -Review Team (Gen-ART) reviews all IETF documents being processed -by the IESG for the IETF Chair. Please treat these comments just -like any other last call comments. - -For more information, please see the FAQ at - -. - -Document: {{ review_req.doc.name }}-?? -Reviewer: {{ review_req.reviewer.person.plain_name }} -Review Date: {{ today }} -IETF LC End Date: {% if review_req.doc.most_recent_ietflc %}{{ review_req.doc.most_recent_ietflc.expires|date:"Y-m-d" }}{% else %}None{% endif %} -IESG Telechat date: {{ review_req.doc.telechat_date|default:'Not scheduled for a telechat' }} - -Summary: - -Major issues: - -Minor issues: - -Nits/editorial comments: - -""") - DBTemplate.objects.filter(id=187).update(content="""I am the assigned Gen-ART reviewer for this draft. The General Area -Review Team (Gen-ART) reviews all IETF documents being processed -by the IESG for the IETF Chair. Please wait for direction from your -document shepherd or AD before posting a new version of the draft. - -For more information, please see the FAQ at - -. - -Document: {{ review_req.doc.name }}-?? -Reviewer: {{ review_req.reviewer.person.plain_name }} -Review Date: {{ today }} -IETF LC End Date: {% if review_req.doc.most_recent_ietflc %}{{ review_req.doc.most_recent_ietflc.expires|date:"Y-m-d" }}{% else %}None{% endif %} -IESG Telechat date: {{ review_req.doc.telechat_date|default:'Not scheduled for a telechat' }} - -Summary: - -Major issues: - -Minor issues: - -Nits/editorial comments: - -""") - -class Migration(migrations.Migration): - - dependencies = [ - ('dbtemplate', '0002_auto_20180220_1052'), - ('doc','0011_reviewassignmentdocevent'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/dbtemplate/migrations/0004_adjust_assignment_email_summary_templates.py b/ietf/dbtemplate/migrations/0004_adjust_assignment_email_summary_templates.py deleted file mode 100644 index f9a2496090..0000000000 --- a/ietf/dbtemplate/migrations/0004_adjust_assignment_email_summary_templates.py +++ /dev/null @@ -1,290 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-03-13 13:41 - - -from django.db import migrations - -def forward(apps, schema_editor): - DBTemplate = apps.get_model('dbtemplate','DBTemplate') - - DBTemplate.objects.filter(pk=182).update(content="""{% autoescape off %}Subject: Open review assignments in {{group.acronym}} - -The following reviewers have assignments:{% for r in review_assignments %}{% ifchanged r.section %} - -{{r.section}} - -{% if r.section == 'Early review requests:' %}Reviewer Due Draft{% else %}Reviewer LC end Draft{% endif %}{% endifchanged %} -{{ r.reviewer.person.plain_name|ljust:"22" }} {% if r.section == 'Early review requests:' %}{{ r.review_request.deadline|date:"Y-m-d" }}{% else %}{{ r.lastcall_ends|default:"None " }}{% endif %} {{ r.review_request.doc_id }}-{% if r.review_request..requested_rev %}{{ r.review_request.requested_rev }}{% else %}{{ r.review_request..doc.rev }}{% endif %} {{ r.earlier_review_mark }}{% endfor %} - -* Other revision previously reviewed -** This revision already reviewed - -{% if rotation_list %}Next in the reviewer rotation: - -{% for p in rotation_list %} {{ p }} -{% endfor %}{% endif %}{% endautoescape %} -""") - - DBTemplate.objects.filter(pk=183).update(content="""{% autoescape off %}Subject: Review Assignments - -Hi all, - -The following reviewers have assignments:{% for r in review_assignments %}{% ifchanged r.section %} - -{{r.section}} - -{% if r.section == 'Early review requests:' %}Reviewer Due Draft{% else %}Reviewer Type LC end Draft{% endif %}{% endifchanged %} -{{ r.reviewer.person.plain_name|ljust:"22" }} {% if r.section == 'Early review requests:' %}{{ r.review_request.deadline|date:"Y-m-d" }}{% else %}{{r.review_request.type.name|ljust:"10"}}{{ r.lastcall_ends|default:"None " }}{% endif %} {{ r.review_request.doc_id }}-{% if r.review_request.requested_rev %}{{ r.review_request.requested_rev }}{% else %}{{ r.review_request.doc.rev }}{% endif %}{% if r.earlier_review_mark %} {{ r.earlier_review_mark }}{% endif %}{% endfor %} - -* Other revision previously reviewed -** This revision already reviewed - -{% if rotation_list %}Next in the reviewer rotation: - -{% for p in rotation_list %} {{ p }} -{% endfor %}{% endif %} -The LC and Telechat review templates are included below: -------------------------------------------------------- - --- Begin LC Template -- -I am the assigned Gen-ART reviewer for this draft. The General Area -Review Team (Gen-ART) reviews all IETF documents being processed -by the IESG for the IETF Chair. Please treat these comments just -like any other last call comments. - -For more information, please see the FAQ at - -. - -Document: -Reviewer: -Review Date: -IETF LC End Date: -IESG Telechat date: (if known) - -Summary: - -Major issues: - -Minor issues: - -Nits/editorial comments: - --- End LC Template -- - --- Begin Telechat Template -- -I am the assigned Gen-ART reviewer for this draft. The General Area -Review Team (Gen-ART) reviews all IETF documents being processed -by the IESG for the IETF Chair. Please wait for direction from your -document shepherd or AD before posting a new version of the draft. - -For more information, please see the FAQ at - -. - -Document: -Reviewer: -Review Date: -IETF LC End Date: -IESG Telechat date: (if known) - -Summary: - -Major issues: - -Minor issues: - -Nits/editorial comments: - --- End Telechat Template -- -{% endautoescape %} - -""") - - DBTemplate.objects.filter(pk=184).update(content="""{% autoescape off %}Subject: Assignments - -Review instructions and related resources are at: -http://tools.ietf.org/area/sec/trac/wiki/SecDirReview{% for r in review_assignments %}{% ifchanged r.section %} - -{{r.section}} - -{% if r.section == 'Early review requests:' %}Reviewer Due Draft{% else %}Reviewer LC end Draft{% endif %}{% endifchanged %} -{{ r.reviewer.person.plain_name|ljust:"22" }}{{ r.earlier_review|yesno:'R, , ' }}{% if r.section == 'Early review requests:' %}{{ r.review_request.deadline|date:"Y-m-d" }}{% else %}{{ r.lastcall_ends|default:"None " }}{% endif %} {{ r.review_request.doc_id }}-{% if r.review_request.requested_rev %}{{ r.review_request.requested_rev }}{% else %}{{ r.review_request.doc.rev }}{% endif %}{% endfor %} - -{% if rotation_list %}Next in the reviewer rotation: - -{% for p in rotation_list %} {{ p }} -{% endfor %}{% endif %}{% endautoescape %} - -""") - - DBTemplate.objects.filter(pk=185).update(content="""{% autoescape off %}Subject: Open review assignments in {{group.acronym}} - -Review instructions and related resources are at: - - -The following reviewers have assignments:{% for r in review_assignments %}{% ifchanged r.section %} - -{{r.section}} - -{% if r.section == 'Early review requests:' %}Reviewer Due Draft{% else %}Reviewer LC end Draft{% endif %}{% endifchanged %} -{{ r.reviewer.person.plain_name|ljust:"22" }} {% if r.section == 'Early review requests:' %}{{ r.review_request.deadline|date:"Y-m-d" }}{% else %}{{ r.lastcall_ends|default:"None " }}{% endif %} {{ r.review_request.doc_id }}-{% if r.review_request.requested_rev %}{{ r.review_request.requested_rev }}{% else %}{{ r.review_request.doc.rev }}{% endif %} {{ r.earlier_review_mark }}{% endfor %} - -* Other revision previously reviewed -** This revision already reviewed - -{% if rotation_list %}Next in the reviewer rotation: - -{% for p in rotation_list %} {{ p }} -{% endfor %}{% endif %}{% endautoescape %} - -""") - - -def reverse(apps, schema_editor): - DBTemplate = apps.get_model('dbtemplate','DBTemplate') - - DBTemplate.objects.filter(pk=182).update(content="""{% autoescape off %}Subject: Open review assignments in {{group.acronym}} - -The following reviewers have assignments:{% for r in review_requests %}{% ifchanged r.section %} - -{{r.section}} - -{% if r.section == 'Early review requests:' %}Reviewer Due Draft{% else %}Reviewer LC end Draft{% endif %}{% endifchanged %} -{{ r.reviewer.person.plain_name|ljust:"22" }} {% if r.section == 'Early review requests:' %}{{ r.deadline|date:"Y-m-d" }}{% else %}{{ r.lastcall_ends|default:"None " }}{% endif %} {{ r.doc_id }}-{% if r.requested_rev %}{{ r.requested_rev }}{% else %}{{ r.doc.rev }}{% endif %} {{ r.earlier_review_mark }}{% endfor %} - -* Other revision previously reviewed -** This revision already reviewed - -{% if rotation_list %}Next in the reviewer rotation: - -{% for p in rotation_list %} {{ p }} -{% endfor %}{% endif %}{% endautoescape %} -""") - - DBTemplate.objects.filter(pk=183).update(content="""{% autoescape off %}Subject: Review Assignments - -Hi all, - -The following reviewers have assignments:{% for r in review_requests %}{% ifchanged r.section %} - -{{r.section}} - -{% if r.section == 'Early review requests:' %}Reviewer Due Draft{% else %}Reviewer Type LC end Draft{% endif %}{% endifchanged %} -{{ r.reviewer.person.plain_name|ljust:"22" }} {% if r.section == 'Early review requests:' %}{{ r.deadline|date:"Y-m-d" }}{% else %}{{r.type.name|ljust:"10"}}{{ r.lastcall_ends|default:"None " }}{% endif %} {{ r.doc_id }}-{% if r.requested_rev %}{{ r.requested_rev }}{% else %}{{ r.doc.rev }}{% endif %}{% if r.earlier_review_mark %} {{ r.earlier_review_mark }}{% endif %}{% endfor %} - -* Other revision previously reviewed -** This revision already reviewed - -{% if rotation_list %}Next in the reviewer rotation: - -{% for p in rotation_list %} {{ p }} -{% endfor %}{% endif %} -The LC and Telechat review templates are included below: -------------------------------------------------------- - --- Begin LC Template -- -I am the assigned Gen-ART reviewer for this draft. The General Area -Review Team (Gen-ART) reviews all IETF documents being processed -by the IESG for the IETF Chair. Please treat these comments just -like any other last call comments. - -For more information, please see the FAQ at - -. - -Document: -Reviewer: -Review Date: -IETF LC End Date: -IESG Telechat date: (if known) - -Summary: - -Major issues: - -Minor issues: - -Nits/editorial comments: - --- End LC Template -- - --- Begin Telechat Template -- -I am the assigned Gen-ART reviewer for this draft. The General Area -Review Team (Gen-ART) reviews all IETF documents being processed -by the IESG for the IETF Chair. Please wait for direction from your -document shepherd or AD before posting a new version of the draft. - -For more information, please see the FAQ at - -. - -Document: -Reviewer: -Review Date: -IETF LC End Date: -IESG Telechat date: (if known) - -Summary: - -Major issues: - -Minor issues: - -Nits/editorial comments: - --- End Telechat Template -- -{% endautoescape %} - -""") - - DBTemplate.objects.filter(pk=184).update(content="""{% autoescape off %}Subject: Assignments - -Review instructions and related resources are at: -http://tools.ietf.org/area/sec/trac/wiki/SecDirReview{% for r in review_requests %}{% ifchanged r.section %} - -{{r.section}} - -{% if r.section == 'Early review requests:' %}Reviewer Due Draft{% else %}Reviewer LC end Draft{% endif %}{% endifchanged %} -{{ r.reviewer.person.plain_name|ljust:"22" }}{{ r.earlier_review|yesno:'R, , ' }}{% if r.section == 'Early review requests:' %}{{ r.deadline|date:"Y-m-d" }}{% else %}{{ r.lastcall_ends|default:"None " }}{% endif %} {{ r.doc_id }}-{% if r.requested_rev %}{{ r.requested_rev }}{% else %}{{ r.doc.rev }}{% endif %}{% endfor %} - -{% if rotation_list %}Next in the reviewer rotation: - -{% for p in rotation_list %} {{ p }} -{% endfor %}{% endif %}{% endautoescape %} - -""") - - DBTemplate.objects.filter(pk=185).update(content="""{% autoescape off %}Subject: Open review assignments in {{group.acronym}} - -Review instructions and related resources are at: - - -The following reviewers have assignments:{% for r in review_requests %}{% ifchanged r.section %} - -{{r.section}} - -{% if r.section == 'Early review requests:' %}Reviewer Due Draft{% else %}Reviewer LC end Draft{% endif %}{% endifchanged %} -{{ r.reviewer.person.plain_name|ljust:"22" }} {% if r.section == 'Early review requests:' %}{{ r.deadline|date:"Y-m-d" }}{% else %}{{ r.lastcall_ends|default:"None " }}{% endif %} {{ r.doc_id }}-{% if r.requested_rev %}{{ r.requested_rev }}{% else %}{{ r.doc.rev }}{% endif %} {{ r.earlier_review_mark }}{% endfor %} - -* Other revision previously reviewed -** This revision already reviewed - -{% if rotation_list %}Next in the reviewer rotation: - -{% for p in rotation_list %} {{ p }} -{% endfor %}{% endif %}{% endautoescape %} - -""") - - -class Migration(migrations.Migration): - - dependencies = [ - ('dbtemplate', '0003_adjust_review_templates'), - ] - - operations = [ - migrations.RunPython(forward,reverse), - ] diff --git a/ietf/dbtemplate/migrations/0005_adjust_assignment_email_summary_templates_2526.py b/ietf/dbtemplate/migrations/0005_adjust_assignment_email_summary_templates_2526.py deleted file mode 100644 index 8a07d89364..0000000000 --- a/ietf/dbtemplate/migrations/0005_adjust_assignment_email_summary_templates_2526.py +++ /dev/null @@ -1,281 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-03-13 13:41 - - -from django.db import migrations - -def forward(apps, schema_editor): - DBTemplate = apps.get_model('dbtemplate','DBTemplate') - - DBTemplate.objects.filter(pk=182).update(content="""{% autoescape off %}Subject: Open review assignments in {{group.acronym}} - -The following reviewers have assignments:{% for r in review_assignments %}{% ifchanged r.section %} - -{{r.section}} - -{% if r.section == 'Early review requests:' %}Reviewer Due Draft{% else %}Reviewer LC end Draft{% endif %}{% endifchanged %} -{{ r.reviewer.person.plain_name|ljust:"22" }} {% if r.section == 'Early review requests:' %}{{ r.review_request.deadline|date:"Y-m-d" }}{% else %}{{ r.lastcall_ends|default:"None " }}{% endif %} {{ r.review_request.doc.name }}-{% if r.review_request.requested_rev %}{{ r.review_request.requested_rev }}{% else %}{{ r.review_request.doc.rev }}{% endif %} {{ r.earlier_reviews }}{% endfor %} - -{% if rotation_list %}Next in the reviewer rotation: - -{% for p in rotation_list %} {{ p }} -{% endfor %}{% endif %}{% endautoescape %} -""") - - DBTemplate.objects.filter(pk=183).update(content="""{% autoescape off %}Subject: Review Assignments - -Hi all, - -The following reviewers have assignments:{% for r in review_assignments %}{% ifchanged r.section %} - -{{r.section}} - -{% if r.section == 'Early review requests:' %}Reviewer Due Draft{% else %}Reviewer Type LC end Draft{% endif %}{% endifchanged %} -{{ r.reviewer.person.plain_name|ljust:"22" }} {% if r.section == 'Early review requests:' %}{{ r.review_request.deadline|date:"Y-m-d" }}{% else %}{{r.review_request.type.name|ljust:"10"}}{{ r.lastcall_ends|default:"None " }}{% endif %} {{ r.review_request.doc.name }}-{% if r.review_request.requested_rev %}{{ r.review_request.requested_rev }}{% else %}{{ r.review_request.doc.rev }}{% endif %}{% if r.earlier_reviews %} {{ r.earlier_reviews }}{% endif %}{% endfor %} - -{% if rotation_list %}Next in the reviewer rotation: - -{% for p in rotation_list %} {{ p }} -{% endfor %}{% endif %} -The LC and Telechat review templates are included below: -------------------------------------------------------- - --- Begin LC Template -- -I am the assigned Gen-ART reviewer for this draft. The General Area -Review Team (Gen-ART) reviews all IETF documents being processed -by the IESG for the IETF Chair. Please treat these comments just -like any other last call comments. - -For more information, please see the FAQ at - -. - -Document: -Reviewer: -Review Date: -IETF LC End Date: -IESG Telechat date: (if known) - -Summary: - -Major issues: - -Minor issues: - -Nits/editorial comments: - --- End LC Template -- - --- Begin Telechat Template -- -I am the assigned Gen-ART reviewer for this draft. The General Area -Review Team (Gen-ART) reviews all IETF documents being processed -by the IESG for the IETF Chair. Please wait for direction from your -document shepherd or AD before posting a new version of the draft. - -For more information, please see the FAQ at - -. - -Document: -Reviewer: -Review Date: -IETF LC End Date: -IESG Telechat date: (if known) - -Summary: - -Major issues: - -Minor issues: - -Nits/editorial comments: - --- End Telechat Template -- -{% endautoescape %} - -""") - - DBTemplate.objects.filter(pk=184).update(content="""{% autoescape off %}Subject: Assignments - -Review instructions and related resources are at: -http://tools.ietf.org/area/sec/trac/wiki/SecDirReview{% for r in review_assignments %}{% ifchanged r.section %} - -{{r.section}} - -{% if r.section == 'Early review requests:' %}Reviewer Due Draft{% else %}Reviewer LC end Draft{% endif %}{% endifchanged %} -{{ r.reviewer.person.plain_name|ljust:"22" }}{{ r.earlier_review|yesno:'R, , ' }}{% if r.section == 'Early review requests:' %}{{ r.review_request.deadline|date:"Y-m-d" }}{% else %}{{ r.lastcall_ends|default:"None " }}{% endif %} {{ r.review_request.doc.name }}-{% if r.review_request.requested_rev %}{{ r.review_request.requested_rev }}{% else %}{{ r.review_request.doc.rev }}{% endif %}{% endfor %} - -{% if rotation_list %}Next in the reviewer rotation: - -{% for p in rotation_list %} {{ p }} -{% endfor %}{% endif %}{% endautoescape %} - -""") - - DBTemplate.objects.filter(pk=185).update(content="""{% autoescape off %}Subject: Open review assignments in {{group.acronym}} - -Review instructions and related resources are at: - - -The following reviewers have assignments:{% for r in review_assignments %}{% ifchanged r.section %} - -{{r.section}} - -{% if r.section == 'Early review requests:' %}Reviewer Due Draft{% else %}Reviewer LC end Draft{% endif %}{% endifchanged %} -{{ r.reviewer.person.plain_name|ljust:"22" }} {% if r.section == 'Early review requests:' %}{{ r.review_request.deadline|date:"Y-m-d" }}{% else %}{{ r.lastcall_ends|default:"None " }}{% endif %} {{ r.review_request.doc.name }}-{% if r.review_request.requested_rev %}{{ r.review_request.requested_rev }}{% else %}{{ r.review_request.doc.rev }}{% endif %} {{ r.earlier_reviews }}{% endfor %} - -{% if rotation_list %}Next in the reviewer rotation: - -{% for p in rotation_list %} {{ p }} -{% endfor %}{% endif %}{% endautoescape %} - -""") - - -def reverse(apps, schema_editor): - DBTemplate = apps.get_model('dbtemplate','DBTemplate') - - DBTemplate.objects.filter(pk=182).update(content="""{% autoescape off %}Subject: Open review assignments in {{group.acronym}} - -The following reviewers have assignments:{% for r in review_assignments %}{% ifchanged r.section %} - -{{r.section}} - -{% if r.section == 'Early review requests:' %}Reviewer Due Draft{% else %}Reviewer LC end Draft{% endif %}{% endifchanged %} -{{ r.reviewer.person.plain_name|ljust:"22" }} {% if r.section == 'Early review requests:' %}{{ r.review_request.deadline|date:"Y-m-d" }}{% else %}{{ r.lastcall_ends|default:"None " }}{% endif %} {{ r.review_request.doc_id }}-{% if r.review_request..requested_rev %}{{ r.review_request.requested_rev }}{% else %}{{ r.review_request..doc.rev }}{% endif %} {{ r.earlier_review_mark }}{% endfor %} - -* Other revision previously reviewed -** This revision already reviewed - -{% if rotation_list %}Next in the reviewer rotation: - -{% for p in rotation_list %} {{ p }} -{% endfor %}{% endif %}{% endautoescape %} -""") - - DBTemplate.objects.filter(pk=183).update(content="""{% autoescape off %}Subject: Review Assignments - -Hi all, - -The following reviewers have assignments:{% for r in review_assignments %}{% ifchanged r.section %} - -{{r.section}} - -{% if r.section == 'Early review requests:' %}Reviewer Due Draft{% else %}Reviewer Type LC end Draft{% endif %}{% endifchanged %} -{{ r.reviewer.person.plain_name|ljust:"22" }} {% if r.section == 'Early review requests:' %}{{ r.review_request.deadline|date:"Y-m-d" }}{% else %}{{r.review_request.type.name|ljust:"10"}}{{ r.lastcall_ends|default:"None " }}{% endif %} {{ r.review_request.doc.name }}-{% if r.review_request.requested_rev %}{{ r.review_request.requested_rev }}{% else %}{{ r.review_request.doc.rev }}{% endif %}{% if r.earlier_review_mark %} {{ r.earlier_review_mark }}{% endif %}{% endfor %} - -* Other revision previously reviewed -** This revision already reviewed - -{% if rotation_list %}Next in the reviewer rotation: - -{% for p in rotation_list %} {{ p }} -{% endfor %}{% endif %} -The LC and Telechat review templates are included below: -------------------------------------------------------- - --- Begin LC Template -- -I am the assigned Gen-ART reviewer for this draft. The General Area -Review Team (Gen-ART) reviews all IETF documents being processed -by the IESG for the IETF Chair. Please treat these comments just -like any other last call comments. - -For more information, please see the FAQ at - -. - -Document: -Reviewer: -Review Date: -IETF LC End Date: -IESG Telechat date: (if known) - -Summary: - -Major issues: - -Minor issues: - -Nits/editorial comments: - --- End LC Template -- - --- Begin Telechat Template -- -I am the assigned Gen-ART reviewer for this draft. The General Area -Review Team (Gen-ART) reviews all IETF documents being processed -by the IESG for the IETF Chair. Please wait for direction from your -document shepherd or AD before posting a new version of the draft. - -For more information, please see the FAQ at - -. - -Document: -Reviewer: -Review Date: -IETF LC End Date: -IESG Telechat date: (if known) - -Summary: - -Major issues: - -Minor issues: - -Nits/editorial comments: - --- End Telechat Template -- -{% endautoescape %} - -""") - - DBTemplate.objects.filter(pk=184).update(content="""{% autoescape off %}Subject: Assignments - -Review instructions and related resources are at: -http://tools.ietf.org/area/sec/trac/wiki/SecDirReview{% for r in review_assignments %}{% ifchanged r.section %} - -{{r.section}} - -{% if r.section == 'Early review requests:' %}Reviewer Due Draft{% else %}Reviewer LC end Draft{% endif %}{% endifchanged %} -{{ r.reviewer.person.plain_name|ljust:"22" }}{{ r.earlier_review|yesno:'R, , ' }}{% if r.section == 'Early review requests:' %}{{ r.review_request.deadline|date:"Y-m-d" }}{% else %}{{ r.lastcall_ends|default:"None " }}{% endif %} {{ r.review_request.doc.name }}-{% if r.review_request.requested_rev %}{{ r.review_request.requested_rev }}{% else %}{{ r.review_request.doc.rev }}{% endif %}{% endfor %} - -{% if rotation_list %}Next in the reviewer rotation: - -{% for p in rotation_list %} {{ p }} -{% endfor %}{% endif %}{% endautoescape %} - -""") - - DBTemplate.objects.filter(pk=185).update(content="""{% autoescape off %}Subject: Open review assignments in {{group.acronym}} - -Review instructions and related resources are at: - - -The following reviewers have assignments:{% for r in review_assignments %}{% ifchanged r.section %} - -{{r.section}} - -{% if r.section == 'Early review requests:' %}Reviewer Due Draft{% else %}Reviewer LC end Draft{% endif %}{% endifchanged %} -{{ r.reviewer.person.plain_name|ljust:"22" }} {% if r.section == 'Early review requests:' %}{{ r.review_request.deadline|date:"Y-m-d" }}{% else %}{{ r.lastcall_ends|default:"None " }}{% endif %} {{ r.review_request.doc.name }}-{% if r.review_request.requested_rev %}{{ r.review_request.requested_rev }}{% else %}{{ r.review_request.doc.rev }}{% endif %} {{ r.earlier_review_mark }}{% endfor %} - -* Other revision previously reviewed -** This revision already reviewed - -{% if rotation_list %}Next in the reviewer rotation: - -{% for p in rotation_list %} {{ p }} -{% endfor %}{% endif %}{% endautoescape %} - -""") - - -class Migration(migrations.Migration): - - dependencies = [ - ('dbtemplate', '0004_adjust_assignment_email_summary_templates'), - ] - - operations = [ - migrations.RunPython(forward,reverse), - ] diff --git a/ietf/dbtemplate/migrations/0006_add_review_assigned_template.py b/ietf/dbtemplate/migrations/0006_add_review_assigned_template.py deleted file mode 100644 index cedd843220..0000000000 --- a/ietf/dbtemplate/migrations/0006_add_review_assigned_template.py +++ /dev/null @@ -1,33 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -from django.db import migrations - -def forward(apps, schema_editor): - DBTemplate = apps.get_model('dbtemplate', 'DBTemplate') - - DBTemplate.objects.create(path='/group/defaults/email/review_assigned.txt', type_id='django', - content="""{{ assigner.ascii }} has assigned you as a reviewer for this document. - -{% if prev_team_reviews %}This team has completed other reviews of this document:{% endif %}{% for assignment in prev_team_reviews %} -- {{ assignment.completed_on }} {{ assignment.reviewer.person.ascii }} -{% if assignment.reviewed_rev %}{{ assignment.reviewed_rev }}{% else %}{{ assignment.review_request.requested_rev }}{% endif %} {{ assignment.result.name }} -{% endfor %} -""") - - -def reverse(apps, schema_editor): - DBTemplate = apps.get_model('dbtemplate', 'DBTemplate') - - DBTemplate.objects.get(path='/group/defaults/email/review_assigned.txt').delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('dbtemplate', '0005_adjust_assignment_email_summary_templates_2526'), - ] - - operations = [ - migrations.RunPython(forward,reverse), - ] diff --git a/ietf/dbtemplate/migrations/0007_adjust_review_assigned.py b/ietf/dbtemplate/migrations/0007_adjust_review_assigned.py deleted file mode 100644 index c1e8324ef9..0000000000 --- a/ietf/dbtemplate/migrations/0007_adjust_review_assigned.py +++ /dev/null @@ -1,38 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.26 on 2019-11-19 11:47 - - -from django.db import migrations - -def forward(apps, schema_editor): - DBTemplate = apps.get_model('dbtemplate','DBTemplate') - qs = DBTemplate.objects.filter(path='/group/defaults/email/review_assigned.txt') - qs.update(content="""{{ assigner.ascii }} has assigned {{ reviewer.person.ascii }} as a reviewer for this document. - -{% if prev_team_reviews %}This team has completed other reviews of this document:{% endif %}{% for assignment in prev_team_reviews %} -- {{ assignment.completed_on }} {{ assignment.reviewer.person.ascii }} -{% if assignment.reviewed_rev %}{{ assignment.reviewed_rev }}{% else %}{{ assignment.review_request.requested_rev }}{% endif %} {{ assignment.result.name }} -{% endfor %} -""") - qs.update(title="Default template for review assignment email") - -def reverse(apps, schema_editor): - DBTemplate = apps.get_model('dbtemplate','DBTemplate') - qs = DBTemplate.objects.filter(path='/group/defaults/email/review_assigned.txt') - qs.update(content="""{{ assigner.ascii }} has assigned you as a reviewer for this document. - -{% if prev_team_reviews %}This team has completed other reviews of this document:{% endif %}{% for assignment in prev_team_reviews %} -- {{ assignment.completed_on }} {{ assignment.reviewer.person.ascii }} -{% if assignment.reviewed_rev %}{{ assignment.reviewed_rev }}{% else %}{{ assignment.review_request.requested_rev }}{% endif %} {{ assignment.result.name }} -{% endfor %} -""") - - -class Migration(migrations.Migration): - - dependencies = [ - ('dbtemplate', '0006_add_review_assigned_template'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/dbtemplate/migrations/0008_add_default_iesg_req_template.py b/ietf/dbtemplate/migrations/0008_add_default_iesg_req_template.py deleted file mode 100644 index 3ca9a6fc8e..0000000000 --- a/ietf/dbtemplate/migrations/0008_add_default_iesg_req_template.py +++ /dev/null @@ -1,32 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.27 on 2020-01-07 09:25 - - -from django.db import migrations - -def forward(apps, schema_editor): - DBTemplate = apps.get_model('dbtemplate', 'DBTemplate') - DBTemplate.objects.create(path='/nomcom/defaults/iesg_requirements', type_id='rst', title='Generic IESG requirements', - content="""============================= -IESG MEMBER DESIRED EXPERTISE -============================= - -Place this year's Generic IESG Member Desired Expertise here. - -This template uses reStructured text for formatting. Feel free to use it (to change the above header for example). -""") - -def reverse(apps, schema_editor): - DBTemplate = apps.get_model('dbtemplate', 'DBTemplate') - DBTemplate.objects.filter(path='/nomcom/defaults/iesg_requirements').delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('dbtemplate', '0007_adjust_review_assigned'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/doc/admin.py b/ietf/doc/admin.py index 64b9d9eff8..0d04e8db3a 100644 --- a/ietf/doc/admin.py +++ b/ietf/doc/admin.py @@ -1,19 +1,22 @@ -# Copyright The IETF Trust 2010-2021, All Rights Reserved +# Copyright The IETF Trust 2010-2025, All Rights Reserved # -*- coding: utf-8 -*- from django.contrib import admin from django.db import models from django import forms +from rangefilter.filters import DateRangeQuickSelectListFilterBuilder from .models import (StateType, State, RelatedDocument, DocumentAuthor, Document, RelatedDocHistory, - DocHistoryAuthor, DocHistory, DocAlias, DocReminder, DocEvent, NewRevisionDocEvent, + DocHistoryAuthor, DocHistory, DocReminder, DocEvent, NewRevisionDocEvent, StateDocEvent, ConsensusDocEvent, BallotType, BallotDocEvent, WriteupDocEvent, LastCallDocEvent, TelechatDocEvent, BallotPositionDocEvent, ReviewRequestDocEvent, InitialReviewDocEvent, AddedMessageEvent, SubmissionDocEvent, DeletedEvent, EditedAuthorsDocEvent, DocumentURL, ReviewAssignmentDocEvent, IanaExpertDocEvent, IRSGBallotDocEvent, DocExtResource, DocumentActionHolder, - BofreqEditorDocEvent, BofreqResponsibleDocEvent ) + BofreqEditorDocEvent, BofreqResponsibleDocEvent, StoredObject, RfcAuthor, + EditedRfcAuthorsDocEvent) +from ietf.utils.admin import SaferTabularInline from ietf.utils.validators import validate_external_resource_value class StateTypeAdmin(admin.ModelAdmin): @@ -21,36 +24,33 @@ class StateTypeAdmin(admin.ModelAdmin): admin.site.register(StateType, StateTypeAdmin) class StateAdmin(admin.ModelAdmin): - list_display = ["slug", "type", 'name', 'order', 'desc'] - list_filter = ["type", ] + list_display = ["slug", "type", 'name', 'order', 'desc', "used"] + list_filter = ["type", "used"] search_fields = ["slug", "type__label", "type__slug", "name", "desc"] filter_horizontal = ["next_states"] admin.site.register(State, StateAdmin) -# class DocAliasInline(admin.TabularInline): -# model = DocAlias -# extra = 1 - -class DocAuthorInline(admin.TabularInline): +class DocAuthorInline(SaferTabularInline): model = DocumentAuthor raw_id_fields = ['person', 'email'] extra = 1 -class DocActionHolderInline(admin.TabularInline): +class DocActionHolderInline(SaferTabularInline): model = DocumentActionHolder raw_id_fields = ['person'] extra = 1 -class RelatedDocumentInline(admin.TabularInline): +class RelatedDocumentInline(SaferTabularInline): model = RelatedDocument + fk_name= 'source' def this(self, instance): - return instance.source.canonical_name() + return instance.source.name readonly_fields = ['this', ] fields = ['this', 'relationship', 'target', ] raw_id_fields = ['target'] extra = 1 -class AdditionalUrlInLine(admin.TabularInline): +class AdditionalUrlInLine(SaferTabularInline): model = DocumentURL fields = ['tag','desc','url',] extra = 1 @@ -70,7 +70,7 @@ class Meta: class DocumentAuthorAdmin(admin.ModelAdmin): list_display = ['id', 'document', 'person', 'email', 'affiliation', 'country', 'order'] - search_fields = ['document__docalias__name', 'person__name', 'email__address', 'affiliation', 'country'] + search_fields = ['document__name', 'person__name', 'email__address', 'affiliation', 'country'] raw_id_fields = ["document", "person", "email"] admin.site.register(DocumentAuthor, DocumentAuthorAdmin) @@ -108,14 +108,6 @@ def state(self, instance): admin.site.register(DocHistory, DocHistoryAdmin) -class DocAliasAdmin(admin.ModelAdmin): - list_display = ['name', 'targets'] - search_fields = ['name', 'docs__name'] - raw_id_fields = ['docs'] - def targets(self, obj): - return ', '.join([o.name for o in obj.docs.all()]) -admin.site.register(DocAlias, DocAliasAdmin) - class DocReminderAdmin(admin.ModelAdmin): list_display = ['id', 'event', 'type', 'due', 'active'] list_filter = ['type', 'due', 'active'] @@ -125,7 +117,7 @@ class DocReminderAdmin(admin.ModelAdmin): class RelatedDocumentAdmin(admin.ModelAdmin): list_display = ['source', 'target', 'relationship', ] list_filter = ['relationship', ] - search_fields = ['source__name', 'target__name', 'target__docs__name', ] + search_fields = ['source__name', 'target__name', ] raw_id_fields = ['source', 'target', ] admin.site.register(RelatedDocument, RelatedDocumentAdmin) @@ -153,6 +145,13 @@ class DocumentActionHolderAdmin(admin.ModelAdmin): # events +class DeletedEventAdmin(admin.ModelAdmin): + list_display = ['id', 'content_type', 'json', 'by', 'time'] + list_filter = ['time'] + raw_id_fields = ['content_type', 'by'] +admin.site.register(DeletedEvent, DeletedEventAdmin) + + class DocEventAdmin(admin.ModelAdmin): def event_type(self, obj): return str(obj.type) @@ -170,39 +169,43 @@ def short_desc(self, obj): admin.site.register(StateDocEvent, DocEventAdmin) admin.site.register(ConsensusDocEvent, DocEventAdmin) admin.site.register(BallotDocEvent, DocEventAdmin) +admin.site.register(IRSGBallotDocEvent, DocEventAdmin) admin.site.register(WriteupDocEvent, DocEventAdmin) admin.site.register(LastCallDocEvent, DocEventAdmin) admin.site.register(TelechatDocEvent, DocEventAdmin) -admin.site.register(ReviewRequestDocEvent, DocEventAdmin) -admin.site.register(ReviewAssignmentDocEvent, DocEventAdmin) admin.site.register(InitialReviewDocEvent, DocEventAdmin) -admin.site.register(AddedMessageEvent, DocEventAdmin) -admin.site.register(SubmissionDocEvent, DocEventAdmin) admin.site.register(EditedAuthorsDocEvent, DocEventAdmin) +admin.site.register(EditedRfcAuthorsDocEvent, DocEventAdmin) admin.site.register(IanaExpertDocEvent, DocEventAdmin) -class DeletedEventAdmin(admin.ModelAdmin): - list_display = ['id', 'content_type', 'json', 'by', 'time'] - list_filter = ['time'] - raw_id_fields = ['content_type', 'by'] -admin.site.register(DeletedEvent, DeletedEventAdmin) - class BallotPositionDocEventAdmin(DocEventAdmin): - raw_id_fields = ["doc", "by", "balloter", "ballot"] + raw_id_fields = DocEventAdmin.raw_id_fields + ["balloter", "ballot"] admin.site.register(BallotPositionDocEvent, BallotPositionDocEventAdmin) - -class IRSGBallotDocEventAdmin(DocEventAdmin): - raw_id_fields = ["doc", "by"] -admin.site.register(IRSGBallotDocEvent, IRSGBallotDocEventAdmin) class BofreqEditorDocEventAdmin(DocEventAdmin): - raw_id_fields = ["doc", "by", "editors" ] + raw_id_fields = DocEventAdmin.raw_id_fields + ["editors"] admin.site.register(BofreqEditorDocEvent, BofreqEditorDocEventAdmin) class BofreqResponsibleDocEventAdmin(DocEventAdmin): - raw_id_fields = ["doc", "by", "responsible" ] + raw_id_fields = DocEventAdmin.raw_id_fields + ["responsible"] admin.site.register(BofreqResponsibleDocEvent, BofreqResponsibleDocEventAdmin) +class ReviewRequestDocEventAdmin(DocEventAdmin): + raw_id_fields = DocEventAdmin.raw_id_fields + ["review_request"] +admin.site.register(ReviewRequestDocEvent, ReviewRequestDocEventAdmin) + +class ReviewAssignmentDocEventAdmin(DocEventAdmin): + raw_id_fields = DocEventAdmin.raw_id_fields + ["review_assignment"] +admin.site.register(ReviewAssignmentDocEvent, ReviewAssignmentDocEventAdmin) + +class AddedMessageEventAdmin(DocEventAdmin): + raw_id_fields = DocEventAdmin.raw_id_fields + ["message"] +admin.site.register(AddedMessageEvent, AddedMessageEventAdmin) + +class SubmissionDocEventAdmin(DocEventAdmin): + raw_id_fields = DocEventAdmin.raw_id_fields + ["submission"] +admin.site.register(SubmissionDocEvent, SubmissionDocEventAdmin) + class DocumentUrlAdmin(admin.ModelAdmin): list_display = ['id', 'doc', 'tag', 'url', 'desc', ] search_fields = ['doc__name', 'url', ] @@ -219,3 +222,28 @@ class DocExtResourceAdmin(admin.ModelAdmin): search_fields = ['doc__name', 'value', 'display_name', 'name__slug',] raw_id_fields = ['doc', ] admin.site.register(DocExtResource, DocExtResourceAdmin) + +class StoredObjectAdmin(admin.ModelAdmin): + list_display = ['store', 'name', 'doc_name', 'modified', 'is_deleted'] + list_filter = [ + 'store', + ('modified', DateRangeQuickSelectListFilterBuilder()), + ('deleted', DateRangeQuickSelectListFilterBuilder()), + ] + search_fields = ['name', 'doc_name', 'doc_rev'] + list_display_links = ['name'] + + @admin.display(boolean=True, description="Deleted?", ordering="deleted") + def is_deleted(self, instance): + return instance.deleted is not None + + +admin.site.register(StoredObject, StoredObjectAdmin) + +class RfcAuthorAdmin(admin.ModelAdmin): + # the email field in the list_display/readonly_fields works through a @property + list_display = ['id', 'document', 'titlepage_name', 'person', 'email', 'affiliation', 'country', 'order'] + search_fields = ['document__name', 'titlepage_name', 'person__name', 'person__email__address', 'affiliation', 'country'] + raw_id_fields = ["document", "person"] + readonly_fields = ["email"] +admin.site.register(RfcAuthor, RfcAuthorAdmin) diff --git a/ietf/doc/api.py b/ietf/doc/api.py new file mode 100644 index 0000000000..73fff6b27f --- /dev/null +++ b/ietf/doc/api.py @@ -0,0 +1,213 @@ +# Copyright The IETF Trust 2024-2026, All Rights Reserved +"""Doc API implementations""" + +from django.db.models import ( + BooleanField, + Count, + OuterRef, + Prefetch, + Q, + QuerySet, + Subquery, +) +from django.db.models.functions import TruncDate +from django_filters import rest_framework as filters +from rest_framework import filters as drf_filters +from rest_framework.mixins import ListModelMixin, RetrieveModelMixin +from rest_framework.pagination import LimitOffsetPagination +from rest_framework.viewsets import GenericViewSet + +from ietf.group.models import Group +from ietf.name.models import StreamName, DocTypeName +from ietf.utils.timezone import RPC_TZINFO +from .models import ( + Document, + DocEvent, + RelatedDocument, + DocumentAuthor, + SUBSERIES_DOC_TYPE_IDS, +) +from .serializers import ( + RfcMetadataSerializer, + RfcStatus, + RfcSerializer, + SubseriesDocSerializer, +) + + +class RfcLimitOffsetPagination(LimitOffsetPagination): + default_limit = 10 + max_limit = 500 + + +class NumberInFilter(filters.BaseInFilter, filters.NumberFilter): + """Filter against a comma-separated list of numbers""" + pass + + +class RfcFilter(filters.FilterSet): + published = filters.DateFromToRangeFilter() + stream = filters.ModelMultipleChoiceFilter( + queryset=StreamName.objects.filter(used=True) + ) + number = NumberInFilter( + field_name="rfc_number" + ) + group = filters.ModelMultipleChoiceFilter( + queryset=Group.objects.all(), + field_name="group__acronym", + to_field_name="acronym", + ) + area = filters.ModelMultipleChoiceFilter( + queryset=Group.objects.areas(), + field_name="group__parent__acronym", + to_field_name="acronym", + ) + status = filters.MultipleChoiceFilter( + choices=[(slug, slug) for slug in RfcStatus.status_slugs], + method=RfcStatus.filter, + ) + sort = filters.OrderingFilter( + fields=( + ("rfc_number", "number"), # ?sort=number / ?sort=-number + ("published", "published"), # ?sort=published / ?sort=-published + ), + ) + + +class PrefetchRelatedDocument(Prefetch): + """Prefetch via a RelatedDocument + + Prefetches following RelatedDocument relationships to other docs. By default, includes + those for which the current RFC is the `source`. If `reverse` is True, includes those + for which it is the `target` instead. Defaults to only "rfc" documents. + """ + + @staticmethod + def _get_queryset(relationship_id, reverse, doc_type_ids): + """Get queryset to use for the prefetch""" + if isinstance(doc_type_ids, str): + doc_type_ids = (doc_type_ids,) + + return RelatedDocument.objects.filter( + **{ + "relationship_id": relationship_id, + f"{'source' if reverse else 'target'}__type_id__in": doc_type_ids, + } + ).select_related("source" if reverse else "target") + + def __init__(self, to_attr, relationship_id, reverse=False, doc_type_ids="rfc"): + super().__init__( + lookup="targets_related" if reverse else "relateddocument_set", + queryset=self._get_queryset(relationship_id, reverse, doc_type_ids), + to_attr=to_attr, + ) + + +def augment_rfc_queryset(queryset: QuerySet[Document]): + return ( + queryset.select_related("std_level", "stream") + .prefetch_related( + Prefetch( + "group", + Group.objects.select_related("parent"), + ), + Prefetch( + "documentauthor_set", + DocumentAuthor.objects.select_related("email", "person"), + ), + PrefetchRelatedDocument( + to_attr="drafts", + relationship_id="became_rfc", + doc_type_ids="draft", + reverse=True, + ), + PrefetchRelatedDocument(to_attr="obsoletes", relationship_id="obs"), + PrefetchRelatedDocument( + to_attr="obsoleted_by", relationship_id="obs", reverse=True + ), + PrefetchRelatedDocument(to_attr="updates", relationship_id="updates"), + PrefetchRelatedDocument( + to_attr="updated_by", relationship_id="updates", reverse=True + ), + PrefetchRelatedDocument( + to_attr="subseries", + relationship_id="contains", + reverse=True, + doc_type_ids=SUBSERIES_DOC_TYPE_IDS, + ), + ) + .annotate( + published_datetime=Subquery( + DocEvent.objects.filter( + doc_id=OuterRef("pk"), + type="published_rfc", + ) + .order_by("-time") + .values("time")[:1] + ), + ) + .annotate(published=TruncDate("published_datetime", tzinfo=RPC_TZINFO)) + .annotate( + # Count of "verified-errata" tags will be 1 or 0, convert to Boolean + has_errata=Count( + "tags", + filter=Q( + tags__slug="verified-errata", + ), + output_field=BooleanField(), + ) + ) + ) + + +class RfcViewSet(ListModelMixin, RetrieveModelMixin, GenericViewSet): + api_key_endpoint = "ietf.api.red_api" # matches prefix in ietf/api/urls.py + lookup_field = "rfc_number" + queryset = augment_rfc_queryset( + Document.objects.filter(type_id="rfc", rfc_number__isnull=False) + ).order_by("-rfc_number") + + pagination_class = RfcLimitOffsetPagination + filter_backends = [filters.DjangoFilterBackend, drf_filters.SearchFilter] + filterset_class = RfcFilter + search_fields = ["title", "abstract"] + + def get_serializer_class(self): + if self.action == "retrieve": + return RfcSerializer + return RfcMetadataSerializer + + +class PrefetchSubseriesContents(Prefetch): + def __init__(self, to_attr): + super().__init__( + lookup="relateddocument_set", + queryset=RelatedDocument.objects.filter( + relationship_id="contains", + target__type_id="rfc", + ).prefetch_related( + Prefetch( + "target", + queryset=augment_rfc_queryset(Document.objects.all()), + ) + ), + to_attr=to_attr, + ) + + +class SubseriesFilter(filters.FilterSet): + type = filters.ModelMultipleChoiceFilter( + queryset=DocTypeName.objects.filter(pk__in=SUBSERIES_DOC_TYPE_IDS) + ) + + +class SubseriesViewSet(ListModelMixin, RetrieveModelMixin, GenericViewSet): + api_key_endpoint = "ietf.api.red_api" # matches prefix in ietf/api/urls.py + lookup_field = "name" + serializer_class = SubseriesDocSerializer + queryset = Document.objects.subseries_docs().prefetch_related( + PrefetchSubseriesContents(to_attr="contents") + ) + filter_backends = [filters.DjangoFilterBackend] + filterset_class = SubseriesFilter diff --git a/ietf/doc/expire.py b/ietf/doc/expire.py index 999d5022df..d42af628f8 100644 --- a/ietf/doc/expire.py +++ b/ietf/doc/expire.py @@ -1,8 +1,10 @@ # Copyright The IETF Trust 2010-2020, All Rights Reserved # -*- coding: utf-8 -*- -# expiry of Internet Drafts +# expiry of Internet-Drafts +import debug # pyflakes:ignore + from django.conf import settings from django.utils import timezone @@ -11,12 +13,13 @@ from typing import List, Optional # pyflakes:ignore +from ietf.doc.storage_utils import exists_in_storage, remove_from_storage +from ietf.doc.utils import update_action_holders from ietf.utils import log from ietf.utils.mail import send_mail -from ietf.doc.models import Document, DocEvent, State, IESG_SUBSTATE_TAGS +from ietf.doc.models import Document, DocEvent, State from ietf.person.models import Person from ietf.meeting.models import Meeting -from ietf.doc.utils import add_state_change_event, update_action_holders from ietf.mailtrigger.utils import gather_address_lists from ietf.utils.timezone import date_today, datetime_today, DEADLINE_TZINFO @@ -34,23 +37,47 @@ def expirable_drafts(queryset=None): # Populate this first time through (but after django has been set up) if nonexpirable_states is None: - # all IESG states except I-D Exists, AD Watching, and Dead block expiry - nonexpirable_states = list(State.objects.filter(used=True, type="draft-iesg").exclude(slug__in=("idexists","watching", "dead"))) + # all IESG states except I-D Exists and Dead block expiry + nonexpirable_states = list( + State.objects.filter(used=True, type="draft-iesg").exclude( + slug__in=("idexists", "dead") + ) + ) # sent to RFC Editor and RFC Published block expiry (the latter # shouldn't be possible for an active draft, though) - nonexpirable_states += list(State.objects.filter(used=True, type__in=("draft-stream-iab", "draft-stream-irtf", "draft-stream-ise"), slug__in=("rfc-edit", "pub"))) + nonexpirable_states += list( + State.objects.filter( + used=True, + type__in=( + "draft-stream-iab", + "draft-stream-irtf", + "draft-stream-ise", + "draft-stream-editorial", + ), + slug__in=("rfc-edit", "pub"), + ) + ) # other IRTF states that block expiration - nonexpirable_states += list(State.objects.filter(used=True, type_id="draft-stream-irtf", slug__in=("irsgpoll", "iesg-rev",))) - - return queryset.filter( - states__type="draft", states__slug="active" - ).exclude( - expires=None - ).exclude( - states__in=nonexpirable_states - ).exclude( - tags="rfc-rev" # under review by the RFC Editor blocks expiry - ).distinct() + nonexpirable_states += list( + State.objects.filter( + used=True, + type_id="draft-stream-irtf", + slug__in=( + "irsgpoll", + "iesg-rev", + ), + ) + ) + + return ( + queryset.filter(states__type="draft", states__slug="active") + .exclude(expires=None) + .exclude(states__in=nonexpirable_states) + .exclude( + tags="rfc-rev" # under review by the RFC Editor blocks expiry + ) + .distinct() + ) def get_soon_to_expire_drafts(days_of_warning): @@ -139,16 +166,32 @@ def move_file(f): if os.path.exists(src): try: + # ghostlinkd would keep this in the combined all archive since it would + # be sourced from a different place. But when ghostlinkd is removed, nothing + # new is needed here - the file will already exist in the combined archive shutil.move(src, dst) except IOError as e: if "No such file or directory" in str(e): pass else: raise - + + def remove_ftp_copy(f): + mark = Path(settings.FTP_DIR) / "internet-drafts" / f + if mark.exists(): + mark.unlink() + + def remove_from_active_draft_storage(file): + # Assumes the glob will never find a file with no suffix + ext = file.suffix[1:] + remove_from_storage("active-draft", f"{ext}/{file.name}", warn_if_missing=False) + + # Note that the object is already in the "draft" storage. src_dir = Path(settings.INTERNET_DRAFT_PATH) for file in src_dir.glob("%s-%s.*" % (doc.name, rev)): move_file(str(file.name)) + remove_ftp_copy(str(file.name)) + remove_from_active_draft_storage(file) def expire_draft(doc): # clean up files @@ -158,28 +201,15 @@ def expire_draft(doc): events = [] - # change the state - if doc.latest_event(type='started_iesg_process'): - new_state = State.objects.get(used=True, type="draft-iesg", slug="dead") - prev_state = doc.get_state(new_state.type_id) - prev_tags = doc.tags.filter(slug__in=IESG_SUBSTATE_TAGS) - if new_state != prev_state: - doc.set_state(new_state) - doc.tags.remove(*prev_tags) - e = add_state_change_event(doc, system, prev_state, new_state, prev_tags=prev_tags, new_tags=[]) - if e: - events.append(e) - e = update_action_holders(doc, prev_state, new_state, prev_tags=prev_tags, new_tags=[]) - if e: - events.append(e) - events.append(DocEvent.objects.create(doc=doc, rev=doc.rev, by=system, type="expired_document", desc="Document has expired")) + prev_draft_state=doc.get_state("draft") doc.set_state(State.objects.get(used=True, type="draft", slug="expired")) + events.append(update_action_holders(doc, prev_draft_state, doc.get_state("draft"),[],[])) doc.save_with_history(events) def clean_up_draft_files(): - """Move unidentified and old files out of the Internet Draft directory.""" + """Move unidentified and old files out of the Internet-Draft directory.""" cut_off = date_today(DEADLINE_TZINFO) pattern = os.path.join(settings.INTERNET_DRAFT_PATH, "draft-*.*") @@ -213,8 +243,19 @@ def splitext(fn): filename, revision = match.groups() def move_file_to(subdir): + # Similar to move_draft_files_to_archive shutil.move(path, os.path.join(settings.INTERNET_DRAFT_ARCHIVE_DIR, subdir, basename)) + mark = Path(settings.FTP_DIR) / "internet-drafts" / basename + if mark.exists(): + mark.unlink() + if ext: + # Note that we're not moving these strays anywhere - the assumption + # is that the active-draft blobstore will not get strays. + # See, however, the note about "major system failures" at "unknown_ids" + blobname = f"{ext[1:]}/{basename}" + if exists_in_storage("active-draft", blobname): + remove_from_storage("active-draft", blobname) try: doc = Document.objects.get(name=filename, rev=revision) @@ -229,4 +270,6 @@ def move_file_to(subdir): move_file_to("") except Document.DoesNotExist: + # All uses of this past 2014 seem related to major system failures. move_file_to("unknown_ids") + diff --git a/ietf/doc/factories.py b/ietf/doc/factories.py index a8b7b4656b..1a178c6f31 100644 --- a/ietf/doc/factories.py +++ b/ietf/doc/factories.py @@ -1,4 +1,4 @@ -# Copyright The IETF Trust 2016-2020, All Rights Reserved +# Copyright The IETF Trust 2016-2023, All Rights Reserved # -*- coding: utf-8 -*- @@ -7,14 +7,14 @@ import factory.fuzzy import datetime -from typing import Optional # pyflakes:ignore +from typing import Any # pyflakes:ignore from django.conf import settings from django.utils import timezone -from ietf.doc.models import ( Document, DocEvent, NewRevisionDocEvent, DocAlias, State, DocumentAuthor, +from ietf.doc.models import ( Document, DocEvent, NewRevisionDocEvent, State, DocumentAuthor, StateDocEvent, BallotPositionDocEvent, BallotDocEvent, BallotType, IRSGBallotDocEvent, TelechatDocEvent, - DocumentActionHolder, BofreqEditorDocEvent, BofreqResponsibleDocEvent, DocExtResource ) + DocumentActionHolder, BofreqEditorDocEvent, BofreqResponsibleDocEvent, DocExtResource, RfcAuthor ) from ietf.group.models import Group from ietf.person.factories import PersonFactory from ietf.group.factories import RoleFactory @@ -23,7 +23,6 @@ from ietf.utils.timezone import date_today - def draft_name_generator(type_id,group,n): return '%s-%s-%s-%s%d'%( type_id, @@ -36,14 +35,18 @@ def draft_name_generator(type_id,group,n): class BaseDocumentFactory(factory.django.DjangoModelFactory): class Meta: model = Document + skip_postgeneration_save = True + # n.b., a few attributes are typed as Any so mypy won't complain when we override in subclasses title = factory.Faker('sentence',nb_words=5) - abstract = factory.Faker('paragraph', nb_sentences=5) + abstract: Any = factory.Faker('paragraph', nb_sentences=5) rev = '00' - std_level_id = None # type: Optional[str] + std_level_id: Any = None intended_std_level_id = None time = timezone.now() - expires = factory.LazyAttribute(lambda o: o.time+datetime.timedelta(days=settings.INTERNET_DRAFT_DAYS_TO_EXPIRE)) + expires: Any = factory.LazyAttribute( + lambda o: o.time+datetime.timedelta(days=settings.INTERNET_DRAFT_DAYS_TO_EXPIRE) + ) pages = factory.fuzzy.FuzzyInteger(2,400) @@ -51,16 +54,11 @@ class Meta: def name(self, n): return draft_name_generator(self.type_id,self.group,n) - newrevisiondocevent = factory.RelatedFactory('ietf.doc.factories.NewRevisionDocEventFactory','doc') - @factory.post_generation - def other_aliases(obj, create, extracted, **kwargs): # pylint: disable=no-self-argument - alias = DocAliasFactory(name=obj.name) - alias.docs.add(obj) - if create and extracted: - for name in extracted: - alias = DocAliasFactory(name=name) - alias.docs.add(obj) + def newrevisiondocevent(obj, create, extracted, **kwargs): # pylint: disable=no-self-argument + if create: + if obj.type_id != "rfc": + NewRevisionDocEventFactory(doc=obj) @factory.post_generation def states(obj, create, extracted, **kwargs): # pylint: disable=no-self-argument @@ -83,13 +81,7 @@ def authors(obj, create, extracted, **kwargs): # pylint: disable=no-self-argumen def relations(obj, create, extracted, **kwargs): # pylint: disable=no-self-argument if create and extracted: for (rel_id, doc) in extracted: - if isinstance(doc, Document): - docalias = doc.docalias.first() - elif isinstance(doc, DocAlias): - docalias = doc - else: - continue - obj.relateddocument_set.create(relationship_id=rel_id, target=docalias) + obj.relateddocument_set.create(relationship_id=rel_id, target=doc) @factory.post_generation def create_revisions(obj, create, extracted, **kwargs): # pylint: disable=no-self-argument @@ -119,10 +111,12 @@ class DocumentFactory(BaseDocumentFactory): group = factory.SubFactory('ietf.group.factories.GroupFactory',acronym='none') -class IndividualDraftFactory(BaseDocumentFactory): - - type_id = 'draft' - group = factory.SubFactory('ietf.group.factories.GroupFactory',acronym='none') +class RfcFactory(BaseDocumentFactory): + type_id = "rfc" + rev = "" + rfc_number = factory.Sequence(lambda n: n + 1000) + name = factory.LazyAttribute(lambda o: f"rfc{o.rfc_number:d}") + expires = None @factory.post_generation def states(obj, create, extracted, **kwargs): @@ -131,15 +125,14 @@ def states(obj, create, extracted, **kwargs): if extracted: for (state_type_id,state_slug) in extracted: obj.set_state(State.objects.get(type_id=state_type_id,slug=state_slug)) - if not obj.get_state('draft-iesg'): - obj.set_state(State.objects.get(type_id='draft-iesg',slug='idexists')) else: - obj.set_state(State.objects.get(type_id='draft',slug='active')) - obj.set_state(State.objects.get(type_id='draft-iesg',slug='idexists')) + obj.set_state(State.objects.get(type_id='rfc',slug='published')) -class IndividualRfcFactory(IndividualDraftFactory): - alias2 = factory.RelatedFactory('ietf.doc.factories.DocAliasFactory','document',name=factory.Sequence(lambda n: 'rfc%04d'%(n+1000))) +class IndividualDraftFactory(BaseDocumentFactory): + + type_id = 'draft' + group = factory.SubFactory('ietf.group.factories.GroupFactory',acronym='none') @factory.post_generation def states(obj, create, extracted, **kwargs): @@ -148,17 +141,17 @@ def states(obj, create, extracted, **kwargs): if extracted: for (state_type_id,state_slug) in extracted: obj.set_state(State.objects.get(type_id=state_type_id,slug=state_slug)) + if not obj.get_state('draft-iesg'): + obj.set_state(State.objects.get(type_id='draft-iesg',slug='idexists')) else: - obj.set_state(State.objects.get(type_id='draft',slug='rfc')) + obj.set_state(State.objects.get(type_id='draft',slug='active')) + obj.set_state(State.objects.get(type_id='draft-iesg',slug='idexists')) - @factory.post_generation - def reset_canonical_name(obj, create, extracted, **kwargs): - if hasattr(obj, '_canonical_name'): - del obj._canonical_name - return None +class IndividualRfcFactory(RfcFactory): + group = factory.SubFactory('ietf.group.factories.GroupFactory',acronym='none') -class WgDraftFactory(BaseDocumentFactory): +class WgDraftFactory(BaseDocumentFactory): type_id = 'draft' group = factory.SubFactory('ietf.group.factories.GroupFactory',type_id='wg') stream_id = 'ietf' @@ -177,30 +170,12 @@ def states(obj, create, extracted, **kwargs): obj.set_state(State.objects.get(type_id='draft-stream-ietf',slug='wg-doc')) obj.set_state(State.objects.get(type_id='draft-iesg',slug='idexists')) -class WgRfcFactory(WgDraftFactory): - - alias2 = factory.RelatedFactory('ietf.doc.factories.DocAliasFactory','document',name=factory.Sequence(lambda n: 'rfc%04d'%(n+1000))) +class WgRfcFactory(RfcFactory): + group = factory.SubFactory('ietf.group.factories.GroupFactory',type_id='wg') + stream_id = 'ietf' std_level_id = 'ps' - @factory.post_generation - def states(obj, create, extracted, **kwargs): - if not create: - return - if extracted: - for (state_type_id,state_slug) in extracted: - obj.set_state(State.objects.get(type_id=state_type_id,slug=state_slug)) - if not obj.get_state('draft-iesg'): - obj.set_state(State.objects.get(type_id='draft-iesg', slug='pub')) - else: - obj.set_state(State.objects.get(type_id='draft',slug='rfc')) - obj.set_state(State.objects.get(type_id='draft-iesg', slug='pub')) - - @factory.post_generation - def reset_canonical_name(obj, create, extracted, **kwargs): - if hasattr(obj, '_canonical_name'): - del obj._canonical_name - return None class RgDraftFactory(BaseDocumentFactory): @@ -223,34 +198,11 @@ def states(obj, create, extracted, **kwargs): obj.set_state(State.objects.get(type_id='draft-iesg',slug='idexists')) -class RgRfcFactory(RgDraftFactory): - - alias2 = factory.RelatedFactory('ietf.doc.factories.DocAliasFactory','document',name=factory.Sequence(lambda n: 'rfc%04d'%(n+1000))) - +class RgRfcFactory(RfcFactory): + group = factory.SubFactory('ietf.group.factories.GroupFactory',type_id='rg') + stream_id = 'irtf' std_level_id = 'inf' - @factory.post_generation - def states(obj, create, extracted, **kwargs): - if not create: - return - if extracted: - for (state_type_id,state_slug) in extracted: - obj.set_state(State.objects.get(type_id=state_type_id,slug=state_slug)) - if not obj.get_state('draft-stream-irtf'): - obj.set_state(State.objects.get(type_id='draft-stream-irtf', slug='pub')) - if not obj.get_state('draft-iesg'): - obj.set_state(State.objects.get(type_id='draft-iesg',slug='idexists')) - else: - obj.set_state(State.objects.get(type_id='draft',slug='rfc')) - obj.set_state(State.objects.get(type_id='draft-stream-irtf', slug='pub')) - obj.set_state(State.objects.get(type_id='draft-iesg',slug='idexists')) - - @factory.post_generation - def reset_canonical_name(obj, create, extracted, **kwargs): - if hasattr(obj, '_canonical_name'): - del obj._canonical_name - return None - class CharterFactory(BaseDocumentFactory): @@ -279,7 +231,7 @@ def changes_status_of(obj, create, extracted, **kwargs): for (rel, target) in extracted: obj.relateddocument_set.create(relationship_id=rel,target=target) else: - obj.relateddocument_set.create(relationship_id='tobcp', target=WgRfcFactory().docalias.first()) + obj.relateddocument_set.create(relationship_id='tobcp', target=WgRfcFactory()) @factory.post_generation def states(obj, create, extracted, **kwargs): @@ -306,9 +258,9 @@ def review_of(obj, create, extracted, **kwargs): if not create: return if extracted: - obj.relateddocument_set.create(relationship_id='conflrev',target=extracted.docalias.first()) + obj.relateddocument_set.create(relationship_id='conflrev',target=extracted) else: - obj.relateddocument_set.create(relationship_id='conflrev',target=DocumentFactory(name=obj.name.replace('conflict-review-','draft-'),type_id='draft',group=Group.objects.get(type_id='individ')).docalias.first()) + obj.relateddocument_set.create(relationship_id='conflrev',target=DocumentFactory(name=obj.name.replace('conflict-review-','draft-'),type_id='draft',group=Group.objects.get(type_id='individ'))) @factory.post_generation @@ -327,30 +279,13 @@ class ReviewFactory(BaseDocumentFactory): name = factory.LazyAttribute(lambda o: 'review-doesnotexist-00-%s-%s'%(o.group.acronym,date_today().isoformat())) group = factory.SubFactory('ietf.group.factories.GroupFactory',type_id='review') -class DocAliasFactory(factory.django.DjangoModelFactory): - class Meta: - model = DocAlias - - @factory.post_generation - def document(self, create, extracted, **kwargs): - if create and extracted: - self.docs.add(extracted) - - @factory.post_generation - def docs(self, create, extracted, **kwargs): - if create and extracted: - for doc in extracted: - if not doc in self.docs.all(): - self.docs.add(doc) - - class DocEventFactory(factory.django.DjangoModelFactory): class Meta: model = DocEvent type = 'added_comment' by = factory.SubFactory('ietf.person.factories.PersonFactory') - doc = factory.SubFactory(DocumentFactory) + doc: Any = factory.SubFactory(DocumentFactory) # `Any` to appease mypy when a subclass overrides doc desc = factory.Faker('sentence',nb_words=6) @factory.lazy_attribute @@ -376,9 +311,16 @@ class Meta: def desc(self): return 'New version available %s-%s'%(self.doc.name,self.rev) +class PublishedRfcDocEventFactory(DocEventFactory): + class Meta: + model = DocEvent + type = "published_rfc" + doc = factory.SubFactory(WgRfcFactory) + class StateDocEventFactory(DocEventFactory): class Meta: model = StateDocEvent + skip_postgeneration_save = True type = 'changed_state' state_type_id = 'draft-iesg' @@ -446,12 +388,25 @@ class Meta: country = factory.Faker('country') order = factory.LazyAttribute(lambda o: o.document.documentauthor_set.count() + 1) +class RfcAuthorFactory(factory.django.DjangoModelFactory): + class Meta: + model = RfcAuthor + + document = factory.SubFactory(DocumentFactory) + titlepage_name = factory.LazyAttribute( + lambda obj: " ".join([obj.person.initials(), obj.person.last_name()]) + ) + person = factory.SubFactory('ietf.person.factories.PersonFactory') + affiliation = factory.Faker('company') + order = factory.LazyAttribute(lambda o: o.document.rfcauthor_set.count() + 1) + class WgDocumentAuthorFactory(DocumentAuthorFactory): document = factory.SubFactory(WgDraftFactory) class BofreqEditorDocEventFactory(DocEventFactory): class Meta: model = BofreqEditorDocEvent + skip_postgeneration_save = True type = "changed_editors" doc = factory.SubFactory('ietf.doc.factories.BofreqFactory') @@ -466,10 +421,12 @@ def editors(obj, create, extracted, **kwargs): else: obj.editors.set(PersonFactory.create_batch(3)) obj.desc = f'Changed editors to {", ".join(obj.editors.values_list("name",flat=True)) or "(None)"}' + obj.save() class BofreqResponsibleDocEventFactory(DocEventFactory): class Meta: model = BofreqResponsibleDocEvent + skip_postgeneration_save = True type = "changed_responsible" doc = factory.SubFactory('ietf.doc.factories.BofreqFactory') @@ -484,7 +441,8 @@ def responsible(obj, create, extracted, **kwargs): else: ad = RoleFactory(group__type_id='area',name_id='ad').person obj.responsible.set([ad]) - obj.desc = f'Changed responsible leadership to {", ".join(obj.responsible.values_list("name",flat=True)) or "(None)"}' + obj.desc = f'Changed responsible leadership to {", ".join(obj.responsible.values_list("name",flat=True)) or "(None)"}' + obj.save() class BofreqFactory(BaseDocumentFactory): type_id = 'bofreq' @@ -531,3 +489,80 @@ class DocExtResourceFactory(factory.django.DjangoModelFactory): class Meta: model = DocExtResource +class EditorialDraftFactory(BaseDocumentFactory): + + type_id = 'draft' + group = factory.SubFactory('ietf.group.factories.GroupFactory',acronym='rswg', type_id='edwg') + stream_id = 'editorial' + + @factory.post_generation + def states(obj, create, extracted, **kwargs): + if not create: + return + if extracted: + for (state_type_id,state_slug) in extracted: + obj.set_state(State.objects.get(type_id=state_type_id,slug=state_slug)) + if not obj.get_state('draft-iesg'): + obj.set_state(State.objects.get(type_id='draft-iesg',slug='idexists')) + else: + obj.set_state(State.objects.get(type_id='draft',slug='active')) + obj.set_state(State.objects.get(type_id='draft-stream-editorial',slug='active')) + obj.set_state(State.objects.get(type_id='draft-iesg',slug='idexists')) + +class EditorialRfcFactory(RgRfcFactory): + pass + +class StatementFactory(BaseDocumentFactory): + type_id = "statement" + title = factory.Faker("sentence") + group = factory.SubFactory("ietf.group.factories.GroupFactory", acronym="iab") + + name = factory.LazyAttribute( + lambda o: "statement-%s-%s" % (xslugify(o.group.acronym), xslugify(o.title)) + ) + uploaded_filename = factory.LazyAttribute(lambda o: f"{o.name}-{o.rev}.md") + + published_statement_event = factory.RelatedFactory( + "ietf.doc.factories.DocEventFactory", + "doc", + type="published_statement", + time=timezone.now() - datetime.timedelta(days=1), + ) + + @factory.post_generation + def states(obj, create, extracted, **kwargs): + if not create: + return + if extracted: + for state_type_id, state_slug in extracted: + obj.set_state(State.objects.get(type_id=state_type_id, slug=state_slug)) + else: + obj.set_state(State.objects.get(type_id="statement", slug="active")) + +class SubseriesFactory(factory.django.DjangoModelFactory): + class Meta: + model = Document + skip_postgeneration_save = True + + @factory.lazy_attribute_sequence + def name(self, n): + return f"{self.type_id}{n}" + + @factory.post_generation + def contains(obj, create, extracted, **kwargs): + if not create: + return + if extracted: + for doc in extracted: + obj.relateddocument_set.create(relationship_id="contains",target=doc) + else: + obj.relateddocument_set.create(relationship_id="contains", target=RfcFactory()) + +class BcpFactory(SubseriesFactory): + type_id="bcp" + +class StdFactory(SubseriesFactory): + type_id="std" + +class FyiFactory(SubseriesFactory): + type_id="fyi" diff --git a/ietf/doc/feeds.py b/ietf/doc/feeds.py index 7885e75e31..0269906fcf 100644 --- a/ietf/doc/feeds.py +++ b/ietf/doc/feeds.py @@ -1,15 +1,20 @@ -# Copyright The IETF Trust 2007-2020, All Rights Reserved -# -*- coding: utf-8 -*- +# Copyright The IETF Trust 2007-2026, All Rights Reserved +import debug # pyflakes:ignore import datetime import unicodedata +from django.conf import settings from django.contrib.syndication.views import Feed, FeedDoesNotExist from django.utils.feedgenerator import Atom1Feed, Rss201rev2Feed from django.urls import reverse as urlreverse -from django.template.defaultfilters import truncatewords, truncatewords_html, date as datefilter -from django.template.defaultfilters import linebreaks # type: ignore +from django.template.defaultfilters import ( + truncatewords, + truncatewords_html, + date as datefilter, +) +from django.template.defaultfilters import linebreaks # type: ignore from django.utils import timezone from django.utils.html import strip_tags @@ -21,17 +26,17 @@ def strip_control_characters(s): """Remove Unicode control / non-printing characters from a string""" - replacement_char = unicodedata.lookup('REPLACEMENT CHARACTER') - return ''.join( - replacement_char if unicodedata.category(c)[0] == 'C' else c - for c in s + replacement_char = unicodedata.lookup("REPLACEMENT CHARACTER") + return "".join( + replacement_char if unicodedata.category(c)[0] == "C" else c for c in s ) + class DocumentChangesFeed(Feed): feed_type = Atom1Feed def get_object(self, request, name): - return Document.objects.get(docalias__name=name) + return Document.objects.get(name=name) def title(self, obj): return "Changes for %s" % obj.display_name() @@ -39,25 +44,37 @@ def title(self, obj): def link(self, obj): if obj is None: raise FeedDoesNotExist - return urlreverse('ietf.doc.views_doc.document_history', kwargs=dict(name=obj.canonical_name())) + return urlreverse( + "ietf.doc.views_doc.document_history", + kwargs=dict(name=obj.name), + ) def subtitle(self, obj): return "History of change entries for %s." % obj.display_name() def items(self, obj): - events = obj.docevent_set.all().order_by("-time","-id") + events = ( + obj.docevent_set.all() + .order_by("-time", "-id") + .select_related("by", "newrevisiondocevent", "submissiondocevent") + ) augment_events_with_revision(obj, events) return events def item_title(self, item): - return strip_control_characters("[%s] %s [rev. %s]" % ( - item.by, - truncatewords(strip_tags(item.desc), 15), - item.rev, - )) + return strip_control_characters( + "[%s] %s [rev. %s]" + % ( + item.by, + truncatewords(strip_tags(item.desc), 15), + item.rev, + ) + ) def item_description(self, item): - return strip_control_characters(truncatewords_html(format_textarea(item.desc), 20)) + return strip_control_characters( + truncatewords_html(format_textarea(item.desc), 20) + ) def item_pubdate(self, item): return item.time @@ -66,17 +83,28 @@ def item_author_name(self, item): return str(item.by) def item_link(self, item): - return urlreverse('ietf.doc.views_doc.document_history', kwargs=dict(name=item.doc.canonical_name())) + "#history-%s" % item.pk + return ( + urlreverse( + "ietf.doc.views_doc.document_history", + kwargs=dict(name=item.doc.name), + ) + + "#history-%s" % item.pk + ) + class InLastCallFeed(Feed): title = "Documents in Last Call" subtitle = "Announcements for documents in last call." feed_type = Atom1Feed - author_name = 'IESG Secretary' + author_name = "IESG Secretary" link = "/doc/iesg/last-call/" def items(self): - docs = list(Document.objects.filter(type="draft", states=State.objects.get(type="draft-iesg", slug="lc"))) + docs = list( + Document.objects.filter( + type="draft", states=State.objects.get(type="draft-iesg", slug="lc") + ) + ) for d in docs: d.lc_event = d.latest_event(LastCallDocEvent, type="sent_last_call") @@ -86,9 +114,11 @@ def items(self): return docs def item_title(self, item): - return "%s (%s - %s)" % (item.name, - datefilter(item.lc_event.time, "F j"), - datefilter(item.lc_event.expires, "F j, Y")) + return "%s (%s - %s)" % ( + item.name, + datefilter(item.lc_event.time, "F j"), + datefilter(item.lc_event.expires, "F j, Y"), + ) def item_description(self, item): return strip_control_characters(linebreaks(item.lc_event.desc)) @@ -96,33 +126,55 @@ def item_description(self, item): def item_pubdate(self, item): return item.lc_event.time + class Rss201WithNamespacesFeed(Rss201rev2Feed): def root_attributes(self): attrs = super(Rss201WithNamespacesFeed, self).root_attributes() - attrs['xmlns:dcterms'] = 'http://purl.org/dc/terms/' - attrs['xmlns:media'] = 'http://search.yahoo.com/mrss/' - attrs['xmlns:xsi'] = 'http://www.w3.org/2001/XMLSchema-instance' + attrs["xmlns:dcterms"] = "http://purl.org/dc/terms/" + attrs["xmlns:media"] = "http://search.yahoo.com/mrss/" + attrs["xmlns:xsi"] = "http://www.w3.org/2001/XMLSchema-instance" return attrs def add_item_elements(self, handler, item): super(Rss201WithNamespacesFeed, self).add_item_elements(handler, item) - for element_name in ['abstract','accessRights', 'format', 'publisher',]: - dc_item_name = 'dcterms_%s' % element_name - dc_element_name = 'dcterms:%s' % element_name - attrs= {'xsi:type':'dcterms:local'} if element_name == 'publisher' else {} + for element_name in [ + "abstract", + "accessRights", + "format", + "publisher", + ]: + dc_item_name = "dcterms_%s" % element_name + dc_element_name = "dcterms:%s" % element_name + attrs = {"xsi:type": "dcterms:local"} if element_name == "publisher" else {} if dc_item_name in item and item[dc_item_name] is not None: - handler.addQuickElement(dc_element_name,item[dc_item_name],attrs) + handler.addQuickElement(dc_element_name, item[dc_item_name], attrs) + + if "doi" in item and item["doi"] is not None: + handler.addQuickElement( + "dcterms:identifier", item["doi"], {"xsi:type": "dcterms:doi"} + ) + if "doiuri" in item and item["doiuri"] is not None: + handler.addQuickElement( + "dcterms:identifier", item["doiuri"], {"xsi:type": "dcterms:uri"} + ) + + # TODO: consider using media:group + if "media_contents" in item and item["media_contents"] is not None: + for media_content in item["media_contents"]: + handler.startElement( + "media:content", + { + "url": media_content["url"], + "type": media_content["media_type"], + }, + ) + if "is_format_of" in media_content: + handler.addQuickElement( + "dcterms:isFormatOf", media_content["is_format_of"] + ) + handler.endElement("media:content") - if 'doi' in item and item['doi'] is not None: - handler.addQuickElement('dcterms:identifier',item['doi'],{'xsi:type':'dcterms:doi'}) - if 'doiuri' in item and item['doiuri'] is not None: - handler.addQuickElement('dcterms:identifier',item['doiuri'],{'xsi:type':'dcterms:uri'}) - - if 'media_content' in item and item['media_content'] is not None: - handler.startElement('media:content',{'url':item['media_content']['url'],'type':'text/plain'}) - handler.addQuickElement('dcterms:isFormatOf',item['media_content']['link_url']) - handler.endElement('media:content') class RfcFeed(Feed): feed_type = Rss201WithNamespacesFeed @@ -130,55 +182,98 @@ class RfcFeed(Feed): author_name = "RFC Editor" link = "https://www.rfc-editor.org/rfc-index2.html" - def get_object(self,request,year=None): + def get_object(self, request, year=None): self.year = year - + def items(self): if self.year: # Find published RFCs based on their official publication year start_of_year = datetime.datetime(int(self.year), 1, 1, tzinfo=RPC_TZINFO) - start_of_next_year = datetime.datetime(int(self.year) + 1, 1, 1, tzinfo=RPC_TZINFO) + start_of_next_year = datetime.datetime( + int(self.year) + 1, 1, 1, tzinfo=RPC_TZINFO + ) rfc_events = DocEvent.objects.filter( - type='published_rfc', + type="published_rfc", time__gte=start_of_year, time__lt=start_of_next_year, - ).order_by('-time') + ).order_by("-time") else: cutoff = timezone.now() - datetime.timedelta(days=8) - rfc_events = DocEvent.objects.filter(type='published_rfc',time__gte=cutoff).order_by('-time') + rfc_events = DocEvent.objects.filter( + type="published_rfc", time__gte=cutoff + ).order_by("-time") results = [(e.doc, e.time) for e in rfc_events] - for doc,time in results: + for doc, time in results: doc.publication_time = time - return [doc for doc,time in results] - + return [doc for doc, time in results] + def item_title(self, item): - return "%s : %s" % (item.canonical_name(),item.title) + return "%s : %s" % (item.name, item.title) def item_description(self, item): return item.abstract def item_link(self, item): - return "https://rfc-editor.org/info/%s"%item.canonical_name() + return "https://rfc-editor.org/info/%s" % item.name def item_pubdate(self, item): return item.publication_time def item_extra_kwargs(self, item): extra = super(RfcFeed, self).item_extra_kwargs(item) - extra.update({'dcterms_accessRights': 'gratis'}) - extra.update({'dcterms_format': 'text/html'}) - extra.update({'media_content': {'url': 'https://rfc-editor.org/rfc/%s.txt' % item.canonical_name(), - 'link_url': self.item_link(item) - } - }) - extra.update({'doi':'10.17487/%s' % item.canonical_name().upper()}) - extra.update({'doiuri':'http://dx.doi.org/10.17487/%s' % item.canonical_name().upper()}) - - #TODO + extra.update({"dcterms_accessRights": "gratis"}) + extra.update({"dcterms_format": "text/html"}) + media_contents = [] + if item.rfc_number < settings.FIRST_V3_RFC: + if item.rfc_number not in [8, 9, 51, 418, 500, 530, 589]: + for fmt, media_type in [("txt", "text/plain"), ("html", "text/html")]: + media_contents.append( + { + "url": f"https://rfc-editor.org/rfc/{item.name}.{fmt}", + "media_type": media_type, + "is_format_of": self.item_link(item), + } + ) + if item.rfc_number not in [571, 587]: + media_contents.append( + { + "url": f"https://www.rfc-editor.org/rfc/pdfrfc/{item.name}.txt.pdf", + "media_type": "application/pdf", + "is_format_of": self.item_link(item), + } + ) + else: + media_contents.append( + { + "url": f"https://www.rfc-editor.org/rfc/{item.name}.xml", + "media_type": "application/rfc+xml", + } + ) + for fmt, media_type in [ + ("txt", "text/plain"), + ("html", "text/html"), + ("pdf", "application/pdf"), + ]: + media_contents.append( + { + "url": f"https://rfc-editor.org/rfc/{item.name}.{fmt}", + "media_type": media_type, + "is_format_of": f"https://www.rfc-editor.org/rfc/{item.name}.xml", + } + ) + extra.update({"media_contents": media_contents}) + + extra.update( + { + "doi": item.doi, + "doiuri": f"https://doi.org/{item.doi}", + } + ) + # R104 Publisher (Mandatory - but we need a string from them first) - extra.update({'dcterms_publisher':'rfc-editor.org'}) + extra.update({"dcterms_publisher": "rfc-editor.org"}) - #TODO MAYBE (Optional stuff) + # TODO MAYBE (Optional stuff) # R108 License # R115 Creator/Contributor (which would we use?) # F305 Checksum (do they use it?) (or should we put the our digital signature in here somewhere?) @@ -188,4 +283,3 @@ def item_extra_kwargs(self, item): # R118 Keyword return extra - diff --git a/ietf/doc/fields.py b/ietf/doc/fields.py index fde5199509..4a6922bf34 100644 --- a/ietf/doc/fields.py +++ b/ietf/doc/fields.py @@ -13,7 +13,7 @@ import debug # pyflakes:ignore -from ietf.doc.models import Document, DocAlias +from ietf.doc.models import Document from ietf.doc.utils import uppercase_std_abbreviated_name from ietf.utils.fields import SearchableField @@ -69,19 +69,3 @@ def ajax_url(self): class SearchableDocumentField(SearchableDocumentsField): """Specialized to only return one Document""" max_entries = 1 - - -class SearchableDocAliasesField(SearchableDocumentsField): - """Search DocAliases instead of Documents""" - model = DocAlias # type: Type[models.Model] - - def doc_type_filter(self, queryset): - """Filter to include only desired doc type - - For DocAlias, pass through to the docs to check type. - """ - return queryset.filter(docs__type=self.doc_type) - -class SearchableDocAliasField(SearchableDocAliasesField): - """Specialized to only return one DocAlias""" - max_entries = 1 \ No newline at end of file diff --git a/ietf/doc/forms.py b/ietf/doc/forms.py index 8a480dcb8a..768d6f96af 100644 --- a/ietf/doc/forms.py +++ b/ietf/doc/forms.py @@ -1,4 +1,4 @@ -# Copyright The IETF Trust 2013-2020, All Rights Reserved +# Copyright The IETF Trust 2013-2025, All Rights Reserved # -*- coding: utf-8 -*- @@ -8,8 +8,8 @@ from django.core.exceptions import ObjectDoesNotExist, ValidationError from django.core.validators import validate_email -from ietf.doc.fields import SearchableDocAliasesField, SearchableDocAliasField -from ietf.doc.models import RelatedDocument, DocExtResource +from ietf.doc.fields import SearchableDocumentField, SearchableDocumentsField +from ietf.doc.models import RelatedDocument, DocExtResource, State from ietf.iesg.models import TelechatDate from ietf.iesg.utils import telechat_page_count from ietf.person.fields import SearchablePersonField, SearchablePersonsField @@ -61,7 +61,7 @@ class DocAuthorChangeBasisForm(forms.Form): basis = forms.CharField(max_length=255, label='Reason for change', help_text='What is the source or reasoning for the changes to the author list?') - + class AdForm(forms.Form): ad = forms.ModelChoiceField(Person.objects.filter(role__name="ad", role__group__state="active", role__group__type='area').order_by('name'), label="Shepherding AD", empty_label="(None)", required=True) @@ -134,13 +134,14 @@ class ActionHoldersForm(forms.Form): IESG_APPROVED_STATE_LIST = ("ann", "rfcqueue", "pub") class AddDownrefForm(forms.Form): - rfc = SearchableDocAliasField( + rfc = SearchableDocumentField( label="Referenced RFC", help_text="The RFC that is approved for downref", - required=True) - drafts = SearchableDocAliasesField( + required=True, + doc_type="rfc") + drafts = SearchableDocumentsField( label="Internet-Drafts that makes the reference", - help_text="The drafts that approve the downref in their Last Call", + help_text="The Internet-Drafts that approve the downref in their Last Call", required=True) def clean_rfc(self): @@ -148,7 +149,7 @@ def clean_rfc(self): raise forms.ValidationError("Please provide a referenced RFC and a referencing Internet-Draft") rfc = self.cleaned_data['rfc'] - if not rfc.document.is_rfc(): + if rfc.type_id != "rfc": raise forms.ValidationError("Cannot find the RFC: " + rfc.name) return rfc @@ -158,12 +159,12 @@ def clean_drafts(self): v_err_names = [] drafts = self.cleaned_data['drafts'] - for da in drafts: - state = da.document.get_state("draft-iesg") + for d in drafts: + state = d.get_state("draft-iesg") if not state or state.slug not in IESG_APPROVED_STATE_LIST: - v_err_names.append(da.name) + v_err_names.append(d.name) if v_err_names: - raise forms.ValidationError("Draft is not yet approved: " + ", ".join(v_err_names)) + raise forms.ValidationError("Internet-Draft is not yet approved: " + ", ".join(v_err_names)) return drafts def clean(self): @@ -173,23 +174,23 @@ def clean(self): v_err_pairs = [] rfc = self.cleaned_data['rfc'] drafts = self.cleaned_data['drafts'] - for da in drafts: - if RelatedDocument.objects.filter(source=da.document, target=rfc, relationship_id='downref-approval'): - v_err_pairs.append(da.name + " --> RFC " + rfc.document.rfc_number()) + for d in drafts: + if RelatedDocument.objects.filter(source=d, target=rfc, relationship_id='downref-approval'): + v_err_pairs.append(f"{d.name} --> RFC {rfc.rfc_number}") if v_err_pairs: raise forms.ValidationError("Downref is already in the registry: " + ", ".join(v_err_pairs)) if 'save_downref_anyway' not in self.data: # this check is skipped if the save_downref_anyway button is used v_err_refnorm = "" - for da in drafts: - if not RelatedDocument.objects.filter(source=da.document, target=rfc, relationship_id='refnorm'): + for d in drafts: + if not RelatedDocument.objects.filter(source=d, target=rfc, relationship_id='refnorm'): if v_err_refnorm: - v_err_refnorm = v_err_refnorm + " or " + da.name + v_err_refnorm = v_err_refnorm + " or " + d.name else: - v_err_refnorm = da.name + v_err_refnorm = d.name if v_err_refnorm: - v_err_refnorm_prefix = "There does not seem to be a normative reference to RFC " + rfc.document.rfc_number() + " by " + v_err_refnorm_prefix = f"There does not seem to be a normative reference to RFC {rfc.rfc_number} by " raise forms.ValidationError(v_err_refnorm_prefix + v_err_refnorm) @@ -265,3 +266,32 @@ def clean(self): @staticmethod def valid_resource_tags(): return ExtResourceName.objects.all().order_by('slug').values_list('slug', flat=True) + +class InvestigateForm(forms.Form): + name_fragment = forms.CharField( + label="File name or fragment to investigate", + required=True, + help_text=( + "Enter a filename such as draft-ietf-some-draft-00.txt or a fragment like draft-ietf-some-draft using at least 8 characters. The search will also work for files that are not necessarily drafts." + ), + min_length=8, + ) + task_id = forms.CharField(required=False, widget=forms.HiddenInput) + + def clean_name_fragment(self): + disallowed_characters = ["%", "/", "\\", "*"] + name_fragment = self.cleaned_data["name_fragment"] + # Manual inspection of the directories at the time of this writing shows + # looking for files with less than 8 characters in the name is not useful + # Requiring this will help protect against the secretariat unintentionally + # matching every draft. + if any(c in name_fragment for c in disallowed_characters): + raise ValidationError(f"The following characters are disallowed: {', '.join(disallowed_characters)}") + return name_fragment + + +class ChangeStatementStateForm(forms.Form): + state = forms.ModelChoiceField( + State.objects.filter(used=True, type="statement"), + empty_label=None, + ) diff --git a/ietf/doc/lastcall.py b/ietf/doc/lastcall.py index 463311587d..dd38fd3909 100644 --- a/ietf/doc/lastcall.py +++ b/ietf/doc/lastcall.py @@ -1,4 +1,4 @@ -# helpers for handling last calls on Internet Drafts +# helpers for handling last calls on Internet-Drafts from django.db.models import Q @@ -73,4 +73,4 @@ def expire_last_call(doc): if doc.type_id == 'draft': lc_text = doc.latest_event(LastCallDocEvent, type="sent_last_call").desc if "document makes the following downward references" in lc_text: - email_last_call_expired_with_downref(doc, lc_text) \ No newline at end of file + email_last_call_expired_with_downref(doc, lc_text) diff --git a/ietf/doc/mails.py b/ietf/doc/mails.py index 54e0f47e20..ddecbb6b54 100644 --- a/ietf/doc/mails.py +++ b/ietf/doc/mails.py @@ -11,7 +11,7 @@ from django.conf import settings from django.urls import reverse as urlreverse from django.utils import timezone -from django.utils.encoding import force_text +from django.utils.encoding import force_str import debug # pyflakes:ignore from ietf.doc.templatetags.mail_filters import std_level_prompt @@ -19,7 +19,7 @@ from ietf.utils import log from ietf.utils.mail import send_mail, send_mail_text from ietf.ipr.utils import iprs_from_docs, related_docs -from ietf.doc.models import WriteupDocEvent, LastCallDocEvent, DocAlias, ConsensusDocEvent +from ietf.doc.models import WriteupDocEvent, LastCallDocEvent, ConsensusDocEvent from ietf.doc.utils import needed_ballot_positions from ietf.doc.utils_bofreq import bofreq_editors, bofreq_responsible from ietf.group.models import Role @@ -54,7 +54,7 @@ def email_ad_approved_doc(request, doc, text): def email_ad_approved_conflict_review(request, review, ok_to_publish): """Email notification when AD approves a conflict review""" - conflictdoc = review.relateddocument_set.get(relationship__slug='conflrev').target.document + conflictdoc = review.relateddocument_set.get(relationship__slug='conflrev').target (to, cc) = gather_address_lists("ad_approved_conflict_review") frm = request.user.person.formatted_email() send_mail(request, @@ -98,7 +98,7 @@ def email_stream_changed(request, doc, old_stream, new_stream, text=""): text = strip_tags(text) send_mail(request, to, None, - "ID Tracker Stream Change Notice: %s" % doc.file_tag(), + "I-D Tracker Stream Change Notice: %s" % doc.file_tag(), "doc/mail/stream_changed_email.txt", dict(text=text, url=settings.IDTRACKER_BASE_URL + doc.get_absolute_url()), @@ -175,7 +175,7 @@ def generate_ballot_writeup(request, doc): e.doc = doc e.rev = doc.rev e.desc = "Ballot writeup was generated" - e.text = force_text(render_to_string("doc/mail/ballot_writeup.txt", {'iana': iana})) + e.text = force_str(render_to_string("doc/mail/ballot_writeup.txt", {'iana': iana, 'doc': doc })) # caller is responsible for saving, if necessary return e @@ -187,7 +187,7 @@ def generate_ballot_rfceditornote(request, doc): e.doc = doc e.rev = doc.rev e.desc = "RFC Editor Note for ballot was generated" - e.text = force_text(render_to_string("doc/mail/ballot_rfceditornote.txt")) + e.text = force_str(render_to_string("doc/mail/ballot_rfceditornote.txt")) e.save() return e @@ -202,7 +202,7 @@ def generate_last_call_announcement(request, doc): doc.filled_title = textwrap.fill(doc.title, width=70, subsequent_indent=" " * 3) - iprs = iprs_from_docs(related_docs(DocAlias.objects.get(name=doc.canonical_name()))) + iprs = iprs_from_docs(related_docs(Document.objects.get(name=doc.name))) if iprs: ipr_links = [ urlreverse("ietf.ipr.views.show", kwargs=dict(id=i.id)) for i in iprs] ipr_links = [ settings.IDTRACKER_BASE_URL+url if not url.startswith("http") else url for url in ipr_links ] @@ -232,7 +232,7 @@ def generate_last_call_announcement(request, doc): e.doc = doc e.rev = doc.rev e.desc = "Last call announcement was generated" - e.text = force_text(mail) + e.text = force_str(mail) # caller is responsible for saving, if necessary return e @@ -252,7 +252,7 @@ def generate_approval_mail(request, doc): e.doc = doc e.rev = doc.rev e.desc = "Ballot approval text was generated" - e.text = force_text(mail) + e.text = force_str(mail) # caller is responsible for saving, if necessary return e @@ -288,7 +288,7 @@ def generate_approval_mail_approved(request, doc): else: contacts = "The IESG contact person is %s." % responsible_directors[0] - doc_type = "RFC" if doc.get_state_slug() == "rfc" else "Internet Draft" + doc_type = "RFC" if doc.get_state_slug() == "rfc" else "Internet-Draft" addrs = gather_address_lists('ballot_approved_ietf_stream',doc=doc).as_strings() return render_to_string("doc/mail/approval_mail.txt", @@ -308,7 +308,7 @@ def generate_approval_mail_rfc_editor(request, doc): # This is essentially dead code - it is only exercised if the IESG ballots on some other stream's document, # which does not happen now that we have conflict reviews. disapproved = doc.get_state_slug("draft-iesg") in DO_NOT_PUBLISH_IESG_STATES - doc_type = "RFC" if doc.get_state_slug() == "rfc" else "Internet Draft" + doc_type = "RFC" if doc.get_state_slug() == "rfc" else "Internet-Draft" addrs = gather_address_lists('ballot_approved_conflrev', doc=doc).as_strings() return render_to_string("doc/mail/approval_mail_rfc_editor.txt", @@ -334,6 +334,9 @@ def generate_publication_request(request, doc): if doc.stream_id == "irtf": approving_body = "IRSG" consensus_body = doc.group.acronym.upper() + if doc.stream_id == "editorial": + approving_body = "RSAB" + consensus_body = doc.group.acronym.upper() else: approving_body = str(doc.stream) consensus_body = approving_body @@ -486,6 +489,54 @@ def email_irsg_ballot_closed(request, doc, ballot): "doc/mail/close_irsg_ballot_mail.txt", ) +def _send_rsab_ballot_email(request, doc, ballot, subject, template): + """Send email notification when IRSG ballot is issued""" + (to, cc) = gather_address_lists('rsab_ballot_issued', doc=doc) + sender = 'IESG Secretary ' + + active_ballot = doc.active_ballot() + if active_ballot is None: + needed_bps = '' + else: + needed_bps = needed_ballot_positions( + doc, + list(active_ballot.active_balloter_positions().values()) + ) + + return send_mail( + request=request, + frm=sender, + to=to, + cc=cc, + subject=subject, + extra={'Reply-To': [sender]}, + template=template, + context=dict( + doc=doc, + doc_url=settings.IDTRACKER_BASE_URL + doc.get_absolute_url(), + needed_ballot_positions=needed_bps, + )) + +def email_rsab_ballot_issued(request, doc, ballot): + """Send email notification when RSAB ballot is issued""" + return _send_rsab_ballot_email( + request, + doc, + ballot, + 'RSAB ballot issued: %s to %s'%(doc.file_tag(), std_level_prompt(doc)), + 'doc/mail/issue_rsab_ballot_mail.txt', + ) + +def email_rsab_ballot_closed(request, doc, ballot): + """Send email notification when RSAB ballot is closed""" + return _send_rsab_ballot_email( + request, + doc, + ballot, + 'RSAB ballot closed: %s to %s'%(doc.file_tag(), std_level_prompt(doc)), + "doc/mail/close_rsab_ballot_mail.txt", + ) + def email_iana(request, doc, to, msg, cc=None): # fix up message and send it with extra info on doc in headers import email @@ -517,7 +568,7 @@ def email_last_call_expired(doc): send_mail(None, addrs.to, "DraftTracker Mail System ", - "Last Call Expired: %s" % doc.file_tag(), + "IETF Last Call Expired: %s" % doc.file_tag(), "doc/mail/change_notice.txt", dict(text=text, doc=doc, @@ -619,7 +670,7 @@ def send_review_possibly_replaces_request(request, doc, submitter_info): to = set(addrs.to) cc = set(addrs.cc) - possibly_replaces = Document.objects.filter(name__in=[alias.name for alias in doc.related_that_doc("possibly-replaces")]) + possibly_replaces = Document.objects.filter(name__in=[related.name for related in doc.related_that_doc("possibly-replaces")]) for other_doc in possibly_replaces: (other_to, other_cc) = gather_address_lists('doc_replacement_suggested',doc=other_doc) to.update(other_to) diff --git a/ietf/doc/management/commands/find_github_backup_info.py b/ietf/doc/management/commands/find_github_backup_info.py deleted file mode 100644 index f1f71452df..0000000000 --- a/ietf/doc/management/commands/find_github_backup_info.py +++ /dev/null @@ -1,74 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved - - -import github3 - -from collections import Counter -from urllib.parse import urlparse - -from django.conf import settings -from django.core.management.base import BaseCommand, CommandError - -from ietf.doc.models import DocExtResource -from ietf.group.models import GroupExtResource -from ietf.person.models import PersonExtResource - -# TODO: Think more about submodules. This currently will only take top level repos, with the assumption that the clone will include arguments to grab all the submodules. -# As a consequence, we might end up pulling more than we need (or that the org or user expected) -# Make sure this is what we want. - -class Command(BaseCommand): - help = ('Locate information about github repositories to backup') - - def add_arguments(self, parser): - parser.add_argument('--verbose', dest='verbose', action='store_true', help='Show counts of types of repositories') - - def handle(self, *args, **options): - - if not (hasattr(settings,'GITHUB_BACKUP_API_KEY') and settings.GITHUB_BACKUP_API_KEY): - raise CommandError("ERROR: can't find GITHUB_BACKUP_API_KEY") # TODO: at >= py3.1, use returncode - - github = github3.login(token = settings.GITHUB_BACKUP_API_KEY) - owners = dict() - repos = set() - - for cls in (DocExtResource, GroupExtResource, PersonExtResource): - for res in cls.objects.filter(name_id__in=('github_repo','github_org')): - path_parts = urlparse(res.value).path.strip('/').split('/') - if not path_parts or not path_parts[0]: - continue - - owner = path_parts[0] - - if owner not in owners: - try: - gh_owner = github.user(username=owner) - owners[owner] = gh_owner - except github3.exceptions.NotFoundError: - continue - - if gh_owner.type in ('User', 'Organization'): - if len(path_parts) > 1: - repo = path_parts[1] - if (owner, repo) not in repos: - try: - github.repository(owner,repo) - repos.add( (owner, repo) ) - except github3.exceptions.NotFoundError: - continue - else: - for repo in github.repositories_by(owner): - repos.add( (owner, repo.name) ) - - owner_types = Counter([owners[owner].type for owner in owners]) - if options['verbose']: - self.stdout.write("Owners:") - for key in owner_types: - self.stdout.write(" %s: %s"%(key,owner_types[key])) - self.stdout.write("Repositories: %d" % len(repos)) - for repo in sorted(repos): - self.stdout.write(" https://github.com/%s/%s" % repo ) - else: - for repo in sorted(repos): - self.stdout.write("%s/%s" % repo ) - diff --git a/ietf/doc/management/commands/fix_105_slides.py b/ietf/doc/management/commands/fix_105_slides.py deleted file mode 100644 index b8689482e8..0000000000 --- a/ietf/doc/management/commands/fix_105_slides.py +++ /dev/null @@ -1,64 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -import os - -from collections import Counter - -from django.core.management.base import BaseCommand - -from ietf.doc.models import DocEvent -from ietf.meeting.models import Meeting, SessionPresentation -from ietf.person.models import Person - -from ietf.secr.proceedings.proc_utils import is_powerpoint, post_process - -class Command(BaseCommand): - help = ('Fix uploaded_filename and generate pdf from pptx') - - def add_arguments(self, parser): - parser.add_argument('--dry-run', action='store_true', dest='dry-run', default=False, help='Report on changes that would be made without making them') - - def handle(self, *args, **options): - ietf105 = Meeting.objects.get(number=105) - slides_path = os.path.join(ietf105.get_materials_path(),'slides') - system_person = Person.objects.get(name="(System)") - counts = Counter() - - for sp in SessionPresentation.objects.filter(session__meeting__number=105,document__type='slides'): #.filter(document__name='slides-105-manet-dlep-multicast-support-discussion'): - slides = sp.document - if not os.path.exists(os.path.join(slides_path,slides.uploaded_filename)): - name, ext = os.path.splitext(slides.uploaded_filename) - target_filename = '%s-%s%s' % (name[:name.rfind('-ss')], slides.rev,ext) - if os.path.exists(os.path.join(slides_path,target_filename)): - slides.uploaded_filename = target_filename - if not options['dry-run']: - e = DocEvent.objects.create(doc=slides, rev=slides.rev, by=system_person, type='changed_document', desc='Corrected uploaded_filename') - slides.save_with_history([e]) - counts['uploaded_filename repair succeeded'] += 1 - - else: - self.stderr.write("Unable to repair %s" % slides) - counts['uploaded_filename repair failed'] += 1 - continue - else: - counts['uploaded_filename already ok'] += 1 - - if is_powerpoint(slides): - base, _ = os.path.splitext(slides.uploaded_filename) - if os.path.exists(os.path.join(slides_path,base+'.pdf')): - self.stderr.write("PDF already exists for %s " % slides) - counts['PDF already exists for a repaired file'] += 1 - else: - if not options['dry-run']: - post_process(slides) - counts['PDF conversions'] += 1 - - if options['dry-run']: - self.stdout.write("This is a dry-run. Nothing has actually changed. In a normal run, the output would say the following:") - - for label,count in counts.iteritems(): - self.stdout.write("%s : %d" % (label,count) ) - - diff --git a/ietf/doc/management/commands/generate_draft_aliases.py b/ietf/doc/management/commands/generate_draft_aliases.py deleted file mode 100755 index 88f4aa98cb..0000000000 --- a/ietf/doc/management/commands/generate_draft_aliases.py +++ /dev/null @@ -1,180 +0,0 @@ -# Copyright The IETF Trust 2012-2021, All Rights Reserved -# -*- coding: utf-8 -*- - -# This was written as a script by Markus Stenberg . -# It was turned into a management command by Russ Housley . - -import datetime -import io -import os -import re -import shutil -import stat -import time - -from tempfile import mkstemp - -from django.conf import settings -from django.core.management.base import BaseCommand -from django.utils import timezone - -import debug # pyflakes:ignore - -from ietf.doc.models import Document -from ietf.group.utils import get_group_role_emails, get_group_ad_emails -from ietf.utils.aliases import dump_sublist -from utils.mail import parseaddr - -DEFAULT_YEARS = 2 - - -def get_draft_ad_emails(doc): - """Get AD email addresses for the given draft, if any.""" - ad_emails = set() - # If working group document, return current WG ADs - if doc.group and doc.group.acronym != 'none': - ad_emails.update(get_group_ad_emails(doc.group)) - # Document may have an explicit AD set - if doc.ad: - ad_emails.add(doc.ad.email_address()) - return ad_emails - - -def get_draft_chair_emails(doc): - """Get chair email addresses for the given draft, if any.""" - chair_emails = set() - if doc.group: - chair_emails.update(get_group_role_emails(doc.group, ['chair', 'secr'])) - return chair_emails - - -def get_draft_shepherd_email(doc): - """Get shepherd email addresses for the given draft, if any.""" - shepherd_email = set() - if doc.shepherd: - shepherd_email.add(doc.shepherd.email_address()) - return shepherd_email - - -def get_draft_authors_emails(doc): - """Get list of authors for the given draft.""" - author_emails = set() - for author in doc.documentauthor_set.all(): - if author.email and author.email.email_address(): - author_emails.add(author.email.email_address()) - return author_emails - - -def get_draft_notify_emails(doc): - """Get list of email addresses to notify for the given draft.""" - ad_email_alias_regex = r"^%s.ad@(%s|%s)$" % (doc.name, settings.DRAFT_ALIAS_DOMAIN, settings.TOOLS_SERVER) - all_email_alias_regex = r"^%s.all@(%s|%s)$" % (doc.name, settings.DRAFT_ALIAS_DOMAIN, settings.TOOLS_SERVER) - author_email_alias_regex = r"^%s@(%s|%s)$" % (doc.name, settings.DRAFT_ALIAS_DOMAIN, settings.TOOLS_SERVER) - notify_email_alias_regex = r"^%s.notify@(%s|%s)$" % (doc.name, settings.DRAFT_ALIAS_DOMAIN, settings.TOOLS_SERVER) - shepherd_email_alias_regex = r"^%s.shepherd@(%s|%s)$" % (doc.name, settings.DRAFT_ALIAS_DOMAIN, settings.TOOLS_SERVER) - notify_emails = set() - if doc.notify: - for e in doc.notify.split(','): - e = e.strip() - if re.search(ad_email_alias_regex, e): - notify_emails.update(get_draft_ad_emails(doc)) - elif re.search(author_email_alias_regex, e): - notify_emails.update(get_draft_authors_emails(doc)) - elif re.search(shepherd_email_alias_regex, e): - notify_emails.update(get_draft_shepherd_email(doc)) - elif re.search(all_email_alias_regex, e): - notify_emails.update(get_draft_ad_emails(doc)) - notify_emails.update(get_draft_authors_emails(doc)) - notify_emails.update(get_draft_shepherd_email(doc)) - elif re.search(notify_email_alias_regex, e): - pass - else: - (name, email) = parseaddr(e) - notify_emails.add(email) - return notify_emails - - -class Command(BaseCommand): - help = ('Generate the draft-aliases and draft-virtual files for Internet-Draft ' - 'mail aliases, placing them in the files configured in ' - 'settings.DRAFT_ALIASES_PATH and settings.DRAFT_VIRTUAL_PATH, ' - 'respectively. The generation includes aliases for Internet-Drafts ' - 'that have seen activity in the last %s years.' % (DEFAULT_YEARS)) - - def handle(self, *args, **options): - show_since = timezone.now() - datetime.timedelta(DEFAULT_YEARS*365) - - date = time.strftime("%Y-%m-%d_%H:%M:%S") - signature = '# Generated by %s at %s\n' % (os.path.abspath(__file__), date) - - ahandle, aname = mkstemp() - os.close(ahandle) - afile = io.open(aname,"w") - - vhandle, vname = mkstemp() - os.close(vhandle) - vfile = io.open(vname,"w") - - afile.write(signature) - vfile.write(signature) - vfile.write("%s anything\n" % settings.DRAFT_VIRTUAL_DOMAIN) - - # Internet-Drafts with active status or expired within DEFAULT_YEARS - drafts = Document.objects.filter(name__startswith='draft-') - active_drafts = drafts.filter(states__slug='active') - inactive_recent_drafts = drafts.exclude(states__slug='active').filter(expires__gte=show_since) - interesting_drafts = active_drafts | inactive_recent_drafts - - alias_domains = ['ietf.org', ] - for draft in interesting_drafts.distinct().iterator(): - # Omit RFCs, unless they were published in the last DEFAULT_YEARS - if draft.docalias.filter(name__startswith='rfc'): - if draft.latest_event(type='published_rfc').time < show_since: - continue - - alias = draft.name - all = set() - - # no suffix and .authors are the same list - emails = get_draft_authors_emails(draft) - all.update(emails) - dump_sublist(afile, vfile, alias, alias_domains, settings.DRAFT_VIRTUAL_DOMAIN, emails) - dump_sublist(afile, vfile, alias+'.authors', alias_domains, settings.DRAFT_VIRTUAL_DOMAIN, emails) - - # .chairs = group chairs - emails = get_draft_chair_emails(draft) - if emails: - all.update(emails) - dump_sublist(afile, vfile, alias+'.chairs', alias_domains, settings.DRAFT_VIRTUAL_DOMAIN, emails) - - # .ad = sponsoring AD / WG AD (WG document) - emails = get_draft_ad_emails(draft) - if emails: - all.update(emails) - dump_sublist(afile, vfile, alias+'.ad', alias_domains, settings.DRAFT_VIRTUAL_DOMAIN, emails) - - # .notify = notify email list from the Document - emails = get_draft_notify_emails(draft) - if emails: - all.update(emails) - dump_sublist(afile, vfile, alias+'.notify', alias_domains, settings.DRAFT_VIRTUAL_DOMAIN, emails) - - # .shepherd = shepherd email from the Document - emails = get_draft_shepherd_email(draft) - if emails: - all.update(emails) - dump_sublist(afile, vfile, alias+'.shepherd', alias_domains, settings.DRAFT_VIRTUAL_DOMAIN, emails) - - # .all = everything from above - dump_sublist(afile, vfile, alias+'.all', alias_domains, settings.DRAFT_VIRTUAL_DOMAIN, all) - - afile.close() - vfile.close() - - os.chmod(aname, stat.S_IWUSR|stat.S_IRUSR|stat.S_IRGRP|stat.S_IROTH) - os.chmod(vname, stat.S_IWUSR|stat.S_IRUSR|stat.S_IRGRP|stat.S_IROTH) - - shutil.move(aname, settings.DRAFT_ALIASES_PATH) - shutil.move(vname, settings.DRAFT_VIRTUAL_PATH) - - \ No newline at end of file diff --git a/ietf/doc/management/commands/generate_draft_bibxml_files.py b/ietf/doc/management/commands/generate_draft_bibxml_files.py deleted file mode 100644 index f2dc508b9b..0000000000 --- a/ietf/doc/management/commands/generate_draft_bibxml_files.py +++ /dev/null @@ -1,84 +0,0 @@ -# Copyright The IETF Trust 2012-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -import datetime -import io -import os -import re -import sys - -from django.conf import settings -from django.core.management.base import BaseCommand -from django.utils import timezone - -import debug # pyflakes:ignore - -from ietf.doc.models import NewRevisionDocEvent -from ietf.doc.utils import bibxml_for_draft - -DEFAULT_DAYS = 7 - -class Command(BaseCommand): - help = ('Generate draft bibxml files for xml2rfc references, placing them in the ' - 'directory configured in settings.BIBXML_BASE_PATH: %s. ' - 'By default, generate files as needed for new draft revisions from the ' - 'last %s days.' % (settings.BIBXML_BASE_PATH, DEFAULT_DAYS)) - - def add_arguments(self, parser): - parser.add_argument('--all', action='store_true', default=False, help="Process all documents, not only recent submissions") - parser.add_argument('--days', type=int, default=DEFAULT_DAYS, help="Look submissions from the last DAYS days, instead of %s" % DEFAULT_DAYS) - - def say(self, msg): - if self.verbosity > 0: - sys.stdout.write(msg) - sys.stdout.write('\n') - - def note(self, msg): - if self.verbosity > 1: - sys.stdout.write(msg) - sys.stdout.write('\n') - - def mutter(self, msg): - if self.verbosity > 2: - sys.stdout.write(msg) - sys.stdout.write('\n') - - def write(self, fn, new): - # normalize new - new = re.sub(r'\r\n?', r'\n', new) - try: - with io.open(fn, encoding='utf-8') as f: - old = f.read() - except IOError: - old = "" - if old.strip() != new.strip(): - self.note('Writing %s' % os.path.basename(fn)) - with io.open(fn, "w", encoding='utf-8') as f: - f.write(new) - - def handle(self, *args, **options): - self.verbosity = options.get("verbosity", 1) - process_all = options.get("all") - days = options.get("days") - # - bibxmldir = os.path.join(settings.BIBXML_BASE_PATH, 'bibxml-ids') - if not os.path.exists(bibxmldir): - os.makedirs(bibxmldir) - # - if process_all: - doc_events = NewRevisionDocEvent.objects.filter(type='new_revision', doc__type_id='draft') - else: - start = timezone.now() - datetime.timedelta(days=days) - doc_events = NewRevisionDocEvent.objects.filter(type='new_revision', doc__type_id='draft', time__gte=start) - doc_events = doc_events.order_by('time') - - for e in doc_events: - self.mutter('%s %s' % (e.time, e.doc.name)) - try: - doc = e.doc - bibxml = bibxml_for_draft(doc, e.rev) - ref_rev_file_name = os.path.join(bibxmldir, 'reference.I-D.%s-%s.xml' % (doc.name, e.rev)) - self.write(ref_rev_file_name, bibxml) - except Exception as ee: - sys.stderr.write('\n%s-%s: %s\n' % (doc.name, doc.rev, ee)) diff --git a/ietf/doc/management/commands/generate_idnits2_rfc_status.py b/ietf/doc/management/commands/generate_idnits2_rfc_status.py deleted file mode 100644 index 45be188018..0000000000 --- a/ietf/doc/management/commands/generate_idnits2_rfc_status.py +++ /dev/null @@ -1,23 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -import os - -from django.conf import settings -from django.core.management.base import BaseCommand - -from ietf.doc.utils import generate_idnits2_rfc_status -from ietf.utils.log import log - -class Command(BaseCommand): - help = ('Generate the rfc_status blob used by idnits2') - - def handle(self, *args, **options): - filename=os.path.join(settings.DERIVED_DIR,'idnits2-rfc-status') - blob = generate_idnits2_rfc_status() - try: - bytes = blob.encode('utf-8') - with open(filename,'wb') as f: - f.write(bytes) - except Exception as e: - log('failed to write idnits2-rfc-status: '+str(e)) - raise e diff --git a/ietf/doc/management/commands/generate_idnits2_rfcs_obsoleted.py b/ietf/doc/management/commands/generate_idnits2_rfcs_obsoleted.py deleted file mode 100644 index 8bd122e87e..0000000000 --- a/ietf/doc/management/commands/generate_idnits2_rfcs_obsoleted.py +++ /dev/null @@ -1,23 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -import os - -from django.conf import settings -from django.core.management.base import BaseCommand - -from ietf.doc.utils import generate_idnits2_rfcs_obsoleted -from ietf.utils.log import log - -class Command(BaseCommand): - help = ('Generate the rfcs-obsoleted file used by idnits2') - - def handle(self, *args, **options): - filename=os.path.join(settings.DERIVED_DIR,'idnits2-rfcs-obsoleted') - blob = generate_idnits2_rfcs_obsoleted() - try: - bytes = blob.encode('utf-8') - with open(filename,'wb') as f: - f.write(bytes) - except Exception as e: - log('failed to write idnits2-rfcs-obsoleted: '+str(e)) - raise e diff --git a/ietf/doc/migrations/0001_initial.py b/ietf/doc/migrations/0001_initial.py index fd50d34fc7..2823abfe63 100644 --- a/ietf/doc/migrations/0001_initial.py +++ b/ietf/doc/migrations/0001_initial.py @@ -1,12 +1,9 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-20 10:52 +# Generated by Django 2.2.28 on 2023-03-20 19:22 - -import datetime import django.core.validators from django.db import migrations, models import django.db.models.deletion +import django.utils.timezone import ietf.utils.models @@ -39,13 +36,14 @@ class Migration(migrations.Migration): fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('json', models.TextField(help_text='Deleted object in JSON format, with attribute names chosen to be suitable for passing into the relevant create method.')), - ('time', models.DateTimeField(default=datetime.datetime.now)), + ('time', models.DateTimeField(default=django.utils.timezone.now)), ], ), migrations.CreateModel( name='DocAlias', fields=[ - ('name', models.CharField(max_length=255, primary_key=True, serialize=False)), + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('name', models.CharField(max_length=255, unique=True)), ], options={ 'verbose_name': 'document alias', @@ -56,8 +54,8 @@ class Migration(migrations.Migration): name='DocEvent', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('time', models.DateTimeField(db_index=True, default=datetime.datetime.now, help_text='When the event happened')), - ('type', models.CharField(choices=[(b'new_revision', b'Added new revision'), (b'new_submission', b'Uploaded new revision'), (b'changed_document', b'Changed document metadata'), (b'added_comment', b'Added comment'), (b'added_message', b'Added message'), (b'edited_authors', b'Edited the documents author list'), (b'deleted', b'Deleted document'), (b'changed_state', b'Changed state'), (b'changed_stream', b'Changed document stream'), (b'expired_document', b'Expired document'), (b'extended_expiry', b'Extended expiry of document'), (b'requested_resurrect', b'Requested resurrect'), (b'completed_resurrect', b'Completed resurrect'), (b'changed_consensus', b'Changed consensus'), (b'published_rfc', b'Published RFC'), (b'added_suggested_replaces', b'Added suggested replacement relationships'), (b'reviewed_suggested_replaces', b'Reviewed suggested replacement relationships'), (b'changed_group', b'Changed group'), (b'changed_protocol_writeup', b'Changed protocol writeup'), (b'changed_charter_milestone', b'Changed charter milestone'), (b'initial_review', b'Set initial review time'), (b'changed_review_announcement', b'Changed WG Review text'), (b'changed_action_announcement', b'Changed WG Action text'), (b'started_iesg_process', b'Started IESG process on document'), (b'created_ballot', b'Created ballot'), (b'closed_ballot', b'Closed ballot'), (b'sent_ballot_announcement', b'Sent ballot announcement'), (b'changed_ballot_position', b'Changed ballot position'), (b'changed_ballot_approval_text', b'Changed ballot approval text'), (b'changed_ballot_writeup_text', b'Changed ballot writeup text'), (b'changed_rfc_editor_note_text', b'Changed RFC Editor Note text'), (b'changed_last_call_text', b'Changed last call text'), (b'requested_last_call', b'Requested last call'), (b'sent_last_call', b'Sent last call'), (b'scheduled_for_telechat', b'Scheduled for telechat'), (b'iesg_approved', b'IESG approved document (no problem)'), (b'iesg_disapproved', b'IESG disapproved document (do not publish)'), (b'approved_in_minute', b'Approved in minute'), (b'iana_review', b'IANA review comment'), (b'rfc_in_iana_registry', b'RFC is in IANA registry'), (b'rfc_editor_received_announcement', b'Announcement was received by RFC Editor'), (b'requested_publication', b'Publication at RFC Editor requested'), (b'sync_from_rfc_editor', b'Received updated information from RFC Editor'), (b'requested_review', b'Requested review'), (b'assigned_review_request', b'Assigned review request'), (b'closed_review_request', b'Closed review request'), (b'downref_approved', b'Downref approved')], max_length=50)), + ('time', models.DateTimeField(db_index=True, default=django.utils.timezone.now, help_text='When the event happened')), + ('type', models.CharField(choices=[('new_revision', 'Added new revision'), ('new_submission', 'Uploaded new revision'), ('changed_document', 'Changed document metadata'), ('added_comment', 'Added comment'), ('added_message', 'Added message'), ('edited_authors', 'Edited the documents author list'), ('deleted', 'Deleted document'), ('changed_state', 'Changed state'), ('changed_stream', 'Changed document stream'), ('expired_document', 'Expired document'), ('extended_expiry', 'Extended expiry of document'), ('requested_resurrect', 'Requested resurrect'), ('completed_resurrect', 'Completed resurrect'), ('changed_consensus', 'Changed consensus'), ('published_rfc', 'Published RFC'), ('added_suggested_replaces', 'Added suggested replacement relationships'), ('reviewed_suggested_replaces', 'Reviewed suggested replacement relationships'), ('changed_action_holders', 'Changed action holders for document'), ('changed_group', 'Changed group'), ('changed_protocol_writeup', 'Changed protocol writeup'), ('changed_charter_milestone', 'Changed charter milestone'), ('initial_review', 'Set initial review time'), ('changed_review_announcement', 'Changed WG Review text'), ('changed_action_announcement', 'Changed WG Action text'), ('started_iesg_process', 'Started IESG process on document'), ('created_ballot', 'Created ballot'), ('closed_ballot', 'Closed ballot'), ('sent_ballot_announcement', 'Sent ballot announcement'), ('changed_ballot_position', 'Changed ballot position'), ('changed_ballot_approval_text', 'Changed ballot approval text'), ('changed_ballot_writeup_text', 'Changed ballot writeup text'), ('changed_rfc_editor_note_text', 'Changed RFC Editor Note text'), ('changed_last_call_text', 'Changed last call text'), ('requested_last_call', 'Requested last call'), ('sent_last_call', 'Sent last call'), ('scheduled_for_telechat', 'Scheduled for telechat'), ('iesg_approved', 'IESG approved document (no problem)'), ('iesg_disapproved', 'IESG disapproved document (do not publish)'), ('approved_in_minute', 'Approved in minute'), ('iana_review', 'IANA review comment'), ('rfc_in_iana_registry', 'RFC is in IANA registry'), ('rfc_editor_received_announcement', 'Announcement was received by RFC Editor'), ('requested_publication', 'Publication at RFC Editor requested'), ('sync_from_rfc_editor', 'Received updated information from RFC Editor'), ('requested_review', 'Requested review'), ('assigned_review_request', 'Assigned review request'), ('closed_review_request', 'Closed review request'), ('closed_review_assignment', 'Closed review assignment'), ('downref_approved', 'Downref approved'), ('posted_related_ipr', 'Posted related IPR'), ('removed_related_ipr', 'Removed related IPR'), ('changed_editors', 'Changed BOF Request editors')], max_length=50)), ('rev', models.CharField(blank=True, max_length=16, null=True, verbose_name='revision')), ('desc', models.TextField()), ], @@ -65,11 +63,22 @@ class Migration(migrations.Migration): 'ordering': ['-time', '-id'], }, ), + migrations.CreateModel( + name='DocExtResource', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('display_name', models.CharField(blank=True, default='', max_length=255)), + ('value', models.CharField(max_length=2083)), + ], + options={ + 'abstract': False, + }, + ), migrations.CreateModel( name='DocHistory', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('time', models.DateTimeField(default=datetime.datetime.now)), + ('time', models.DateTimeField(default=django.utils.timezone.now)), ('title', models.CharField(max_length=255, validators=[django.core.validators.RegexValidator(message='Please enter a string without control characters.', regex='^[^\x00-\x1f]*$')])), ('abstract', models.TextField(blank=True)), ('rev', models.CharField(blank=True, max_length=16, verbose_name='revision')), @@ -77,8 +86,9 @@ class Migration(migrations.Migration): ('words', models.IntegerField(blank=True, null=True)), ('order', models.IntegerField(blank=True, default=1)), ('expires', models.DateTimeField(blank=True, null=True)), - ('notify', models.CharField(blank=True, max_length=255)), + ('notify', models.TextField(blank=True, max_length=1023)), ('external_url', models.URLField(blank=True)), + ('uploaded_filename', models.TextField(blank=True)), ('note', models.TextField(blank=True)), ('internal_comments', models.TextField(blank=True)), ('name', models.CharField(max_length=255)), @@ -112,7 +122,8 @@ class Migration(migrations.Migration): migrations.CreateModel( name='Document', fields=[ - ('time', models.DateTimeField(default=datetime.datetime.now)), + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('time', models.DateTimeField(default=django.utils.timezone.now)), ('title', models.CharField(max_length=255, validators=[django.core.validators.RegexValidator(message='Please enter a string without control characters.', regex='^[^\x00-\x1f]*$')])), ('abstract', models.TextField(blank=True)), ('rev', models.CharField(blank=True, max_length=16, verbose_name='revision')), @@ -120,77 +131,17 @@ class Migration(migrations.Migration): ('words', models.IntegerField(blank=True, null=True)), ('order', models.IntegerField(blank=True, default=1)), ('expires', models.DateTimeField(blank=True, null=True)), - ('notify', models.CharField(blank=True, max_length=255)), + ('notify', models.TextField(blank=True, max_length=1023)), ('external_url', models.URLField(blank=True)), + ('uploaded_filename', models.TextField(blank=True)), ('note', models.TextField(blank=True)), ('internal_comments', models.TextField(blank=True)), - ('name', models.CharField(max_length=255, primary_key=True, serialize=False, validators=[django.core.validators.RegexValidator(b'^[-a-z0-9]+$', b'Provide a valid document name consisting of lowercase letters, numbers and hyphens.', b'invalid')])), - ('ad', ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='ad_document_set', to='person.Person', verbose_name='area director')), - ('formal_languages', models.ManyToManyField(blank=True, help_text='Formal languages used in document', to='name.FormalLanguageName')), + ('name', models.CharField(max_length=255, unique=True, validators=[django.core.validators.RegexValidator('^[-a-z0-9]+$', 'Provide a valid document name consisting of lowercase letters, numbers and hyphens.', 'invalid')])), ], options={ 'abstract': False, }, ), - migrations.CreateModel( - name='DocumentAuthor', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('affiliation', models.CharField(blank=True, help_text='Organization/company used by author for submission', max_length=100)), - ('country', models.CharField(blank=True, help_text='Country used by author for submission', max_length=255)), - ('order', models.IntegerField(default=1)), - ('document', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document')), - ('email', ietf.utils.models.ForeignKey(blank=True, help_text='Email address used by author for submission', null=True, on_delete=django.db.models.deletion.CASCADE, to='person.Email')), - ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), - ], - options={ - 'ordering': ['document', 'order'], - 'abstract': False, - }, - ), - migrations.CreateModel( - name='DocumentURL', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('desc', models.CharField(blank=True, default='', max_length=255)), - ('url', models.URLField(max_length=512)), - ('doc', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document')), - ('tag', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.DocUrlTagName')), - ], - ), - migrations.CreateModel( - name='RelatedDocHistory', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('relationship', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.DocRelationshipName')), - ('source', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.DocHistory')), - ('target', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='reversely_related_document_history_set', to='doc.DocAlias')), - ], - ), - migrations.CreateModel( - name='RelatedDocument', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('relationship', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.DocRelationshipName')), - ('source', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document')), - ('target', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.DocAlias')), - ], - ), - migrations.CreateModel( - name='State', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('slug', models.SlugField()), - ('name', models.CharField(max_length=255)), - ('used', models.BooleanField(default=True)), - ('desc', models.TextField(blank=True)), - ('order', models.IntegerField(default=0)), - ('next_states', models.ManyToManyField(blank=True, related_name='previous_states', to='doc.State')), - ], - options={ - 'ordering': ['type', 'order'], - }, - ), migrations.CreateModel( name='StateType', fields=[ @@ -221,6 +172,21 @@ class Migration(migrations.Migration): ('discuss_time', models.DateTimeField(blank=True, help_text='Time discuss text was written', null=True)), ('comment', models.TextField(blank=True, help_text='Optional comment')), ('comment_time', models.DateTimeField(blank=True, help_text='Time optional comment was written', null=True)), + ('send_email', models.BooleanField(default=None, null=True)), + ], + bases=('doc.docevent',), + ), + migrations.CreateModel( + name='BofreqEditorDocEvent', + fields=[ + ('docevent_ptr', models.OneToOneField(auto_created=True, on_delete=django.db.models.deletion.CASCADE, parent_link=True, primary_key=True, serialize=False, to='doc.DocEvent')), + ], + bases=('doc.docevent',), + ), + migrations.CreateModel( + name='BofreqResponsibleDocEvent', + fields=[ + ('docevent_ptr', models.OneToOneField(auto_created=True, on_delete=django.db.models.deletion.CASCADE, parent_link=True, primary_key=True, serialize=False, to='doc.DocEvent')), ], bases=('doc.docevent',), ), @@ -228,7 +194,7 @@ class Migration(migrations.Migration): name='ConsensusDocEvent', fields=[ ('docevent_ptr', models.OneToOneField(auto_created=True, on_delete=django.db.models.deletion.CASCADE, parent_link=True, primary_key=True, serialize=False, to='doc.DocEvent')), - ('consensus', models.NullBooleanField(default=None)), + ('consensus', models.BooleanField(default=None, null=True)), ], bases=('doc.docevent',), ), @@ -240,6 +206,13 @@ class Migration(migrations.Migration): ], bases=('doc.docevent',), ), + migrations.CreateModel( + name='IanaExpertDocEvent', + fields=[ + ('docevent_ptr', models.OneToOneField(auto_created=True, on_delete=django.db.models.deletion.CASCADE, parent_link=True, primary_key=True, serialize=False, to='doc.DocEvent')), + ], + bases=('doc.docevent',), + ), migrations.CreateModel( name='InitialReviewDocEvent', fields=[ @@ -263,6 +236,13 @@ class Migration(migrations.Migration): ], bases=('doc.docevent',), ), + migrations.CreateModel( + name='ReviewAssignmentDocEvent', + fields=[ + ('docevent_ptr', models.OneToOneField(auto_created=True, on_delete=django.db.models.deletion.CASCADE, parent_link=True, primary_key=True, serialize=False, to='doc.DocEvent')), + ], + bases=('doc.docevent',), + ), migrations.CreateModel( name='ReviewRequestDocEvent', fields=[ @@ -301,9 +281,88 @@ class Migration(migrations.Migration): ], bases=('doc.docevent',), ), + migrations.CreateModel( + name='State', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('slug', models.SlugField()), + ('name', models.CharField(max_length=255)), + ('used', models.BooleanField(default=True)), + ('desc', models.TextField(blank=True)), + ('order', models.IntegerField(default=0)), + ('next_states', models.ManyToManyField(blank=True, related_name='previous_states', to='doc.State')), + ('type', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.StateType')), + ], + options={ + 'ordering': ['type', 'order'], + }, + ), + migrations.CreateModel( + name='RelatedDocument', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('relationship', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.DocRelationshipName')), + ('source', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document')), + ('target', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.DocAlias')), + ], + ), + migrations.CreateModel( + name='RelatedDocHistory', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('relationship', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.DocRelationshipName')), + ('source', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.DocHistory')), + ('target', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='reversely_related_document_history_set', to='doc.DocAlias')), + ], + ), + migrations.CreateModel( + name='DocumentURL', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('desc', models.CharField(blank=True, default='', max_length=255)), + ('url', models.URLField(max_length=2083)), + ('doc', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document')), + ('tag', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.DocUrlTagName')), + ], + ), + migrations.CreateModel( + name='DocumentAuthor', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('affiliation', models.CharField(blank=True, help_text='Organization/company used by author for submission', max_length=100)), + ('country', models.CharField(blank=True, help_text='Country used by author for submission', max_length=255)), + ('order', models.IntegerField(default=1)), + ('document', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document')), + ('email', ietf.utils.models.ForeignKey(blank=True, help_text='Email address used by author for submission', null=True, on_delete=django.db.models.deletion.CASCADE, to='person.Email')), + ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), + ], + options={ + 'ordering': ['document', 'order'], + 'abstract': False, + }, + ), + migrations.CreateModel( + name='DocumentActionHolder', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('time_added', models.DateTimeField(default=django.utils.timezone.now)), + ('document', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document')), + ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), + ], + ), + migrations.AddField( + model_name='document', + name='action_holders', + field=models.ManyToManyField(blank=True, through='doc.DocumentActionHolder', to='person.Person'), + ), + migrations.AddField( + model_name='document', + name='ad', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='ad_document_set', to='person.Person', verbose_name='area director'), + ), migrations.AddField( - model_name='state', - name='type', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.StateType'), + model_name='document', + name='formal_languages', + field=models.ManyToManyField(blank=True, help_text='Formal languages used in document', to='name.FormalLanguageName'), ), ] diff --git a/ietf/doc/migrations/0002_auto_20180220_1052.py b/ietf/doc/migrations/0002_auto_20180220_1052.py deleted file mode 100644 index 811e9eb811..0000000000 --- a/ietf/doc/migrations/0002_auto_20180220_1052.py +++ /dev/null @@ -1,237 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-20 10:52 - - -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - initial = True - - dependencies = [ - ('review', '0001_initial'), - ('contenttypes', '0002_remove_content_type_name'), - ('name', '0001_initial'), - ('submit', '0001_initial'), - ('person', '0001_initial'), - ('message', '0001_initial'), - ('doc', '0001_initial'), - ('group', '0001_initial'), - ] - - operations = [ - migrations.AddField( - model_name='document', - name='group', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='group.Group'), - ), - migrations.AddField( - model_name='document', - name='intended_std_level', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.IntendedStdLevelName', verbose_name='Intended standardization level'), - ), - migrations.AddField( - model_name='document', - name='shepherd', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='shepherd_document_set', to='person.Email'), - ), - migrations.AddField( - model_name='document', - name='states', - field=models.ManyToManyField(blank=True, to='doc.State'), - ), - migrations.AddField( - model_name='document', - name='std_level', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.StdLevelName', verbose_name='Standardization level'), - ), - migrations.AddField( - model_name='document', - name='stream', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.StreamName'), - ), - migrations.AddField( - model_name='document', - name='tags', - field=models.ManyToManyField(blank=True, to='name.DocTagName'), - ), - migrations.AddField( - model_name='document', - name='type', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.DocTypeName'), - ), - migrations.AddField( - model_name='docreminder', - name='event', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.DocEvent'), - ), - migrations.AddField( - model_name='docreminder', - name='type', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.DocReminderTypeName'), - ), - migrations.AddField( - model_name='dochistoryauthor', - name='document', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='documentauthor_set', to='doc.DocHistory'), - ), - migrations.AddField( - model_name='dochistoryauthor', - name='email', - field=ietf.utils.models.ForeignKey(blank=True, help_text='Email address used by author for submission', null=True, on_delete=django.db.models.deletion.CASCADE, to='person.Email'), - ), - migrations.AddField( - model_name='dochistoryauthor', - name='person', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person'), - ), - migrations.AddField( - model_name='dochistory', - name='ad', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='ad_dochistory_set', to='person.Person', verbose_name='area director'), - ), - migrations.AddField( - model_name='dochistory', - name='doc', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='history_set', to='doc.Document'), - ), - migrations.AddField( - model_name='dochistory', - name='formal_languages', - field=models.ManyToManyField(blank=True, help_text='Formal languages used in document', to='name.FormalLanguageName'), - ), - migrations.AddField( - model_name='dochistory', - name='group', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='group.Group'), - ), - migrations.AddField( - model_name='dochistory', - name='intended_std_level', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.IntendedStdLevelName', verbose_name='Intended standardization level'), - ), - migrations.AddField( - model_name='dochistory', - name='shepherd', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='shepherd_dochistory_set', to='person.Email'), - ), - migrations.AddField( - model_name='dochistory', - name='states', - field=models.ManyToManyField(blank=True, to='doc.State'), - ), - migrations.AddField( - model_name='dochistory', - name='std_level', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.StdLevelName', verbose_name='Standardization level'), - ), - migrations.AddField( - model_name='dochistory', - name='stream', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.StreamName'), - ), - migrations.AddField( - model_name='dochistory', - name='tags', - field=models.ManyToManyField(blank=True, to='name.DocTagName'), - ), - migrations.AddField( - model_name='dochistory', - name='type', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.DocTypeName'), - ), - migrations.AddField( - model_name='docevent', - name='by', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person'), - ), - migrations.AddField( - model_name='docevent', - name='doc', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document'), - ), - migrations.AddField( - model_name='docalias', - name='document', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document'), - ), - migrations.AddField( - model_name='deletedevent', - name='by', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person'), - ), - migrations.AddField( - model_name='deletedevent', - name='content_type', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='contenttypes.ContentType'), - ), - migrations.AddField( - model_name='ballottype', - name='doc_type', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.DocTypeName'), - ), - migrations.AddField( - model_name='ballottype', - name='positions', - field=models.ManyToManyField(blank=True, to='name.BallotPositionName'), - ), - migrations.AddField( - model_name='submissiondocevent', - name='submission', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='submit.Submission'), - ), - migrations.AddField( - model_name='statedocevent', - name='state', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.State'), - ), - migrations.AddField( - model_name='statedocevent', - name='state_type', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.StateType'), - ), - migrations.AddField( - model_name='reviewrequestdocevent', - name='review_request', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='review.ReviewRequest'), - ), - migrations.AddField( - model_name='reviewrequestdocevent', - name='state', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.ReviewRequestStateName'), - ), - migrations.AddField( - model_name='ballotpositiondocevent', - name='ad', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person'), - ), - migrations.AddField( - model_name='ballotpositiondocevent', - name='ballot', - field=ietf.utils.models.ForeignKey(default=None, null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.BallotDocEvent'), - ), - migrations.AddField( - model_name='ballotpositiondocevent', - name='pos', - field=ietf.utils.models.ForeignKey(default='norecord', on_delete=django.db.models.deletion.CASCADE, to='name.BallotPositionName', verbose_name='position'), - ), - migrations.AddField( - model_name='ballotdocevent', - name='ballot_type', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.BallotType'), - ), - migrations.AddField( - model_name='addedmessageevent', - name='in_reply_to', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='doc_irtomanual', to='message.Message'), - ), - migrations.AddField( - model_name='addedmessageevent', - name='message', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='doc_manualevents', to='message.Message'), - ), - ] diff --git a/ietf/doc/migrations/0002_auto_20230320_1222.py b/ietf/doc/migrations/0002_auto_20230320_1222.py new file mode 100644 index 0000000000..90b2d11a25 --- /dev/null +++ b/ietf/doc/migrations/0002_auto_20230320_1222.py @@ -0,0 +1,292 @@ +# Generated by Django 2.2.28 on 2023-03-20 19:22 + +from django.db import migrations, models +import django.db.models.deletion +import ietf.utils.models + + +class Migration(migrations.Migration): + + initial = True + + dependencies = [ + ('message', '0001_initial'), + ('name', '0001_initial'), + ('person', '0001_initial'), + ('review', '0001_initial'), + ('group', '0001_initial'), + ('contenttypes', '0002_remove_content_type_name'), + ('submit', '0001_initial'), + ('doc', '0001_initial'), + ] + + operations = [ + migrations.AddField( + model_name='document', + name='group', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='group.Group'), + ), + migrations.AddField( + model_name='document', + name='intended_std_level', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.IntendedStdLevelName', verbose_name='Intended standardization level'), + ), + migrations.AddField( + model_name='document', + name='shepherd', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='shepherd_document_set', to='person.Email'), + ), + migrations.AddField( + model_name='document', + name='states', + field=models.ManyToManyField(blank=True, to='doc.State'), + ), + migrations.AddField( + model_name='document', + name='std_level', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.StdLevelName', verbose_name='Standardization level'), + ), + migrations.AddField( + model_name='document', + name='stream', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.StreamName'), + ), + migrations.AddField( + model_name='document', + name='tags', + field=models.ManyToManyField(blank=True, to='name.DocTagName'), + ), + migrations.AddField( + model_name='document', + name='type', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.DocTypeName'), + ), + migrations.AddField( + model_name='docreminder', + name='event', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.DocEvent'), + ), + migrations.AddField( + model_name='docreminder', + name='type', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.DocReminderTypeName'), + ), + migrations.AddField( + model_name='dochistoryauthor', + name='document', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='documentauthor_set', to='doc.DocHistory'), + ), + migrations.AddField( + model_name='dochistoryauthor', + name='email', + field=ietf.utils.models.ForeignKey(blank=True, help_text='Email address used by author for submission', null=True, on_delete=django.db.models.deletion.CASCADE, to='person.Email'), + ), + migrations.AddField( + model_name='dochistoryauthor', + name='person', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person'), + ), + migrations.AddField( + model_name='dochistory', + name='ad', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='ad_dochistory_set', to='person.Person', verbose_name='area director'), + ), + migrations.AddField( + model_name='dochistory', + name='doc', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='history_set', to='doc.Document'), + ), + migrations.AddField( + model_name='dochistory', + name='formal_languages', + field=models.ManyToManyField(blank=True, help_text='Formal languages used in document', to='name.FormalLanguageName'), + ), + migrations.AddField( + model_name='dochistory', + name='group', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='group.Group'), + ), + migrations.AddField( + model_name='dochistory', + name='intended_std_level', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.IntendedStdLevelName', verbose_name='Intended standardization level'), + ), + migrations.AddField( + model_name='dochistory', + name='shepherd', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='shepherd_dochistory_set', to='person.Email'), + ), + migrations.AddField( + model_name='dochistory', + name='states', + field=models.ManyToManyField(blank=True, to='doc.State'), + ), + migrations.AddField( + model_name='dochistory', + name='std_level', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.StdLevelName', verbose_name='Standardization level'), + ), + migrations.AddField( + model_name='dochistory', + name='stream', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.StreamName'), + ), + migrations.AddField( + model_name='dochistory', + name='tags', + field=models.ManyToManyField(blank=True, to='name.DocTagName'), + ), + migrations.AddField( + model_name='dochistory', + name='type', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.DocTypeName'), + ), + migrations.AddField( + model_name='docextresource', + name='doc', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document'), + ), + migrations.AddField( + model_name='docextresource', + name='name', + field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ExtResourceName'), + ), + migrations.AddField( + model_name='docevent', + name='by', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person'), + ), + migrations.AddField( + model_name='docevent', + name='doc', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document'), + ), + migrations.AddField( + model_name='docalias', + name='docs', + field=models.ManyToManyField(related_name='docalias', to='doc.Document'), + ), + migrations.AddField( + model_name='deletedevent', + name='by', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person'), + ), + migrations.AddField( + model_name='deletedevent', + name='content_type', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='contenttypes.ContentType'), + ), + migrations.AddField( + model_name='ballottype', + name='doc_type', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.DocTypeName'), + ), + migrations.AddField( + model_name='ballottype', + name='positions', + field=models.ManyToManyField(blank=True, to='name.BallotPositionName'), + ), + migrations.CreateModel( + name='IRSGBallotDocEvent', + fields=[ + ('ballotdocevent_ptr', models.OneToOneField(auto_created=True, on_delete=django.db.models.deletion.CASCADE, parent_link=True, primary_key=True, serialize=False, to='doc.BallotDocEvent')), + ('duedate', models.DateTimeField(blank=True, null=True)), + ], + bases=('doc.ballotdocevent',), + ), + migrations.AddField( + model_name='submissiondocevent', + name='submission', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='submit.Submission'), + ), + migrations.AddField( + model_name='statedocevent', + name='state', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.State'), + ), + migrations.AddField( + model_name='statedocevent', + name='state_type', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.StateType'), + ), + migrations.AddField( + model_name='reviewrequestdocevent', + name='review_request', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='review.ReviewRequest'), + ), + migrations.AddField( + model_name='reviewrequestdocevent', + name='state', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.ReviewRequestStateName'), + ), + migrations.AddField( + model_name='reviewassignmentdocevent', + name='review_assignment', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='review.ReviewAssignment'), + ), + migrations.AddField( + model_name='reviewassignmentdocevent', + name='state', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.ReviewAssignmentStateName'), + ), + migrations.AddIndex( + model_name='documentauthor', + index=models.Index(fields=['document', 'order'], name='doc_documen_documen_7fabe2_idx'), + ), + migrations.AddConstraint( + model_name='documentactionholder', + constraint=models.UniqueConstraint(fields=('document', 'person'), name='unique_action_holder'), + ), + migrations.AddIndex( + model_name='dochistoryauthor', + index=models.Index(fields=['document', 'order'], name='doc_dochist_documen_7e2441_idx'), + ), + migrations.AddIndex( + model_name='docevent', + index=models.Index(fields=['type', 'doc'], name='doc_doceven_type_43e53e_idx'), + ), + migrations.AddIndex( + model_name='docevent', + index=models.Index(fields=['-time', '-id'], name='doc_doceven_time_1a258f_idx'), + ), + migrations.AddField( + model_name='bofreqresponsibledocevent', + name='responsible', + field=models.ManyToManyField(blank=True, to='person.Person'), + ), + migrations.AddField( + model_name='bofreqeditordocevent', + name='editors', + field=models.ManyToManyField(blank=True, to='person.Person'), + ), + migrations.AddField( + model_name='ballotpositiondocevent', + name='ballot', + field=ietf.utils.models.ForeignKey(default=None, null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.BallotDocEvent'), + ), + migrations.AddField( + model_name='ballotpositiondocevent', + name='balloter', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person'), + ), + migrations.AddField( + model_name='ballotpositiondocevent', + name='pos', + field=ietf.utils.models.ForeignKey(default='norecord', on_delete=django.db.models.deletion.CASCADE, to='name.BallotPositionName', verbose_name='position'), + ), + migrations.AddField( + model_name='ballotdocevent', + name='ballot_type', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.BallotType'), + ), + migrations.AddField( + model_name='addedmessageevent', + name='in_reply_to', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='doc_irtomanual', to='message.Message'), + ), + migrations.AddField( + model_name='addedmessageevent', + name='message', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='doc_manualevents', to='message.Message'), + ), + ] diff --git a/ietf/doc/migrations/0003_auto_20180401_1231.py b/ietf/doc/migrations/0003_auto_20180401_1231.py deleted file mode 100644 index 7760b25112..0000000000 --- a/ietf/doc/migrations/0003_auto_20180401_1231.py +++ /dev/null @@ -1,20 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.11 on 2018-04-01 12:31 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0002_auto_20180220_1052'), - ] - - operations = [ - migrations.AddIndex( - model_name='docevent', - index=models.Index(fields=['type', 'doc'], name='doc_doceven_type_43e53e_idx'), - ), - ] diff --git a/ietf/doc/migrations/0003_remove_document_info_order.py b/ietf/doc/migrations/0003_remove_document_info_order.py new file mode 100644 index 0000000000..dcd324b71f --- /dev/null +++ b/ietf/doc/migrations/0003_remove_document_info_order.py @@ -0,0 +1,20 @@ +# Copyright The IETF Trust 2023, All Rights Reserved +from django.db import migrations + + +class Migration(migrations.Migration): + + dependencies = [ + ('doc', '0002_auto_20230320_1222'), + ] + + operations = [ + migrations.RemoveField( + model_name='dochistory', + name='order', + ), + migrations.RemoveField( + model_name='document', + name='order', + ), + ] diff --git a/ietf/doc/migrations/0004_add_draft_stream_replaced_states.py b/ietf/doc/migrations/0004_add_draft_stream_replaced_states.py deleted file mode 100644 index 9d57cb476f..0000000000 --- a/ietf/doc/migrations/0004_add_draft_stream_replaced_states.py +++ /dev/null @@ -1,30 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.13 on 2018-05-03 11:50 - - -from django.db import migrations - -def forward(apps, schema_editor): - State = apps.get_model('doc','State') - for type_id in ('draft-stream-iab','draft-stream-ise','draft-stream-irtf'): - State.objects.create(type_id=type_id, - slug='repl', - name='Replaced', - desc='Replaced', - ) - - -def reverse(apps, schema_editor): - State = apps.get_model('doc','State') - State.objects.filter(type_id__in=('draft-stream-iab','draft-stream-ise','draft-stream-irtf'), slug='repl').delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0003_auto_20180401_1231'), - ] - - operations = [ - migrations.RunPython(forward,reverse) - ] diff --git a/ietf/doc/migrations/0004_alter_dochistory_ad_alter_dochistory_shepherd_and_more.py b/ietf/doc/migrations/0004_alter_dochistory_ad_alter_dochistory_shepherd_and_more.py new file mode 100644 index 0000000000..adc0e69627 --- /dev/null +++ b/ietf/doc/migrations/0004_alter_dochistory_ad_alter_dochistory_shepherd_and_more.py @@ -0,0 +1,36 @@ +# Generated by Django 4.0.10 on 2023-05-16 20:36 + +from django.db import migrations +import django.db.models.deletion +import ietf.utils.models + + +class Migration(migrations.Migration): + + dependencies = [ + ('person', '0001_initial'), + ('doc', '0003_remove_document_info_order'), + ] + + operations = [ + migrations.AlterField( + model_name='dochistory', + name='ad', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='ad_%(class)s_set', to='person.person', verbose_name='area director'), + ), + migrations.AlterField( + model_name='dochistory', + name='shepherd', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='shepherd_%(class)s_set', to='person.email'), + ), + migrations.AlterField( + model_name='document', + name='ad', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='ad_%(class)s_set', to='person.person', verbose_name='area director'), + ), + migrations.AlterField( + model_name='document', + name='shepherd', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='shepherd_%(class)s_set', to='person.email'), + ), + ] diff --git a/ietf/doc/migrations/0005_alter_docevent_type.py b/ietf/doc/migrations/0005_alter_docevent_type.py new file mode 100644 index 0000000000..f8a3cfc795 --- /dev/null +++ b/ietf/doc/migrations/0005_alter_docevent_type.py @@ -0,0 +1,86 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations, models + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0004_alter_dochistory_ad_alter_dochistory_shepherd_and_more"), + ] + + operations = [ + migrations.AlterField( + model_name="docevent", + name="type", + field=models.CharField( + choices=[ + ("new_revision", "Added new revision"), + ("new_submission", "Uploaded new revision"), + ("changed_document", "Changed document metadata"), + ("added_comment", "Added comment"), + ("added_message", "Added message"), + ("edited_authors", "Edited the documents author list"), + ("deleted", "Deleted document"), + ("changed_state", "Changed state"), + ("changed_stream", "Changed document stream"), + ("expired_document", "Expired document"), + ("extended_expiry", "Extended expiry of document"), + ("requested_resurrect", "Requested resurrect"), + ("completed_resurrect", "Completed resurrect"), + ("changed_consensus", "Changed consensus"), + ("published_rfc", "Published RFC"), + ( + "added_suggested_replaces", + "Added suggested replacement relationships", + ), + ( + "reviewed_suggested_replaces", + "Reviewed suggested replacement relationships", + ), + ("changed_action_holders", "Changed action holders for document"), + ("changed_group", "Changed group"), + ("changed_protocol_writeup", "Changed protocol writeup"), + ("changed_charter_milestone", "Changed charter milestone"), + ("initial_review", "Set initial review time"), + ("changed_review_announcement", "Changed WG Review text"), + ("changed_action_announcement", "Changed WG Action text"), + ("started_iesg_process", "Started IESG process on document"), + ("created_ballot", "Created ballot"), + ("closed_ballot", "Closed ballot"), + ("sent_ballot_announcement", "Sent ballot announcement"), + ("changed_ballot_position", "Changed ballot position"), + ("changed_ballot_approval_text", "Changed ballot approval text"), + ("changed_ballot_writeup_text", "Changed ballot writeup text"), + ("changed_rfc_editor_note_text", "Changed RFC Editor Note text"), + ("changed_last_call_text", "Changed last call text"), + ("requested_last_call", "Requested last call"), + ("sent_last_call", "Sent last call"), + ("scheduled_for_telechat", "Scheduled for telechat"), + ("iesg_approved", "IESG approved document (no problem)"), + ("iesg_disapproved", "IESG disapproved document (do not publish)"), + ("approved_in_minute", "Approved in minute"), + ("iana_review", "IANA review comment"), + ("rfc_in_iana_registry", "RFC is in IANA registry"), + ( + "rfc_editor_received_announcement", + "Announcement was received by RFC Editor", + ), + ("requested_publication", "Publication at RFC Editor requested"), + ( + "sync_from_rfc_editor", + "Received updated information from RFC Editor", + ), + ("requested_review", "Requested review"), + ("assigned_review_request", "Assigned review request"), + ("closed_review_request", "Closed review request"), + ("closed_review_assignment", "Closed review assignment"), + ("downref_approved", "Downref approved"), + ("posted_related_ipr", "Posted related IPR"), + ("removed_related_ipr", "Removed related IPR"), + ("changed_editors", "Changed BOF Request editors"), + ("published_statement", "Published statement"), + ], + max_length=50, + ), + ), + ] diff --git a/ietf/doc/migrations/0005_fix_replaced_iab_irtf_stream_docs.py b/ietf/doc/migrations/0005_fix_replaced_iab_irtf_stream_docs.py deleted file mode 100644 index 3529baf980..0000000000 --- a/ietf/doc/migrations/0005_fix_replaced_iab_irtf_stream_docs.py +++ /dev/null @@ -1,112 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.13 on 2018-05-03 12:16 - - -from django.db import migrations - -def forward(apps, schema_editor): - Document = apps.get_model('doc','Document') - State = apps.get_model('doc','State') - - iab_active = State.objects.get(type_id='draft-stream-iab',slug='active') - iab_replaced = State.objects.get(type_id='draft-stream-iab',slug='repl') - - irtf_active = State.objects.get(type_id='draft-stream-irtf',slug='active') - irtf_candidate = State.objects.get(type_id='draft-stream-irtf',slug='candidat') - irtf_replaced = State.objects.get(type_id='draft-stream-irtf',slug='repl') - irtf_dead = State.objects.get(type_id='draft-stream-irtf',slug='dead') - - doc = Document.objects.get(name='draft-flanagan-rfc-preservation') - doc.states.remove(iab_active) - doc.states.add(iab_replaced) - - doc = Document.objects.get(name='draft-trammell-semi-report') - doc.states.remove(iab_active) - doc.states.add(iab_replaced) - - doc = Document.objects.get(name='draft-nir-cfrg-chacha20-poly1305') - doc.states.remove(irtf_candidate) - doc.states.add(irtf_replaced) - - doc = Document.objects.get(name='draft-ladd-spake2') - doc.states.remove(irtf_candidate) - doc.states.add(irtf_replaced) - - doc = Document.objects.get(name='draft-lee-nfvrg-resource-management-service-chain') - doc.states.remove(irtf_candidate) - doc.states.add(irtf_replaced) - - doc = Document.objects.get(name='draft-keranen-t2trg-rest-iot') - doc.states.remove(irtf_candidate) - doc.states.add(irtf_replaced) - - doc = Document.objects.get(name='draft-josefsson-argon2') - doc.states.remove(irtf_active) - doc.states.add(irtf_replaced) - - doc = Document.objects.get(name='draft-tenoever-hrpc-research') - doc.states.remove(irtf_active) - doc.states.add(irtf_replaced) - - doc = Document.objects.get(name='draft-kutscher-icnrg-challenges') - doc.states.remove(irtf_dead) - doc.states.add(irtf_replaced) - -def reverse(apps, schema_editor): - Document = apps.get_model('doc','Document') - State = apps.get_model('doc','State') - - iab_active = State.objects.get(type_id='draft-stream-iab',slug='active') - iab_replaced = State.objects.get(type_id='draft-stream-iab',slug='repl') - - irtf_active = State.objects.get(type_id='draft-stream-irtf',slug='active') - irtf_candidate = State.objects.get(type_id='draft-stream-irtf',slug='candidat') - irtf_replaced = State.objects.get(type_id='draft-stream-irtf',slug='repl') - irtf_dead = State.objects.get(type_id='draft-stream-irtf',slug='dead') - - doc = Document.objects.get(name='draft-flanagan-rfc-preservation') - doc.states.add(iab_active) - doc.states.remove(iab_replaced) - - doc = Document.objects.get(name='draft-trammell-semi-report') - doc.states.add(iab_active) - doc.states.remove(iab_replaced) - - doc = Document.objects.get(name='draft-nir-cfrg-chacha20-poly1305') - doc.states.add(irtf_candidate) - doc.states.remove(irtf_replaced) - - doc = Document.objects.get(name='draft-ladd-spake2') - doc.states.add(irtf_candidate) - doc.states.remove(irtf_replaced) - - doc = Document.objects.get(name='draft-lee-nfvrg-resource-management-service-chain') - doc.states.add(irtf_candidate) - doc.states.remove(irtf_replaced) - - doc = Document.objects.get(name='draft-keranen-t2trg-rest-iot') - doc.states.add(irtf_candidate) - doc.states.remove(irtf_replaced) - - doc = Document.objects.get(name='draft-josefsson-argon2') - doc.states.add(irtf_active) - doc.states.remove(irtf_replaced) - - doc = Document.objects.get(name='draft-tenoever-hrpc-research') - doc.states.add(irtf_active) - doc.states.remove(irtf_replaced) - - doc = Document.objects.get(name='draft-kutscher-icnrg-challenges') - doc.states.add(irtf_dead) - doc.states.remove(irtf_replaced) - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0004_add_draft_stream_replaced_states'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/doc/migrations/0006_ballotpositiondocevent_send_email.py b/ietf/doc/migrations/0006_ballotpositiondocevent_send_email.py deleted file mode 100644 index 42f58ca02c..0000000000 --- a/ietf/doc/migrations/0006_ballotpositiondocevent_send_email.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.15 on 2018-10-03 06:39 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0005_fix_replaced_iab_irtf_stream_docs'), - ] - - operations = [ - migrations.AddField( - model_name='ballotpositiondocevent', - name='send_email', - field=models.NullBooleanField(default=None), - ), - ] diff --git a/ietf/doc/migrations/0006_statements.py b/ietf/doc/migrations/0006_statements.py new file mode 100644 index 0000000000..9a074292e5 --- /dev/null +++ b/ietf/doc/migrations/0006_statements.py @@ -0,0 +1,43 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations + + +def forward(apps, schema_editor): + StateType = apps.get_model("doc", "StateType") + State = apps.get_model("doc", "State") + + StateType.objects.create(slug="statement", label="Statement State") + State.objects.create( + slug="active", + type_id="statement", + name="Active", + order=0, + desc="The statement is active", + ) + State.objects.create( + slug="replaced", + type_id="statement", + name="Replaced", + order=0, + desc="The statement has been replaced", + ) + + +def reverse(apps, schema_editor): + StateType = apps.get_model("doc", "StateType") + State = apps.get_model("doc", "State") + + State.objects.filter(type_id="statement").delete() + StateType.objects.filter(slug="statement").delete() + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0005_alter_docevent_type"), + ("name", "0004_statements"), + ] + + operations = [ + migrations.RunPython(forward, reverse), + ] diff --git a/ietf/doc/migrations/0007_alter_docevent_type.py b/ietf/doc/migrations/0007_alter_docevent_type.py new file mode 100644 index 0000000000..c98144d70f --- /dev/null +++ b/ietf/doc/migrations/0007_alter_docevent_type.py @@ -0,0 +1,90 @@ +# Generated by Django 4.2.4 on 2023-08-23 21:38 + +from django.db import migrations, models + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0006_statements"), + ] + + operations = [ + migrations.AlterField( + model_name="docevent", + name="type", + field=models.CharField( + choices=[ + ("new_revision", "Added new revision"), + ("new_submission", "Uploaded new revision"), + ("changed_document", "Changed document metadata"), + ("added_comment", "Added comment"), + ("added_message", "Added message"), + ("edited_authors", "Edited the documents author list"), + ("deleted", "Deleted document"), + ("changed_state", "Changed state"), + ("changed_stream", "Changed document stream"), + ("expired_document", "Expired document"), + ("extended_expiry", "Extended expiry of document"), + ("requested_resurrect", "Requested resurrect"), + ("completed_resurrect", "Completed resurrect"), + ("changed_consensus", "Changed consensus"), + ("published_rfc", "Published RFC"), + ( + "added_suggested_replaces", + "Added suggested replacement relationships", + ), + ( + "reviewed_suggested_replaces", + "Reviewed suggested replacement relationships", + ), + ("changed_action_holders", "Changed action holders for document"), + ("changed_group", "Changed group"), + ("changed_protocol_writeup", "Changed protocol writeup"), + ("changed_charter_milestone", "Changed charter milestone"), + ("initial_review", "Set initial review time"), + ("changed_review_announcement", "Changed WG Review text"), + ("changed_action_announcement", "Changed WG Action text"), + ("started_iesg_process", "Started IESG process on document"), + ("created_ballot", "Created ballot"), + ("closed_ballot", "Closed ballot"), + ("sent_ballot_announcement", "Sent ballot announcement"), + ("changed_ballot_position", "Changed ballot position"), + ("changed_ballot_approval_text", "Changed ballot approval text"), + ("changed_ballot_writeup_text", "Changed ballot writeup text"), + ("changed_rfc_editor_note_text", "Changed RFC Editor Note text"), + ("changed_last_call_text", "Changed last call text"), + ("requested_last_call", "Requested last call"), + ("sent_last_call", "Sent last call"), + ("scheduled_for_telechat", "Scheduled for telechat"), + ("iesg_approved", "IESG approved document (no problem)"), + ("iesg_disapproved", "IESG disapproved document (do not publish)"), + ("approved_in_minute", "Approved in minute"), + ("iana_review", "IANA review comment"), + ("rfc_in_iana_registry", "RFC is in IANA registry"), + ( + "rfc_editor_received_announcement", + "Announcement was received by RFC Editor", + ), + ("requested_publication", "Publication at RFC Editor requested"), + ( + "sync_from_rfc_editor", + "Received updated information from RFC Editor", + ), + ("requested_review", "Requested review"), + ("assigned_review_request", "Assigned review request"), + ("closed_review_request", "Closed review request"), + ("closed_review_assignment", "Closed review assignment"), + ("downref_approved", "Downref approved"), + ("posted_related_ipr", "Posted related IPR"), + ("removed_related_ipr", "Removed related IPR"), + ( + "removed_objfalse_related_ipr", + "Removed Objectively False related IPR", + ), + ("changed_editors", "Changed BOF Request editors"), + ("published_statement", "Published statement"), + ], + max_length=50, + ), + ), + ] diff --git a/ietf/doc/migrations/0007_idexists.py b/ietf/doc/migrations/0007_idexists.py deleted file mode 100644 index d8df81d5fd..0000000000 --- a/ietf/doc/migrations/0007_idexists.py +++ /dev/null @@ -1,51 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.16 on 2018-11-04 10:56 - - -from tqdm import tqdm - -from django.db import migrations - - -def forward(apps, schema_editor): - State = apps.get_model('doc','State') - Document = apps.get_model('doc','Document') - DocHistory = apps.get_model('doc','DocHistory') - - idexists = State.objects.create( - type_id = 'draft-iesg', - slug = 'idexists', - name = 'I-D Exists', - used = True, - desc = 'The IESG has not started processing this draft, or has stopped processing it without publicastion.', - order = 0, - ) - idexists.next_states.set(State.objects.filter(type='draft-iesg',slug__in=['pub-req','watching'])) - - #for doc in tqdm(Document.objects.filter(type='draft'): - # if not doc.states.filter(type='draft-iesg').exists(): - # doc.states.add(idexists) - # for dh in doc.history_set.all(): - # if not dh.states.filter(type='draft-iesg').exists(): - # dh.states.add(idexists) - - for doc in tqdm(Document.objects.filter(type_id='draft').exclude(states__type_id='draft-iesg')): - doc.states.add(idexists) - for history in tqdm(DocHistory.objects.filter(type_id='draft').exclude(states__type_id='draft-iesg')): - history.states.add(idexists) - - -def reverse(apps, schema_editor): - State = apps.get_model('doc','State') - State.objects.filter(slug='idexists').delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0006_ballotpositiondocevent_send_email'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/doc/migrations/0008_add_uploaded_filename.py b/ietf/doc/migrations/0008_add_uploaded_filename.py deleted file mode 100644 index bf36901080..0000000000 --- a/ietf/doc/migrations/0008_add_uploaded_filename.py +++ /dev/null @@ -1,26 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.17 on 2018-12-28 13:11 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0007_idexists'), - ] - - operations = [ - migrations.AddField( - model_name='dochistory', - name='uploaded_filename', - field=models.TextField(blank=True), - ), - migrations.AddField( - model_name='document', - name='uploaded_filename', - field=models.TextField(blank=True), - ), - ] diff --git a/ietf/doc/migrations/0008_alter_docevent_type.py b/ietf/doc/migrations/0008_alter_docevent_type.py new file mode 100644 index 0000000000..52a75f074b --- /dev/null +++ b/ietf/doc/migrations/0008_alter_docevent_type.py @@ -0,0 +1,91 @@ +# Generated by Django 4.2.7 on 2023-11-04 13:02 + +from django.db import migrations, models + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0007_alter_docevent_type"), + ] + + operations = [ + migrations.AlterField( + model_name="docevent", + name="type", + field=models.CharField( + choices=[ + ("new_revision", "Added new revision"), + ("new_submission", "Uploaded new revision"), + ("changed_document", "Changed document metadata"), + ("added_comment", "Added comment"), + ("added_message", "Added message"), + ("edited_authors", "Edited the documents author list"), + ("deleted", "Deleted document"), + ("changed_state", "Changed state"), + ("changed_stream", "Changed document stream"), + ("expired_document", "Expired document"), + ("extended_expiry", "Extended expiry of document"), + ("requested_resurrect", "Requested resurrect"), + ("completed_resurrect", "Completed resurrect"), + ("changed_consensus", "Changed consensus"), + ("published_rfc", "Published RFC"), + ( + "added_suggested_replaces", + "Added suggested replacement relationships", + ), + ( + "reviewed_suggested_replaces", + "Reviewed suggested replacement relationships", + ), + ("changed_action_holders", "Changed action holders for document"), + ("changed_group", "Changed group"), + ("changed_protocol_writeup", "Changed protocol writeup"), + ("changed_charter_milestone", "Changed charter milestone"), + ("initial_review", "Set initial review time"), + ("changed_review_announcement", "Changed WG Review text"), + ("changed_action_announcement", "Changed WG Action text"), + ("started_iesg_process", "Started IESG process on document"), + ("created_ballot", "Created ballot"), + ("closed_ballot", "Closed ballot"), + ("sent_ballot_announcement", "Sent ballot announcement"), + ("changed_ballot_position", "Changed ballot position"), + ("changed_ballot_approval_text", "Changed ballot approval text"), + ("changed_ballot_writeup_text", "Changed ballot writeup text"), + ("changed_rfc_editor_note_text", "Changed RFC Editor Note text"), + ("changed_last_call_text", "Changed last call text"), + ("requested_last_call", "Requested last call"), + ("sent_last_call", "Sent last call"), + ("scheduled_for_telechat", "Scheduled for telechat"), + ("iesg_approved", "IESG approved document (no problem)"), + ("iesg_disapproved", "IESG disapproved document (do not publish)"), + ("approved_in_minute", "Approved in minute"), + ("iana_review", "IANA review comment"), + ("rfc_in_iana_registry", "RFC is in IANA registry"), + ( + "rfc_editor_received_announcement", + "Announcement was received by RFC Editor", + ), + ("requested_publication", "Publication at RFC Editor requested"), + ( + "sync_from_rfc_editor", + "Received updated information from RFC Editor", + ), + ("requested_review", "Requested review"), + ("assigned_review_request", "Assigned review request"), + ("closed_review_request", "Closed review request"), + ("closed_review_assignment", "Closed review assignment"), + ("downref_approved", "Downref approved"), + ("posted_related_ipr", "Posted related IPR"), + ("removed_related_ipr", "Removed related IPR"), + ( + "removed_objfalse_related_ipr", + "Removed Objectively False related IPR", + ), + ("changed_editors", "Changed BOF Request editors"), + ("published_statement", "Published statement"), + ("approved_slides", "Slides approved"), + ], + max_length=50, + ), + ), + ] diff --git a/ietf/doc/migrations/0009_add_rfc_states.py b/ietf/doc/migrations/0009_add_rfc_states.py new file mode 100644 index 0000000000..07a6ac0205 --- /dev/null +++ b/ietf/doc/migrations/0009_add_rfc_states.py @@ -0,0 +1,23 @@ +# Generated by Django 4.2.2 on 2023-06-14 20:57 + +from django.db import migrations + + +def forward(apps, schema_editor): + StateType = apps.get_model("doc", "StateType") + rfc_statetype, _ = StateType.objects.get_or_create(slug="rfc", label="State") + + State = apps.get_model("doc", "State") + State.objects.get_or_create( + type=rfc_statetype, slug="published", name="Published", used=True, order=1 + ) + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0008_alter_docevent_type"), + ] + + operations = [ + migrations.RunPython(forward), + ] diff --git a/ietf/doc/migrations/0009_move_non_url_externalurls_to_uploaded_filename.py b/ietf/doc/migrations/0009_move_non_url_externalurls_to_uploaded_filename.py deleted file mode 100644 index f97ebbdffc..0000000000 --- a/ietf/doc/migrations/0009_move_non_url_externalurls_to_uploaded_filename.py +++ /dev/null @@ -1,49 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.17 on 2018-12-28 13:33 - - -from django.db import migrations -from django.db.models import F - -def forward(apps, schema_editor): - Document = apps.get_model('doc','Document') - Document.objects.exclude(type_id__in=('review','recording')).update(uploaded_filename = F('external_url')) - Document.objects.exclude(type_id__in=('review','recording')).update(external_url="") - - Document.objects.filter(name='slides-100-edu-sessf-patents-at-ietf-an-overview-of-bcp79bis').update(uploaded_filename='slides-100-edu-sessf-patents-at-ietf-an-overview-of-bcp79bis-00.pdf') - - DocHistory = apps.get_model('doc','DocHistory') - DocHistory.objects.exclude(type_id__in=('review','recording')).update(uploaded_filename = F('external_url')) - DocHistory.objects.exclude(type_id__in=('review','recording')).update(external_url="") - - DocHistory.objects.filter(uploaded_filename='https://www.ietf.org/proceedings/97/slides/slides-97-edu-sessb-local-version-of-newcomers-training-in-korean-00.pdf').update(uploaded_filename='slides-97-edu-sessb-local-version-of-newcomers-training-in-korean-00.pdf') - DocHistory.objects.filter(uploaded_filename='http://materials-98-codec-opus-newvectors-00.tar.gz').update(uploaded_filename='materials-98-codec-opus-newvectors-00.tar.gz') - DocHistory.objects.filter(uploaded_filename='http://materials-98-codec-opus-update-00.patch').update(uploaded_filename='materials-98-codec-opus-update-00.patch') - DocHistory.objects.filter(uploaded_filename='http://slides-100-edu-sessf-patents-at-ietf-an-overview-of-bcp79bis-00.pdf').update(uploaded_filename='slides-100-edu-sessf-patents-at-ietf-an-overview-of-bcp79bis-00.pdf') - DocHistory.objects.filter(uploaded_filename='http://bluesheets-97-6man-201611150930-00.pdf/').update(uploaded_filename='bluesheets-97-6man-201611150930-00.pdf') - DocHistory.objects.filter(uploaded_filename='http://agenda-interim-2017-stir-01-stir-01-01.txt').update(uploaded_filename='agenda-interim-2017-stir-01-stir-01-01.txt') - DocHistory.objects.filter(uploaded_filename='http://agenda-interim-2017-icnrg-02-icnrg-01-05.html').update(uploaded_filename='agenda-interim-2017-icnrg-02-icnrg-01-05.html') - - -def reverse(apps, schema_editor): - Document = apps.get_model('doc','Document') - Document.objects.exclude(type_id__in=('review','recording')).update(external_url = F('uploaded_filename')) - Document.objects.exclude(type_id__in=('review','recording')).update(uploaded_filename="") - - DocHistory = apps.get_model('doc','DocHistory') - DocHistory.objects.exclude(type_id__in=('review','recording')).update(external_url = F('uploaded_filename')) - DocHistory.objects.exclude(type_id__in=('review','recording')).update(uploaded_filename="") - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0008_add_uploaded_filename'), - ('review', '0008_remove_reviewrequest_old_id'), - ('meeting', '0011_auto_20190114_0550'), - ] - - operations = [ - migrations.RunPython(forward,reverse), - ] diff --git a/ietf/doc/migrations/0010_auto_20190225_1302.py b/ietf/doc/migrations/0010_auto_20190225_1302.py deleted file mode 100644 index 2f8dfef459..0000000000 --- a/ietf/doc/migrations/0010_auto_20190225_1302.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-02-25 13:02 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0009_move_non_url_externalurls_to_uploaded_filename'), - ] - - operations = [ - migrations.AlterField( - model_name='documenturl', - name='url', - field=models.URLField(max_length=2083), - ), - ] diff --git a/ietf/doc/migrations/0010_dochistory_rfc_number_document_rfc_number.py b/ietf/doc/migrations/0010_dochistory_rfc_number_document_rfc_number.py new file mode 100644 index 0000000000..26b2a85c62 --- /dev/null +++ b/ietf/doc/migrations/0010_dochistory_rfc_number_document_rfc_number.py @@ -0,0 +1,22 @@ +# Generated by Django 4.2.2 on 2023-06-14 22:28 + +from django.db import migrations, models + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0009_add_rfc_states"), + ] + + operations = [ + migrations.AddField( + model_name="dochistory", + name="rfc_number", + field=models.PositiveIntegerField(blank=True, null=True), + ), + migrations.AddField( + model_name="document", + name="rfc_number", + field=models.PositiveIntegerField(blank=True, null=True), + ), + ] diff --git a/ietf/doc/migrations/0011_create_rfc_documents.py b/ietf/doc/migrations/0011_create_rfc_documents.py new file mode 100644 index 0000000000..466ff81bb0 --- /dev/null +++ b/ietf/doc/migrations/0011_create_rfc_documents.py @@ -0,0 +1,76 @@ +# Generated by Django 4.2.2 on 2023-06-15 15:27 + +from django.db import migrations + + +def forward(apps, schema_editor): + Document = apps.get_model("doc", "Document") + DocAlias = apps.get_model("doc", "DocAlias") + DocumentAuthor = apps.get_model("doc", "DocumentAuthor") + + State = apps.get_model("doc", "State") + draft_rfc_state = State.objects.get(type_id="draft", slug="rfc") + rfc_published_state = State.objects.get(type_id="rfc", slug="published") + + # Find draft Documents in the "rfc" state + found_by_state = Document.objects.filter(states=draft_rfc_state).distinct() + + # Find Documents with an "rfc..." alias and confirm they're the same set + rfc_docaliases = DocAlias.objects.filter(name__startswith="rfc") + found_by_name = Document.objects.filter(docalias__in=rfc_docaliases).distinct() + assert set(found_by_name) == set(found_by_state), "mismatch between rfcs identified by state and docalias" + + # As of 2023-06-15, there is one Document with two rfc aliases: rfc6312 and rfc6342 are the same Document. This + # was due to a publication error. Because we go alias-by-alias, no special handling is needed in this migration. + + for rfc_alias in rfc_docaliases.order_by("name"): + assert rfc_alias.docs.count() == 1, f"DocAlias {rfc_alias} is linked to more than 1 Document" + draft = rfc_alias.docs.first() + if draft.name.startswith("rfc"): + rfc = draft + rfc.type_id = "rfc" + rfc.rfc_number = int(draft.name[3:]) + rfc.save() + rfc.states.set([rfc_published_state]) + else: + rfc = Document.objects.create( + type_id="rfc", + name=rfc_alias.name, + rfc_number=int(rfc_alias.name[3:]), + time=draft.time, + title=draft.title, + stream=draft.stream, + group=draft.group, + abstract=draft.abstract, + pages=draft.pages, + words=draft.words, + std_level=draft.std_level, + ad=draft.ad, + external_url=draft.external_url, + uploaded_filename=draft.uploaded_filename, + note=draft.note, + ) + rfc.states.set([rfc_published_state]) + rfc.formal_languages.set(draft.formal_languages.all()) + + # Copy Authors + for da in draft.documentauthor_set.all(): + DocumentAuthor.objects.create( + document=rfc, + person=da.person, + email=da.email, + affiliation=da.affiliation, + country=da.country, + order=da.order, + ) + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0010_dochistory_rfc_number_document_rfc_number"), + ("name", "0010_rfc_doctype_names"), + ] + + operations = [ + migrations.RunPython(forward), + ] diff --git a/ietf/doc/migrations/0011_reviewassignmentdocevent.py b/ietf/doc/migrations/0011_reviewassignmentdocevent.py deleted file mode 100644 index e143b53397..0000000000 --- a/ietf/doc/migrations/0011_reviewassignmentdocevent.py +++ /dev/null @@ -1,29 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.18 on 2019-01-11 11:22 - - -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0009_refactor_review_request'), - ('name', '0005_reviewassignmentstatename'), - ('doc', '0010_auto_20190225_1302'), - ] - - operations = [ - migrations.CreateModel( - name='ReviewAssignmentDocEvent', - fields=[ - ('docevent_ptr', models.OneToOneField(auto_created=True, on_delete=django.db.models.deletion.CASCADE, parent_link=True, primary_key=True, serialize=False, to='doc.DocEvent')), - ('review_assignment', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='review.ReviewAssignment')), - ('state', ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.ReviewAssignmentStateName')), - ], - bases=('doc.docevent',), - ), - ] diff --git a/ietf/doc/migrations/0012_add_event_type_closed_review_assignment.py b/ietf/doc/migrations/0012_add_event_type_closed_review_assignment.py deleted file mode 100644 index ae9bb3ce83..0000000000 --- a/ietf/doc/migrations/0012_add_event_type_closed_review_assignment.py +++ /dev/null @@ -1,24 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-01 04:43 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0011_reviewassignmentdocevent'), - # present to facilitate migration to just before review.0010: - ('name', '0006_adjust_statenames'), - ('dbtemplate', '0004_adjust_assignment_email_summary_templates'), - ] - - operations = [ - migrations.AlterField( - model_name='docevent', - name='type', - field=models.CharField(choices=[('new_revision', 'Added new revision'), ('new_submission', 'Uploaded new revision'), ('changed_document', 'Changed document metadata'), ('added_comment', 'Added comment'), ('added_message', 'Added message'), ('edited_authors', 'Edited the documents author list'), ('deleted', 'Deleted document'), ('changed_state', 'Changed state'), ('changed_stream', 'Changed document stream'), ('expired_document', 'Expired document'), ('extended_expiry', 'Extended expiry of document'), ('requested_resurrect', 'Requested resurrect'), ('completed_resurrect', 'Completed resurrect'), ('changed_consensus', 'Changed consensus'), ('published_rfc', 'Published RFC'), ('added_suggested_replaces', 'Added suggested replacement relationships'), ('reviewed_suggested_replaces', 'Reviewed suggested replacement relationships'), ('changed_group', 'Changed group'), ('changed_protocol_writeup', 'Changed protocol writeup'), ('changed_charter_milestone', 'Changed charter milestone'), ('initial_review', 'Set initial review time'), ('changed_review_announcement', 'Changed WG Review text'), ('changed_action_announcement', 'Changed WG Action text'), ('started_iesg_process', 'Started IESG process on document'), ('created_ballot', 'Created ballot'), ('closed_ballot', 'Closed ballot'), ('sent_ballot_announcement', 'Sent ballot announcement'), ('changed_ballot_position', 'Changed ballot position'), ('changed_ballot_approval_text', 'Changed ballot approval text'), ('changed_ballot_writeup_text', 'Changed ballot writeup text'), ('changed_rfc_editor_note_text', 'Changed RFC Editor Note text'), ('changed_last_call_text', 'Changed last call text'), ('requested_last_call', 'Requested last call'), ('sent_last_call', 'Sent last call'), ('scheduled_for_telechat', 'Scheduled for telechat'), ('iesg_approved', 'IESG approved document (no problem)'), ('iesg_disapproved', 'IESG disapproved document (do not publish)'), ('approved_in_minute', 'Approved in minute'), ('iana_review', 'IANA review comment'), ('rfc_in_iana_registry', 'RFC is in IANA registry'), ('rfc_editor_received_announcement', 'Announcement was received by RFC Editor'), ('requested_publication', 'Publication at RFC Editor requested'), ('sync_from_rfc_editor', 'Received updated information from RFC Editor'), ('requested_review', 'Requested review'), ('assigned_review_request', 'Assigned review request'), ('closed_review_request', 'Closed review request'), ('closed_review_assignment', 'Closed review assignment'), ('downref_approved', 'Downref approved')], max_length=50), - ), - ] diff --git a/ietf/doc/migrations/0012_move_rfc_docevents.py b/ietf/doc/migrations/0012_move_rfc_docevents.py new file mode 100644 index 0000000000..9969a8f0ad --- /dev/null +++ b/ietf/doc/migrations/0012_move_rfc_docevents.py @@ -0,0 +1,88 @@ +# Generated by Django 4.2.2 on 2023-06-20 18:36 + +from django.db import migrations +from django.db.models import Q + + +def forward(apps, schema_editor): + """Move RFC events from the draft to the rfc Document""" + DocAlias = apps.get_model("doc", "DocAlias") + DocEvent = apps.get_model("doc", "DocEvent") + Document = apps.get_model("doc", "Document") + + # queryset with events migrated regardless of whether before or after the "published_rfc" event + events_always_migrated = DocEvent.objects.filter( + Q( + type__in=[ + "published_rfc", # do not remove this one! + ] + ) + ) + + # queryset with events migrated only after the "published_rfc" event + events_migrated_after_pub = DocEvent.objects.exclude( + type__in=[ + "created_ballot", + "closed_ballot", + "sent_ballot_announcement", + "changed_ballot_position", + "changed_ballot_approval_text", + "changed_ballot_writeup_text", + ] + ).exclude( + type="added_comment", + desc__contains="ballot set", # excludes 311 comments that all apply to drafts + ) + + # special case for rfc 6312/6342 draft, which has two published_rfc events + ignore = ["rfc6312", "rfc6342"] # do not reprocess these later + rfc6312 = Document.objects.get(name="rfc6312") + rfc6342 = Document.objects.get(name="rfc6342") + draft = DocAlias.objects.get(name="rfc6312").docs.first() + assert draft == DocAlias.objects.get(name="rfc6342").docs.first() + published_events = list( + DocEvent.objects.filter(doc=draft, type="published_rfc").order_by("time") + ) + assert len(published_events) == 2 + ( + pub_event_6312, + pub_event_6342, + ) = published_events # order matches pub dates at rfc-editor.org + + pub_event_6312.doc = rfc6312 + pub_event_6312.save() + events_migrated_after_pub.filter( + doc=draft, + time__gte=pub_event_6312.time, + time__lt=pub_event_6342.time, + ).update(doc=rfc6312) + + pub_event_6342.doc = rfc6342 + pub_event_6342.save() + events_migrated_after_pub.filter( + doc=draft, + time__gte=pub_event_6342.time, + ).update(doc=rfc6342) + + # Now handle all the rest + for rfc in Document.objects.filter(type_id="rfc").exclude(name__in=ignore): + draft = DocAlias.objects.get(name=rfc.name).docs.first() + assert draft is not None + published_event = DocEvent.objects.get(doc=draft, type="published_rfc") + events_always_migrated.filter( + doc=draft, + ).update(doc=rfc) + events_migrated_after_pub.filter( + doc=draft, + time__gte=published_event.time, + ).update(doc=rfc) + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0011_create_rfc_documents"), + ] + + operations = [ + migrations.RunPython(forward), + ] diff --git a/ietf/doc/migrations/0013_add_document_docalias_id.py b/ietf/doc/migrations/0013_add_document_docalias_id.py deleted file mode 100644 index 1b9d0ab91b..0000000000 --- a/ietf/doc/migrations/0013_add_document_docalias_id.py +++ /dev/null @@ -1,26 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-08 08:41 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0012_add_event_type_closed_review_assignment'), - ] - - operations = [ - migrations.AddField( - model_name='docalias', - name='id', - field=models.IntegerField(default=0), - ), - migrations.AddField( - model_name='document', - name='id', - field=models.IntegerField(default=0), - ), - ] diff --git a/ietf/doc/migrations/0013_rfc_relateddocuments.py b/ietf/doc/migrations/0013_rfc_relateddocuments.py new file mode 100644 index 0000000000..9baddaebdb --- /dev/null +++ b/ietf/doc/migrations/0013_rfc_relateddocuments.py @@ -0,0 +1,45 @@ +# Generated by Django 4.2.3 on 2023-07-05 22:40 + +from django.db import migrations + + +def forward(apps, schema_editor): + DocAlias = apps.get_model("doc", "DocAlias") + Document = apps.get_model("doc", "Document") + RelatedDocument = apps.get_model("doc", "RelatedDocument") + for rfc_alias in DocAlias.objects.filter(name__startswith="rfc").exclude( + docs__type_id="rfc" + ): + # Move these over to the RFC + RelatedDocument.objects.filter( + relationship__slug__in=( + "tobcp", + "toexp", + "tohist", + "toinf", + "tois", + "tops", + "obs", + "updates", + ), + source__docalias=rfc_alias, + ).update(source=Document.objects.get(name=rfc_alias.name)) + # Duplicate references on the RFC but keep the ones on the draft as well + originals = list( + RelatedDocument.objects.filter( + relationship__slug__in=("refinfo", "refnorm", "refold", "refunk"), + source__docalias=rfc_alias, + ) + ) + for o in originals: + o.pk = None + o.source = Document.objects.get(name=rfc_alias.name) + RelatedDocument.objects.bulk_create(originals) + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0012_move_rfc_docevents"), + ] + + operations = [migrations.RunPython(forward)] diff --git a/ietf/doc/migrations/0014_move_rfc_docaliases.py b/ietf/doc/migrations/0014_move_rfc_docaliases.py new file mode 100644 index 0000000000..c82a98e052 --- /dev/null +++ b/ietf/doc/migrations/0014_move_rfc_docaliases.py @@ -0,0 +1,38 @@ +# Generated by Django 4.2.2 on 2023-06-20 18:36 + +from django.db import migrations + + +def forward(apps, schema_editor): + """Point "rfc..." DocAliases at the rfc-type Document + + Creates a became_rfc RelatedDocument to preserve the connection between the draft and the rfc. + """ + DocAlias = apps.get_model("doc", "DocAlias") + Document = apps.get_model("doc", "Document") + RelatedDocument = apps.get_model("doc", "RelatedDocument") + + for rfc_alias in DocAlias.objects.filter(name__startswith="rfc"): + rfc = Document.objects.get(name=rfc_alias.name) + aliased_doc = rfc_alias.docs.get() # implicitly confirms only one value in rfc_alias.docs + if aliased_doc != rfc: + # If the DocAlias was not already pointing at the rfc, it was pointing at the draft + # it came from. Create the relationship between draft and rfc Documents. + assert aliased_doc.type_id == "draft", f"Alias for {rfc.name} should be pointing at a draft" + RelatedDocument.objects.create( + source=aliased_doc, + target=rfc_alias, + relationship_id="became_rfc", + ) + # Now move the alias from the draft to the rfc + rfc_alias.docs.set([rfc]) + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0013_rfc_relateddocuments"), + ] + + operations = [ + migrations.RunPython(forward), + ] diff --git a/ietf/doc/migrations/0014_set_document_docalias_id.py b/ietf/doc/migrations/0014_set_document_docalias_id.py deleted file mode 100644 index a97f6d7e47..0000000000 --- a/ietf/doc/migrations/0014_set_document_docalias_id.py +++ /dev/null @@ -1,36 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-08 08:42 - - -import sys - -from tqdm import tqdm - -from django.db import migrations - - -def forward(apps, schema_editor): - Document = apps.get_model('doc','Document') - sys.stderr.write('\n') - for i, d in enumerate(tqdm(Document.objects.all()), start=1): - d.id = i - d.save() - - DocAlias = apps.get_model('doc','DocAlias') - for i, d in enumerate(tqdm(DocAlias.objects.all()), start=1): - d.id = i - d.save() - -def reverse(apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0013_add_document_docalias_id'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/doc/migrations/0015_1_add_fk_to_document_id.py b/ietf/doc/migrations/0015_1_add_fk_to_document_id.py deleted file mode 100644 index 6d2e7415f2..0000000000 --- a/ietf/doc/migrations/0015_1_add_fk_to_document_id.py +++ /dev/null @@ -1,122 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-08 10:29 - - -import django.core.validators -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0014_set_document_docalias_id'), - ] - - operations = [ - # Fix name and id fields first - migrations.AlterField( - model_name='docalias', - name='name', - field=models.CharField(max_length=255, unique=True), - ), - migrations.AlterField( - model_name='docalias', - name='id', - field=models.IntegerField(primary_key=True, serialize=False), - ), - migrations.AlterField( - model_name='document', - name='name', - field=models.CharField(max_length=255, unique=True, validators=[django.core.validators.RegexValidator('^[-a-z0-9]+$', 'Provide a valid document name consisting of lowercase letters, numbers and hyphens.', 'invalid')]), - ), - migrations.AlterField( - model_name='document', - name='id', - field=models.IntegerField(primary_key=True, serialize=False), - ), - - # Then remaining fields - migrations.AddField( - model_name='docalias', - name='document2', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field=b'id'), - ), - migrations.AddField( - model_name='dochistory', - name='doc2', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, related_name='history_set', to='doc.Document', to_field=b'id'), - ), - migrations.AddField( - model_name='documentauthor', - name='document2', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field=b'id'), - ), - migrations.AddField( - model_name='documenturl', - name='doc2', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field=b'id'), - ), - migrations.AddField( - model_name='relateddochistory', - name='target2', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, related_name='reversely_related_document_history_set', to='doc.DocAlias', to_field=b'id'), - ), - migrations.AddField( - model_name='relateddocument', - name='source2', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field=b'id'), - ), - migrations.AddField( - model_name='relateddocument', - name='target2', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.DocAlias', to_field=b'id'), - ), - migrations.AddField( - model_name='docevent', - name='doc2', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field='id'), - ), - migrations.AlterField( - model_name='docalias', - name='document', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='old_docalias', to='doc.Document', to_field=b'name'), - ), - migrations.AlterField( - model_name='dochistory', - name='doc', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='old_hist', to='doc.Document', to_field=b'name'), - ), - migrations.AlterField( - model_name='documentauthor', - name='document', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='old_doc_auth', to='doc.Document', to_field=b'name'), - ), - migrations.AlterField( - model_name='documenturl', - name='doc', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='old_doc_url', to='doc.Document', to_field=b'name'), - ), - migrations.AlterField( - model_name='relateddochistory', - name='target', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='old_hist_target', to='doc.DocAlias', to_field=b'name'), - ), - migrations.AlterField( - model_name='relateddocument', - name='source', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='old_rel_source', to='doc.Document', to_field=b'name'), - ), - migrations.AlterField( - model_name='relateddocument', - name='target', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='old_rel_target', to='doc.DocAlias', to_field=b'name'), - ), - migrations.AlterField( - model_name='docevent', - name='doc', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='old_docevent', to='doc.Document', to_field=b'name'), - ), - ] diff --git a/ietf/doc/migrations/0015_2_add_doc_document_m2m_fields.py b/ietf/doc/migrations/0015_2_add_doc_document_m2m_fields.py deleted file mode 100644 index e49a40396f..0000000000 --- a/ietf/doc/migrations/0015_2_add_doc_document_m2m_fields.py +++ /dev/null @@ -1,94 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-28 12:42 - - -import sys, time - -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -def timestamp(apps, schema_editor): - sys.stderr.write('\n %s' % time.strftime('%Y-%m-%d %H:%M:%S')) - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0006_adjust_statenames'), - ('doc', '0015_1_add_fk_to_document_id'), - ] - - operations = [ - migrations.CreateModel( - name='DocumentLanguages', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('document', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field='name', related_name='doclanguages')), - ('formallanguagename', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.FormalLanguageName')), - ], - ), - migrations.CreateModel( - name='DocumentStates', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('document', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field='name', related_name='docstates')), - ('state', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.State')), - ], - ), - migrations.CreateModel( - name='DocumentTags', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('document', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field='name', related_name='doctags')), - ('doctagname', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.DocTagName')), - ], - ), - migrations.AddField( - model_name='document', - name='formal_languages2', - field=models.ManyToManyField(blank=True, related_name='languagedocs', through='doc.DocumentLanguages', to='name.FormalLanguageName'), - ), - migrations.AddField( - model_name='document', - name='states2', - field=models.ManyToManyField(blank=True, related_name='statedocs', through='doc.DocumentStates', to='doc.State'), - ), - migrations.AddField( - model_name='document', - name='tags2', - field=models.ManyToManyField(blank=True, related_name='tagdocs', through='doc.DocumentTags', to='name.DocTagName'), - ), - # Here we copy the content of the existing implicit m2m tables for - # the Document m2m fields into the explicit through tables, in order - # to be able to later set the correct id from name - migrations.RunPython(timestamp, timestamp), - migrations.RunSQL( - "INSERT INTO doc_documentlanguages SELECT * FROM doc_document_formal_languages;", - ""), - migrations.RunPython(timestamp, timestamp), - migrations.RunSQL( - "INSERT INTO doc_documentstates SELECT * FROM doc_document_states;", - ""), - migrations.RunPython(timestamp, timestamp), - migrations.RunSQL( - "INSERT INTO doc_documenttags SELECT * FROM doc_document_tags;", - ""), - migrations.RunPython(timestamp, timestamp), - migrations.AddField( - model_name='documentlanguages', - name='document2', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field='id', null=True, default=None), - ), - migrations.AddField( - model_name='documentstates', - name='document2', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field='id', null=True, default=None), ), - migrations.AddField( - model_name='documenttags', - name='document2', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field='id', null=True, default=None), - - ), - ] diff --git a/ietf/doc/migrations/0015_relate_no_aliases.py b/ietf/doc/migrations/0015_relate_no_aliases.py new file mode 100644 index 0000000000..4ba3dd9607 --- /dev/null +++ b/ietf/doc/migrations/0015_relate_no_aliases.py @@ -0,0 +1,84 @@ +# Generated by Django 4.2.2 on 2023-06-16 13:40 + +from django.db import migrations +import django.db.models.deletion +from django.db.models import F, Subquery, OuterRef, CharField +import ietf.utils.models + +def forward(apps, schema_editor): + RelatedDocument = apps.get_model("doc", "RelatedDocument") + DocAlias = apps.get_model("doc", "DocAlias") + target_subquery = Subquery(DocAlias.objects.filter(pk=OuterRef("deprecated_target")).values("docs")[:1]) + name_subquery = Subquery(DocAlias.objects.filter(pk=OuterRef("deprecated_target")).values("name")[:1]) + RelatedDocument.objects.annotate(firstdoc=target_subquery).annotate(aliasname=name_subquery).update(target=F("firstdoc"),originaltargetaliasname=F("aliasname")) + +def reverse(apps, schema_editor): + pass + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0014_move_rfc_docaliases"), + ] + + operations = [ + migrations.AlterField( + model_name='relateddocument', + name='target', + field=ietf.utils.models.ForeignKey( + db_index=False, + on_delete=django.db.models.deletion.CASCADE, + to='doc.docalias', + ), + ), + migrations.RenameField( + model_name="relateddocument", + old_name="target", + new_name="deprecated_target" + ), + migrations.AlterField( + model_name='relateddocument', + name='deprecated_target', + field=ietf.utils.models.ForeignKey( + db_index=True, + on_delete=django.db.models.deletion.CASCADE, + to='doc.docalias', + ), + ), + migrations.AddField( + model_name="relateddocument", + name="target", + field=ietf.utils.models.ForeignKey( + default=1, # A lie, but a convenient one - no relations point here. + on_delete=django.db.models.deletion.CASCADE, + related_name="targets_related", + to="doc.document", + db_index=False, + ), + preserve_default=False, + ), + migrations.AddField( + model_name="relateddocument", + name="originaltargetaliasname", + field=CharField(max_length=255,null=True,blank=True), + preserve_default=True, + ), + migrations.RunPython(forward, reverse), + migrations.AlterField( + model_name="relateddocument", + name="target", + field=ietf.utils.models.ForeignKey( + on_delete=django.db.models.deletion.CASCADE, + related_name="targets_related", + to="doc.document", + db_index=True, + ), + ), + migrations.RemoveField( + model_name="relateddocument", + name="deprecated_target", + field=ietf.utils.models.ForeignKey( + on_delete=django.db.models.deletion.CASCADE, + to='doc.DocAlias', + ), + ), + ] diff --git a/ietf/doc/migrations/0016_relate_hist_no_aliases.py b/ietf/doc/migrations/0016_relate_hist_no_aliases.py new file mode 100644 index 0000000000..df5fb3c325 --- /dev/null +++ b/ietf/doc/migrations/0016_relate_hist_no_aliases.py @@ -0,0 +1,87 @@ +# Generated by Django 4.2.2 on 2023-06-16 13:40 + +from django.db import migrations +import django.db.models.deletion +from django.db.models import F, Subquery, OuterRef, CharField +import ietf.utils.models + +def forward(apps, schema_editor): + RelatedDocHistory = apps.get_model("doc", "RelatedDocHistory") + DocAlias = apps.get_model("doc", "DocAlias") + target_subquery = Subquery(DocAlias.objects.filter(pk=OuterRef("deprecated_target")).values("docs")[:1]) + name_subquery = Subquery(DocAlias.objects.filter(pk=OuterRef("deprecated_target")).values("name")[:1]) + RelatedDocHistory.objects.annotate(firstdoc=target_subquery).annotate(aliasname=name_subquery).update(target=F("firstdoc"),originaltargetaliasname=F("aliasname")) + +def reverse(apps, schema_editor): + pass + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0015_relate_no_aliases"), + ] + + operations = [ + migrations.AlterField( + model_name='relateddochistory', + name='target', + field=ietf.utils.models.ForeignKey( + db_index=False, + on_delete=django.db.models.deletion.CASCADE, + to='doc.docalias', + related_name='reversely_related_document_history_set', + ), + ), + migrations.RenameField( + model_name="relateddochistory", + old_name="target", + new_name="deprecated_target" + ), + migrations.AlterField( + model_name='relateddochistory', + name='deprecated_target', + field=ietf.utils.models.ForeignKey( + db_index=True, + on_delete=django.db.models.deletion.CASCADE, + to='doc.docalias', + related_name='deprecated_reversely_related_document_history_set', + ), + ), + migrations.AddField( + model_name="relateddochistory", + name="target", + field=ietf.utils.models.ForeignKey( + default=1, # A lie, but a convenient one - no relations point here. + on_delete=django.db.models.deletion.CASCADE, + to="doc.document", + db_index=False, + related_name='reversely_related_document_history_set', + ), + preserve_default=False, + ), + migrations.AddField( + model_name="relateddochistory", + name="originaltargetaliasname", + field=CharField(max_length=255,null=True,blank=True), + preserve_default=True, + ), + migrations.RunPython(forward, reverse), + migrations.AlterField( + model_name="relateddochistory", + name="target", + field=ietf.utils.models.ForeignKey( + on_delete=django.db.models.deletion.CASCADE, + to="doc.document", + db_index=True, + related_name='reversely_related_document_history_set', + ), + ), + migrations.RemoveField( + model_name="relateddochistory", + name="deprecated_target", + field=ietf.utils.models.ForeignKey( + on_delete=django.db.models.deletion.CASCADE, + to='doc.DocAlias', + related_name='deprecated_reversely_related_document_history_set', + ), + ), + ] diff --git a/ietf/doc/migrations/0016_set_document_docalias_fk.py b/ietf/doc/migrations/0016_set_document_docalias_fk.py deleted file mode 100644 index 67d1333f38..0000000000 --- a/ietf/doc/migrations/0016_set_document_docalias_fk.py +++ /dev/null @@ -1,113 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-08 14:04 - - -import sys - -from tqdm import tqdm - -from django.db import migrations - -def forward(apps, schema_editor): - - def add_id_fk(o, a, nameid): - n = getattr(o, a+'_id') - if n: - i = nameid[n] - if not isinstance(i, int): - raise ValueError("Inappropriate value: %s: nameid[%s]: %s" % (o.__class__.__name__, n, i)) - if getattr(o, a+'2_id') != i: - setattr(o, a+'2_id', i) - o.save() - - DocAlias = apps.get_model('doc','DocAlias') - DocEvent = apps.get_model('doc', 'DocEvent') - DocHistory = apps.get_model('doc', 'DocHistory') - Document = apps.get_model('doc', 'Document') - DocumentAuthor = apps.get_model('doc', 'DocumentAuthor') - DocumentLanguages = apps.get_model('doc', 'DocumentLanguages') - DocumentStates = apps.get_model('doc', 'DocumentStates') - DocumentTags = apps.get_model('doc', 'DocumentTags') - DocumentURL = apps.get_model('doc', 'DocumentURL') - Group = apps.get_model('group', 'Group') - IprDocRel = apps.get_model('ipr', 'IprDocRel') - LiaisonStatementAttachment = apps.get_model('liaisons', 'LiaisonStatementAttachment') - RelatedDocHistory = apps.get_model('doc', 'RelatedDocHistory') - RelatedDocument = apps.get_model('doc', 'RelatedDocument') - ReviewAssignment = apps.get_model('review', 'ReviewAssignment') - ReviewRequest = apps.get_model('review', 'ReviewRequest') - ReviewWish = apps.get_model('review', 'ReviewWish') - SessionPresentation = apps.get_model('meeting', 'SessionPresentation') - Submission = apps.get_model('submit', 'Submission') - - # Document id fixup ------------------------------------------------------------ - - objs = Document.objects.in_bulk() - nameid = { o.name: o.id for id, o in objs.items() } - - sys.stderr.write('\n') - - sys.stderr.write('Setting Document FKs:\n') - - for C, a in [ - ( DocAlias , 'document'), - ( DocEvent , 'doc'), - ( DocHistory , 'doc'), - ( DocumentAuthor , 'document'), - ( DocumentLanguages , 'document'), - ( DocumentStates , 'document'), - ( DocumentTags , 'document'), - ( DocumentURL , 'doc'), - ( Group , 'charter'), - ( LiaisonStatementAttachment , 'document'), - ( RelatedDocument , 'source'), - ( ReviewAssignment , 'review'), - ( ReviewRequest , 'doc'), - ( ReviewRequest , 'unused_review'), - ( ReviewWish , 'doc'), - ( SessionPresentation , 'document'), - ( Submission , 'draft'), - ]: - sys.stderr.write(' %s.%s:\n' % (C.__name__, a)) - for o in tqdm(C.objects.all()): - add_id_fk(o, a, nameid) - - # DocAlias id fixup ------------------------------------------------------------ - - sys.stderr.write('\n') - - objs = DocAlias.objects.in_bulk() - nameid = { o.name: o.id for id, o in objs.items() } - - sys.stderr.write('Setting DocAlias FKs:\n') - - for C, a in [ - ( IprDocRel , 'document'), - ( RelatedDocument , 'target'), - ( RelatedDocHistory , 'target'), - ]: - sys.stderr.write(' %s.%s:\n' % (C.__name__, a)) - for o in tqdm(C.objects.all()): - add_id_fk(o, a, nameid) - -def reverse(apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('community', '0004_set_document_m2m_keys'), - ('doc', '0015_2_add_doc_document_m2m_fields'), - ('group', '0014_set_document_m2m_keys'), - ('ipr', '0003_add_ipdocrel_document2_fk'), - ('liaisons', '0003_liaison_document2_fk'), - ('meeting', '0015_sessionpresentation_document2_fk'), - ('message', '0003_set_document_m2m_keys'), - ('review', '0011_review_document2_fk'), - ('submit', '0002_submission_document2_fk'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/doc/migrations/0017_delete_docalias.py b/ietf/doc/migrations/0017_delete_docalias.py new file mode 100644 index 0000000000..207ca81e15 --- /dev/null +++ b/ietf/doc/migrations/0017_delete_docalias.py @@ -0,0 +1,16 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations + + +class Migration(migrations.Migration): + dependencies = [ + ("ipr", "0002_iprdocrel_no_aliases"), + ("doc", "0016_relate_hist_no_aliases"), + ] + + operations = [ + migrations.DeleteModel( + name="DocAlias", + ), + ] diff --git a/ietf/doc/migrations/0017_make_document_id_primary_key.py b/ietf/doc/migrations/0017_make_document_id_primary_key.py deleted file mode 100644 index baadd6f725..0000000000 --- a/ietf/doc/migrations/0017_make_document_id_primary_key.py +++ /dev/null @@ -1,26 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-09 05:46 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0016_set_document_docalias_fk'), - ] - - operations = [ - migrations.AlterField( - model_name='docalias', - name='id', - field=models.AutoField(primary_key=True, serialize=False), - ), - migrations.AlterField( - model_name='document', - name='id', - field=models.AutoField(primary_key=True, serialize=False), - ), - ] diff --git a/ietf/doc/migrations/0018_move_dochistory.py b/ietf/doc/migrations/0018_move_dochistory.py new file mode 100644 index 0000000000..0bc29b0bc4 --- /dev/null +++ b/ietf/doc/migrations/0018_move_dochistory.py @@ -0,0 +1,45 @@ +# Generated by Django 4.2.5 on 2023-09-11 17:52 + +from django.db import migrations + +from django.db.models import Subquery, OuterRef, F + + +def forward(apps, schema_editor): + DocHistory = apps.get_model("doc", "DocHistory") + RelatedDocument = apps.get_model("doc", "RelatedDocument") + Document = apps.get_model("doc", "Document") + DocHistory.objects.filter(type_id="draft", doc__type_id="rfc").update(type_id="rfc") + DocHistory.objects.filter( + type_id="draft", doc__type_id="draft", name__startswith="rfc" + ).annotate( + rfc_id=Subquery( + RelatedDocument.objects.filter( + source_id=OuterRef("doc_id"), relationship_id="became_rfc" + ).values_list("target_id", flat=True)[:1] + ) + ).update( + doc_id=F("rfc_id"), type_id="rfc" + ) + DocHistory.objects.filter(type_id="rfc").annotate( + rfcno=Subquery( + Document.objects.filter(pk=OuterRef("doc_id")).values_list( + "rfc_number", flat=True + )[:1] + ) + ).update(rfc_number=F("rfcno")) + assert not DocHistory.objects.filter( + name__startswith="rfc", type_id="draft" + ).exists() + assert not DocHistory.objects.filter( + type_id="rfc", rfc_number__isnull=True + ).exists() + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0017_delete_docalias"), + ] + + # There is no going back + operations = [migrations.RunPython(forward)] diff --git a/ietf/doc/migrations/0018_remove_old_document_field.py b/ietf/doc/migrations/0018_remove_old_document_field.py deleted file mode 100644 index 887dcc7707..0000000000 --- a/ietf/doc/migrations/0018_remove_old_document_field.py +++ /dev/null @@ -1,125 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-20 09:53 - - -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0017_make_document_id_primary_key'), - ] - - operations = [ - migrations.AlterModelOptions( - name='documentauthor', - options={'ordering': ['document2', 'order']}, - ), - migrations.RemoveIndex( - model_name='docevent', - name='doc_doceven_type_43e53e_idx', - ), - migrations.RemoveField( - model_name='docalias', - name='document', - ), - migrations.RemoveField( - model_name='docevent', - name='doc', - ), - migrations.RemoveField( - model_name='dochistory', - name='doc', - ), - migrations.RemoveField( - model_name='documentauthor', - name='document', - ), - migrations.RemoveField( - model_name='documenturl', - name='doc', - ), - migrations.RemoveField( - model_name='relateddochistory', - name='target', - ), - migrations.RemoveField( - model_name='relateddocument', - name='source', - ), - migrations.RemoveField( - model_name='relateddocument', - name='target', - ), - migrations.AddIndex( - model_name='docevent', - index=models.Index(fields=['type', 'doc2'], name='doc_doceven_type_ac7748_idx'), - ), - # The following 9 migrations are related to the m2m fields on Document - # Remove the intermediary model field pointing to Document.name - migrations.RemoveField( - model_name='documentlanguages', - name='document', - ), - migrations.RemoveField( - model_name='documentstates', - name='document', - ), - migrations.RemoveField( - model_name='documenttags', - name='document', - ), - # Rename the intermediary model field pointing to Document.id, to - # match the implicit m2m table - migrations.RenameField( - model_name='documentlanguages', - old_name='document2', - new_name='document', - ), - migrations.RenameField( - model_name='documentstates', - old_name='document2', - new_name='document', - ), - migrations.RenameField( - model_name='documenttags', - old_name='document2', - new_name='document', - ), - # Alter the fields to point to Document.pk instead of Document.name - migrations.AlterField( - model_name='documentlanguages', - name='document', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.Document'), - ), - migrations.AlterField( - model_name='documentstates', - name='document', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.Document'), - ), - migrations.AlterField( - model_name='documenttags', - name='document', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.Document'), - ), - # Remove the implicit m2m tables which point to Document.name - migrations.RemoveField( - model_name='document', - name='formal_languages', - ), - migrations.RemoveField( - model_name='document', - name='states', - ), - migrations.RemoveField( - model_name='document', - name='tags', - ), - # Next (in a separate migration, in order to commit the above before - # we proceed) we'll create the implicit m2m tables again, this time - # pointing to Document.id since that's now the primary key. - ] diff --git a/ietf/doc/migrations/0019_rename_field_document2.py b/ietf/doc/migrations/0019_rename_field_document2.py deleted file mode 100644 index 517ae9ee73..0000000000 --- a/ietf/doc/migrations/0019_rename_field_document2.py +++ /dev/null @@ -1,84 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-21 05:31 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0018_remove_old_document_field'), - ] - - operations = [ - migrations.AlterModelOptions( - name='documentauthor', - options={'ordering': ['document', 'order']}, - ), - migrations.RemoveIndex( - model_name='docevent', - name='doc_doceven_type_ac7748_idx', - ), - migrations.RenameField( - model_name='docalias', - old_name='document2', - new_name='document', - ), - migrations.RenameField( - model_name='docevent', - old_name='doc2', - new_name='doc', - ), - migrations.RenameField( - model_name='dochistory', - old_name='doc2', - new_name='doc', - ), - migrations.RenameField( - model_name='documentauthor', - old_name='document2', - new_name='document', - ), - migrations.RenameField( - model_name='documenturl', - old_name='doc2', - new_name='doc', - ), - migrations.RenameField( - model_name='relateddochistory', - old_name='target2', - new_name='target', - ), - migrations.RenameField( - model_name='relateddocument', - old_name='source2', - new_name='source', - ), - migrations.RenameField( - model_name='relateddocument', - old_name='target2', - new_name='target', - ), - migrations.AddIndex( - model_name='docevent', - index=models.Index(fields=['type', 'doc'], name='doc_doceven_type_43e53e_idx'), - ), - # Add back the m2m field we removed in 0018_... - migrations.AddField( - model_name='document', - name='formal_languages', - field=models.ManyToManyField(blank=True, help_text='Formal languages used in document', to='name.FormalLanguageName'), - ), - migrations.AddField( - model_name='document', - name='states', - field=models.ManyToManyField(blank=True, to='doc.State'), - ), - migrations.AddField( - model_name='document', - name='tags', - field=models.ManyToManyField(blank=True, to='name.DocTagName'), - ), - ] diff --git a/ietf/doc/migrations/0019_subseries.py b/ietf/doc/migrations/0019_subseries.py new file mode 100644 index 0000000000..be2c612ac0 --- /dev/null +++ b/ietf/doc/migrations/0019_subseries.py @@ -0,0 +1,21 @@ +# Copyright The IETF Trust 2023, All Rights Reserved +from django.db import migrations + + +def forward(apps, schema_editor): + StateType = apps.get_model("doc", "StateType") + for slug in ["bcp", "std", "fyi"]: + StateType.objects.create(slug=slug, label=f"{slug} state") + + +def reverse(apps, schema_editor): + StateType = apps.get_model("doc", "StateType") + StateType.objects.filter(slug__in=["bcp", "std", "fyi"]).delete() + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0018_move_dochistory"), + ] + + operations = [migrations.RunPython(forward, reverse)] diff --git a/ietf/doc/migrations/0020_copy_docs_m2m_table.py b/ietf/doc/migrations/0020_copy_docs_m2m_table.py deleted file mode 100644 index 482c4d7852..0000000000 --- a/ietf/doc/migrations/0020_copy_docs_m2m_table.py +++ /dev/null @@ -1,34 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-21 05:31 - - -import sys, time - -from django.db import migrations - -def timestamp(apps, schema_editor): - sys.stderr.write('\n %s' % time.strftime('%Y-%m-%d %H:%M:%S')) - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0019_rename_field_document2'), - ] - - operations = [ - # Copy the doc IDs from the explicit m2m table to the implicit table - migrations.RunPython(timestamp, timestamp), - migrations.RunSQL( - "INSERT INTO doc_document_formal_languages SELECT id,document_id,formallanguagename_id FROM doc_documentlanguages;", - ""), - migrations.RunPython(timestamp, timestamp), - migrations.RunSQL( - "INSERT INTO doc_document_states SELECT id,document_id,state_id FROM doc_documentstates;", - ""), - migrations.RunPython(timestamp, timestamp), - migrations.RunSQL( - "INSERT INTO doc_document_tags SELECT id,document_id,doctagname_id FROM doc_documenttags;", - ""), - migrations.RunPython(timestamp, timestamp), - ] diff --git a/ietf/doc/migrations/0020_move_errata_tags.py b/ietf/doc/migrations/0020_move_errata_tags.py new file mode 100644 index 0000000000..897b88f467 --- /dev/null +++ b/ietf/doc/migrations/0020_move_errata_tags.py @@ -0,0 +1,29 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations + +from django.db.models import Subquery, OuterRef, F + + +def forward(apps, schema_editor): + Document = apps.get_model("doc", "Document") + RelatedDocument = apps.get_model("doc", "RelatedDocument") + Document.tags.through.objects.filter( + doctagname_id__in=["errata", "verified-errata"], document__type_id="draft" + ).annotate( + rfcdoc=Subquery( + RelatedDocument.objects.filter( + relationship_id="became_rfc", source_id=OuterRef("document__pk") + ).values_list("target__pk", flat=True)[:1] + ) + ).update( + document_id=F("rfcdoc") + ) + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0019_subseries"), + ] + + operations = [migrations.RunPython(forward)] diff --git a/ietf/doc/migrations/0021_narrativeminutes.py b/ietf/doc/migrations/0021_narrativeminutes.py new file mode 100644 index 0000000000..0f330bd053 --- /dev/null +++ b/ietf/doc/migrations/0021_narrativeminutes.py @@ -0,0 +1,39 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations + + +def forward(apps, schema_editor): + StateType = apps.get_model("doc", "StateType") + State = apps.get_model("doc", "State") + + StateType.objects.create( + slug="narrativeminutes", + label="State", + ) + for order, slug in enumerate(["active", "deleted"]): + State.objects.create( + slug=slug, + type_id="narrativeminutes", + name=slug.capitalize(), + order=order, + desc="", + used=True, + ) + + +def reverse(apps, schema_editor): + StateType = apps.get_model("doc", "StateType") + State = apps.get_model("doc", "State") + + State.objects.filter(type_id="narrativeminutes").delete() + StateType.objects.filter(slug="narrativeminutes").delete() + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0020_move_errata_tags"), + ("name", "0013_narrativeminutes"), + ] + + operations = [migrations.RunPython(forward, reverse)] diff --git a/ietf/doc/migrations/0021_remove_docs2_m2m.py b/ietf/doc/migrations/0021_remove_docs2_m2m.py deleted file mode 100644 index c659b65deb..0000000000 --- a/ietf/doc/migrations/0021_remove_docs2_m2m.py +++ /dev/null @@ -1,63 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-30 03:36 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0020_copy_docs_m2m_table'), - ] - - operations = [ - # Get rid of the explicit m2m tables which we needed only to be - # able to convert from Document.name to Document.id - migrations.RemoveField( - model_name='documentlanguages', - name='document', - ), - migrations.RemoveField( - model_name='documentlanguages', - name='formallanguagename', - ), - migrations.RemoveField( - model_name='documentstates', - name='document', - ), - migrations.RemoveField( - model_name='documentstates', - name='state', - ), - migrations.RemoveField( - model_name='documenttags', - name='doctagname', - ), - migrations.RemoveField( - model_name='documenttags', - name='document', - ), - migrations.RemoveField( - model_name='document', - name='formal_languages2', - ), - migrations.RemoveField( - model_name='document', - name='states2', - ), - migrations.RemoveField( - model_name='document', - name='tags2', - ), - migrations.DeleteModel( - name='DocumentLanguages', - ), - migrations.DeleteModel( - name='DocumentStates', - ), - migrations.DeleteModel( - name='DocumentTags', - ), - ] diff --git a/ietf/doc/migrations/0022_document_primary_key_cleanup.py b/ietf/doc/migrations/0022_document_primary_key_cleanup.py deleted file mode 100644 index 9ddc908f5c..0000000000 --- a/ietf/doc/migrations/0022_document_primary_key_cleanup.py +++ /dev/null @@ -1,68 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-06-10 03:47 - - -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0021_remove_docs2_m2m'), - ] - - operations = [ - migrations.AlterField( - model_name='docalias', - name='document', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document'), - ), - migrations.AlterField( - model_name='docalias', - name='id', - field=models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID'), - ), - migrations.AlterField( - model_name='docevent', - name='doc', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document'), - ), - migrations.AlterField( - model_name='dochistory', - name='doc', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='history_set', to='doc.Document'), - ), - migrations.AlterField( - model_name='document', - name='id', - field=models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID'), - ), - migrations.AlterField( - model_name='documentauthor', - name='document', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document'), - ), - migrations.AlterField( - model_name='documenturl', - name='doc', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document'), - ), - migrations.AlterField( - model_name='relateddochistory', - name='target', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='reversely_related_document_history_set', to='doc.DocAlias'), - ), - migrations.AlterField( - model_name='relateddocument', - name='source', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document'), - ), - migrations.AlterField( - model_name='relateddocument', - name='target', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.DocAlias'), - ), - ] diff --git a/ietf/doc/migrations/0022_remove_dochistory_internal_comments_and_more.py b/ietf/doc/migrations/0022_remove_dochistory_internal_comments_and_more.py new file mode 100644 index 0000000000..ad27793a83 --- /dev/null +++ b/ietf/doc/migrations/0022_remove_dochistory_internal_comments_and_more.py @@ -0,0 +1,21 @@ +# Generated by Django 4.2.15 on 2024-08-16 16:43 + +from django.db import migrations + + +class Migration(migrations.Migration): + + dependencies = [ + ("doc", "0021_narrativeminutes"), + ] + + operations = [ + migrations.RemoveField( + model_name="dochistory", + name="internal_comments", + ), + migrations.RemoveField( + model_name="document", + name="internal_comments", + ), + ] diff --git a/ietf/doc/migrations/0023_bofreqspamstate.py b/ietf/doc/migrations/0023_bofreqspamstate.py new file mode 100644 index 0000000000..dbbaf996e9 --- /dev/null +++ b/ietf/doc/migrations/0023_bofreqspamstate.py @@ -0,0 +1,30 @@ +# Copyright The IETF Trust 2024, All Rights Reserved + +from django.db import migrations + + +def forward(apps, schema_editor): + State = apps.get_model("doc", "State") + State.objects.get_or_create( + type_id="bofreq", + slug="spam", + defaults={"name": "Spam", "desc": "The BOF request is spam", "order": 5}, + ) + + +def reverse(apps, schema_editor): + State = apps.get_model("doc", "State") + Document = apps.get_model("doc", "Document") + assert not Document.objects.filter( + states__type="bofreq", states__slug="spam" + ).exists() + State.objects.filter(type_id="bofreq", slug="spam").delete() + + +class Migration(migrations.Migration): + + dependencies = [ + ("doc", "0022_remove_dochistory_internal_comments_and_more"), + ] + + operations = [migrations.RunPython(forward, reverse)] diff --git a/ietf/doc/migrations/0023_one_to_many_docalias.py b/ietf/doc/migrations/0023_one_to_many_docalias.py deleted file mode 100644 index cf3c8330a6..0000000000 --- a/ietf/doc/migrations/0023_one_to_many_docalias.py +++ /dev/null @@ -1,43 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-06-10 04:36 - - -import sys - -from tqdm import tqdm - -from django.db import migrations, models - - -def forward(apps, schema_editor): - DocAlias = apps.get_model('doc','DocAlias') - sys.stderr.write('\n') - for a in tqdm(DocAlias.objects.all()): - a.docs.add(a.document) - -def reverse(apps, schema_editor): - DocAlias = apps.get_model('doc','DocAlias') - sys.stderr.write('\n') - for a in tqdm(DocAlias.objects.all()): - a.document = a.document - a.save() - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0022_document_primary_key_cleanup'), - ] - - operations = [ - migrations.AddField( - model_name='docalias', - name='docs', - field=models.ManyToManyField(related_name='docalias', to='doc.Document'), - ), - migrations.RunPython(forward, reverse), - migrations.RemoveField( - model_name='docalias', - name='document', - ), - ] diff --git a/ietf/doc/migrations/0024_iana_experts.py b/ietf/doc/migrations/0024_iana_experts.py deleted file mode 100644 index 0922b9c243..0000000000 --- a/ietf/doc/migrations/0024_iana_experts.py +++ /dev/null @@ -1,59 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.23 on 2019-08-07 12:07 - - -from django.db import migrations - -def forward(apps, schema_editor): - StateType = apps.get_model('doc','StateType') - State = apps.get_model('doc','State') - - StateType.objects.create(slug='draft-iana-experts',label='IANA Experts State') - State.objects.create(type_id='draft-iana-experts', - slug='need-experts', - name='Need IANA Expert(s)', - used=True, - desc='One or more registries need experts assigned', - order=0 - ) - State.objects.create(type_id='draft-iana-experts', - slug='reviews-assigned', - name='Reviews assigned', - used=True, - desc='One or more expert reviews have been assigned', - order=1 - ) - State.objects.create(type_id='draft-iana-experts', - slug='expert-issues', - name='Issues identified', - used=True, - desc='Some expert reviewers have identified issues', - order=2 - ) - State.objects.create(type_id='draft-iana-experts', - slug='reviewers-ok', - name='Expert Reviews OK', - used=True, - desc='All expert reviews have been completed with no blocking issues', - order=2 - ) - -def reverse(apps, schema_editor): - StateType = apps.get_model('doc','StateType') - State = apps.get_model('doc','State') - - State.objects.filter(type_id='draft-iana-experts', slug__in=('need-experts','reviews-assigned','reviews-complete')).delete() - StateType.objects.filter(slug='draft-iana-experts').delete() - - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0023_one_to_many_docalias'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/doc/migrations/0024_remove_ad_is_watching_states.py b/ietf/doc/migrations/0024_remove_ad_is_watching_states.py new file mode 100644 index 0000000000..0c0fb0ad25 --- /dev/null +++ b/ietf/doc/migrations/0024_remove_ad_is_watching_states.py @@ -0,0 +1,121 @@ +# Copyright The IETF Trust 2024, All Rights Reserved + +from django.db import migrations + + +def get_helper(DocHistory, RelatedDocument, RelatedDocHistory, DocumentAuthor, DocHistoryAuthor): + """Dependency injection wrapper""" + + def save_document_in_history(doc): + """Save a snapshot of document and related objects in the database. + + Local copy of ietf.doc.utils.save_document_in_history() to avoid depending on the + code base in a migration. + """ + + def get_model_fields_as_dict(obj): + return dict((field.name, getattr(obj, field.name)) + for field in obj._meta.fields + if field is not obj._meta.pk) + + # copy fields + fields = get_model_fields_as_dict(doc) + fields["doc"] = doc + fields["name"] = doc.name + + dochist = DocHistory(**fields) + dochist.save() + + # copy many to many + for field in doc._meta.many_to_many: + if field.remote_field.through and field.remote_field.through._meta.auto_created: + hist_field = getattr(dochist, field.name) + hist_field.clear() + hist_field.set(getattr(doc, field.name).all()) + + # copy remaining tricky many to many + def transfer_fields(obj, HistModel): + mfields = get_model_fields_as_dict(item) + # map doc -> dochist + for k, v in mfields.items(): + if v == doc: + mfields[k] = dochist + HistModel.objects.create(**mfields) + + for item in RelatedDocument.objects.filter(source=doc): + transfer_fields(item, RelatedDocHistory) + + for item in DocumentAuthor.objects.filter(document=doc): + transfer_fields(item, DocHistoryAuthor) + + return dochist + + return save_document_in_history + + +def forward(apps, schema_editor): + """Mark watching draft-iesg state unused after removing it from Documents""" + StateDocEvent = apps.get_model("doc", "StateDocEvent") + Document = apps.get_model("doc", "Document") + State = apps.get_model("doc", "State") + StateType = apps.get_model("doc", "StateType") + Person = apps.get_model("person", "Person") + + save_document_in_history = get_helper( + DocHistory=apps.get_model("doc", "DocHistory"), + RelatedDocument=apps.get_model("doc", "RelatedDocument"), + RelatedDocHistory=apps.get_model("doc", "RelatedDocHistory"), + DocumentAuthor=apps.get_model("doc", "DocumentAuthor"), + DocHistoryAuthor=apps.get_model("doc", "DocHistoryAuthor"), + ) + + draft_iesg_state_type = StateType.objects.get(slug="draft-iesg") + idexists_state = State.objects.get(type=draft_iesg_state_type, slug="idexists") + watching_state = State.objects.get(type=draft_iesg_state_type, slug="watching") + system_person = Person.objects.get(name="(System)") + + # Remove state from documents that currently have it + for doc in Document.objects.filter(states=watching_state): + assert doc.type_id == "draft" + doc.states.remove(watching_state) + doc.states.add(idexists_state) + e = StateDocEvent.objects.create( + type="changed_state", + by=system_person, + doc=doc, + rev=doc.rev, + desc=f"{draft_iesg_state_type.label} changed to {idexists_state.name} from {watching_state.name}", + state_type=draft_iesg_state_type, + state=idexists_state, + ) + doc.time = e.time + doc.save() + save_document_in_history(doc) + assert not Document.objects.filter(states=watching_state).exists() + + # Mark state as unused + watching_state.used = False + watching_state.save() + + +def reverse(apps, schema_editor): + """Mark watching draft-iesg state as used + + Does not try to re-apply the state to Documents modified by the forward migration. This + could be done in theory, but would either require dangerous history rewriting or add a + lot of history junk. + """ + State = apps.get_model("doc", "State") + StateType = apps.get_model("doc", "StateType") + State.objects.filter( + type=StateType.objects.get(slug="draft-iesg"), slug="watching" + ).update(used=True) + + +class Migration(migrations.Migration): + + dependencies = [ + ("doc", "0023_bofreqspamstate"), + ] + + operations = [migrations.RunPython(forward, reverse)] diff --git a/ietf/doc/migrations/0025_ianaexpertdocevent.py b/ietf/doc/migrations/0025_ianaexpertdocevent.py deleted file mode 100644 index 282d506d9c..0000000000 --- a/ietf/doc/migrations/0025_ianaexpertdocevent.py +++ /dev/null @@ -1,24 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.23 on 2019-08-07 12:27 - - -from django.db import migrations, models -import django.db.models.deletion - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0024_iana_experts'), - ] - - operations = [ - migrations.CreateModel( - name='IanaExpertDocEvent', - fields=[ - ('docevent_ptr', models.OneToOneField(auto_created=True, on_delete=django.db.models.deletion.CASCADE, parent_link=True, primary_key=True, serialize=False, to='doc.DocEvent')), - ], - bases=('doc.docevent',), - ), - ] diff --git a/ietf/doc/migrations/0025_storedobject_storedobject_unique_name_per_store.py b/ietf/doc/migrations/0025_storedobject_storedobject_unique_name_per_store.py new file mode 100644 index 0000000000..e948ca3011 --- /dev/null +++ b/ietf/doc/migrations/0025_storedobject_storedobject_unique_name_per_store.py @@ -0,0 +1,66 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +from django.db import migrations, models + + +class Migration(migrations.Migration): + + dependencies = [ + ("doc", "0024_remove_ad_is_watching_states"), + ] + + operations = [ + migrations.CreateModel( + name="StoredObject", + fields=[ + ( + "id", + models.AutoField( + auto_created=True, + primary_key=True, + serialize=False, + verbose_name="ID", + ), + ), + ("store", models.CharField(max_length=256)), + ("name", models.CharField(max_length=1024)), + ("sha384", models.CharField(max_length=96)), + ("len", models.PositiveBigIntegerField()), + ( + "store_created", + models.DateTimeField( + help_text="The instant the object ws first placed in the store" + ), + ), + ( + "created", + models.DateTimeField( + help_text="Instant object became known. May not be the same as the storage's created value for the instance. It will hold ctime for objects imported from older disk storage" + ), + ), + ( + "modified", + models.DateTimeField( + help_text="Last instant object was modified. May not be the same as the storage's modified value for the instance. It will hold mtime for objects imported from older disk storage unless they've actually been overwritten more recently" + ), + ), + ("doc_name", models.CharField(blank=True, max_length=255, null=True)), + ("doc_rev", models.CharField(blank=True, max_length=16, null=True)), + ("deleted", models.DateTimeField(null=True)), + ], + options={ + "indexes": [ + models.Index( + fields=["doc_name", "doc_rev"], + name="doc_storedo_doc_nam_d04465_idx", + ) + ], + }, + ), + migrations.AddConstraint( + model_name="storedobject", + constraint=models.UniqueConstraint( + fields=("store", "name"), name="unique_name_per_store" + ), + ), + ] diff --git a/ietf/doc/migrations/0026_add_draft_rfceditor_state.py b/ietf/doc/migrations/0026_add_draft_rfceditor_state.py deleted file mode 100644 index 0490f15ba9..0000000000 --- a/ietf/doc/migrations/0026_add_draft_rfceditor_state.py +++ /dev/null @@ -1,24 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -from django.db import migrations - -def forward(apps, schema_editor): - State = apps.get_model('doc','State') - State.objects.get_or_create(type_id='draft-rfceditor', slug='tooling-issue', name='TI', - desc='Tooling Issue; an update is needed to one or more of the tools in the publication pipeline before this document can be published') - -def reverse(apps, schema_editor): - State = apps.get_model('doc','State') - State.objects.filter(type_id='draft-rfceditor', slug='tooling-issue').delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0025_ianaexpertdocevent'), - ] - - operations = [ - migrations.RunPython(forward,reverse) - ] diff --git a/ietf/doc/migrations/0026_change_wg_state_descriptions.py b/ietf/doc/migrations/0026_change_wg_state_descriptions.py new file mode 100644 index 0000000000..b02b12c97e --- /dev/null +++ b/ietf/doc/migrations/0026_change_wg_state_descriptions.py @@ -0,0 +1,117 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +from django.db import migrations + +def forward(apps, schema_editor): + State = apps.get_model("doc","State") + for name, desc in [ + ("WG Document","The document has been adopted by the Working Group (WG) and is under development. A document can only be adopted by one WG at a time. However, a document may be transferred between WGs."), + ("Parked WG Document","The Working Group (WG) document is in a temporary state where it will not be actively developed. The reason for the pause is explained via a datatracker comments section."), + ("Dead WG Document","The Working Group (WG) document has been abandoned by the WG. No further development is planned in this WG. A decision to resume work on this document and move it out of this state is possible."), + ("In WG Last Call","The Working Group (WG) document is currently subject to an active WG Last Call (WGLC) review per Section 7.4 of RFC2418."), + ("Waiting for Implementation","The progression of this Working Group (WG) document towards publication is paused as it awaits implementation. The process governing the approach to implementations is WG-specific."), + ("Held by WG","Held by Working Group (WG) chairs for administrative reasons. See document history for details."), + ("Waiting for WG Chair Go-Ahead","The Working Group (WG) document has completed Working Group Last Call (WGLC), but the WG chair(s) are not yet ready to call consensus on the document. The reasons for this may include comments from the WGLC need to be responded to, or a revision to the document is needed"), + ("WG Consensus: Waiting for Write-Up","The Working Group (WG) document has consensus to proceed to publication. However, the document is waiting for a document shepherd write-up per RFC4858."), + ("Submitted to IESG for Publication","The Working Group (WG) document has left the WG and been submitted to the Internet Engineering Steering Group (IESG) for evaluation and publication. See the “IESG State” or “RFC Editor State” for further details on the state of the document."), + ("Candidate for WG Adoption","The individual submission document has been marked by the Working Group (WG) chairs as a candidate for adoption by the WG, but no adoption call has been started."), + ("Call For Adoption By WG Issued","A call for adoption of the individual submission document has been issued by the Working Group (WG) chairs. This call is still running but the WG has not yet reached consensus for adoption."), + ("Adopted by a WG","The individual submission document has been adopted by the Working Group (WG), but a WG document replacing this document with the typical naming convention of 'draft- ietf-wgname-topic-nn' has not yet been submitted."), + ("Adopted for WG Info Only","The document is adopted by the Working Group (WG) for its internal use. The WG has decided that it will not pursue publication of it as an RFC."), + ]: + State.objects.filter(name=name).update(desc=desc) + +def reverse(apps, schema_editor): + State = apps.get_model("doc","State") + for name, desc in [ + ("WG Document","""4.2.4. WG Document + + The "WG Document" state describes an I-D that has been adopted by an IETF WG and is being actively developed. + + A WG Chair may transition an I-D into the "WG Document" state at any time as long as the I-D is not being considered or developed in any other WG. + + Alternatively, WG Chairs may rely upon new functionality to be added to the Datatracker to automatically move version-00 drafts into the "WG Document" state as described in Section 4.1. + + Under normal conditions, it should not be possible for an I-D to be in the "WG Document" state in more than one WG at a time. This said, I-Ds may be transferred from one WG to another with the consent of the WG Chairs and the responsible ADs."""), + ("Parked WG Document","""4.2.5. Parked WG Document + + A "Parked WG Document" is an I-D that has lost its author or editor, is waiting for another document to be written or for a review to be completed, or cannot be progressed by the working group for some other reason. + + Some of the annotation tags described in Section 4.3 may be used in conjunction with this state to indicate why an I-D has been parked, and/or what may need to happen for the I-D to be un-parked. + + Parking a WG draft will not prevent it from expiring; however, this state can be used to indicate why the I-D has stopped progressing in the WG. + + A "Parked WG Document" that is not expired may be transferred from one WG to another with the consent of the WG Chairs and the responsible ADs."""), + ("Dead WG Document","""4.2.6. Dead WG Document + + A "Dead WG Document" is an I-D that has been abandoned. Note that 'Dead' is not always a final state for a WG I-D. If consensus is subsequently achieved, a "Dead WG Document" may be resurrected. A "Dead WG Document" that is not resurrected will eventually expire. + + Note that an I-D that is declared to be "Dead" in one WG and that is not expired may be transferred to a non-dead state in another WG with the consent of the WG Chairs and the responsible ADs."""), + ("In WG Last Call","""4.2.7. In WG Last Call + + A document "In WG Last Call" is an I-D for which a WG Last Call (WGLC) has been issued and is in progress. + + Note that conducting a WGLC is an optional part of the IETF WG process, per Section 7.4 of RFC 2418 [RFC2418]. + + If a WG Chair decides to conduct a WGLC on an I-D, the "In WG Last Call" state can be used to track the progress of the WGLC. The Chair may configure the Datatracker to send a WGLC message to one or more mailing lists when the Chair moves the I-D into this state. The WG Chair may also be able to select a different set of mailing lists for a different document undergoing a WGLC; some documents may deserve coordination with other WGs. + + A WG I-D in this state should remain "In WG Last Call" until the WG Chair moves it to another state. The WG Chair may configure the Datatracker to send an e-mail after a specified period of time to remind or 'nudge' the Chair to conclude the WGLC and to determine the next state for the document. + + It is possible for one WGLC to lead into another WGLC for the same document. For example, an I-D that completed a WGLC as an "Informational" document may need another WGLC if a decision is taken to convert the I-D into a Standards Track document."""), + ("Waiting for Implementation","""In some areas, it can be desirable to wait for multiple interoperable implementations before progressing a draft to be an RFC, and in some WGs this is required. This state should be entered after WG Last Call has completed."""), + ("Held by WG","""Held by WG, see document history for details."""), + ("Waiting for WG Chair Go-Ahead","""4.2.8. Waiting for WG Chair Go-Ahead + + A WG Chair may wish to place an I-D that receives a lot of comments during a WGLC into the "Waiting for WG Chair Go-Ahead" state. This state describes an I-D that has undergone a WGLC; however, the Chair is not yet ready to call consensus on the document. + + If comments from the WGLC need to be responded to, or a revision to the I-D is needed, the Chair may place an I-D into this state until all of the WGLC comments are adequately addressed and the (possibly revised) document is in the I-D repository."""), + ("WG Consensus: Waiting for Write-Up","""4.2.9. WG Consensus: Waiting for Writeup + + A document in the "WG Consensus: Waiting for Writeup" state has essentially completed its development within the working group, and is nearly ready to be sent to the IESG for publication. The last thing to be done is the preparation of a protocol writeup by a Document Shepherd. The IESG requires that a document shepherd writeup be completed before publication of the I-D is requested. The IETF document shepherding process and the role of a WG Document Shepherd is described in RFC 4858 [RFC4858] + + A WG Chair may call consensus on an I-D without a formal WGLC and transition an I-D that was in the "WG Document" state directly into this state. + + The name of this state includes the words "Waiting for Writeup" because a good document shepherd writeup takes time to prepare."""), + ("Submitted to IESG for Publication","""4.2.10. Submitted to IESG for Publication + + This state describes a WG document that has been submitted to the IESG for publication and that has not been sent back to the working group for revision. + + An I-D in this state may be under review by the IESG, it may have been approved and be in the RFC Editor's queue, or it may have been published as an RFC. Other possibilities exist too. The document may be "Dead" (in the IESG state machine) or in a "Do Not Publish" state."""), + ("Candidate for WG Adoption","""The document has been marked as a candidate for WG adoption by the WG Chair. This state can be used before a call for adoption is issued (and the document is put in the "Call For Adoption By WG Issued" state), to indicate that the document is in the queue for a call for adoption, even if none has been issued yet."""), + ("Call For Adoption By WG Issued","""4.2.1. Call for Adoption by WG Issued + + The "Call for Adoption by WG Issued" state should be used to indicate when an I-D is being considered for adoption by an IETF WG. An I-D that is in this state is actively being considered for adoption and has not yet achieved consensus, preference, or selection in the WG. + + This state may be used to describe an I-D that someone has asked a WG to consider for adoption, if the WG Chair has agreed with the request. This state may also be used to identify an I-D that a WG Chair asked an author to write specifically for consideration as a candidate WG item [WGDTSPEC], and/or an I-D that is listed as a 'candidate draft' in the WG's charter. + + Under normal conditions, it should not be possible for an I-D to be in the "Call for Adoption by WG Issued" state in more than one working group at the same time. This said, it is not uncommon for authors to "shop" their I-Ds to more than one WG at a time, with the hope of getting their documents adopted somewhere. + + After this state is implemented in the Datatracker, an I-D that is in the "Call for Adoption by WG Issued" state will not be able to be "shopped" to any other WG without the consent of the WG Chairs and the responsible ADs impacted by the shopping. + + Note that Figure 1 includes an arc leading from this state to outside of the WG state machine. This illustrates that some I-Ds that are considered do not get adopted as WG drafts. An I-D that is not adopted as a WG draft will transition out of the WG state machine and revert back to having no stream-specific state; however, the status change history log of the I-D will record that the I-D was previously in the "Call for Adoption by WG Issued" state."""), + ("Adopted by a WG","""4.2.2. Adopted by a WG + + The "Adopted by a WG" state describes an individual submission I-D that an IETF WG has agreed to adopt as one of its WG drafts. + + WG Chairs who use this state will be able to clearly indicate when their WGs adopt individual submission I-Ds. This will facilitate the Datatracker's ability to correctly capture "Replaces" information for WG drafts and correct "Replaced by" information for individual submission I-Ds that have been replaced by WG drafts. + + This state is needed because the Datatracker uses the filename of an I-D as a key to search its database for status information about the I-D, and because the filename of a WG I-D is supposed to be different from the filename of an individual submission I-D. The filename of an individual submission I-D will typically be formatted as 'draft-author-wgname-topic-nn'. + + The filename of a WG document is supposed to be formatted as 'draft- ietf-wgname-topic-nn'. + + An individual I-D that is adopted by a WG may take weeks or months to be resubmitted by the author as a new (version-00) WG draft. If the "Adopted by a WG" state is not used, the Datatracker has no way to determine that an I-D has been adopted until a new version of the I-D is submitted to the WG by the author and until the I-D is approved for posting by a WG Chair."""), + ("Adopted for WG Info Only","""4.2.3. Adopted for WG Info Only + + The "Adopted for WG Info Only" state describes a document that contains useful information for the WG that adopted it, but the document is not intended to be published as an RFC. The WG will not actively develop the contents of the I-D or progress it for publication as an RFC. The only purpose of the I-D is to provide information for internal use by the WG."""), + ]: + State.objects.filter(name=name).update(desc=desc) + +class Migration(migrations.Migration): + + dependencies = [ + ("doc", "0025_storedobject_storedobject_unique_name_per_store"), + ] + + operations = [ + migrations.RunPython(forward, reverse) + ] diff --git a/ietf/doc/migrations/0027_add_irsg_doc_positions.py b/ietf/doc/migrations/0027_add_irsg_doc_positions.py deleted file mode 100644 index e88ac9dda7..0000000000 --- a/ietf/doc/migrations/0027_add_irsg_doc_positions.py +++ /dev/null @@ -1,62 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.22 on 2019-08-03 10:09 - - -from django.db import migrations - -# forward, reverse initially copied from migration 0004 -def forward(apps, schema_editor): - State = apps.get_model('doc','State') - State.objects.create(type_id='draft-stream-irtf', - slug='irsg_review', - name='IRSG Review', - desc='IRSG Review', - used=True, - ) - BallotPositionName = apps.get_model('name','BallotPositionName') - # desc, used, order, and blocking all have suitable defaults - BallotPositionName.objects.create(slug="moretime", - name="Need More Time", - ) - BallotPositionName.objects.create(slug="notready", - name="Not Ready", - ) - - # Create a new ballot type for IRSG ballot - # include positions for the ballot type - BallotType = apps.get_model('doc','BallotType') - bt = BallotType.objects.create(doc_type_id="draft", - slug="irsg-approve", - name="IRSG Approve", - question="Is this draft ready for publication in the IRTF stream?", - ) - bt.positions.set(['yes','noobj','recuse','notready','moretime']) - -def reverse(apps, schema_editor): - State = apps.get_model('doc','State') - State.objects.filter(type_id__in=('draft-stream-irtf',), slug='irsg_review').delete() - - Position = apps.get_model('name','BallotPositionName') - for pos in ("moretime", "notready"): - Position.objects.filter(slug=pos).delete() - - IRSGBallot = apps.get_model('doc','BallotType') - IRSGBallot.objects.filter(slug="irsg-approve").delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0026_add_draft_rfceditor_state'), - ('name', '0007_fix_m2m_slug_id_length'), - ] - - operations = [ - migrations.RunPython(forward,reverse), - migrations.RenameField( - model_name='ballotpositiondocevent', - old_name='ad', - new_name='balloter', - ), - - ] diff --git a/ietf/doc/migrations/0027_alter_dochistory_title_alter_document_title.py b/ietf/doc/migrations/0027_alter_dochistory_title_alter_document_title.py new file mode 100644 index 0000000000..e0d8560e6f --- /dev/null +++ b/ietf/doc/migrations/0027_alter_dochistory_title_alter_document_title.py @@ -0,0 +1,41 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +import django.core.validators +from django.db import migrations, models + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0026_change_wg_state_descriptions"), + ] + + operations = [ + migrations.AlterField( + model_name="dochistory", + name="title", + field=models.CharField( + max_length=255, + validators=[ + django.core.validators.ProhibitNullCharactersValidator, # type:ignore + django.core.validators.RegexValidator( + message="Please enter a string without control characters.", + regex="^[^\x01-\x1f]*$", + ), + ], + ), + ), + migrations.AlterField( + model_name="document", + name="title", + field=models.CharField( + max_length=255, + validators=[ + django.core.validators.ProhibitNullCharactersValidator, # type:ignore + django.core.validators.RegexValidator( + message="Please enter a string without control characters.", + regex="^[^\x01-\x1f]*$", + ), + ], + ), + ), + ] diff --git a/ietf/doc/migrations/0028_irsgballotdocevent.py b/ietf/doc/migrations/0028_irsgballotdocevent.py deleted file mode 100644 index 92f4d52d73..0000000000 --- a/ietf/doc/migrations/0028_irsgballotdocevent.py +++ /dev/null @@ -1,25 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.25 on 2019-10-10 10:37 - - -from django.db import migrations, models -import django.db.models.deletion - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0027_add_irsg_doc_positions'), - ] - - operations = [ - migrations.CreateModel( - name='IRSGBallotDocEvent', - fields=[ - ('ballotdocevent_ptr', models.OneToOneField(auto_created=True, on_delete=django.db.models.deletion.CASCADE, parent_link=True, primary_key=True, serialize=False, to='doc.BallotDocEvent')), - ('duedate', models.DateTimeField(blank=True, null=True)), - ], - bases=('doc.ballotdocevent',), - ), - ] diff --git a/ietf/doc/migrations/0028_rfcauthor.py b/ietf/doc/migrations/0028_rfcauthor.py new file mode 100644 index 0000000000..776dc22eb1 --- /dev/null +++ b/ietf/doc/migrations/0028_rfcauthor.py @@ -0,0 +1,84 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +from django.db import migrations, models +import django.db.models.deletion +import ietf.utils.models + + +class Migration(migrations.Migration): + dependencies = [ + ("person", "0005_alter_historicalperson_pronouns_selectable_and_more"), + ("doc", "0027_alter_dochistory_title_alter_document_title"), + ] + + operations = [ + migrations.CreateModel( + name="RfcAuthor", + fields=[ + ( + "id", + models.AutoField( + auto_created=True, + primary_key=True, + serialize=False, + verbose_name="ID", + ), + ), + ("titlepage_name", models.CharField(max_length=128)), + ("is_editor", models.BooleanField(default=False)), + ( + "affiliation", + models.CharField( + blank=True, + help_text="Organization/company used by author for submission", + max_length=100, + ), + ), + ( + "country", + models.CharField( + blank=True, + help_text="Country used by author for submission", + max_length=255, + ), + ), + ("order", models.IntegerField(default=1)), + ( + "document", + ietf.utils.models.ForeignKey( + limit_choices_to={"type_id": "rfc"}, + on_delete=django.db.models.deletion.CASCADE, + to="doc.document", + ), + ), + ( + "email", + ietf.utils.models.ForeignKey( + blank=True, + help_text="Email address used by author for submission", + null=True, + on_delete=django.db.models.deletion.PROTECT, + to="person.email", + ), + ), + ( + "person", + ietf.utils.models.ForeignKey( + blank=True, + null=True, + on_delete=django.db.models.deletion.PROTECT, + to="person.person", + ), + ), + ], + options={ + "ordering": ["document", "order"], + "indexes": [ + models.Index( + fields=["document", "order"], + name="doc_rfcauth_documen_6b5dc4_idx", + ) + ], + }, + ), + ] diff --git a/ietf/doc/migrations/0029_add_ipr_event_types.py b/ietf/doc/migrations/0029_add_ipr_event_types.py deleted file mode 100644 index 08073b16e8..0000000000 --- a/ietf/doc/migrations/0029_add_ipr_event_types.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.27 on 2020-01-17 11:54 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0028_irsgballotdocevent'), - ] - - operations = [ - migrations.AlterField( - model_name='docevent', - name='type', - field=models.CharField(choices=[('new_revision', 'Added new revision'), ('new_submission', 'Uploaded new revision'), ('changed_document', 'Changed document metadata'), ('added_comment', 'Added comment'), ('added_message', 'Added message'), ('edited_authors', 'Edited the documents author list'), ('deleted', 'Deleted document'), ('changed_state', 'Changed state'), ('changed_stream', 'Changed document stream'), ('expired_document', 'Expired document'), ('extended_expiry', 'Extended expiry of document'), ('requested_resurrect', 'Requested resurrect'), ('completed_resurrect', 'Completed resurrect'), ('changed_consensus', 'Changed consensus'), ('published_rfc', 'Published RFC'), ('added_suggested_replaces', 'Added suggested replacement relationships'), ('reviewed_suggested_replaces', 'Reviewed suggested replacement relationships'), ('changed_group', 'Changed group'), ('changed_protocol_writeup', 'Changed protocol writeup'), ('changed_charter_milestone', 'Changed charter milestone'), ('initial_review', 'Set initial review time'), ('changed_review_announcement', 'Changed WG Review text'), ('changed_action_announcement', 'Changed WG Action text'), ('started_iesg_process', 'Started IESG process on document'), ('created_ballot', 'Created ballot'), ('closed_ballot', 'Closed ballot'), ('sent_ballot_announcement', 'Sent ballot announcement'), ('changed_ballot_position', 'Changed ballot position'), ('changed_ballot_approval_text', 'Changed ballot approval text'), ('changed_ballot_writeup_text', 'Changed ballot writeup text'), ('changed_rfc_editor_note_text', 'Changed RFC Editor Note text'), ('changed_last_call_text', 'Changed last call text'), ('requested_last_call', 'Requested last call'), ('sent_last_call', 'Sent last call'), ('scheduled_for_telechat', 'Scheduled for telechat'), ('iesg_approved', 'IESG approved document (no problem)'), ('iesg_disapproved', 'IESG disapproved document (do not publish)'), ('approved_in_minute', 'Approved in minute'), ('iana_review', 'IANA review comment'), ('rfc_in_iana_registry', 'RFC is in IANA registry'), ('rfc_editor_received_announcement', 'Announcement was received by RFC Editor'), ('requested_publication', 'Publication at RFC Editor requested'), ('sync_from_rfc_editor', 'Received updated information from RFC Editor'), ('requested_review', 'Requested review'), ('assigned_review_request', 'Assigned review request'), ('closed_review_request', 'Closed review request'), ('closed_review_assignment', 'Closed review assignment'), ('downref_approved', 'Downref approved'), ('posted_related_ipr', 'Posted related IPR'), ('removed_related_ipr', 'Removed related IPR')], max_length=50), - ), - ] diff --git a/ietf/doc/migrations/0029_editedrfcauthorsdocevent.py b/ietf/doc/migrations/0029_editedrfcauthorsdocevent.py new file mode 100644 index 0000000000..60837c5cb2 --- /dev/null +++ b/ietf/doc/migrations/0029_editedrfcauthorsdocevent.py @@ -0,0 +1,30 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +from django.db import migrations, models +import django.db.models.deletion + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0028_rfcauthor"), + ] + + operations = [ + migrations.CreateModel( + name="EditedRfcAuthorsDocEvent", + fields=[ + ( + "docevent_ptr", + models.OneToOneField( + auto_created=True, + on_delete=django.db.models.deletion.CASCADE, + parent_link=True, + primary_key=True, + serialize=False, + to="doc.docevent", + ), + ), + ], + bases=("doc.docevent",), + ), + ] diff --git a/ietf/doc/migrations/0030_alter_dochistory_title_alter_document_title.py b/ietf/doc/migrations/0030_alter_dochistory_title_alter_document_title.py new file mode 100644 index 0000000000..9ee858b2e8 --- /dev/null +++ b/ietf/doc/migrations/0030_alter_dochistory_title_alter_document_title.py @@ -0,0 +1,41 @@ +# Copyright The IETF Trust 2026, All Rights Reserved + +import django.core.validators +from django.db import migrations, models + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0029_editedrfcauthorsdocevent"), + ] + + operations = [ + migrations.AlterField( + model_name="dochistory", + name="title", + field=models.CharField( + max_length=255, + validators=[ + django.core.validators.ProhibitNullCharactersValidator(), + django.core.validators.RegexValidator( + message="Please enter a string without control characters.", + regex="^[^\x01-\x1f]*$", + ), + ], + ), + ), + migrations.AlterField( + model_name="document", + name="title", + field=models.CharField( + max_length=255, + validators=[ + django.core.validators.ProhibitNullCharactersValidator(), + django.core.validators.RegexValidator( + message="Please enter a string without control characters.", + regex="^[^\x01-\x1f]*$", + ), + ], + ), + ), + ] diff --git a/ietf/doc/migrations/0030_fix_bytes_mailarch_url.py b/ietf/doc/migrations/0030_fix_bytes_mailarch_url.py deleted file mode 100644 index 3c3ad2aa6f..0000000000 --- a/ietf/doc/migrations/0030_fix_bytes_mailarch_url.py +++ /dev/null @@ -1,37 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-21 14:27 - - -from __future__ import absolute_import, print_function, unicode_literals - -import re - -from django.conf import settings -from django.db import migrations - - -def forward(apps, schema_editor): - - Document = apps.get_model('doc', 'Document') - - print('') - for d in Document.objects.filter(external_url__contains="/b'"): - match = re.search("^(%s/arch/msg/[^/]+/)b'([^']+)'$" % settings.MAILING_LIST_ARCHIVE_URL, d.external_url) - if match: - d.external_url = "%s%s" % (match.group(1), match.group(2)) - d.save() - print('Fixed url #%s: %s' % (d.id, d.external_url)) - -def reverse(apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0029_add_ipr_event_types'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/doc/migrations/0031_change_draft_stream_ietf_state_descriptions.py b/ietf/doc/migrations/0031_change_draft_stream_ietf_state_descriptions.py new file mode 100644 index 0000000000..c664126da3 --- /dev/null +++ b/ietf/doc/migrations/0031_change_draft_stream_ietf_state_descriptions.py @@ -0,0 +1,57 @@ +# Copyright The IETF Trust 2026, All Rights Reserved + +from django.db import migrations + + +def forward(apps, schema_editor): + State = apps.get_model("doc", "State") + for name, desc in [ + ( + "Adopted by a WG", + "The individual submission document has been adopted by the Working Group (WG), but some administrative matter still needs to be completed (e.g., a WG document replacing this document with the typical naming convention of 'draft-ietf-wgname-topic-nn' has not yet been submitted).", + ), + ( + "WG Document", + "The document has been identified as a Working Group (WG) document and is under development per Section 7.2 of RFC2418.", + ), + ( + "Waiting for WG Chair Go-Ahead", + "The Working Group (WG) document has completed Working Group Last Call (WGLC), but the WG chairs are not yet ready to call consensus on the document. The reasons for this may include comments from the WGLC need to be responded to, or a revision to the document is needed.", + ), + ( + "Submitted to IESG for Publication", + "The Working Group (WG) document has been submitted to the Internet Engineering Steering Group (IESG) for evaluation and publication per Section 7.4 of RFC2418. See the “IESG State” or “RFC Editor State” for further details on the state of the document.", + ), + ]: + State.objects.filter(name=name).update(desc=desc, type="draft-stream-ietf") + + +def reverse(apps, schema_editor): + State = apps.get_model("doc", "State") + for name, desc in [ + ( + "Adopted by a WG", + "The individual submission document has been adopted by the Working Group (WG), but a WG document replacing this document with the typical naming convention of 'draft- ietf-wgname-topic-nn' has not yet been submitted.", + ), + ( + "WG Document", + "The document has been adopted by the Working Group (WG) and is under development. A document can only be adopted by one WG at a time. However, a document may be transferred between WGs.", + ), + ( + "Waiting for WG Chair Go-Ahead", + "The Working Group (WG) document has completed Working Group Last Call (WGLC), but the WG chair(s) are not yet ready to call consensus on the document. The reasons for this may include comments from the WGLC need to be responded to, or a revision to the document is needed", + ), + ( + "Submitted to IESG for Publication", + "The Working Group (WG) document has left the WG and been submitted to the Internet Engineering Steering Group (IESG) for evaluation and publication. See the “IESG State” or “RFC Editor State” for further details on the state of the document.", + ), + ]: + State.objects.filter(name=name).update(desc=desc, type="draft-stream-ietf") + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0030_alter_dochistory_title_alter_document_title"), + ] + + operations = [migrations.RunPython(forward, reverse)] diff --git a/ietf/doc/migrations/0031_set_state_for_charters_of_replaced_groups.py b/ietf/doc/migrations/0031_set_state_for_charters_of_replaced_groups.py deleted file mode 100644 index 1bf96f9a4d..0000000000 --- a/ietf/doc/migrations/0031_set_state_for_charters_of_replaced_groups.py +++ /dev/null @@ -1,42 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright The IETF Trust 2020, All Rights Reserved -# Generated by Django 1.11.28 on 2020-03-03 13:54 -from __future__ import unicode_literals - -from django.db import migrations - -def forward(apps, schema_editor): - - Person = apps.get_model('person', 'Person') - - Document = apps.get_model('doc','Document') - State = apps.get_model('doc','State') - BallotDocEvent = apps.get_model('doc','BallotDocEvent') - - replaced_state = State.objects.create(type_id='charter', slug='replaced', name='Replaced', used=True, desc="This charter's group was replaced.", order = 0) - by = Person.objects.get(name='(System)') - - for doc in Document.objects.filter(type_id='charter',states__type_id='charter',states__slug__in=['intrev','extrev'],group__state='replaced'): - doc.states.remove(*list(doc.states.filter(type_id='charter'))) - doc.states.add(replaced_state) - ballot = BallotDocEvent.objects.filter(doc=doc, type__in=('created_ballot', 'closed_ballot')).order_by('-time', '-id').first() - if ballot and ballot.type == 'created_ballot': - e = BallotDocEvent(type="closed_ballot", doc=doc, rev=doc.rev, by=by) - e.ballot_type = ballot.ballot_type - e.desc = 'Closed "%s" ballot' % e.ballot_type.name - e.save() - - -def reverse(apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0030_fix_bytes_mailarch_url'), - ('person', '0009_auto_20190118_0725'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/doc/migrations/0032_auto_20200624_1332.py b/ietf/doc/migrations/0032_auto_20200624_1332.py deleted file mode 100644 index 1dd656a31f..0000000000 --- a/ietf/doc/migrations/0032_auto_20200624_1332.py +++ /dev/null @@ -1,23 +0,0 @@ -# Generated by Django 2.1.15 on 2020-06-24 13:32 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0031_set_state_for_charters_of_replaced_groups'), - ] - - operations = [ - migrations.AlterField( - model_name='ballotpositiondocevent', - name='send_email', - field=models.BooleanField(default=None, null=True), - ), - migrations.AlterField( - model_name='consensusdocevent', - name='consensus', - field=models.BooleanField(default=None, null=True), - ), - ] diff --git a/ietf/doc/migrations/0032_remove_rfcauthor_email.py b/ietf/doc/migrations/0032_remove_rfcauthor_email.py new file mode 100644 index 0000000000..a0e147da59 --- /dev/null +++ b/ietf/doc/migrations/0032_remove_rfcauthor_email.py @@ -0,0 +1,16 @@ +# Copyright The IETF Trust 2026, All Rights Reserved + +from django.db import migrations + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0031_change_draft_stream_ietf_state_descriptions"), + ] + + operations = [ + migrations.RemoveField( + model_name="rfcauthor", + name="email", + ), + ] diff --git a/ietf/doc/migrations/0033_dochistory_keywords_document_keywords.py b/ietf/doc/migrations/0033_dochistory_keywords_document_keywords.py new file mode 100644 index 0000000000..5e2513e15a --- /dev/null +++ b/ietf/doc/migrations/0033_dochistory_keywords_document_keywords.py @@ -0,0 +1,31 @@ +# Copyright The IETF Trust 2026, All Rights Reserved + +from django.db import migrations, models +import ietf.doc.models + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0032_remove_rfcauthor_email"), + ] + + operations = [ + migrations.AddField( + model_name="dochistory", + name="keywords", + field=models.JSONField( + default=list, + max_length=1000, + validators=[ietf.doc.models.validate_doc_keywords], + ), + ), + migrations.AddField( + model_name="document", + name="keywords", + field=models.JSONField( + default=list, + max_length=1000, + validators=[ietf.doc.models.validate_doc_keywords], + ), + ), + ] diff --git a/ietf/doc/migrations/0033_populate_auth48_urls.py b/ietf/doc/migrations/0033_populate_auth48_urls.py deleted file mode 100644 index d6092b630d..0000000000 --- a/ietf/doc/migrations/0033_populate_auth48_urls.py +++ /dev/null @@ -1,67 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- -from __future__ import unicode_literals - -from django.db import migrations -from django.db.models import OuterRef, Subquery - -from re import match - - -def forward(apps, schema_editor): - """Add DocumentURLs for docs in the Auth48 state - - Checks the latest StateDocEvent; if it is in the auth48 state and the - event desc has an AUTH48 link, creates an auth48 DocumentURL for that doc. - """ - Document = apps.get_model('doc', 'Document') - StateDocEvent = apps.get_model('doc', 'StateDocEvent') - DocumentURL = apps.get_model('doc', 'DocumentURL') - - # Regex - extracts auth48 URL as first match group - pattern = r'RFC Editor state changed to AUTH48.*.*' - - # To avoid 100k queries, set up a subquery to find the latest StateDocEvent for each doc... - latest_events = StateDocEvent.objects.filter(doc=OuterRef('pk')).order_by('-time', '-id') - # ... then annotate the doc list with that and select only those in the auth48 state... - auth48_docs = Document.objects.annotate( - current_state_slug=Subquery(latest_events.values('state__slug')[:1]) - ).filter(current_state_slug='auth48') - # ... and add an auth48 DocumentURL if one is found. - for doc in auth48_docs: - # Retrieve the full StateDocEvent. Results in a query per doc, but - # only for the few few in the auth48 state. - sde = StateDocEvent.objects.filter(doc=doc).order_by('-time', '-id').first() - urlmatch = match(pattern, sde.desc) # Auth48 URL is usually in the event desc - if urlmatch is not None: - DocumentURL.objects.create(doc=doc, tag_id='auth48', url=urlmatch[1]) - - # Validate the migration using a different approach to find auth48 docs. - # This is slower than above, but still avoids querying for every Document. - auth48_events = StateDocEvent.objects.filter(state__slug='auth48') - for a48_event in auth48_events: - doc = a48_event.doc - latest_sde = StateDocEvent.objects.filter(doc=doc).order_by('-time', '-id').first() - if latest_sde.state and latest_sde.state.slug == 'auth48' and match(pattern, latest_sde.desc) is not None: - # Currently in the auth48 state with a URL - assert doc.documenturl_set.filter(tag_id='auth48').count() == 1 - else: - # Either no longer in auth48 state or had no URL - assert doc.documenturl_set.filter(tag_id='auth48').count() == 0 - - -def reverse(apps, schema_editor): - """Remove any auth48 DocumentURLs - these did not exist before""" - DocumentURL = apps.get_model('doc', 'DocumentURL') - DocumentURL.objects.filter(tag_id='auth48').delete() - - -class Migration(migrations.Migration): - dependencies = [ - ('doc', '0032_auto_20200624_1332'), - ('name', '0013_add_auth48_docurltagname'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/doc/migrations/0034_extres.py b/ietf/doc/migrations/0034_extres.py deleted file mode 100644 index 2157d0e863..0000000000 --- a/ietf/doc/migrations/0034_extres.py +++ /dev/null @@ -1,28 +0,0 @@ -# -*- coding: utf-8 -*- -# Generated by Django 1.11.29 on 2020-04-15 10:20 -from __future__ import unicode_literals - -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0014_extres'), - ('doc', '0033_populate_auth48_urls'), - ] - - operations = [ - migrations.CreateModel( - name='DocExtResource', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('display_name', models.CharField(blank=True, default='', max_length=255)), - ('value', models.CharField(max_length=2083)), - ('doc', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document')), - ('name', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ExtResourceName')), - ], - ), - ] diff --git a/ietf/doc/migrations/0035_populate_docextresources.py b/ietf/doc/migrations/0035_populate_docextresources.py deleted file mode 100644 index 04b396a963..0000000000 --- a/ietf/doc/migrations/0035_populate_docextresources.py +++ /dev/null @@ -1,125 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.29 on 2020-03-19 13:06 -from __future__ import unicode_literals - -import re - -import debug # pyflakes:ignore - -from collections import OrderedDict, Counter -from io import StringIO - -from django.db import migrations - -from ietf.utils.validators import validate_external_resource_value -from django.core.exceptions import ValidationError - - -name_map = { - "Issue.*": "tracker", - ".*FAQ.*": "faq", - ".*Area Web Page": "webpage", - ".*Wiki": "wiki", - "Home Page": "webpage", - "Slack.*": "slack", - "Additional .* Web Page": "webpage", - "Additional .* Page": "webpage", - "Yang catalog entry.*": "yc_entry", - "Yang impact analysis.*": "yc_impact", - "GitHub": "github_repo", - "Github page": "github_repo", - "GitHub repo.*": "github_repo", - "Github repository.*": "github_repo", - "GitHub org.*": "github_org", - "GitHub User.*": "github_username", - "GitLab User": "gitlab_username", - "GitLab User Name": "gitlab_username", -} - -url_map = OrderedDict({ - "https?://github\\.com": "github_repo", - "https://git.sr.ht/": "repo", - "https://todo.sr.ht/": "tracker", - "https?://trac\\.ietf\\.org/.*/wiki": "wiki", - "ietf\\.org.*/trac/wiki": "wiki", - "trac.*wiki": "wiki", - "www\\.ietf\\.org/mailman" : None, - "www\\.ietf\\.org/mail-archive" : None, - "mailarchive\\.ietf\\.org" : None, - "ietf\\.org/logs": "jabber_log", - "ietf\\.org/jabber/logs": "jabber_log", - "xmpp:.*?join": "jabber_room", - "bell-labs\\.com": None, - "html\\.charters": None, - "datatracker\\.ietf\\.org": None, -}) - -def forward(apps, schema_editor): - DocExtResource = apps.get_model('doc', 'DocExtResource') - ExtResourceName = apps.get_model('name', 'ExtResourceName') - DocumentUrl = apps.get_model('doc', 'DocumentUrl') - - stats = Counter() - stats_file = StringIO() - - for doc_url in DocumentUrl.objects.all(): - doc_url.url = doc_url.url.strip() - match_found = False - for regext,slug in name_map.items(): - if re.fullmatch(regext, doc_url.desc): - match_found = True - stats['mapped'] += 1 - name = ExtResourceName.objects.get(slug=slug) - try: - validate_external_resource_value(name, doc_url.url) - DocExtResource.objects.create(doc=doc_url.doc, name_id=slug, value=doc_url.url, display_name=doc_url.desc) - except ValidationError as e: # pyflakes:ignore - print("Failed validation:", doc_url.url, e, file=stats_file) - stats['failed_validation'] +=1 - break - if not match_found: - for regext, slug in url_map.items(): - if re.search(regext, doc_url.url): - match_found = True - if slug: - stats['mapped'] +=1 - name = ExtResourceName.objects.get(slug=slug) - # Munge the URL if it's the first github repo match - # Remove "/tree/master" substring if it exists - # Remove trailing "/issues" substring if it exists - # Remove "/blob/master/.*" pattern if present - if regext == "https?://github\\.com": - doc_url.url = doc_url.url.replace("/tree/master","") - doc_url.url = re.sub('/issues$', '', doc_url.url) - doc_url.url = re.sub('/blob/master.*$', '', doc_url.url) - try: - validate_external_resource_value(name, doc_url.url) - DocExtResource.objects.create(doc=doc_url.doc, name=name, value=doc_url.url, display_name=doc_url.desc) - except ValidationError as e: # pyflakes:ignore - print("Failed validation:", doc_url.url, e, file=stats_file) - stats['failed_validation'] +=1 - else: - stats['ignored'] +=1 - break - if not match_found: - print("Not Mapped:", doc_url.desc, doc_url.tag.slug, doc_url.doc.name, doc_url.url, file=stats_file) - stats['not_mapped'] += 1 - print('') - print(stats_file.getvalue()) - print (stats) - -def reverse(apps, schema_editor): - DocExtResource = apps.get_model('doc', 'DocExtResource') - DocExtResource.objects.all().delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0034_extres'), - ('name', '0015_populate_extres'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/doc/migrations/0036_orgs_vs_repos.py b/ietf/doc/migrations/0036_orgs_vs_repos.py deleted file mode 100644 index 37a0db0d26..0000000000 --- a/ietf/doc/migrations/0036_orgs_vs_repos.py +++ /dev/null @@ -1,34 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved - -from urllib.parse import urlparse -from django.db import migrations - -def categorize(url): - # This will categorize a few urls pointing into files in a repo as a repo, but that's better than calling them an org - element_count = len(urlparse(url).path.strip('/').split('/')) - if element_count < 1: - print("Bad github resource:",url) - return 'github_org' if element_count == 1 else 'github_repo' - -def forward(apps, schema_editor): - DocExtResource = apps.get_model('doc','DocExtResource') - - for resource in DocExtResource.objects.filter(name__slug__in=('github_org','github_repo')): - category = categorize(resource.value) - if resource.name_id != category: - resource.name_id = category - resource.save() - -def reverse(apps, schema_editor): - # Intentionally don't try to return to former worse state - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0035_populate_docextresources'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/doc/migrations/0037_clean_up_missing_docaliases.py b/ietf/doc/migrations/0037_clean_up_missing_docaliases.py deleted file mode 100644 index 6ce350c3f8..0000000000 --- a/ietf/doc/migrations/0037_clean_up_missing_docaliases.py +++ /dev/null @@ -1,29 +0,0 @@ -# Generated by Django 2.2.16 on 2020-09-22 07:58 - -from django.db import migrations - - -def forward(apps, schema_editor): - Document = apps.get_model('doc', 'Document') - DocAlias = apps.get_model('doc', 'DocAlias') - - docs_without_alias = Document.objects.filter(docalias__isnull=True) - - bad_aliases = DocAlias.objects.filter(name__in=docs_without_alias.values_list('name')) - bad_aliases.delete() - - for doc in docs_without_alias: - DocAlias.objects.create(name=doc.name).docs.add(doc) - -def reverse(apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0036_orgs_vs_repos'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/doc/migrations/0038_auto_20201109_0429.py b/ietf/doc/migrations/0038_auto_20201109_0429.py deleted file mode 100644 index 1335032b28..0000000000 --- a/ietf/doc/migrations/0038_auto_20201109_0429.py +++ /dev/null @@ -1,17 +0,0 @@ -# Generated by Django 2.2.17 on 2020-11-09 04:29 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0037_clean_up_missing_docaliases'), - ] - - operations = [ - migrations.AddIndex( - model_name='docevent', - index=models.Index(fields=['-time', '-id'], name='doc_doceven_time_1a258f_idx'), - ), - ] diff --git a/ietf/doc/migrations/0039_auto_20201109_0439.py b/ietf/doc/migrations/0039_auto_20201109_0439.py deleted file mode 100644 index e6e3364568..0000000000 --- a/ietf/doc/migrations/0039_auto_20201109_0439.py +++ /dev/null @@ -1,21 +0,0 @@ -# Generated by Django 2.2.17 on 2020-11-09 04:39 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0038_auto_20201109_0429'), - ] - - operations = [ - migrations.AddIndex( - model_name='dochistoryauthor', - index=models.Index(fields=['document', 'order'], name='doc_dochist_documen_7e2441_idx'), - ), - migrations.AddIndex( - model_name='documentauthor', - index=models.Index(fields=['document', 'order'], name='doc_documen_documen_7fabe2_idx'), - ), - ] diff --git a/ietf/doc/migrations/0040_add_changed_action_holders_docevent_type.py b/ietf/doc/migrations/0040_add_changed_action_holders_docevent_type.py deleted file mode 100644 index 3aa1278712..0000000000 --- a/ietf/doc/migrations/0040_add_changed_action_holders_docevent_type.py +++ /dev/null @@ -1,19 +0,0 @@ -# Generated by Django 2.2.17 on 2021-01-15 12:50 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0040_lengthen_used_roles_fields'), # only needed for schema vs data ordering - ('doc', '0039_auto_20201109_0439'), - ] - - operations = [ - migrations.AlterField( - model_name='docevent', - name='type', - field=models.CharField(choices=[('new_revision', 'Added new revision'), ('new_submission', 'Uploaded new revision'), ('changed_document', 'Changed document metadata'), ('added_comment', 'Added comment'), ('added_message', 'Added message'), ('edited_authors', 'Edited the documents author list'), ('deleted', 'Deleted document'), ('changed_state', 'Changed state'), ('changed_stream', 'Changed document stream'), ('expired_document', 'Expired document'), ('extended_expiry', 'Extended expiry of document'), ('requested_resurrect', 'Requested resurrect'), ('completed_resurrect', 'Completed resurrect'), ('changed_consensus', 'Changed consensus'), ('published_rfc', 'Published RFC'), ('added_suggested_replaces', 'Added suggested replacement relationships'), ('reviewed_suggested_replaces', 'Reviewed suggested replacement relationships'), ('changed_action_holders', 'Changed action holders for document'), ('changed_group', 'Changed group'), ('changed_protocol_writeup', 'Changed protocol writeup'), ('changed_charter_milestone', 'Changed charter milestone'), ('initial_review', 'Set initial review time'), ('changed_review_announcement', 'Changed WG Review text'), ('changed_action_announcement', 'Changed WG Action text'), ('started_iesg_process', 'Started IESG process on document'), ('created_ballot', 'Created ballot'), ('closed_ballot', 'Closed ballot'), ('sent_ballot_announcement', 'Sent ballot announcement'), ('changed_ballot_position', 'Changed ballot position'), ('changed_ballot_approval_text', 'Changed ballot approval text'), ('changed_ballot_writeup_text', 'Changed ballot writeup text'), ('changed_rfc_editor_note_text', 'Changed RFC Editor Note text'), ('changed_last_call_text', 'Changed last call text'), ('requested_last_call', 'Requested last call'), ('sent_last_call', 'Sent last call'), ('scheduled_for_telechat', 'Scheduled for telechat'), ('iesg_approved', 'IESG approved document (no problem)'), ('iesg_disapproved', 'IESG disapproved document (do not publish)'), ('approved_in_minute', 'Approved in minute'), ('iana_review', 'IANA review comment'), ('rfc_in_iana_registry', 'RFC is in IANA registry'), ('rfc_editor_received_announcement', 'Announcement was received by RFC Editor'), ('requested_publication', 'Publication at RFC Editor requested'), ('sync_from_rfc_editor', 'Received updated information from RFC Editor'), ('requested_review', 'Requested review'), ('assigned_review_request', 'Assigned review request'), ('closed_review_request', 'Closed review request'), ('closed_review_assignment', 'Closed review assignment'), ('downref_approved', 'Downref approved'), ('posted_related_ipr', 'Posted related IPR'), ('removed_related_ipr', 'Removed related IPR')], max_length=50), - ), - ] diff --git a/ietf/doc/migrations/0041_add_documentactionholder.py b/ietf/doc/migrations/0041_add_documentactionholder.py deleted file mode 100644 index 3832a5603f..0000000000 --- a/ietf/doc/migrations/0041_add_documentactionholder.py +++ /dev/null @@ -1,35 +0,0 @@ -# Generated by Django 2.2.17 on 2021-01-15 12:50 - -import datetime -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0018_auto_20201109_0439'), - ('doc', '0040_add_changed_action_holders_docevent_type'), - ] - - operations = [ - migrations.CreateModel( - name='DocumentActionHolder', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('time_added', models.DateTimeField(default=datetime.datetime.now)), - ('document', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document')), - ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), - ], - ), - migrations.AddField( - model_name='document', - name='action_holders', - field=models.ManyToManyField(blank=True, through='doc.DocumentActionHolder', to='person.Person'), - ), - migrations.AddConstraint( - model_name='documentactionholder', - constraint=models.UniqueConstraint(fields=('document', 'person'), name='unique_action_holder'), - ), - ] diff --git a/ietf/doc/migrations/0042_bofreq_states.py b/ietf/doc/migrations/0042_bofreq_states.py deleted file mode 100644 index 95119f8147..0000000000 --- a/ietf/doc/migrations/0042_bofreq_states.py +++ /dev/null @@ -1,36 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -# Generated by Django 2.2.23 on 2021-05-21 13:29 - -from django.db import migrations - -def forward(apps, schema_editor): - StateType = apps.get_model('doc', 'StateType') - State = apps.get_model('doc', 'State') - - StateType.objects.create(slug='bofreq', label='BOF Request State') - proposed = State.objects.create(type_id='bofreq', slug='proposed', name='Proposed', used=True, desc='The BOF request is proposed', order=0) - approved = State.objects.create(type_id='bofreq', slug='approved', name='Approved', used=True, desc='The BOF request is approved', order=1) - declined = State.objects.create(type_id='bofreq', slug='declined', name='Declined', used=True, desc='The BOF request is declined', order=2) - replaced = State.objects.create(type_id='bofreq', slug='replaced', name='Replaced', used=True, desc='The BOF request is proposed', order=3) - abandoned = State.objects.create(type_id='bofreq', slug='abandoned', name='Abandoned', used=True, desc='The BOF request is abandoned', order=4) - - proposed.next_states.set([approved,declined,replaced,abandoned]) - -def reverse(apps, schema_editor): - StateType = apps.get_model('doc', 'StateType') - State = apps.get_model('doc', 'State') - State.objects.filter(type_id='bofreq').delete() - StateType.objects.filter(slug='bofreq').delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0041_add_documentactionholder'), - ('name', '0027_add_bofrequest'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/doc/migrations/0043_bofreq_docevents.py b/ietf/doc/migrations/0043_bofreq_docevents.py deleted file mode 100644 index 7a300426b3..0000000000 --- a/ietf/doc/migrations/0043_bofreq_docevents.py +++ /dev/null @@ -1,36 +0,0 @@ -# Generated by Django 2.2.24 on 2021-07-06 13:34 - -from django.db import migrations, models -import django.db.models.deletion - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0019_auto_20210604_1443'), - ('doc', '0042_bofreq_states'), - ] - - operations = [ - migrations.AlterField( - model_name='docevent', - name='type', - field=models.CharField(choices=[('new_revision', 'Added new revision'), ('new_submission', 'Uploaded new revision'), ('changed_document', 'Changed document metadata'), ('added_comment', 'Added comment'), ('added_message', 'Added message'), ('edited_authors', 'Edited the documents author list'), ('deleted', 'Deleted document'), ('changed_state', 'Changed state'), ('changed_stream', 'Changed document stream'), ('expired_document', 'Expired document'), ('extended_expiry', 'Extended expiry of document'), ('requested_resurrect', 'Requested resurrect'), ('completed_resurrect', 'Completed resurrect'), ('changed_consensus', 'Changed consensus'), ('published_rfc', 'Published RFC'), ('added_suggested_replaces', 'Added suggested replacement relationships'), ('reviewed_suggested_replaces', 'Reviewed suggested replacement relationships'), ('changed_action_holders', 'Changed action holders for document'), ('changed_group', 'Changed group'), ('changed_protocol_writeup', 'Changed protocol writeup'), ('changed_charter_milestone', 'Changed charter milestone'), ('initial_review', 'Set initial review time'), ('changed_review_announcement', 'Changed WG Review text'), ('changed_action_announcement', 'Changed WG Action text'), ('started_iesg_process', 'Started IESG process on document'), ('created_ballot', 'Created ballot'), ('closed_ballot', 'Closed ballot'), ('sent_ballot_announcement', 'Sent ballot announcement'), ('changed_ballot_position', 'Changed ballot position'), ('changed_ballot_approval_text', 'Changed ballot approval text'), ('changed_ballot_writeup_text', 'Changed ballot writeup text'), ('changed_rfc_editor_note_text', 'Changed RFC Editor Note text'), ('changed_last_call_text', 'Changed last call text'), ('requested_last_call', 'Requested last call'), ('sent_last_call', 'Sent last call'), ('scheduled_for_telechat', 'Scheduled for telechat'), ('iesg_approved', 'IESG approved document (no problem)'), ('iesg_disapproved', 'IESG disapproved document (do not publish)'), ('approved_in_minute', 'Approved in minute'), ('iana_review', 'IANA review comment'), ('rfc_in_iana_registry', 'RFC is in IANA registry'), ('rfc_editor_received_announcement', 'Announcement was received by RFC Editor'), ('requested_publication', 'Publication at RFC Editor requested'), ('sync_from_rfc_editor', 'Received updated information from RFC Editor'), ('requested_review', 'Requested review'), ('assigned_review_request', 'Assigned review request'), ('closed_review_request', 'Closed review request'), ('closed_review_assignment', 'Closed review assignment'), ('downref_approved', 'Downref approved'), ('posted_related_ipr', 'Posted related IPR'), ('removed_related_ipr', 'Removed related IPR'), ('changed_editors', 'Changed BOF Request editors')], max_length=50), - ), - migrations.CreateModel( - name='BofreqResponsibleDocEvent', - fields=[ - ('docevent_ptr', models.OneToOneField(auto_created=True, on_delete=django.db.models.deletion.CASCADE, parent_link=True, primary_key=True, serialize=False, to='doc.DocEvent')), - ('responsible', models.ManyToManyField(blank=True, to='person.Person')), - ], - bases=('doc.docevent',), - ), - migrations.CreateModel( - name='BofreqEditorDocEvent', - fields=[ - ('docevent_ptr', models.OneToOneField(auto_created=True, on_delete=django.db.models.deletion.CASCADE, parent_link=True, primary_key=True, serialize=False, to='doc.DocEvent')), - ('editors', models.ManyToManyField(blank=True, to='person.Person')), - ], - bases=('doc.docevent',), - ), - ] diff --git a/ietf/doc/migrations/0044_procmaterials_states.py b/ietf/doc/migrations/0044_procmaterials_states.py deleted file mode 100644 index 2892935881..0000000000 --- a/ietf/doc/migrations/0044_procmaterials_states.py +++ /dev/null @@ -1,34 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -# Generated by Django 2.2.23 on 2021-05-21 13:29 - -from django.db import migrations - -def forward(apps, schema_editor): - StateType = apps.get_model('doc', 'StateType') - State = apps.get_model('doc', 'State') - - StateType.objects.create(slug='procmaterials', label='Proceedings Materials State') - active = State.objects.create(type_id='procmaterials', slug='active', name='Active', used=True, desc='The material is active', order=0) - removed = State.objects.create(type_id='procmaterials', slug='removed', name='Removed', used=True, desc='The material is removed', order=1) - - active.next_states.set([removed]) - removed.next_states.set([active]) - -def reverse(apps, schema_editor): - StateType = apps.get_model('doc', 'StateType') - State = apps.get_model('doc', 'State') - State.objects.filter(type_id='procmaterials').delete() - StateType.objects.filter(slug='procmaterials').delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0043_bofreq_docevents'), - ('name', '0031_add_procmaterials'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/doc/migrations/0045_docstates_chatlogs_polls.py b/ietf/doc/migrations/0045_docstates_chatlogs_polls.py deleted file mode 100644 index 044cc60de4..0000000000 --- a/ietf/doc/migrations/0045_docstates_chatlogs_polls.py +++ /dev/null @@ -1,34 +0,0 @@ -# Copyright The IETF Trust 2022, All Rights Reserved -from django.db import migrations - -def forward(apps, schema_editor): - StateType = apps.get_model("doc", "StateType") - State = apps.get_model("doc", "State") - for slug in ("chatlog", "polls"): - StateType.objects.create(slug=slug, label="State") - for state_slug in ("active", "deleted"): - State.objects.create( - type_id = slug, - slug = state_slug, - name = state_slug.capitalize(), - used = True, - desc = "", - order = 0, - ) - -def reverse(apps, schema_editor): - StateType = apps.get_model("doc", "StateType") - State = apps.get_model("doc", "State") - State.objects.filter(type_id__in=("chatlog", "polls")).delete() - StateType.objects.filter(slug__in=("chatlog", "polls")).delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0044_procmaterials_states'), - ('name', '0045_polls_and_chatlogs'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/doc/migrations/0046_use_timezone_now_for_doc_models.py b/ietf/doc/migrations/0046_use_timezone_now_for_doc_models.py deleted file mode 100644 index 3749fd973f..0000000000 --- a/ietf/doc/migrations/0046_use_timezone_now_for_doc_models.py +++ /dev/null @@ -1,39 +0,0 @@ -# Generated by Django 2.2.28 on 2022-07-12 11:24 - -from django.db import migrations, models -import django.utils.timezone - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0045_docstates_chatlogs_polls'), - ] - - operations = [ - migrations.AlterField( - model_name='deletedevent', - name='time', - field=models.DateTimeField(default=django.utils.timezone.now), - ), - migrations.AlterField( - model_name='docevent', - name='time', - field=models.DateTimeField(db_index=True, default=django.utils.timezone.now, help_text='When the event happened'), - ), - migrations.AlterField( - model_name='dochistory', - name='time', - field=models.DateTimeField(default=django.utils.timezone.now), - ), - migrations.AlterField( - model_name='document', - name='time', - field=models.DateTimeField(default=django.utils.timezone.now), - ), - migrations.AlterField( - model_name='documentactionholder', - name='time_added', - field=models.DateTimeField(default=django.utils.timezone.now), - ), - ] diff --git a/ietf/doc/migrations/0047_tzaware_deletedevents.py b/ietf/doc/migrations/0047_tzaware_deletedevents.py deleted file mode 100644 index bf258de6e1..0000000000 --- a/ietf/doc/migrations/0047_tzaware_deletedevents.py +++ /dev/null @@ -1,61 +0,0 @@ -# Generated by Django 2.2.28 on 2022-08-31 20:26 - -import datetime -import json - -from zoneinfo import ZoneInfo - -from django.db import migrations - - -TZ_BEFORE = ZoneInfo('PST8PDT') - - -def forward(apps, schema_editor): - DeletedEvent = apps.get_model('doc', 'DeletedEvent') - for deleted_event in DeletedEvent.objects.all(): - fields = json.loads(deleted_event.json) - replacements = {} - for k, v in fields.items(): - if isinstance(v, str): - try: - dt = datetime.datetime.strptime(v, '%Y-%m-%d %H:%M:%S') - except: - pass - else: - replacements[k] = dt.replace(tzinfo=TZ_BEFORE).astimezone(datetime.timezone.utc).isoformat() - if len(replacements) > 0: - fields.update(replacements) - deleted_event.json = json.dumps(fields) - deleted_event.save() - - -def reverse(apps, schema_editor): - DeletedEvent = apps.get_model('doc', 'DeletedEvent') - for deleted_event in DeletedEvent.objects.all(): - fields = json.loads(deleted_event.json) - replacements = {} - for k, v in fields.items(): - if isinstance(v, str) and 'T' in v: - try: - dt = datetime.datetime.fromisoformat(v) - except: - pass - else: - replacements[k] = dt.astimezone(TZ_BEFORE).replace(tzinfo=None).strftime('%Y-%m-%d %H:%M:%S') - if len(replacements) > 0: - fields.update(replacements) - deleted_event.json = json.dumps(fields) - deleted_event.save() - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0046_use_timezone_now_for_doc_models'), - ('utils', '0003_pause_to_change_use_tz'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/doc/migrations/0048_allow_longer_notify.py b/ietf/doc/migrations/0048_allow_longer_notify.py deleted file mode 100644 index f2c6959d8d..0000000000 --- a/ietf/doc/migrations/0048_allow_longer_notify.py +++ /dev/null @@ -1,23 +0,0 @@ -# Generated by Django 2.2.28 on 2022-12-05 17:53 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0047_tzaware_deletedevents'), - ] - - operations = [ - migrations.AlterField( - model_name='dochistory', - name='notify', - field=models.TextField(blank=True, max_length=1023), - ), - migrations.AlterField( - model_name='document', - name='notify', - field=models.TextField(blank=True, max_length=1023), - ), - ] diff --git a/ietf/doc/models.py b/ietf/doc/models.py index 42086721b9..cc79b73831 100644 --- a/ietf/doc/models.py +++ b/ietf/doc/models.py @@ -1,41 +1,58 @@ -# Copyright The IETF Trust 2010-2020, All Rights Reserved +# Copyright The IETF Trust 2010-2026, All Rights Reserved # -*- coding: utf-8 -*- +from collections import namedtuple import datetime import logging -import io import os + +import django.db import rfc2html +from io import BufferedReader from pathlib import Path + +from django.core.exceptions import ValidationError +from django.db.models import Q from lxml import etree -from typing import Optional, TYPE_CHECKING +from typing import Optional, Protocol, TYPE_CHECKING, Union from weasyprint import HTML as wpHTML +from weasyprint.text.fonts import FontConfiguration from django.db import models from django.core import checks +from django.core.files.base import File from django.core.cache import caches -from django.core.validators import URLValidator, RegexValidator +from django.core.validators import ( + URLValidator, + RegexValidator, + ProhibitNullCharactersValidator, +) from django.urls import reverse as urlreverse from django.contrib.contenttypes.models import ContentType from django.conf import settings from django.utils import timezone -from django.utils.encoding import force_text +from django.utils.encoding import force_str from django.utils.html import mark_safe # type:ignore from django.contrib.staticfiles import finders import debug # pyflakes:ignore from ietf.group.models import Group +from ietf.doc.storage_utils import ( + store_str as utils_store_str, + store_bytes as utils_store_bytes, + store_file as utils_store_file +) from ietf.name.models import ( DocTypeName, DocTagName, StreamName, IntendedStdLevelName, StdLevelName, DocRelationshipName, DocReminderTypeName, BallotPositionName, ReviewRequestStateName, ReviewAssignmentStateName, FormalLanguageName, DocUrlTagName, ExtResourceName) from ietf.person.models import Email, Person from ietf.person.utils import get_active_balloters from ietf.utils import log -from ietf.utils.admin import admin_link from ietf.utils.decorators import memoize +from ietf.utils.text import decode_document_content from ietf.utils.validators import validate_no_control_chars from ietf.utils.mail import formataddr from ietf.utils.models import ForeignKey @@ -56,16 +73,22 @@ def __str__(self): @checks.register('db-consistency') def check_statetype_slugs(app_configs, **kwargs): errors = [] - state_type_slugs = [ t.slug for t in StateType.objects.all() ] - for type in DocTypeName.objects.all(): - if not type.slug in state_type_slugs: - errors.append(checks.Error( - "The document type '%s (%s)' does not have a corresponding entry in the doc.StateType table" % (type.name, type.slug), - hint="You should add a doc.StateType entry with a slug '%s' to match the DocTypeName slug."%(type.slug), - obj=type, - id='datatracker.doc.E0015', - )) - return errors + try: + state_type_slugs = [ t.slug for t in StateType.objects.all() ] + except django.db.ProgrammingError: + # When running initial migrations on an empty DB, attempting to retrieve StateType will raise a + # ProgrammingError. Until Django 3, there is no option to skip the checks. + return [] + else: + for type in DocTypeName.objects.all(): + if not type.slug in state_type_slugs: + errors.append(checks.Error( + "The document type '%s (%s)' does not have a corresponding entry in the doc.StateType table" % (type.name, type.slug), + hint="You should add a doc.StateType entry with a slug '%s' to match the DocTypeName slug."%(type.slug), + obj=type, + id='datatracker.doc.E0015', + )) + return errors class State(models.Model): type = ForeignKey(StateType) @@ -75,7 +98,7 @@ class State(models.Model): desc = models.TextField(blank=True) order = models.IntegerField(default=0) - next_states = models.ManyToManyField('State', related_name="previous_states", blank=True) + next_states = models.ManyToManyField('doc.State', related_name="previous_states", blank=True) def __str__(self): return self.name @@ -88,16 +111,31 @@ class Meta: IESG_STATCHG_CONFLREV_ACTIVE_STATES = ("iesgeval", "defer") IESG_SUBSTATE_TAGS = ('ad-f-up', 'need-rev', 'extpty') + +def validate_doc_keywords(value): + if ( + not isinstance(value, list | tuple | set) + or not all(isinstance(elt, str) for elt in value) + ): + raise ValidationError("Value must be an array of strings") + + class DocumentInfo(models.Model): """Any kind of document. Draft, RFC, Charter, IPR Statement, Liaison Statement""" time = models.DateTimeField(default=timezone.now) # should probably have auto_now=True type = ForeignKey(DocTypeName, blank=True, null=True) # Draft, Agenda, Minutes, Charter, Discuss, Guideline, Email, Review, Issue, Wiki, External ... - title = models.CharField(max_length=255, validators=[validate_no_control_chars, ]) + title = models.CharField( + max_length=255, + validators=[ + ProhibitNullCharactersValidator(), + validate_no_control_chars, + ], + ) states = models.ManyToManyField(State, blank=True) # plain state (Active/Expired/...), IESG state, stream state tags = models.ManyToManyField(DocTagName, blank=True) # Revised ID Needed, ExternalParty, AD Followup, ... - stream = ForeignKey(StreamName, blank=True, null=True) # IETF, IAB, IRTF, Independent Submission + stream = ForeignKey(StreamName, blank=True, null=True) # IETF, IAB, IRTF, Independent Submission, Editorial group = ForeignKey(Group, blank=True, null=True) # WG, RG, IAB, IESG, Edu, Tools abstract = models.TextField(blank=True) @@ -105,7 +143,6 @@ class DocumentInfo(models.Model): pages = models.IntegerField(blank=True, null=True) words = models.IntegerField(blank=True, null=True) formal_languages = models.ManyToManyField(FormalLanguageName, blank=True, help_text="Formal languages used in document") - order = models.IntegerField(default=1, blank=True) # This is probably obviated by SessionPresentaion.order intended_std_level = ForeignKey(IntendedStdLevelName, verbose_name="Intended standardization level", blank=True, null=True) std_level = ForeignKey(StdLevelName, verbose_name="Standardization level", blank=True, null=True) ad = ForeignKey(Person, verbose_name="area director", related_name='ad_%(class)s_set', blank=True, null=True) @@ -115,7 +152,18 @@ class DocumentInfo(models.Model): external_url = models.URLField(blank=True) uploaded_filename = models.TextField(blank=True) note = models.TextField(blank=True) - internal_comments = models.TextField(blank=True) + rfc_number = models.PositiveIntegerField(blank=True, null=True) # only valid for type="rfc" + keywords = models.JSONField( + default=list, + max_length=1000, + validators=[validate_doc_keywords], + ) + + @property + def doi(self) -> str | None: + if self.type_id == "rfc" and self.rfc_number is not None: + return f"{settings.IETF_DOI_PREFIX}/RFC{self.rfc_number:04d}" + return None def file_extension(self): if not hasattr(self, '_cached_extension'): @@ -128,20 +176,20 @@ def file_extension(self): def get_file_path(self): if not hasattr(self, '_cached_file_path'): - if self.type_id == "draft": + if self.type_id == "rfc": + self._cached_file_path = settings.RFC_PATH + elif self.type_id == "draft": if self.is_dochistory(): self._cached_file_path = settings.INTERNET_ALL_DRAFTS_ARCHIVE_DIR else: - if self.get_state_slug() == "rfc": - self._cached_file_path = settings.RFC_PATH + # This could be simplified since anything in INTERNET_DRAFT_PATH is also already in INTERNET_ALL_DRAFTS_ARCHIVE_DIR + draft_state = self.get_state('draft') + if draft_state and draft_state.slug == 'active': + self._cached_file_path = settings.INTERNET_DRAFT_PATH else: - draft_state = self.get_state('draft') - if draft_state and draft_state.slug == 'active': - self._cached_file_path = settings.INTERNET_DRAFT_PATH - else: - self._cached_file_path = settings.INTERNET_ALL_DRAFTS_ARCHIVE_DIR + self._cached_file_path = settings.INTERNET_ALL_DRAFTS_ARCHIVE_DIR elif self.meeting_related() and self.type_id in ( - "agenda", "minutes", "slides", "bluesheets", "procmaterials", "chatlog", "polls" + "agenda", "minutes", "narrativeminutes", "slides", "bluesheets", "procmaterials", "chatlog", "polls" ): meeting = self.get_related_meeting() if meeting is not None: @@ -154,7 +202,7 @@ def get_file_path(self): self._cached_file_path = settings.CONFLICT_REVIEW_PATH elif self.type_id == "statchg": self._cached_file_path = settings.STATUS_CHANGE_PATH - elif self.type_id == "bofreq": + elif self.type_id == "bofreq": # TODO: This is probably unneeded, as is the separate path setting self._cached_file_path = settings.BOFREQ_PATH else: self._cached_file_path = settings.DOCUMENT_PATH_PATTERN.format(doc=self) @@ -164,27 +212,26 @@ def get_base_name(self): if not hasattr(self, '_cached_base_name'): if self.uploaded_filename: self._cached_base_name = self.uploaded_filename + elif self.type_id == 'rfc': + self._cached_base_name = "%s.txt" % self.name elif self.type_id == 'draft': if self.is_dochistory(): self._cached_base_name = "%s-%s.txt" % (self.doc.name, self.rev) else: - if self.get_state_slug() == 'rfc': - self._cached_base_name = "%s.txt" % self.canonical_name() - else: - self._cached_base_name = "%s-%s.txt" % (self.name, self.rev) + self._cached_base_name = "%s-%s.txt" % (self.name, self.rev) elif self.type_id in ["slides", "agenda", "minutes", "bluesheets", "procmaterials", ] and self.meeting_related(): ext = 'pdf' if self.type_id == 'procmaterials' else 'txt' - self._cached_base_name = f'{self.canonical_name()}-{self.rev}.{ext}' + self._cached_base_name = f'{self.name}-{self.rev}.{ext}' elif self.type_id == 'review': # TODO: This will be wrong if a review is updated on the same day it was created (or updated more than once on the same day) self._cached_base_name = "%s.txt" % self.name - elif self.type_id == 'bofreq': + elif self.type_id in ['bofreq', 'statement']: self._cached_base_name = "%s-%s.md" % (self.name, self.rev) else: if self.rev: - self._cached_base_name = "%s-%s.txt" % (self.canonical_name(), self.rev) + self._cached_base_name = "%s-%s.txt" % (self.name, self.rev) else: - self._cached_base_name = "%s.txt" % (self.canonical_name(), ) + self._cached_base_name = "%s.txt" % (self.name, ) return self._cached_base_name def get_file_name(self): @@ -192,27 +239,38 @@ def get_file_name(self): self._cached_file_name = os.path.join(self.get_file_path(), self.get_base_name()) return self._cached_file_name - def revisions(self): + + def revisions_by_dochistory(self): revisions = [] - doc = self.doc if isinstance(self, DocHistory) else self - for e in doc.docevent_set.filter(type='new_revision').distinct(): - if e.rev and not e.rev in revisions: - revisions.append(e.rev) - if not doc.rev in revisions: - revisions.append(doc.rev) - revisions.sort() + if self.type_id != "rfc": + for h in self.history_set.order_by("time", "id"): + if h.rev and not h.rev in revisions: + revisions.append(h.rev) + if not self.rev in revisions: + revisions.append(self.rev) return revisions + def revisions_by_newrevisionevent(self): + revisions = [] + if self.type_id != "rfc": + doc = self.doc if isinstance(self, DocHistory) else self + for e in doc.docevent_set.filter(type='new_revision').distinct(): + if e.rev and not e.rev in revisions: + revisions.append(e.rev) + if not doc.rev in revisions: + revisions.append(doc.rev) + revisions.sort() + return revisions def get_href(self, meeting=None): - return self._get_ref(meeting=meeting,meeting_doc_refs=settings.MEETING_DOC_HREFS) + return self._get_ref(meeting=meeting, versioned=True) def get_versionless_href(self, meeting=None): - return self._get_ref(meeting=meeting,meeting_doc_refs=settings.MEETING_DOC_GREFS) + return self._get_ref(meeting=meeting, versioned=False) - def _get_ref(self, meeting=None, meeting_doc_refs=settings.MEETING_DOC_HREFS): + def _get_ref(self, meeting=None, versioned=True): """ Returns an url to the document text. This differs from .get_absolute_url(), which returns an url to the datatracker page for the document. @@ -221,12 +279,16 @@ def _get_ref(self, meeting=None, meeting_doc_refs=settings.MEETING_DOC_HREFS): # the earlier resolution order, but there's at the moment one single # instance which matches this (with correct results), so we won't # break things all over the place. - if not hasattr(self, '_cached_href'): + cache_attr = "_cached_href" if versioned else "_cached_versionless_href" + if not hasattr(self, cache_attr): validator = URLValidator() if self.external_url and self.external_url.split(':')[0] in validator.schemes: validator(self.external_url) return self.external_url + meeting_doc_refs = ( + settings.MEETING_DOC_HREFS if versioned else settings.MEETING_DOC_GREFS + ) if self.type_id in settings.DOC_HREFS and self.type_id in meeting_doc_refs: if self.meeting_related(): self.is_meeting_related = True @@ -236,7 +298,7 @@ def _get_ref(self, meeting=None, meeting_doc_refs=settings.MEETING_DOC_HREFS): format = settings.DOC_HREFS[self.type_id] elif self.type_id in settings.DOC_HREFS: self.is_meeting_related = False - if self.is_rfc(): + if self.type_id == "rfc": format = settings.DOC_HREFS['rfc'] else: format = settings.DOC_HREFS[self.type_id] @@ -263,10 +325,23 @@ def _get_ref(self, meeting=None, meeting_doc_refs=settings.MEETING_DOC_HREFS): info = dict(doc=self) href = format.format(**info) + + # For slides that are not meeting-related, we need to know the file extension. + # Assume we have access to the same files as settings.DOC_HREFS["slides"] and + # see what extension is available + if self.type_id == "slides" and not self.meeting_related() and not href.endswith("/"): + filepath = Path(self.get_file_path()) / self.get_base_name() # start with this + if not filepath.exists(): + # Look for other extensions - grab the first one, sorted for stability + for existing in sorted(filepath.parent.glob(f"{filepath.stem}.*")): + filepath = filepath.with_suffix(existing.suffix) + break + href += filepath.suffix # tack on the extension + if href.startswith('/'): href = settings.IDTRACKER_BASE_URL + href - self._cached_href = href - return self._cached_href + setattr(self, cache_attr, href) + return getattr(self, cache_attr) def set_state(self, state): """Switch state type implicit in state to state. This just @@ -326,7 +401,9 @@ def friendly_state(self): if not state: return "Unknown state" - if self.type_id == 'draft': + if self.type_id == "rfc": + return f"RFC {self.rfc_number} ({self.std_level})" + elif self.type_id == 'draft': iesg_state = self.get_state("draft-iesg") iesg_state_summary = None if iesg_state: @@ -335,13 +412,15 @@ def friendly_state(self): iesg_state_summary = iesg_state.name if iesg_substate: iesg_state_summary = iesg_state_summary + "::"+"::".join(tag.name for tag in iesg_substate) - - if state.slug == "rfc": - return "RFC %s (%s)" % (self.rfc_number(), self.std_level) + + rfc = self.became_rfc() + if rfc: + return f"Became RFC {rfc.rfc_number} ({rfc.std_level})" + elif state.slug == "repl": rs = self.related_that("replaces") if rs: - return mark_safe("Replaced by " + ", ".join("%s" % (urlreverse('ietf.doc.views_doc.document_main', kwargs=dict(name=alias.document.name)), alias.document) for alias in rs)) + return mark_safe("Replaced by " + ", ".join("%s" % (urlreverse('ietf.doc.views_doc.document_main', kwargs=dict(name=related.name)), related) for related in rs)) else: return "Replaced" elif state.slug == "active": @@ -367,30 +446,56 @@ def friendly_state(self): else: return state.name - def is_rfc(self): - if not hasattr(self, '_cached_is_rfc'): - self._cached_is_rfc = self.pk and self.type_id == 'draft' and self.states.filter(type='draft',slug='rfc').exists() - return self._cached_is_rfc - - def rfc_number(self): - if not hasattr(self, '_cached_rfc_number'): - self._cached_rfc_number = None - if self.is_rfc(): - n = self.canonical_name() - if n.startswith("rfc"): - self._cached_rfc_number = n[3:] + def author_names(self): + """Author names as a list of strings""" + names = [] + if self.type_id == "rfc" and self.rfcauthor_set.exists(): + for author in self.rfcauthor_set.select_related("person"): + if author.person: + names.append(author.person.name) else: - if isinstance(self,Document): - logger.error("Document self.is_rfc() is True but self.canonical_name() is %s" % n) - return self._cached_rfc_number + # titlepage_name cannot be blank + names.append(author.titlepage_name) + else: + names = [ + author.person.name + for author in self.documentauthor_set.select_related("person") + ] + return names + + def author_persons_or_names(self): + """Authors as a list of named tuples with person and/or titlepage_name""" + Author = namedtuple("Author", "person titlepage_name") + persons_or_names = [] + if self.type_id=="rfc" and self.rfcauthor_set.exists(): + for author in self.rfcauthor_set.select_related("person"): + persons_or_names.append(Author(person=author.person, titlepage_name=author.titlepage_name)) + else: + for author in self.documentauthor_set.select_related("person"): + persons_or_names.append(Author(person=author.person, titlepage_name="")) + return persons_or_names - @property - def rfcnum(self): - return self.rfc_number() + def author_persons(self): + """Authors as a list of Persons + + Omits any RfcAuthors with a null person field. + """ + if self.type_id == "rfc" and self.rfcauthor_set.exists(): + authors_qs = self.rfcauthor_set.filter(person__isnull=False) + else: + authors_qs = self.documentauthor_set.all() + return [a.person for a in authors_qs.select_related("person")] def author_list(self): + """List of author emails""" + if self.type_id == "rfc" and self.rfcauthor_set.exists(): + author_qs = self.rfcauthor_set.select_related("person").order_by("order") + else: + author_qs = self.documentauthor_set.select_related("email").order_by( + "order" + ) best_addresses = [] - for author in self.documentauthor_set.all(): + for author in author_qs: if author.email: if author.email.active or not author.email.person: best_addresses.append(author.email.address) @@ -398,9 +503,6 @@ def author_list(self): best_addresses.append(author.email.person.email_address()) return ", ".join(best_addresses) - def authors(self): - return [ a.person for a in self.documentauthor_set.all() ] - # This, and several other ballot related functions here, assume that there is only one active ballot for a document at any point in time. # If that assumption is violated, they will only expose the most recently created ballot def ballot_open(self, ballot_type_slug): @@ -425,7 +527,7 @@ def has_rfc_editor_note(self): return e != None and (e.text != "") def meeting_related(self): - if self.type_id in ("agenda","minutes","bluesheets","slides","recording","procmaterials","chatlog","polls"): + if self.type_id in ("agenda","minutes", "narrativeminutes", "bluesheets","slides","recording","procmaterials","chatlog","polls"): return self.type_id != "slides" or self.get_state_slug('reuse_policy')=='single' return False @@ -460,9 +562,9 @@ def relations_that(self, relationship): if not isinstance(relationship, tuple): raise TypeError("Expected a string or tuple, received %s" % type(relationship)) if isinstance(self, Document): - return RelatedDocument.objects.filter(target__docs=self, relationship__in=relationship).select_related('source') + return RelatedDocument.objects.filter(target=self, relationship__in=relationship).select_related('source') elif isinstance(self, DocHistory): - return RelatedDocHistory.objects.filter(target__docs=self.doc, relationship__in=relationship).select_related('source') + return RelatedDocHistory.objects.filter(target=self.doc, relationship__in=relationship).select_related('source') else: raise TypeError("Expected method called on Document or DocHistory") @@ -496,15 +598,14 @@ def all_relations_that_doc(self, relationship, related=None): for r in rels: if not r in related: related += ( r, ) - for doc in r.target.docs.all(): - related = doc.all_relations_that_doc(relationship, related) + related = r.target.all_relations_that_doc(relationship, related) return related def related_that(self, relationship): - return list(set([x.source.docalias.get(name=x.source.name) for x in self.relations_that(relationship)])) + return list(set([x.source for x in self.relations_that(relationship)])) def all_related_that(self, relationship, related=None): - return list(set([x.source.docalias.get(name=x.source.name) for x in self.all_relations_that(relationship)])) + return list(set([x.source for x in self.all_relations_that(relationship)])) def related_that_doc(self, relationship): return list(set([x.target for x in self.relations_that_doc(relationship)])) @@ -513,42 +614,43 @@ def all_related_that_doc(self, relationship, related=None): return list(set([x.target for x in self.all_relations_that_doc(relationship)])) def replaces(self): - return set([ d for r in self.related_that_doc("replaces") for d in r.docs.all() ]) - - def replaces_canonical_name(self): - s = set([ r.document for r in self.related_that_doc("replaces")]) - first = list(s)[0] if s else None - return None if first is None else first.filename_with_rev() + return self.related_that_doc("replaces") def replaced_by(self): return set([ r.document for r in self.related_that("replaces") ]) - def text(self): + def _text_path(self): path = self.get_file_name() root, ext = os.path.splitext(path) txtpath = root+'.txt' if ext != '.txt' and os.path.exists(txtpath): path = txtpath - try: - with io.open(path, 'rb') as file: - raw = file.read() - except IOError: + return path + + def text_exists(self): + path = Path(self._text_path()) + return path.exists() + + def text(self, size = -1): + path = Path(self._text_path()) + if not path.exists(): return None try: - text = raw.decode('utf-8') - except UnicodeDecodeError: - text = raw.decode('latin-1') - # - return text + with path.open('rb') as file: + raw = file.read(size) + except IOError as e: + log.log(f"Error reading text for {path}: {e}") + return None + return decode_document_content(raw) def text_or_error(self): return self.text() or "Error; cannot read '%s'"%self.get_base_name() def html_body(self, classes=""): - if self.get_state_slug() == "rfc": + if self.type_id == "rfc": try: html = Path( - os.path.join(settings.RFC_PATH, self.canonical_name() + ".html") + os.path.join(settings.RFC_PATH, self.name + ".html") ).read_text() except (IOError, UnicodeDecodeError): return None @@ -619,6 +721,7 @@ def pdfized(self): stylesheets.append(finders.find("ietf/css/document_html_txt.css")) else: text = self.htmlized() + stylesheets.append(f'{settings.STATIC_IETF_ORG_INTERNAL}/fonts/noto-sans-mono/import.css') cache = caches["pdfized"] cache_key = name.split(".")[0] @@ -628,15 +731,20 @@ def pdfized(self): pdf = None if not pdf: try: + font_config = FontConfiguration() pdf = wpHTML( string=text, base_url=settings.IDTRACKER_BASE_URL ).write_pdf( stylesheets=stylesheets, + font_config=font_config, presentational_hints=True, - optimize_size=("fonts", "images"), + optimize_images=True, ) except AssertionError: pdf = None + except Exception as e: + log.log('weasyprint failed:'+str(e)) + raise if pdf: cache.set(cache_key, pdf, settings.PDFIZER_CACHE_TIME) return pdf @@ -645,70 +753,224 @@ def references(self): return self.relations_that_doc(('refnorm','refinfo','refunk','refold')) def referenced_by(self): - return self.relations_that(('refnorm','refinfo','refunk','refold')).filter(source__states__type__slug='draft',source__states__slug__in=['rfc','active']) - + return self.relations_that(("refnorm", "refinfo", "refunk", "refold")).filter( + models.Q( + source__type__slug="draft", + source__states__type__slug="draft", + source__states__slug="active", + ) + | models.Q(source__type__slug="rfc") + ).distinct() + def referenced_by_rfcs(self): - return self.relations_that(('refnorm','refinfo','refunk','refold')).filter(source__states__type__slug='draft',source__states__slug='rfc') + """Get refs to this doc from RFCs""" + return self.relations_that(("refnorm", "refinfo", "refunk", "refold")).filter( + source__type__slug="rfc" + ) + def became_rfc(self): + if not hasattr(self, "_cached_became_rfc"): + doc = self if isinstance(self, Document) else self.doc + self._cached_became_rfc = next(iter(doc.related_that_doc("became_rfc")), None) + return self._cached_became_rfc + + def came_from_draft(self): + if not hasattr(self, "_cached_came_from_draft"): + doc = self if isinstance(self, Document) else self.doc + self._cached_came_from_draft = next(iter(doc.related_that("became_rfc")), None) + return self._cached_came_from_draft + + def contains(self): + return self.related_that_doc("contains") + + def part_of(self): + return self.related_that("contains") + + def referenced_by_rfcs_as_rfc_or_draft(self): + """Get refs to this doc, or a draft/rfc it came from, from an RFC""" + refs_to = self.referenced_by_rfcs() + if self.type_id == "rfc" and self.came_from_draft(): + refs_to |= self.came_from_draft().referenced_by_rfcs() + return refs_to + + def sent_to_rfc_editor_event(self): + if self.stream_id == "ietf": + return self.docevent_set.filter(type="iesg_approved").order_by("-time").first() + elif self.stream_id in ["editorial", "iab", "irtf", "ise"]: + return self.docevent_set.filter(type="requested_publication").order_by("-time").first() + else: + return None class Meta: abstract = True + +class HasNameRevAndTypeIdProtocol(Protocol): + """Typing Protocol describing a class that has name, rev, and type_id properties""" + @property + def name(self) -> str: ... + @property + def rev(self) -> str: ... + @property + def type_id(self) -> str: ... + + +class StorableMixin: + """Mixin that adds storage helpers to a DocumentInfo subclass""" + def store_str( + self: HasNameRevAndTypeIdProtocol, + name: str, + content: str, + allow_overwrite: bool = False + ) -> None: + return utils_store_str(self.type_id, name, content, allow_overwrite, self.name, self.rev) + + def store_bytes( + self: HasNameRevAndTypeIdProtocol, + name: str, + content: bytes, + allow_overwrite: bool = False, + doc_name: Optional[str] = None, + doc_rev: Optional[str] = None + ) -> None: + return utils_store_bytes(self.type_id, name, content, allow_overwrite, self.name, self.rev) + + def store_file( + self: HasNameRevAndTypeIdProtocol, + name: str, + file: Union[File, BufferedReader], + allow_overwrite: bool = False, + doc_name: Optional[str] = None, + doc_rev: Optional[str] = None + ) -> None: + return utils_store_file(self.type_id, name, file, allow_overwrite, self.name, self.rev) + + STATUSCHANGE_RELATIONS = ('tops','tois','tohist','toinf','tobcp','toexp') class RelatedDocument(models.Model): source = ForeignKey('Document') - target = ForeignKey('DocAlias') + target = ForeignKey('Document', related_name='targets_related') relationship = ForeignKey(DocRelationshipName) + originaltargetaliasname = models.CharField(max_length=255, null=True, blank=True) def action(self): return self.relationship.name def __str__(self): return u"%s %s %s" % (self.source.name, self.relationship.name.lower(), self.target.name) def is_downref(self): - - if self.source.type.slug!='draft' or self.relationship.slug not in ['refnorm','refold','refunk']: + if self.source.type_id not in ["draft","rfc"] or self.relationship.slug not in [ + "refnorm", + "refold", + "refunk", + ]: return None - state = self.source.get_state() - if state and state.slug == 'rfc': - source_lvl = self.source.std_level.slug if self.source.std_level else None - elif self.source.intended_std_level: - source_lvl = self.source.intended_std_level.slug + if self.source.type_id == "rfc": + source_lvl = self.source.std_level_id + elif self.source.type_id in ["bcp","std"]: + source_lvl = self.source.type_id else: - source_lvl = None + source_lvl = self.source.intended_std_level_id - if source_lvl not in ['bcp','ps','ds','std']: + if source_lvl not in ["bcp", "ps", "ds", "std", "unkn"]: return None - if self.target.document.get_state().slug == 'rfc': - if not self.target.document.std_level: + if self.target.type_id == 'rfc': + if not self.target.std_level: target_lvl = 'unkn' else: - target_lvl = self.target.document.std_level.slug + target_lvl = self.target.std_level_id + elif self.target.type_id in ["bcp", "std"]: + target_lvl = self.target.type_id else: - if not self.target.document.intended_std_level: + if not self.target.intended_std_level: target_lvl = 'unkn' else: - target_lvl = self.target.document.intended_std_level.slug + target_lvl = self.target.intended_std_level_id - rank = { 'ps':1, 'ds':2, 'std':3, 'bcp':3 } + if self.relationship.slug not in ["refnorm", "refunk"]: + return None - if ( target_lvl not in rank ) or ( rank[target_lvl] < rank[source_lvl] ): - if self.relationship.slug == 'refnorm' and target_lvl!='unkn': - return "Downref" - else: - return "Possible Downref" + if source_lvl in ["inf", "exp"]: + return None + + pos_downref = ( + "Downref" if self.relationship_id != "refunk" else "Possible Downref" + ) + + if source_lvl in ["bcp", "ps", "ds", "std"] and target_lvl in ["inf", "exp"]: + return pos_downref + + if source_lvl == "ds" and target_lvl == "ps": + return pos_downref + + if source_lvl == "std" and target_lvl in ["ps", "ds"]: + return pos_downref + + if source_lvl not in ["inf", "exp"] and target_lvl == "unkn": + return "Possible Downref" + + if source_lvl == "unkn" and target_lvl in ["ps", "ds"]: + return "Possible Downref" return None def is_approved_downref(self): - if self.target.document.get_state().slug == 'rfc': - if RelatedDocument.objects.filter(relationship_id='downref-approval', target=self.target): + if self.target.type_id == 'rfc': + if RelatedDocument.objects.filter(relationship_id='downref-approval', target=self.target).exists(): return "Approved Downref" return False +class RfcAuthor(models.Model): + """Captures the authors of an RFC as represented on the RFC title page. + + This deviates from DocumentAuthor in that it does not get moved into the DocHistory + hierarchy as documents are saved. It will attempt to preserve email, country, and affiliation + from the DocumentAuthor objects associated with the draft leading to this RFC (which + may be wrong if the author moves or changes affiliation while the document is in the + queue). + + It does not, at this time, attempt to capture the authors from anything _but_ the title + page. The datatracker may know more about such authors based on information from the draft + leading to the RFC, and future work may take that into account. + + Once doc.rfcauthor_set.exists() for a doc of type `rfc`, doc.documentauthor_set should be + ignored. + """ + + document = ForeignKey( + "Document", + on_delete=models.CASCADE, + limit_choices_to={"type_id": "rfc"}, # only affects ModelForms (e.g., admin) + ) + titlepage_name = models.CharField(max_length=128, blank=False) + is_editor = models.BooleanField(default=False) + person = ForeignKey(Person, null=True, blank=True, on_delete=models.PROTECT) + affiliation = models.CharField(max_length=100, blank=True, help_text="Organization/company used by author for submission") + country = models.CharField(max_length=255, blank=True, help_text="Country used by author for submission") + order = models.IntegerField(default=1) + + def __str__(self): + return u"%s %s (%s)" % (self.document.name, self.person, self.order) + + class Meta: + ordering=["document", "order"] + indexes=[ + models.Index(fields=["document", "order"]) + ] + + @property + def email(self) -> Email | None: + return self.person.email() if self.person else None + + def format_for_titlepage(self): + if self.is_editor: + return f"{self.titlepage_name}, Ed." + return self.titlepage_name + + class DocumentAuthorInfo(models.Model): person = ForeignKey(Person) # email should only be null for some historic documents @@ -758,7 +1020,7 @@ class Meta: def role_for_doc(self): """Brief string description of this person's relationship to the doc""" roles = [] - if self.person in self.document.authors(): + if self.person in self.document.author_persons(): roles.append('Author') if self.person == self.document.ad: roles.append('Responsible AD') @@ -777,13 +1039,25 @@ def role_for_doc(self): roles.append('Action Holder') return ', '.join(roles) +# N.B., at least a couple dozen documents exist that do not satisfy this validator validate_docname = RegexValidator( r'^[-a-z0-9]+$', "Provide a valid document name consisting of lowercase letters, numbers and hyphens.", 'invalid' ) -class Document(DocumentInfo): + +SUBSERIES_DOC_TYPE_IDS = ("bcp", "fyi", "std") + + +class DocumentQuerySet(models.QuerySet): + def subseries_docs(self): + return self.filter(type_id__in=SUBSERIES_DOC_TYPE_IDS) + + +class Document(StorableMixin, DocumentInfo): + objects = DocumentQuerySet.as_manager() + name = models.CharField(max_length=255, validators=[validate_docname,], unique=True) # immutable action_holders = models.ManyToManyField(Person, through=DocumentActionHolder, blank=True) @@ -800,7 +1074,7 @@ def get_absolute_url(self): name = self.name url = None if self.type_id == "draft" and self.get_state_slug() == "rfc": - name = self.canonical_name() + name = self.name url = urlreverse('ietf.doc.views_doc.document_main', kwargs={ 'name': name }, urlconf="ietf.urls") elif self.type_id in ('slides','bluesheets','recording'): session = self.session_set.first() @@ -838,28 +1112,8 @@ def latest_event(self, *args, **filter_args): e = model.objects.filter(doc=self).filter(**filter_args).order_by('-time', '-id').first() return e - def canonical_name(self): - if not hasattr(self, '_canonical_name'): - name = self.name - if self.type_id == "draft" and self.get_state_slug() == "rfc": - a = self.docalias.filter(name__startswith="rfc").order_by('-name').first() - if a: - name = a.name - elif self.type_id == "charter": - from ietf.doc.utils_charter import charter_name_for_group # Imported locally to avoid circular imports - try: - name = charter_name_for_group(self.chartered_group) - except Group.DoesNotExist: - pass - self._canonical_name = name - return self._canonical_name - - - def canonical_docalias(self): - return self.docalias.get(name=self.name) - def display_name(self): - name = self.canonical_name() + name = self.name if name.startswith('rfc'): name = name.upper() return name @@ -909,6 +1163,22 @@ def request_closed_time(self, review_req): e = self.latest_event(ReviewRequestDocEvent, type="closed_review_request", review_request=review_req) return e.time if e and e.time else None + @property + def area(self) -> Group | None: + """Get area for document, if one exists + + None for non-IETF-stream documents. N.b., this is stricter than Group.area() and + uses different logic from Document.area_acronym(). + """ + if self.stream_id != "ietf": + return None + if self.group is None: + return None + parent = self.group.parent + if parent.type_id == "area": + return parent + return None + def area_acronym(self): g = self.group if g: @@ -952,23 +1222,33 @@ def most_recent_ietflc(self): def displayname_with_link(self): return mark_safe('%s-%s' % (self.get_absolute_url(), self.name , self.rev)) - def ipr(self,states=('posted','removed')): + def ipr(self,states=settings.PUBLISH_IPR_STATES): """Returns the IPR disclosures against this document (as a queryset over IprDocRel).""" - from ietf.ipr.models import IprDocRel - return IprDocRel.objects.filter(document__docs=self, disclosure__state__in=states) + # from ietf.ipr.models import IprDocRel + # return IprDocRel.objects.filter(document__docs=self, disclosure__state__in=states) # TODO - clear these comments away + return self.iprdocrel_set.filter(disclosure__state__in=states) def related_ipr(self): """Returns the IPR disclosures against this document and those documents this document directly or indirectly obsoletes or replaces """ from ietf.ipr.models import IprDocRel - iprs = IprDocRel.objects.filter(document__in=list(self.docalias.all())+self.all_related_that_doc(('obs','replaces'))).filter(disclosure__state__in=('posted','removed')).values_list('disclosure', flat=True).distinct() + iprs = ( + IprDocRel.objects.filter( + document__in=[self] + + self.all_related_that_doc(("obs", "replaces")) + ) + .filter(disclosure__state__in=settings.PUBLISH_IPR_STATES) + .values_list("disclosure", flat=True) + .distinct() + ) return iprs + def future_presentations(self): """ returns related SessionPresentation objects for meetings that have not yet ended. This implementation allows for 2 week meetings """ - candidate_presentations = self.sessionpresentation_set.filter( + candidate_presentations = self.presentations.filter( session__meeting__date__gte=date_today() - datetime.timedelta(days=15) ) return sorted( @@ -981,11 +1261,11 @@ def last_presented(self): """ returns related SessionPresentation objects for the most recent meeting in the past""" # Assumes no two meetings have the same start date - if the assumption is violated, one will be chosen arbitrarily today = date_today() - candidate_presentations = self.sessionpresentation_set.filter(session__meeting__date__lte=today) + candidate_presentations = self.presentations.filter(session__meeting__date__lte=today) candidate_meetings = set([p.session.meeting for p in candidate_presentations if p.session.meeting.end_date()%s" % (self.name, ','.join([force_text(d.name) for d in self.docs.all() if isinstance(d, Document) ])) - document_link = admin_link("document") - class Meta: - verbose_name = "document alias" - verbose_name_plural = "document aliases" class DocReminder(models.Model): event = ForeignKey('DocEvent') @@ -1277,9 +1558,17 @@ class DocReminder(models.Model): # IPR events ("posted_related_ipr", "Posted related IPR"), ("removed_related_ipr", "Removed related IPR"), + ("removed_objfalse_related_ipr", "Removed Objectively False related IPR"), # Bofreq Editor events - ("changed_editors", "Changed BOF Request editors") + ("changed_editors", "Changed BOF Request editors"), + + # Statement events + ("published_statement", "Published statement"), + + # Slide events + ("approved_slides", "Slides approved"), + ] class DocEvent(models.Model): @@ -1341,7 +1630,7 @@ class BallotDocEvent(DocEvent): ballot_type = ForeignKey(BallotType) def active_balloter_positions(self): - """Return dict mapping each active AD or IRSG member to a current ballot position (or None if they haven't voted).""" + """Return dict mapping each active member of the balloting body to a current ballot position (or None if they haven't voted).""" res = {} active_balloters = get_active_balloters(self.ballot_type) @@ -1384,7 +1673,7 @@ def all_positions(self): while p.old_positions and p.old_positions[-1].slug == "norecord": p.old_positions.pop() - # add any missing ADs/IRSGers through fake No Record events + # add any missing balloters through fake No Record events if self.doc.active_ballot() == self: norecord = BallotPositionName.objects.get(slug="norecord") for balloter in active_balloters: @@ -1476,6 +1765,11 @@ class EditedAuthorsDocEvent(DocEvent): """ basis = models.CharField(help_text="What is the source or reasoning for the changes to the author list",max_length=255) + +class EditedRfcAuthorsDocEvent(DocEvent): + """Change to the RfcAuthor list for a document""" + + class BofreqEditorDocEvent(DocEvent): """ Capture the proponents of a BOF Request.""" editors = models.ManyToManyField('person.Person', blank=True) @@ -1483,3 +1777,42 @@ class BofreqEditorDocEvent(DocEvent): class BofreqResponsibleDocEvent(DocEvent): """ Capture the responsible leadership (IAB and IESG members) for a BOF Request """ responsible = models.ManyToManyField('person.Person', blank=True) + + +class StoredObjectQuerySet(models.QuerySet): + def exclude_deleted(self): + return self.filter(deleted__isnull=True) + + +class StoredObject(models.Model): + """Hold metadata about objects placed in object storage""" + + objects = StoredObjectQuerySet.as_manager() + + store = models.CharField(max_length=256) + name = models.CharField(max_length=1024, null=False, blank=False) # N.B. the 1024 limit on name comes from S3 + sha384 = models.CharField(max_length=96) + len = models.PositiveBigIntegerField() + store_created = models.DateTimeField(help_text="The instant the object ws first placed in the store") + created = models.DateTimeField( + null=False, + help_text="Instant object became known. May not be the same as the storage's created value for the instance. It will hold ctime for objects imported from older disk storage" + ) + modified = models.DateTimeField( + null=False, + help_text="Last instant object was modified. May not be the same as the storage's modified value for the instance. It will hold mtime for objects imported from older disk storage unless they've actually been overwritten more recently" + ) + doc_name = models.CharField(max_length=255, null=True, blank=True) + doc_rev = models.CharField(max_length=16, null=True, blank=True) + deleted = models.DateTimeField(null=True) + + class Meta: + constraints = [ + models.UniqueConstraint(fields=['store', 'name'], name='unique_name_per_store'), + ] + indexes = [ + models.Index(fields=["doc_name", "doc_rev"]), + ] + + def __str__(self): + return f"{self.store}:{self.name}" diff --git a/ietf/doc/resources.py b/ietf/doc/resources.py index 99e26ac33d..1d86df78d0 100644 --- a/ietf/doc/resources.py +++ b/ietf/doc/resources.py @@ -12,13 +12,14 @@ from ietf import api from ietf.doc.models import (BallotType, DeletedEvent, StateType, State, Document, - DocumentAuthor, DocEvent, StateDocEvent, DocHistory, ConsensusDocEvent, DocAlias, + DocumentAuthor, DocEvent, StateDocEvent, DocHistory, ConsensusDocEvent, TelechatDocEvent, DocReminder, LastCallDocEvent, NewRevisionDocEvent, WriteupDocEvent, InitialReviewDocEvent, DocHistoryAuthor, BallotDocEvent, RelatedDocument, RelatedDocHistory, BallotPositionDocEvent, AddedMessageEvent, SubmissionDocEvent, ReviewRequestDocEvent, ReviewAssignmentDocEvent, EditedAuthorsDocEvent, DocumentURL, - IanaExpertDocEvent, IRSGBallotDocEvent, DocExtResource, DocumentActionHolder, - BofreqEditorDocEvent,BofreqResponsibleDocEvent) + IanaExpertDocEvent, IRSGBallotDocEvent, DocExtResource, DocumentActionHolder, + BofreqEditorDocEvent, BofreqResponsibleDocEvent, StoredObject, RfcAuthor, + EditedRfcAuthorsDocEvent) from ietf.name.resources import BallotPositionNameResource, DocTypeNameResource class BallotTypeResource(ModelResource): @@ -130,7 +131,6 @@ class Meta: "external_url": ALL, "uploaded_filename": ALL, "note": ALL, - "internal_comments": ALL, "name": ALL, "type": ALL_WITH_RELATIONS, "stream": ALL_WITH_RELATIONS, @@ -247,7 +247,6 @@ class Meta: "external_url": ALL, "uploaded_filename": ALL, "note": ALL, - "internal_comments": ALL, "name": ALL, "type": ALL_WITH_RELATIONS, "stream": ALL_WITH_RELATIONS, @@ -286,21 +285,6 @@ class Meta: } api.doc.register(ConsensusDocEventResource()) -class DocAliasResource(ModelResource): - document = ToOneField(DocumentResource, 'document') - class Meta: - cache = SimpleCache() - queryset = DocAlias.objects.all() - serializer = api.Serializer() - detail_uri_name = 'name' - #resource_name = 'docalias' - ordering = ['id', ] - filtering = { - "name": ALL, - "document": ALL_WITH_RELATIONS, - } -api.doc.register(DocAliasResource()) - from ietf.person.resources import PersonResource class TelechatDocEventResource(ModelResource): by = ToOneField(PersonResource, 'by') @@ -490,7 +474,7 @@ class Meta: from ietf.name.resources import DocRelationshipNameResource class RelatedDocumentResource(ModelResource): source = ToOneField(DocumentResource, 'source') - target = ToOneField(DocAliasResource, 'target') + target = ToOneField(DocumentResource, 'target') relationship = ToOneField(DocRelationshipNameResource, 'relationship') class Meta: cache = SimpleCache() @@ -509,7 +493,7 @@ class Meta: from ietf.name.resources import DocRelationshipNameResource class RelatedDocHistoryResource(ModelResource): source = ToOneField(DocHistoryResource, 'source') - target = ToOneField(DocAliasResource, 'target') + target = ToOneField(DocumentResource, 'target') relationship = ToOneField(DocRelationshipNameResource, 'relationship') class Meta: cache = SimpleCache() @@ -667,6 +651,31 @@ class Meta: api.doc.register(EditedAuthorsDocEventResource()) + +from ietf.person.resources import PersonResource +class EditedRfcAuthorsDocEventResource(ModelResource): + by = ToOneField(PersonResource, 'by') + doc = ToOneField(DocumentResource, 'doc') + docevent_ptr = ToOneField(DocEventResource, 'docevent_ptr') + class Meta: + queryset = EditedRfcAuthorsDocEvent.objects.all() + serializer = api.Serializer() + cache = SimpleCache() + #resource_name = 'editedrfcauthorsdocevent' + ordering = ['id', ] + filtering = { + "id": ALL, + "time": ALL, + "type": ALL, + "rev": ALL, + "desc": ALL, + "by": ALL_WITH_RELATIONS, + "doc": ALL_WITH_RELATIONS, + "docevent_ptr": ALL_WITH_RELATIONS, + } +api.doc.register(EditedRfcAuthorsDocEventResource()) + + from ietf.name.resources import DocUrlTagNameResource class DocumentURLResource(ModelResource): doc = ToOneField(DocumentResource, 'doc') @@ -859,3 +868,51 @@ class Meta: "responsible": ALL_WITH_RELATIONS, } api.doc.register(BofreqResponsibleDocEventResource()) + + +class StoredObjectResource(ModelResource): + class Meta: + queryset = StoredObject.objects.all() + serializer = api.Serializer() + cache = SimpleCache() + #resource_name = 'storedobject' + ordering = ['id', ] + filtering = { + "id": ALL, + "store": ALL, + "name": ALL, + "sha384": ALL, + "len": ALL, + "store_created": ALL, + "created": ALL, + "modified": ALL, + "doc_name": ALL, + "doc_rev": ALL, + "deleted": ALL, + } +api.doc.register(StoredObjectResource()) + + +from ietf.person.resources import EmailResource, PersonResource +class RfcAuthorResource(ModelResource): + document = ToOneField(DocumentResource, 'document') + person = ToOneField(PersonResource, 'person', null=True) + email = ToOneField(EmailResource, 'email', null=True, readonly=True) + class Meta: + queryset = RfcAuthor.objects.all() + serializer = api.Serializer() + cache = SimpleCache() + #resource_name = 'rfcauthor' + ordering = ['id', ] + filtering = { + "id": ALL, + "titlepage_name": ALL, + "is_editor": ALL, + "affiliation": ALL, + "country": ALL, + "order": ALL, + "document": ALL_WITH_RELATIONS, + "person": ALL_WITH_RELATIONS, + "email": ALL_WITH_RELATIONS, + } +api.doc.register(RfcAuthorResource()) diff --git a/ietf/doc/serializers.py b/ietf/doc/serializers.py new file mode 100644 index 0000000000..3651670962 --- /dev/null +++ b/ietf/doc/serializers.py @@ -0,0 +1,360 @@ +# Copyright The IETF Trust 2024-2026, All Rights Reserved +"""django-rest-framework serializers""" + +from dataclasses import dataclass +from typing import Literal, ClassVar + +from django.db.models.manager import BaseManager +from django.db.models.query import QuerySet +from drf_spectacular.utils import extend_schema_field +from rest_framework import serializers + +from ietf.group.serializers import ( + AreaDirectorSerializer, + AreaSerializer, + GroupSerializer, +) +from ietf.name.serializers import StreamNameSerializer +from ietf.utils import log +from .models import Document, DocumentAuthor, RfcAuthor + + +class RfcAuthorSerializer(serializers.ModelSerializer): + """Serializer for an RfcAuthor / DocumentAuthor in a response""" + + email = serializers.EmailField(source="email.address", read_only=True) + datatracker_person_path = serializers.URLField( + source="person.get_absolute_url", + required=False, + help_text="URL for person link (relative to datatracker base URL)", + read_only=True, + ) + + class Meta: + model = RfcAuthor + fields = [ + "titlepage_name", + "is_editor", + "person", + "email", + "affiliation", + "country", + "datatracker_person_path", + ] + + def to_representation(self, instance): + """instance -> primitive data types + + Translates a DocumentAuthor into an equivalent RfcAuthor we can use the same + serializer for either type. + """ + if isinstance(instance, DocumentAuthor): + # create a non-persisted RfcAuthor as a shim - do not save it! + document_author = instance + instance = RfcAuthor( + titlepage_name=document_author.person.plain_name(), + is_editor=False, + person=document_author.person, + affiliation=document_author.affiliation, + country=document_author.country, + order=document_author.order, + ) + return super().to_representation(instance) + + def validate(self, data): + email = data.get("email") + if email is not None: + person = data.get("person") + if person is None: + raise serializers.ValidationError( + { + "email": "cannot have an email without a person", + }, + code="email-without-person", + ) + if email.person_id != person.pk: + raise serializers.ValidationError( + { + "email": "email must belong to person", + }, + code="email-person-mismatch", + ) + return data + + +@dataclass +class DocIdentifier: + type: Literal["doi", "issn"] + value: str + + +class DocIdentifierSerializer(serializers.Serializer): + type = serializers.ChoiceField(choices=["doi", "issn"]) + value = serializers.CharField() + + +type RfcStatusSlugT = Literal[ + "std", + "ps", + "ds", + "bcp", + "inf", + "exp", + "hist", + "unkn", + "not-issued", +] + + +@dataclass +class RfcStatus: + """Helper to extract the 'Status' from an RFC document for serialization""" + + slug: RfcStatusSlugT + + # Names that aren't just the slug itself. ClassVar annotation prevents dataclass from treating this as a field. + fancy_names: ClassVar[dict[RfcStatusSlugT, str]] = { + "std": "internet standard", + "ps": "proposed standard", + "ds": "draft standard", + "bcp": "best current practice", + "inf": "informational", + "exp": "experimental", + "hist": "historic", + "unkn": "unknown", + } + + # ClassVar annotation prevents dataclass from treating this as a field + stdlevelname_slug_map: ClassVar[dict[str, RfcStatusSlugT]] = { + "bcp": "bcp", + "ds": "ds", + "exp": "exp", + "hist": "hist", + "inf": "inf", + "std": "std", + "ps": "ps", + "unkn": "unkn", + } + + # ClassVar annotation prevents dataclass from treating this as a field + status_slugs: ClassVar[list[RfcStatusSlugT]] = sorted( + # TODO implement "not-issued" RFCs + set(stdlevelname_slug_map.values()) | {"not-issued"} + ) + + @property + def name(self): + return RfcStatus.fancy_names.get(self.slug, self.slug) + + @classmethod + def from_document(cls, doc: Document): + """Decide the status that applies to a document""" + return cls( + slug=(cls.stdlevelname_slug_map.get(doc.std_level.slug, "unkn")), + ) + + @classmethod + def filter(cls, queryset, name, value: list[RfcStatusSlugT]): + """Filter a queryset by status + + This is basically the inverse of the from_document() method. Given a status name, filter + the queryset to those in that status. The queryset should be a Document queryset. + """ + interesting_slugs = [ + stdlevelname_slug + for stdlevelname_slug, status_slug in cls.stdlevelname_slug_map.items() + if status_slug in value + ] + if len(interesting_slugs) == 0: + return queryset.none() + return queryset.filter(std_level__slug__in=interesting_slugs) + + +class RfcStatusSerializer(serializers.Serializer): + """Status serializer for a Document instance""" + + slug = serializers.ChoiceField(choices=RfcStatus.status_slugs) + name = serializers.CharField() + + def to_representation(self, instance: Document): + return super().to_representation(instance=RfcStatus.from_document(instance)) + + +class ShepherdSerializer(serializers.Serializer): + email = serializers.EmailField(source="email_address") + + +class RelatedDraftSerializer(serializers.Serializer): + id = serializers.IntegerField(source="source.id") + name = serializers.CharField(source="source.name") + title = serializers.CharField(source="source.title") + shepherd = ShepherdSerializer(source="source.shepherd", allow_null=True) + ad = AreaDirectorSerializer(source="source.ad", allow_null=True) + + +class RelatedRfcSerializer(serializers.Serializer): + id = serializers.IntegerField(source="target.id") + number = serializers.IntegerField(source="target.rfc_number") + title = serializers.CharField(source="target.title") + + +class ReverseRelatedRfcSerializer(serializers.Serializer): + id = serializers.IntegerField(source="source.id") + number = serializers.IntegerField(source="source.rfc_number") + title = serializers.CharField(source="source.title") + + +class ContainingSubseriesSerializer(serializers.Serializer): + name = serializers.CharField(source="source.name") + type = serializers.CharField(source="source.type_id") + + +class RfcFormatSerializer(serializers.Serializer): + RFC_FORMATS = ("xml", "txt", "html", "pdf", "ps", "json", "notprepped") + + fmt = serializers.ChoiceField(choices=RFC_FORMATS) + name = serializers.CharField(help_text="Name of blob in the blob store") + + +class RfcMetadataSerializer(serializers.ModelSerializer): + """Serialize metadata of an RFC + + This needs to be called with a Document queryset that has been processed with + api.augment_rfc_queryset() or it very likely will not work. Some of the typing + refers to Document, but this should really be WithAnnotations[Document, ...]. + However, have not been able to make that work yet. + """ + + number = serializers.IntegerField(source="rfc_number") + published = serializers.DateField() + status = RfcStatusSerializer(source="*") + authors = serializers.SerializerMethodField() + group = GroupSerializer() + area = AreaSerializer(read_only=True) + stream = StreamNameSerializer() + ad = AreaDirectorSerializer(read_only=True, allow_null=True) + group_list_email = serializers.EmailField(source="group.list_email", read_only=True) + identifiers = serializers.SerializerMethodField() + draft = serializers.SerializerMethodField() + obsoletes = RelatedRfcSerializer(many=True, read_only=True) + obsoleted_by = ReverseRelatedRfcSerializer(many=True, read_only=True) + updates = RelatedRfcSerializer(many=True, read_only=True) + updated_by = ReverseRelatedRfcSerializer(many=True, read_only=True) + subseries = ContainingSubseriesSerializer(many=True, read_only=True) + formats = RfcFormatSerializer( + many=True, read_only=True, help_text="Available formats" + ) + keywords = serializers.ListField(child=serializers.CharField(), read_only=True) + has_errata = serializers.BooleanField(read_only=True) + + class Meta: + model = Document + fields = [ + "number", + "title", + "published", + "status", + "pages", + "authors", + "group", + "area", + "stream", + "ad", + "group_list_email", + "identifiers", + "obsoletes", + "obsoleted_by", + "updates", + "updated_by", + "subseries", + "draft", + "abstract", + "formats", + "keywords", + "has_errata", + ] + + @extend_schema_field(RfcAuthorSerializer(many=True)) + def get_authors(self, doc: Document): + # If doc has any RfcAuthors, use those, otherwise fall back to DocumentAuthors + author_queryset: QuerySet[RfcAuthor] | QuerySet[DocumentAuthor] = ( + doc.rfcauthor_set.all() + if doc.rfcauthor_set.exists() + else doc.documentauthor_set.all() + ) + # RfcAuthorSerializer can deal with DocumentAuthor instances + return RfcAuthorSerializer( + instance=author_queryset, + many=True, + ).data + + @extend_schema_field(DocIdentifierSerializer(many=True)) + def get_identifiers(self, doc: Document): + identifiers = [] + if doc.doi: + identifiers.append( + DocIdentifier(type="doi", value=doc.doi) + ) + return DocIdentifierSerializer(instance=identifiers, many=True).data + + @extend_schema_field(RelatedDraftSerializer) + def get_draft(self, doc: Document): + if hasattr(doc, "drafts"): + # This is the expected case - drafts is added by a Prefetch in + # the augment_rfc_queryset() method. + try: + related_doc = doc.drafts[0] + except IndexError: + return None + else: + # Fallback in case augment_rfc_queryset() was not called + log.log( + f"Warning: {self.__class__}.get_draft() called without prefetched draft" + ) + related_doc = doc.came_from_draft() + return RelatedDraftSerializer(related_doc).data + + +class RfcSerializer(RfcMetadataSerializer): + """Serialize an RFC, including its metadata and text content if available""" + + text = serializers.CharField(allow_null=True) + + class Meta: + model = RfcMetadataSerializer.Meta.model + fields = RfcMetadataSerializer.Meta.fields + ["text"] + + +class SubseriesContentListSerializer(serializers.ListSerializer): + """ListSerializer that gets its object from item.target""" + + def to_representation(self, data): + """ + List of object instances -> List of dicts of primitive datatypes. + """ + # Dealing with nested relationships, data can be a Manager, + # so, first get a queryset from the Manager if needed + iterable = data.all() if isinstance(data, BaseManager) else data + # Serialize item.target instead of item itself + return [self.child.to_representation(item.target) for item in iterable] + + +class SubseriesContentSerializer(RfcMetadataSerializer): + """Serialize RFC contained in a subseries doc""" + + class Meta(RfcMetadataSerializer.Meta): + list_serializer_class = SubseriesContentListSerializer + + +class SubseriesDocSerializer(serializers.ModelSerializer): + """Serialize a subseries document (e.g., a BCP or STD)""" + + contents = SubseriesContentSerializer(many=True) + + class Meta: + model = Document + fields = [ + "name", + "type", + "contents", + ] diff --git a/ietf/doc/storage.py b/ietf/doc/storage.py new file mode 100644 index 0000000000..ee1e76c4fa --- /dev/null +++ b/ietf/doc/storage.py @@ -0,0 +1,181 @@ +# Copyright The IETF Trust 2025, All Rights Reserved +from typing import Optional + +import debug # pyflakes:ignore +import json + +from contextlib import contextmanager +from storages.backends.s3 import S3Storage + +from django.core.files.base import File + +from ietf.blobdb.storage import BlobdbStorage +from ietf.doc.models import StoredObject +from ietf.utils.log import log +from ietf.utils.storage import MetadataFile +from ietf.utils.timezone import timezone + + +class StoredObjectFile(MetadataFile): + """Django storage File object that represents a StoredObject""" + def __init__(self, file, name, mtime=None, content_type="", store=None, doc_name=None, doc_rev=None): + super().__init__( + file=file, + name=name, + mtime=mtime, + content_type=content_type, + ) + self.store = store + self.doc_name = doc_name + self.doc_rev = doc_rev + + @classmethod + def from_storedobject(cls, file, name, store): + """Alternate constructor for objects that already exist in the StoredObject table""" + stored_object = StoredObject.objects.exclude_deleted().filter(store=store, name=name).first() + if stored_object is None: + raise FileNotFoundError(f"StoredObject for {store}:{name} does not exist or was deleted") + file = cls(file, name, store, doc_name=stored_object.doc_name, doc_rev=stored_object.doc_rev) + if int(file.custom_metadata["len"]) != stored_object.len: + raise RuntimeError(f"File length changed unexpectedly for {store}:{name}") + if file.custom_metadata["sha384"] != stored_object.sha384: + raise RuntimeError(f"SHA-384 hash changed unexpectedly for {store}:{name}") + return file + + +@contextmanager +def maybe_log_timing(enabled, op, **kwargs): + """If enabled, log elapsed time and additional data from kwargs + + Emits log even if an exception occurs + """ + before = timezone.now() + exception = None + try: + yield + except Exception as err: + exception = err + raise + finally: + if enabled: + dt = timezone.now() - before + log( + json.dumps( + { + "log": "S3Storage_timing", + "seconds": dt.total_seconds(), + "op": op, + "exception": "" if exception is None else repr(exception), + **kwargs, + } + ) + ) + + +class MetadataS3Storage(S3Storage): + def get_default_settings(self): + # add a default for the ietf_log_blob_timing boolean + return super().get_default_settings() | {"ietf_log_blob_timing": False} + + def _save(self, name, content: File): + with maybe_log_timing( + self.ietf_log_blob_timing, "_save", bucket_name=self.bucket_name, name=name + ): + return super()._save(name, content) + + def _open(self, name, mode="rb"): + with maybe_log_timing( + self.ietf_log_blob_timing, + "_open", + bucket_name=self.bucket_name, + name=name, + mode=mode, + ): + return super()._open(name, mode) + + def delete(self, name): + with maybe_log_timing( + self.ietf_log_blob_timing, "delete", bucket_name=self.bucket_name, name=name + ): + super().delete(name) + + def _get_write_parameters(self, name, content=None): + # debug.show('f"getting write parameters for {name}"') + params = super()._get_write_parameters(name, content) + # If we have a non-empty explicit content type, use it + content_type = getattr(content, "content_type", "").strip() + if content_type != "": + params["ContentType"] = content_type + if "Metadata" not in params: + params["Metadata"] = {} + if hasattr(content, "custom_metadata"): + params["Metadata"].update(content.custom_metadata) + return params + + +class StoredObjectBlobdbStorage(BlobdbStorage): + warn_if_missing = True # TODO-BLOBSTORE make this configurable (or remove it) + + def _save_stored_object(self, name, content) -> StoredObject: + now = timezone.now() + record, created = StoredObject.objects.get_or_create( + store=self.bucket_name, + name=name, + defaults=dict( + sha384=content.custom_metadata["sha384"], + len=int(content.custom_metadata["len"]), + store_created=now, + created=now, + modified=now, + doc_name=getattr( + content, + "doc_name", # Note that these are assumed to be invariant + None, # should be blank? + ), + doc_rev=getattr( + content, + "doc_rev", # for a given name + None, # should be blank? + ), + ), + ) + if not created and ( + record.sha384 != content.custom_metadata["sha384"] + or record.len != int(content.custom_metadata["len"]) + or record.deleted is not None + ): + record.sha384 = content.custom_metadata["sha384"] + record.len = int(content.custom_metadata["len"]) + record.modified = now + record.deleted = None + record.save() + return record + + def _delete_stored_object(self, name) -> Optional[StoredObject]: + existing_record = StoredObject.objects.filter(store=self.bucket_name, name=name) + if not existing_record.exists() and self.warn_if_missing: + complaint = ( + f"WARNING: Asked to delete {name} from {self.bucket_name} storage, " + f"but there was no matching StoredObject" + ) + log(complaint) + debug.show("complaint") + else: + now = timezone.now() + # Note that existing_record is a queryset that will have one matching object + existing_record.exclude_deleted().update(deleted=now) + return existing_record.first() + + def _save(self, name, content): + """Perform the save operation + + In principle the name could change on save to the blob store. As of now, BlobdbStorage + will not change it, but allow for that possibility. Callers should be prepared for this. + """ + saved_name = super()._save(name, content) + self._save_stored_object(saved_name, content) + return saved_name + + def delete(self, name): + self._delete_stored_object(name) + super().delete(name) diff --git a/ietf/doc/storage_utils.py b/ietf/doc/storage_utils.py new file mode 100644 index 0000000000..9c18bb8a8a --- /dev/null +++ b/ietf/doc/storage_utils.py @@ -0,0 +1,194 @@ +# Copyright The IETF Trust 2025, All Rights Reserved +import datetime +from io import BufferedReader +from typing import Optional, Union + +import debug # pyflakes ignore + +from django.conf import settings +from django.core.files.base import ContentFile, File +from django.core.files.storage import storages, Storage + +from ietf.utils.log import log +from ietf.utils.text import decode_document_content + + +class StorageUtilsError(Exception): + pass + + +class AlreadyExistsError(StorageUtilsError): + pass + + +def _get_storage(kind: str) -> Storage: + if kind in settings.ARTIFACT_STORAGE_NAMES: + return storages[kind] + else: + debug.say(f"Got into not-implemented looking for {kind}") + raise NotImplementedError(f"Don't know how to store {kind}") + + +def exists_in_storage(kind: str, name: str) -> bool: + if settings.ENABLE_BLOBSTORAGE: + try: + store = _get_storage(kind) + with store.open(name): + return True + except FileNotFoundError: + return False + except Exception as err: + log(f"Blobstore Error: Failed to test existence of {kind}:{name}: {repr(err)}") + if settings.SERVER_MODE == "development": + raise + return False + + +def remove_from_storage(kind: str, name: str, warn_if_missing: bool = True) -> None: + if settings.ENABLE_BLOBSTORAGE: + try: + if exists_in_storage(kind, name): + _get_storage(kind).delete(name) + elif warn_if_missing: + complaint = ( + f"WARNING: Asked to delete non-existent {name} from {kind} storage" + ) + debug.show("complaint") + log(complaint) + except Exception as err: + log(f"Blobstore Error: Failed to remove {kind}:{name}: {repr(err)}") + if settings.SERVER_MODE == "development": + raise + return None + + +def store_file( + kind: str, + name: str, + file: Union[File, BufferedReader], + allow_overwrite: bool = False, + doc_name: Optional[str] = None, + doc_rev: Optional[str] = None, + content_type: str="", + mtime: Optional[datetime.datetime]=None, +) -> None: + from .storage import StoredObjectFile # avoid circular import + if settings.ENABLE_BLOBSTORAGE: + try: + is_new = not exists_in_storage(kind, name) + # debug.show('f"Asked to store {name} in {kind}: is_new={is_new}, allow_overwrite={allow_overwrite}"') + if not allow_overwrite and not is_new: + debug.show('f"Failed to save {kind}:{name} - name already exists in store"') + raise AlreadyExistsError(f"Failed to save {kind}:{name} - name already exists in store") + new_name = _get_storage(kind).save( + name, + StoredObjectFile( + file=file, + name=name, + doc_name=doc_name, + doc_rev=doc_rev, + mtime=mtime, + content_type=content_type, + ), + ) + if new_name != name: + complaint = f"Error encountered saving '{name}' - results stored in '{new_name}' instead." + debug.show("complaint") + raise StorageUtilsError(complaint) + except Exception as err: + log(f"Blobstore Error: Failed to store file {kind}:{name}: {repr(err)}") + if settings.SERVER_MODE == "development": + raise # TODO-BLOBSTORE eventually make this an error for all modes + return None + + +def store_bytes( + kind: str, + name: str, + content: bytes, + allow_overwrite: bool = False, + doc_name: Optional[str] = None, + doc_rev: Optional[str] = None, + content_type: str = "", + mtime: Optional[datetime.datetime] = None, +) -> None: + if settings.ENABLE_BLOBSTORAGE: + try: + store_file( + kind, + name, + ContentFile(content), + allow_overwrite, + doc_name, + doc_rev, + content_type, + mtime, + ) + except Exception as err: + # n.b., not likely to get an exception here because store_file or store_bytes will catch it + log(f"Blobstore Error: Failed to store bytes to {kind}:{name}: {repr(err)}") + if settings.SERVER_MODE == "development": + raise # TODO-BLOBSTORE eventually make this an error for all modes + return None + + +def store_str( + kind: str, + name: str, + content: str, + allow_overwrite: bool = False, + doc_name: Optional[str] = None, + doc_rev: Optional[str] = None, + content_type: str = "", + mtime: Optional[datetime.datetime] = None, +) -> None: + if settings.ENABLE_BLOBSTORAGE: + try: + content_bytes = content.encode("utf-8") + store_bytes( + kind, + name, + content_bytes, + allow_overwrite, + doc_name, + doc_rev, + content_type, + mtime, + ) + except Exception as err: + # n.b., not likely to get an exception here because store_file or store_bytes will catch it + log(f"Blobstore Error: Failed to store string to {kind}:{name}: {repr(err)}") + if settings.SERVER_MODE == "development": + raise # TODO-BLOBSTORE eventually make this an error for all modes + return None + + +def retrieve_bytes(kind: str, name: str) -> bytes: + from ietf.doc.storage import maybe_log_timing + if not settings.ENABLE_BLOBSTORAGE: + return b"" + try: + store = _get_storage(kind) + with store.open(name) as f: + with maybe_log_timing( + hasattr(store, "ietf_log_blob_timing") and store.ietf_log_blob_timing, + "read", + bucket_name=store.bucket_name if hasattr(store, "bucket_name") else "", + name=name, + ): + content = f.read() + except Exception as err: + log(f"Blobstore Error: Failed to read bytes from {kind}:{name}: {repr(err)}") + raise + return content + + +def retrieve_str(kind: str, name: str) -> str: + if not settings.ENABLE_BLOBSTORAGE: + return "" + try: + content = decode_document_content(retrieve_bytes(kind, name)) + except Exception as err: + log(f"Blobstore Error: Failed to read string from {kind}:{name}: {repr(err)}") + raise + return content diff --git a/ietf/doc/tasks.py b/ietf/doc/tasks.py new file mode 100644 index 0000000000..273242e35f --- /dev/null +++ b/ietf/doc/tasks.py @@ -0,0 +1,222 @@ +# Copyright The IETF Trust 2024-2026, All Rights Reserved +# +# Celery task definitions +# +import datetime + +import debug # pyflakes:ignore + +from celery import shared_task +from celery.exceptions import MaxRetriesExceededError +from pathlib import Path + +from django.conf import settings +from django.utils import timezone + +from ietf.doc.utils_r2 import rfcs_are_in_r2 +from ietf.doc.utils_red import trigger_red_precomputer +from ietf.utils import log, searchindex +from ietf.utils.timezone import datetime_today + +from .expire import ( + in_draft_expire_freeze, + get_expired_drafts, + expirable_drafts, + send_expire_notice_for_draft, + expire_draft, + clean_up_draft_files, + get_soon_to_expire_drafts, + send_expire_warning_for_draft, +) +from .lastcall import get_expired_last_calls, expire_last_call +from .models import Document, NewRevisionDocEvent +from .utils import ( + generate_idnits2_rfc_status, + generate_idnits2_rfcs_obsoleted, + rebuild_reference_relations, + update_or_create_draft_bibxml_file, + ensure_draft_bibxml_path_exists, + investigate_fragment, +) +from .utils_bofreq import fixup_bofreq_timestamps +from .utils_errata import signal_update_rfc_metadata + + +@shared_task +def expire_ids_task(): + try: + if not in_draft_expire_freeze(): + log.log("Expiring drafts ...") + for doc in get_expired_drafts(): + # verify expirability -- it might have changed after get_expired_drafts() was run + # (this whole loop took about 2 minutes on 04 Jan 2018) + # N.B., re-running expirable_drafts() repeatedly is fairly expensive. Where possible, + # it's much faster to run it once on a superset query of the objects you are going + # to test and keep its results. That's not desirable here because it would defeat + # the purpose of double-checking that a document is still expirable when it is actually + # being marked as expired. + if expirable_drafts( + Document.objects.filter(pk=doc.pk) + ).exists() and doc.expires < datetime_today() + datetime.timedelta(1): + send_expire_notice_for_draft(doc) + expire_draft(doc) + log.log(f" Expired draft {doc.name}-{doc.rev}") + + log.log("Cleaning up draft files") + clean_up_draft_files() + except Exception as e: + log.log("Exception in expire-ids: %s" % e) + raise + + +@shared_task +def notify_expirations_task(notify_days=14): + for doc in get_soon_to_expire_drafts(notify_days): + send_expire_warning_for_draft(doc) + + +@shared_task +def expire_last_calls_task(): + for doc in get_expired_last_calls(): + try: + expire_last_call(doc) + except Exception: + log.log( + f"ERROR: Failed to expire last call for {doc.file_tag()} (id={doc.pk})" + ) + else: + log.log(f"Expired last call for {doc.file_tag()} (id={doc.pk})") + + +@shared_task +def generate_idnits2_rfc_status_task(): + outpath = Path(settings.DERIVED_DIR) / "idnits2-rfc-status" + blob = generate_idnits2_rfc_status() + try: + outpath.write_text(blob, encoding="utf8") # TODO-BLOBSTORE + except Exception as e: + log.log(f"failed to write idnits2-rfc-status: {e}") + + +@shared_task +def generate_idnits2_rfcs_obsoleted_task(): + outpath = Path(settings.DERIVED_DIR) / "idnits2-rfcs-obsoleted" + blob = generate_idnits2_rfcs_obsoleted() + try: + outpath.write_text(blob, encoding="utf8") # TODO-BLOBSTORE + except Exception as e: + log.log(f"failed to write idnits2-rfcs-obsoleted: {e}") + + +@shared_task +def generate_draft_bibxml_files_task(days=7, process_all=False): + """Generate bibxml files for recently updated docs + + If process_all is False (the default), processes only docs with new revisions + in the last specified number of days. + """ + if not process_all and days < 1: + raise ValueError("Must call with days >= 1 or process_all=True") + ensure_draft_bibxml_path_exists() + doc_events = NewRevisionDocEvent.objects.filter( + type="new_revision", + doc__type_id="draft", + ).order_by("time") + if not process_all: + doc_events = doc_events.filter( + time__gte=timezone.now() - datetime.timedelta(days=days) + ) + for event in doc_events: + try: + update_or_create_draft_bibxml_file(event.doc, event.rev) + except Exception as err: + log.log(f"Error generating bibxml for {event.doc.name}-{event.rev}: {err}") + + +@shared_task(ignore_result=False) +def investigate_fragment_task(name_fragment: str): + return { + "name_fragment": name_fragment, + "results": investigate_fragment(name_fragment), + } + + +@shared_task +def rebuild_reference_relations_task(doc_names: list[str]): + log.log(f"Task: Rebuilding reference relations for {doc_names}") + for doc in Document.objects.filter(name__in=doc_names, type__in=["rfc", "draft"]): + filenames = dict() + base = ( + settings.RFC_PATH + if doc.type_id == "rfc" + else settings.INTERNET_ALL_DRAFTS_ARCHIVE_DIR + ) + stem = doc.name if doc.type_id == "rfc" else f"{doc.name}-{doc.rev}" + for ext in ["xml", "txt"]: + path = Path(base) / f"{stem}.{ext}" + if path.is_file(): + filenames[ext] = str(path) + if len(filenames) > 0: + rebuild_reference_relations(doc, filenames) + else: + log.log(f"Found no content for {stem}") + + +@shared_task +def fixup_bofreq_timestamps_task(): # pragma: nocover + fixup_bofreq_timestamps() + + +@shared_task +def signal_update_rfc_metadata_task(rfc_number_list=()): + signal_update_rfc_metadata(rfc_number_list) + + +@shared_task(bind=True) +def trigger_red_precomputer_task(self, rfc_number_list=()): + if not rfcs_are_in_r2(rfc_number_list): + log.log(f"Objects are not yet in R2 for RFCs {rfc_number_list}") + try: + countdown = getattr(settings, "RED_PRECOMPUTER_TRIGGER_RETRY_DELAY", 10) + max_retries = getattr(settings, "RED_PRECOMPUTER_TRIGGER_MAX_RETRIES", 12) + self.retry(countdown=countdown, max_retries=max_retries) + except MaxRetriesExceededError: + log.log(f"Gave up waiting for objects in R2 for RFCs {rfc_number_list}") + else: + trigger_red_precomputer(rfc_number_list) + + +@shared_task(bind=True) +def update_rfc_searchindex_task(self, rfc_number: int): + """Update the search index for one RFC""" + if not searchindex.enabled(): + log.log("Search indexing is not enabled, skipping") + return + + rfc = Document.objects.filter(type_id="rfc", rfc_number=rfc_number).first() + if rfc is None: + log.log( + f"ERROR: Document for rfc{rfc_number} not found, not updating search index" + ) + return + try: + searchindex.update_or_create_rfc_entry(rfc) + except Exception as err: + log.log(f"Search index update for {rfc.name} failed ({err})") + if isinstance(err, searchindex.RETRYABLE_ERROR_CLASSES): + searchindex_settings = searchindex.get_settings() + self.retry( + countdown=searchindex_settings["TASK_RETRY_DELAY"], + max_retries=searchindex_settings["TASK_MAX_RETRIES"], + ) + + +@shared_task +def rebuild_searchindex_task(*, batchsize=40, drop_collection=False): + if drop_collection: + searchindex.delete_collection() + searchindex.create_collection() + searchindex.update_or_create_rfc_entries( + Document.objects.filter(type_id="rfc").order_by("-rfc_number"), + batchsize=batchsize, + ) diff --git a/ietf/doc/templatetags/active_groups_menu.py b/ietf/doc/templatetags/active_groups_menu.py index dd97c8e45b..c60d6dcd1a 100644 --- a/ietf/doc/templatetags/active_groups_menu.py +++ b/ietf/doc/templatetags/active_groups_menu.py @@ -8,23 +8,19 @@ register = template.Library() -parents = GroupTypeName.objects.filter( - slug__in=["ag", "area", "rag", "team", "dir", "program"] -) - -others = [] -for group in Group.objects.filter(acronym__in=("rsoc",), state_id="active"): - group.menu_url = reverse("ietf.group.views.group_home", kwargs=dict(acronym=group.acronym)) # type: ignore - # could use group.about_url() instead - others.append(group) - @register.simple_tag def active_groups_menu(flavor): - global parents, others + parents = GroupTypeName.objects.filter(slug__in=["ag", "area", "rag", "team", "dir", "program", "iabworkshop"]) + others = [] + for group in Group.objects.filter(acronym__in=("rsoc",), state_id="active"): + group.menu_url = reverse("ietf.group.views.group_home", kwargs=dict(acronym=group.acronym)) # type: ignore + # could use group.about_url() instead + others.append(group) + for p in parents: p.menu_url = "/%s/" % p.slug return render_to_string( "base/menu_active_groups.html", {"parents": parents, "others": others, "flavor": flavor}, - ) \ No newline at end of file + ) diff --git a/ietf/doc/templatetags/ballot_icon.py b/ietf/doc/templatetags/ballot_icon.py index ee7d5f6278..07a6c7f926 100644 --- a/ietf/doc/templatetags/ballot_icon.py +++ b/ietf/doc/templatetags/ballot_icon.py @@ -53,7 +53,9 @@ def showballoticon(doc): if doc.type_id == "draft": if doc.stream_id == 'ietf' and doc.get_state_slug("draft-iesg") not in IESG_BALLOT_ACTIVE_STATES: return False - elif doc.stream_id == 'irtf' and doc.get_state_slug("draft-stream-irtf") not in ['irsgpoll']: + elif doc.stream_id == 'irtf' and doc.get_state_slug("draft-stream-irtf") != "irsgpoll": + return False + elif doc.stream_id == 'editorial' and doc.get_state_slug("draft-stream-rsab") != "rsabpoll": return False elif doc.type_id == "charter": if doc.get_state_slug() not in ("intrev", "extrev", "iesgrev"): @@ -94,9 +96,14 @@ def sort_key(t): positions = list(ballot.active_balloter_positions().items()) positions.sort(key=sort_key) + request = context.get("request") + ballot_edit_return_point_param = f"ballot_edit_return_point={request.path}" + right_click_string = '' if has_role(user, "Area Director"): - right_click_string = 'oncontextmenu="window.location.href=\'%s\';return false;"' % urlreverse('ietf.doc.views_ballot.edit_position', kwargs=dict(name=doc.name, ballot_id=ballot.pk)) + right_click_string = 'oncontextmenu="window.location.href=\'{}?{}\';return false;"'.format( + urlreverse('ietf.doc.views_ballot.edit_position', kwargs=dict(name=doc.name, ballot_id=ballot.pk)), + ballot_edit_return_point_param) my_blocking = False for i, (balloter, pos) in enumerate(positions): @@ -105,14 +112,20 @@ def sort_key(t): break typename = "Unknown" - if ballot.ballot_type.slug=='irsg-approve': + if ballot.ballot_type.slug == "irsg-approve": typename = "IRSG" + elif ballot.ballot_type.slug == "rsab-approve": + typename = "RSAB" else: typename = "IESG" + + modal_url = "{}?{}".format( + urlreverse("ietf.doc.views_doc.ballot_popup", kwargs=dict(name=doc.name, ballot_id=ballot.pk)), + ballot_edit_return_point_param) res = [' goal2: - class_name = "bg-danger" + class_name = "text-bg-danger" elif days > goal1: - class_name = "bg-warning" + class_name = "text-bg-warning" else: # don't show a badge when things are in the green; clutters display # class_name = "text-success" @@ -244,6 +256,6 @@ def auth48_alert_badge(doc): rfced_state = doc.get_state_slug('draft-rfceditor') if rfced_state == 'auth48': - return mark_safe('AUTH48') + return mark_safe('AUTH48') return '' diff --git a/ietf/doc/templatetags/document_type_badge.py b/ietf/doc/templatetags/document_type_badge.py new file mode 100644 index 0000000000..a82c606ff9 --- /dev/null +++ b/ietf/doc/templatetags/document_type_badge.py @@ -0,0 +1,29 @@ +# Copyright The IETF Trust 2015-2020, All Rights Reserved +from django import template +from django.conf import settings +from django.template.loader import render_to_string +from ietf.utils.log import log + +register = template.Library() + + +@register.simple_tag +def document_type_badge(doc, snapshot, submission, resurrected_by): + context = {"doc": doc, "snapshot": snapshot, "submission": submission, "resurrected_by": resurrected_by} + if doc.type_id == "rfc": + return render_to_string( + "doc/badge/doc-badge-rfc.html", + context, + ) + elif doc.type_id == "draft": + return render_to_string( + "doc/badge/doc-badge-draft.html", + context, + ) + else: + error_message = f"Unsupported document type {doc.type_id}." + if settings.SERVER_MODE != 'production': + raise ValueError(error_message) + else: + log(error_message) + return "" diff --git a/ietf/doc/templatetags/ietf_filters.py b/ietf/doc/templatetags/ietf_filters.py index 1137cf636a..ae5df641c2 100644 --- a/ietf/doc/templatetags/ietf_filters.py +++ b/ietf/doc/templatetags/ietf_filters.py @@ -1,9 +1,10 @@ -# Copyright The IETF Trust 2007-2020, All Rights Reserved +# Copyright The IETF Trust 2007-2023, All Rights Reserved # -*- coding: utf-8 -*- import datetime import re +from pathlib import Path from urllib.parse import urljoin from zoneinfo import ZoneInfo @@ -13,8 +14,7 @@ from django.template.defaultfilters import truncatewords_html, linebreaksbr, stringfilter, striptags from django.utils.safestring import mark_safe, SafeData from django.utils.html import strip_tags -from django.utils.encoding import force_text -from django.utils.encoding import force_str # pyflakes:ignore force_str is used in the doctests +from django.utils.encoding import force_str from django.urls import reverse as urlreverse from django.core.cache import cache from django.core.exceptions import ValidationError @@ -23,12 +23,14 @@ import debug # pyflakes:ignore -from ietf.doc.models import BallotDocEvent, DocAlias +from ietf.doc.models import BallotDocEvent, Document from ietf.doc.models import ConsensusDocEvent -from ietf.utils.html import sanitize_fragment +from ietf.ietfauth.utils import can_request_rfc_publication as utils_can_request_rfc_publication from ietf.utils import log from ietf.doc.utils import prettify_std_name -from ietf.utils.text import wordwrap, fill, wrap_text_if_unwrapped, bleach_linker, bleach_cleaner, validate_url +from ietf.utils.html import clean_html +from ietf.utils.text import wordwrap, fill, wrap_text_if_unwrapped, linkify +from ietf.utils.validators import validate_url register = template.Library() @@ -97,7 +99,7 @@ def sanitize(value): attributes to those deemed acceptable. See ietf/utils/html.py for the details. """ - return mark_safe(sanitize_fragment(value)) + return mark_safe(clean_html(value)) # For use with ballot view @@ -131,7 +133,7 @@ def bracketpos(pos,posslug): @register.filter def prettystdname(string, space=" "): from ietf.doc.utils import prettify_std_name - return prettify_std_name(force_text(string or ""), space) + return prettify_std_name(force_str(string or ""), space) @register.filter def rfceditor_info_url(rfcnum : str): @@ -139,15 +141,16 @@ def rfceditor_info_url(rfcnum : str): return urljoin(settings.RFC_EDITOR_INFO_BASE_URL, f'rfc{rfcnum}') -def doc_canonical_name(name): +def doc_name(name): """Check whether a given document exists, and return its canonical name""" def find_unique(n): key = hash(n) found = cache.get(key) if not found: - exact = DocAlias.objects.filter(name=n).first() + exact = Document.objects.filter(name=n).first() found = exact.name if exact else "_" + # TODO review this cache policy (and the need for these entire function) cache.set(key, found, timeout=60*60*24) # cache for one day return None if found == "_" else found @@ -173,7 +176,7 @@ def find_unique(n): def link_charter_doc_match(match): - if not doc_canonical_name(match[0]): + if not doc_name(match[0]): return match[0] url = urlreverse( "ietf.doc.views_doc.document_main", @@ -186,7 +189,7 @@ def link_non_charter_doc_match(match): name = match[0] # handle "I-D.*"" reference-style matches name = re.sub(r"^i-d\.(.*)", r"draft-\1", name, flags=re.IGNORECASE) - cname = doc_canonical_name(name) + cname = doc_name(name) if not cname: return match[0] if name == cname: @@ -201,7 +204,7 @@ def link_non_charter_doc_match(match): url = urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=cname)) return f'{match[0]}' - cname = doc_canonical_name(name) + cname = doc_name(name) if not cname: return match[0] if name == cname: @@ -221,12 +224,11 @@ def link_non_charter_doc_match(match): def link_other_doc_match(match): doc = match[2].strip().lower() rev = match[3] - if not doc_canonical_name(doc + rev): + if not doc_name(doc + rev): return match[0] url = urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=doc + rev)) return f'{match[1]}' - @register.filter(name="urlize_ietf_docs", is_safe=True, needs_autoescape=True) def urlize_ietf_docs(string, autoescape=None): """ @@ -255,6 +257,7 @@ def urlize_ietf_docs(string, autoescape=None): string, flags=re.IGNORECASE | re.ASCII, ) + return mark_safe(string) @@ -267,7 +270,7 @@ def urlize_related_source_list(related, document_html=False): names = set() titles = set() for rel in related: - name=rel.source.canonical_name() + name=rel.source.name title = rel.source.title if name in names and title in titles: continue @@ -282,14 +285,14 @@ def urlize_related_source_list(related, document_html=False): url=url) )) return links - + @register.filter(name='urlize_related_target_list', is_safe=True, document_html=False) def urlize_related_target_list(related, document_html=False): """Convert a list of RelatedDocuments into list of links using the target document's canonical name""" links = [] for rel in related: - name=rel.target.document.canonical_name() - title = rel.target.document.title + name=rel.target.name + title = rel.target.title url = urlreverse('ietf.doc.views_doc.document_main' if document_html is False else 'ietf.doc.views_doc.document_html', kwargs=dict(name=name)) name = escape(name) title = escape(title) @@ -299,7 +302,7 @@ def urlize_related_target_list(related, document_html=False): url=url) )) return links - + @register.filter(name='dashify') def dashify(string): """ @@ -409,9 +412,9 @@ def startswith(x, y): return str(x).startswith(y) -@register.filter(name='removesuffix', is_safe=False) -def removesuffix(value, suffix): - """Remove an exact-match suffix +@register.filter(name='removeprefix', is_safe=False) +def removeprefix(value, prefix): + """Remove an exact-match prefix The is_safe flag is False because indiscriminate use of this could result in non-safe output. See https://docs.djangoproject.com/en/2.2/howto/custom-template-tags/#filters-and-auto-escaping @@ -419,8 +422,8 @@ def removesuffix(value, suffix): HTML-unsafe output. """ base = str(value) - if base.endswith(suffix): - return base[:-len(suffix)] + if base.startswith(prefix): + return base[len(prefix):] else: return base @@ -444,16 +447,16 @@ def ad_area(user): @register.filter def format_history_text(text, trunc_words=25): """Run history text through some cleaning and add ellipsis if it's too long.""" - full = mark_safe(bleach_cleaner.clean(text)) - full = bleach_linker.linkify(urlize_ietf_docs(full)) + full = mark_safe(clean_html(text)) + full = linkify(urlize_ietf_docs(full)) return format_snippet(full, trunc_words) @register.filter def format_snippet(text, trunc_words=25): # urlize if there aren't already links present - text = bleach_linker.linkify(text) - full = keep_spacing(collapsebr(linebreaksbr(mark_safe(sanitize_fragment(text))))) + text = linkify(text) + full = keep_spacing(collapsebr(linebreaksbr(mark_safe(clean_html(text))))) snippet = truncatewords_html(full, trunc_words) if snippet != full: return mark_safe('
%s
%s
' % (snippet, full)) @@ -477,6 +480,19 @@ def state(doc, slug): slug = "%s-stream-%s" % (doc.type_id, doc.stream_id) return doc.get_state(slug) + +@register.filter +def is_unexpected_wg_state(doc): + """Returns a flag indicating whether the document has an unexpected wg state.""" + if not doc.type_id == "draft": + return False + + draft_iesg_state = doc.get_state("draft-iesg") + draft_stream_state = doc.get_state("draft-stream-ietf") + + return draft_iesg_state.slug != "idexists" and draft_stream_state is not None and draft_stream_state.slug != "sub-pub" + + @register.filter def statehelp(state): "Output help icon with tooltip for state." @@ -505,10 +521,52 @@ def plural(text, seq, arg='s'): else: return text + pluralize(len(seq), arg) + +# Translation table to escape ICS characters. The {} | {} construction builds up a dict +# mapping characters to arbitrary-length strings or None. Values in later dicts override +# earlier ones prior to conversion to a translation table, so excluding a char and then +# mapping it to an escape sequence results in its being escaped, not dropped. +rfc5545_text_escapes = str.maketrans( + # text = *(TSAFE-CHAR / ":" / DQUOTE / ESCAPED-CHAR) + # TSAFE-CHAR = WSP / %x21 / %x23-2B / %x2D-39 / %x3C-5B / + # %x5D-7E / NON-US-ASCII + {chr(c): None for c in range(0x00, 0x20)} # strip 0x00-0x20 + | { + # ESCAPED-CHAR = ("\\" / "\;" / "\," / "\N" / "\n") + "\n": r"\n", + ";": r"\;", + ",": r"\,", + "\\": r"\\", # rhs is two backslashes! + "\t": "\t", # htab ok (0x09) + " ": " ", # space ok (0x20) + } +) + + @register.filter def ics_esc(text): - text = re.sub(r"([\n,;\\])", r"\\\1", text) - return text + """Escape a string to use in an iCalendar text context + + >>> ics_esc('simple') + 'simple' + + For the next tests, it helps to know: + chr(0x09) = "\t" + chr(0x0a) = "\n" + chr(0x0d) = "\r" + chr(0x5c) = "\\" + + >>> ics_esc(f'strips{chr(0x0d)}out{chr(0x0d)}LFs') + 'stripsoutLFs' + + + >>> ics_esc(f'escapes;and,and{chr(0x5c)}and{chr(0x0a)}') + 'escapes\\\\;and\\\\,and\\\\\\\\and\\\\n' + + >>> ics_esc(f"keeps spaces : and{chr(0x09)}tabs") + 'keeps spaces : and\\ttabs' + """ + return text.translate(rfc5545_text_escapes) @register.simple_tag @@ -530,15 +588,22 @@ def ics_date_time(dt, tzname): >>> ics_date_time(datetime.datetime(2022,1,2,3,4,5), 'UTC') ':20220102T030405Z' + >>> ics_date_time(datetime.datetime(2022,1,2,3,4,5), 'GmT') + ':20220102T030405Z' + >>> ics_date_time(datetime.datetime(2022,1,2,3,4,5), 'America/Los_Angeles') ';TZID=America/Los_Angeles:20220102T030405' """ timestamp = dt.strftime('%Y%m%dT%H%M%S') - if tzname.lower() == 'utc': + if tzname.lower() in ('gmt', 'utc'): return f':{timestamp}Z' else: return f';TZID={ics_esc(tzname)}:{timestamp}' +@register.filter +def next_day(value): + return value + datetime.timedelta(days=1) + @register.filter def consensus(doc): @@ -556,7 +621,7 @@ def consensus(doc): @register.filter def std_level_to_label_format(doc): """Returns valid Bootstrap classes to label a status level badge.""" - if doc.is_rfc(): + if doc.type_id == "rfc": if doc.related_that("obs"): return "obs" else: @@ -577,6 +642,8 @@ def pos_to_label_format(text): 'Recuse': 'bg-recuse text-light', 'Not Ready': 'bg-discuss text-light', 'Need More Time': 'bg-discuss text-light', + 'Concern': 'bg-discuss text-light', + }.get(str(text), 'bg-norecord text-dark') @register.filter @@ -591,6 +658,7 @@ def pos_to_border_format(text): 'Recuse': 'border-recuse', 'Not Ready': 'border-discuss', 'Need More Time': 'border-discuss', + 'Concern': 'border-discuss', }.get(str(text), 'border-norecord') @register.filter @@ -650,7 +718,7 @@ def rfcbis(s): @stringfilter def urlize(value): raise RuntimeError("Use linkify from textfilters instead of urlize") - + @register.filter @stringfilter def charter_major_rev(rev): @@ -664,17 +732,25 @@ def charter_minor_rev(rev): @register.filter() def can_defer(user,doc): ballot = doc.latest_event(BallotDocEvent, type="created_ballot") - if ballot and (doc.type_id == "draft" or doc.type_id == "conflrev") and doc.stream_id == 'ietf' and has_role(user, 'Area Director,Secretariat'): + if ballot and (doc.type_id == "draft" or doc.type_id == "conflrev" or doc.type_id=="statchg") and doc.stream_id == 'ietf' and has_role(user, 'Area Director,Secretariat'): return True else: return False +@register.filter() +def can_clear_ballot(user, doc): + return can_defer(user, doc) + +@register.filter() +def can_request_rfc_publication(user, doc): + return utils_can_request_rfc_publication(user, doc) + @register.filter() def can_ballot(user,doc): - # Only IRSG members (and the secretariat, handled by code separately) can take positions on IRTF documents - # Otherwise, an AD can take a position on anything that has a ballot open - if doc.type_id == 'draft' and doc.stream_id == 'irtf': - return has_role(user,'IRSG Member') + if doc.stream_id == "irtf" and doc.type_id == "draft": + return has_role(user,"IRSG Member") + elif doc.stream_id == "editorial" and doc.type_id == "draft": + return has_role(user,"RSAB Member") else: return user.person.role_set.filter(name="ad", group__type="area", group__state="active") @@ -693,10 +769,10 @@ def action_holder_badge(action_holder): '' >>> action_holder_badge(DocumentActionHolderFactory(time_added=timezone.now() - datetime.timedelta(days=16))) - ' 16' + ' 16' >>> action_holder_badge(DocumentActionHolderFactory(time_added=timezone.now() - datetime.timedelta(days=30))) - ' 30' + ' 30' >>> settings.DOC_ACTION_HOLDER_AGE_LIMIT_DAYS = old_limit """ @@ -704,7 +780,7 @@ def action_holder_badge(action_holder): age = (timezone.now() - action_holder.time_added).days if age > age_limit: return mark_safe( - ' %d' + ' %d' % (age, "s" if age != 1 else "", age_limit, age) ) else: @@ -831,3 +907,171 @@ def is_valid_url(url): except ValidationError: return False return True + + +@register.filter +def badgeify(blob): + """ + Add an appropriate bootstrap badge around "text", based on its contents. + """ + config = [ + (r"rejected|not ready|serious issues", "danger", "x-lg"), + (r"complete|accepted|ready", "success", ""), + (r"has nits|almost ready", "info", "info-lg"), + (r"has issues|on the right track", "warning", "exclamation-lg"), + (r"assigned", "info", "person-plus-fill"), + (r"will not review|overtaken by events|withdrawn", "secondary", "dash-lg"), + (r"no response", "warning", "question-lg"), + ] + text = str(blob) + + for pattern, color, icon in config: + if re.search(pattern, text, flags=re.IGNORECASE): + # Shorten the badge text + text = re.sub(r"with ", "w/", text, flags=re.IGNORECASE) + text = re.sub(r"document", "doc", text, flags=re.IGNORECASE) + text = re.sub(r"will not", "won't", text, flags=re.IGNORECASE) + + return mark_safe( + f""" + + {text.capitalize()} + + """ + ) + + return text + +@register.filter +def simple_history_delta_changes(history): + """Returns diff between given history and previous entry.""" + prev = history.prev_record + if prev: + delta = history.diff_against(prev) + return delta.changes + return [] + +@register.filter +def simple_history_delta_change_cnt(history): + """Returns number of changes between given history and previous entry.""" + prev = history.prev_record + if prev: + delta = history.diff_against(prev) + return len(delta.changes) + return 0 + +@register.filter +def mtime(path): + """Returns a datetime object representing mtime given a pathlib Path object""" + return datetime.datetime.fromtimestamp(path.stat().st_mtime).astimezone(ZoneInfo(settings.TIME_ZONE)) + +@register.filter +def mtime_is_epoch(path): + return path.stat().st_mtime == 0 + +@register.filter +def url_for_path(path): + """Consructs a 'best' URL for web access to the given pathlib Path object. + + Assumes that the path is into the Internet-Draft archive or the proceedings. + """ + if Path(settings.AGENDA_PATH) in path.parents: + return ( + f"https://www.ietf.org/proceedings/{path.relative_to(settings.AGENDA_PATH)}" + ) + elif any( + [ + pathdir in path.parents + for pathdir in [ + Path(settings.INTERNET_DRAFT_PATH), + Path(settings.INTERNET_DRAFT_ARCHIVE_DIR).parent, + Path(settings.INTERNET_ALL_DRAFTS_ARCHIVE_DIR), + ] + ] + ): + return f"{settings.IETF_ID_ARCHIVE_URL}{path.name}" + else: + return "#" + + +@register.filter +def is_in_stream(doc): + """ + Check if the doc is in one of the states in it stream that + indicate that is actually adopted, i.e., part of the stream. + (There are various "candidate" states that necessitate this + filter.) + """ + if not doc.stream: + return False + stream = doc.stream.slug + state = doc.get_state_slug(f"draft-stream-{doc.stream.slug}") + if not state: + return True + if stream == "ietf": + return state not in ["wg-cand", "c-adopt"] + elif stream == "irtf": + return state != "candidat" + elif stream == "iab": + return state not in ["candidat", "diff-org"] + elif stream == "editorial": + return True + return False + + +@register.filter +def is_doc_ietf_adoptable(doc): + return doc.stream_id is None or all( + [ + doc.stream_id == "ietf", + doc.get_state_slug("draft-stream-ietf") + not in [ + "c-adopt", + "adopt-wg", + "info", + "wg-doc", + "parked", + "dead", + "wg-lc", + "waiting-for-implementation", + "chair-w", + "writeupw", + "sub-pub", + ], + doc.get_state_slug("draft") != "rfc", + doc.became_rfc() is None, + ] + ) + + +@register.filter +def can_issue_ietf_wg_lc(doc): + return all( + [ + doc.stream_id == "ietf", + doc.get_state_slug("draft-stream-ietf") + not in ["wg-cand", "c-adopt", "wg-lc"], + doc.get_state_slug("draft") != "rfc", + doc.became_rfc() is None, + ] + ) + + +@register.filter +def can_submit_to_iesg(doc): + return all( + [ + doc.stream_id == "ietf", + doc.get_state_slug("draft-iesg") == "idexists", + doc.get_state_slug("draft-stream-ietf") not in ["wg-cand", "c-adopt"], + ] + ) + + +@register.filter +def has_had_ietf_wg_lc(doc): + return ( + doc.stream_id == "ietf" + and doc.docevent_set.filter(statedocevent__state__slug="wg-lc").exists() + ) + diff --git a/ietf/doc/templatetags/mail_filters.py b/ietf/doc/templatetags/mail_filters.py index 32e8dd0ca8..6be6620315 100644 --- a/ietf/doc/templatetags/mail_filters.py +++ b/ietf/doc/templatetags/mail_filters.py @@ -9,7 +9,7 @@ def std_level_prompt(doc): to the object's intended_std_level (with the word RFC appended in some cases), or a prompt requesting that the intended_std_level be set.""" - prompt = "*** YOU MUST SELECT AN INTENDED STATUS FOR THIS DRAFT AND REGENERATE THIS TEXT ***" + prompt = "*** YOU MUST SELECT AN INTENDED STATUS FOR THIS INTERNET-DRAFT AND REGENERATE THIS TEXT ***" if doc.intended_std_level: prompt = doc.intended_std_level.name diff --git a/ietf/doc/templatetags/tests_ietf_filters.py b/ietf/doc/templatetags/tests_ietf_filters.py index f791d61530..b5130849ea 100644 --- a/ietf/doc/templatetags/tests_ietf_filters.py +++ b/ietf/doc/templatetags/tests_ietf_filters.py @@ -3,13 +3,26 @@ from django.conf import settings from ietf.doc.factories import ( - WgDraftFactory, + WgRfcFactory, IndividualDraftFactory, CharterFactory, NewRevisionDocEventFactory, + StatusChangeFactory, + RgDraftFactory, + EditorialDraftFactory, + WgDraftFactory, + ConflictReviewFactory, + BofreqFactory, + StatementFactory, + RfcFactory, +) +from ietf.doc.models import DocEvent +from ietf.doc.templatetags.ietf_filters import ( + urlize_ietf_docs, + is_valid_url, + is_in_stream, + is_unexpected_wg_state, ) -from ietf.doc.models import State, DocEvent, DocAlias -from ietf.doc.templatetags.ietf_filters import urlize_ietf_docs, is_valid_url from ietf.person.models import Person from ietf.utils.test_utils import TestCase @@ -19,29 +32,42 @@ class IetfFiltersTests(TestCase): + def test_is_in_stream(self): + for draft in [ + IndividualDraftFactory(), + CharterFactory(), + StatusChangeFactory(), + ConflictReviewFactory(), + StatementFactory(), + BofreqFactory(), + ]: + self.assertFalse(is_in_stream(draft)) + for draft in [RgDraftFactory(), WgDraftFactory(), EditorialDraftFactory()]: + self.assertTrue(is_in_stream(draft)) + for stream in ["iab", "ietf", "irtf", "ise", "editorial"]: + self.assertTrue(is_in_stream(IndividualDraftFactory(stream_id=stream))) + def test_is_valid_url(self): cases = [(settings.IDTRACKER_BASE_URL, True), ("not valid", False)] for url, result in cases: self.assertEqual(is_valid_url(url), result) def test_urlize_ietf_docs(self): - wg_id = WgDraftFactory() - wg_id.set_state(State.objects.get(type="draft", slug="rfc")) - wg_id.std_level_id = "bcp" - wg_id.save_with_history( + rfc = WgRfcFactory(rfc_number=123456, std_level_id="bcp") + rfc.save_with_history( [ DocEvent.objects.create( - doc=wg_id, - rev=wg_id.rev, + doc=rfc, + rev=rfc.rev, type="published_rfc", by=Person.objects.get(name="(System)"), ) ] ) - DocAlias.objects.create(name="rfc123456").docs.add(wg_id) - DocAlias.objects.create(name="bcp123456").docs.add(wg_id) - DocAlias.objects.create(name="std123456").docs.add(wg_id) - DocAlias.objects.create(name="fyi123456").docs.add(wg_id) + # TODO - bring these into existance when subseries are well modeled + # DocAlias.objects.create(name="bcp123456").docs.add(rfc) + # DocAlias.objects.create(name="std123456").docs.add(rfc) + # DocAlias.objects.create(name="fyi123456").docs.add(rfc) id = IndividualDraftFactory(name="draft-me-rfc123456bis") id_num = IndividualDraftFactory(name="draft-rosen-rfcefdp-update-2026") @@ -59,15 +85,16 @@ def test_urlize_ietf_docs(self): cases = [ ("no change", "no change"), - ("bCp123456", 'bCp123456'), - ("Std 00123456", 'Std 00123456'), - ( - "FyI 0123456 changes std 00123456", - 'FyI 0123456 changes std 00123456', - ), + # TODO: rework subseries when we add them + # ("bCp123456", 'bCp123456'), + # ("Std 00123456", 'Std 00123456'), + # ( + # "FyI 0123456 changes std 00123456", + # 'FyI 0123456 changes std 00123456', + # ), ("rfc123456", 'rfc123456'), ("Rfc 0123456", 'Rfc 0123456'), - (wg_id.name, f'{wg_id.name}'), + (rfc.name, f'{rfc.name}'), ( f"{id.name}-{id.rev}.txt", f'{id.name}-{id.rev}.txt', @@ -149,3 +176,17 @@ def test_urlize_ietf_docs(self): for input, output in cases: # debug.show("(input, urlize_ietf_docs(input), output)") self.assertEqual(urlize_ietf_docs(input), output) + + def test_is_unexpected_wg_state(self): + """ + Test that the unexpected_wg_state function works correctly + """ + # test documents with expected wg states + self.assertFalse(is_unexpected_wg_state(RfcFactory())) + self.assertFalse(is_unexpected_wg_state(WgDraftFactory (states=[('draft-stream-ietf', 'sub-pub')]))) + self.assertFalse(is_unexpected_wg_state(WgDraftFactory (states=[('draft-iesg', 'idexists')]))) + self.assertFalse(is_unexpected_wg_state(WgDraftFactory (states=[('draft-stream-ietf', 'wg-cand'), ('draft-iesg','idexists')]))) + + # test documents with unexpected wg states due to invalid combination of states + self.assertTrue(is_unexpected_wg_state(WgDraftFactory (states=[('draft-stream-ietf', 'wg-cand'), ('draft-iesg','lc-req')]))) + self.assertTrue(is_unexpected_wg_state(WgDraftFactory (states=[('draft-stream-ietf', 'chair-w'), ('draft-iesg','pub-req')]))) diff --git a/ietf/doc/templatetags/wg_menu.py b/ietf/doc/templatetags/wg_menu.py index e82f125c07..3e8d209448 100644 --- a/ietf/doc/templatetags/wg_menu.py +++ b/ietf/doc/templatetags/wg_menu.py @@ -1,4 +1,4 @@ -# Copyright The IETF Trust 2009-2022, All Rights Reserved +# Copyright The IETF Trust 2009-2023, All Rights Reserved # Copyright (C) 2009-2010 Nokia Corporation and/or its subsidiary(-ies). # All rights reserved. Contact: Pasi Eronen @@ -32,6 +32,8 @@ # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE # OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. +import debug # pyflakes: ignore + from django import template from django.template.loader import render_to_string from django.db import models @@ -60,15 +62,13 @@ @register.simple_tag def wg_menu(flavor): - global parents - for p in parents: p.short_name = parent_short_names.get(p.acronym) or p.name if p.short_name.endswith(" Area"): p.short_name = p.short_name[: -len(" Area")] if p.type_id == "area": - p.menu_url = "/wg/#" + p.acronym + p.menu_url = "/wg/#" + p.acronym.upper() elif p.acronym == "irtf": p.menu_url = "/rg/" elif p.acronym == "iab": diff --git a/ietf/doc/tests.py b/ietf/doc/tests.py index c60ee8b3d4..f92c9648e6 100644 --- a/ietf/doc/tests.py +++ b/ietf/doc/tests.py @@ -1,13 +1,16 @@ -# Copyright The IETF Trust 2012-2020, All Rights Reserved +# Copyright The IETF Trust 2012-2024, All Rights Reserved # -*- coding: utf-8 -*- import os import datetime import io +from hashlib import sha384 + +from django.http import HttpRequest import lxml import bibtexparser -import mock +from unittest import mock import json import copy import random @@ -16,11 +19,9 @@ from pathlib import Path from pyquery import PyQuery from urllib.parse import urlparse, parse_qs -from tempfile import NamedTemporaryFile from collections import defaultdict from zoneinfo import ZoneInfo -from django.core.management import call_command from django.urls import reverse as urlreverse from django.conf import settings from django.forms import Form @@ -31,20 +32,34 @@ from tastypie.test import ResourceTestCaseMixin +from weasyprint.urls import URLFetchingError + import debug # pyflakes:ignore -from ietf.doc.models import ( Document, DocAlias, DocRelationshipName, RelatedDocument, State, +from ietf.doc.models import ( Document, DocRelationshipName, RelatedDocument, State, DocEvent, BallotPositionDocEvent, LastCallDocEvent, WriteupDocEvent, NewRevisionDocEvent, BallotType, - EditedAuthorsDocEvent ) -from ietf.doc.factories import ( DocumentFactory, DocEventFactory, CharterFactory, - ConflictReviewFactory, WgDraftFactory, IndividualDraftFactory, WgRfcFactory, - IndividualRfcFactory, StateDocEventFactory, BallotPositionDocEventFactory, - BallotDocEventFactory, DocumentAuthorFactory, NewRevisionDocEventFactory, - StatusChangeFactory, BofreqFactory, DocExtResourceFactory) + EditedAuthorsDocEvent, StateType) +from ietf.doc.factories import (DocumentFactory, DocEventFactory, CharterFactory, + ConflictReviewFactory, WgDraftFactory, + IndividualDraftFactory, WgRfcFactory, + IndividualRfcFactory, StateDocEventFactory, + BallotPositionDocEventFactory, + BallotDocEventFactory, DocumentAuthorFactory, + NewRevisionDocEventFactory, + StatusChangeFactory, DocExtResourceFactory, + RgDraftFactory, BcpFactory, RfcAuthorFactory) from ietf.doc.forms import NotifyForm from ietf.doc.fields import SearchableDocumentsField -from ietf.doc.utils import create_ballot_if_not_open, uppercase_std_abbreviated_name -from ietf.doc.views_search import ad_dashboard_group, ad_dashboard_group_type, shorten_group_name # TODO: red flag that we're importing from views in tests. Move these to utils. +from ietf.doc.utils import ( + create_ballot_if_not_open, + investigate_fragment, + uppercase_std_abbreviated_name, + DraftAliasGenerator, + generate_idnits2_rfc_status, + generate_idnits2_rfcs_obsoleted, + get_doc_email_aliases, +) +from ietf.doc.views_doc import get_diff_revisions from ietf.group.models import Group, Role from ietf.group.factories import GroupFactory, RoleFactory from ietf.ipr.factories import HolderIprDisclosureFactory @@ -52,20 +67,22 @@ from ietf.meeting.factories import ( MeetingFactory, SessionFactory, SessionPresentationFactory, ProceedingsMaterialFactory ) -from ietf.name.models import SessionStatusName, BallotPositionName, DocTypeName +from ietf.name.models import SessionStatusName, BallotPositionName, DocTypeName, RoleName from ietf.person.models import Person from ietf.person.factories import PersonFactory, EmailFactory -from ietf.utils.mail import outbox, empty_outbox -from ietf.utils.test_utils import login_testing_unauthorized, unicontent, reload_db_objects +from ietf.utils.mail import get_payload_text, outbox, empty_outbox +from ietf.utils.test_utils import login_testing_unauthorized, unicontent from ietf.utils.test_utils import TestCase from ietf.utils.text import normalize_text from ietf.utils.timezone import date_today, datetime_today, DEADLINE_TZINFO, RPC_TZINFO +from ietf.doc.utils_search import AD_WORKLOAD class SearchTests(TestCase): def test_search(self): draft = WgDraftFactory(name='draft-ietf-mars-test',group=GroupFactory(acronym='mars',parent=Group.objects.get(acronym='farfut')),authors=[PersonFactory()],ad=PersonFactory()) + rfc = WgRfcFactory() draft.set_state(State.objects.get(used=True, type="draft-iesg", slug="pub-req")) old_draft = IndividualDraftFactory(name='draft-foo-mars-test',authors=[PersonFactory()],title="Optimizing Martian Network Topologies") old_draft.set_state(State.objects.get(used=True, type="draft", slug="expired")) @@ -93,11 +110,16 @@ def test_search(self): self.assertEqual(r.status_code, 200) self.assertContains(r, "draft-foo-mars-test") - # find by rfc/active/inactive - draft.set_state(State.objects.get(type="draft", slug="rfc")) - r = self.client.get(base_url + "?rfcs=on&name=%s" % draft.name) + r = self.client.get(base_url + "?olddrafts=on&name=FoO") # mixed case self.assertEqual(r.status_code, 200) - self.assertContains(r, draft.title) + self.assertContains(r, "draft-foo-mars-test") + + # find by RFC + r = self.client.get(base_url + "?rfcs=on&name=%s" % rfc.name) + self.assertEqual(r.status_code, 200) + self.assertContains(r, rfc.title) + + # find by active/inactive draft.set_state(State.objects.get(type="draft", slug="active")) r = self.client.get(base_url + "?activedrafts=on&name=%s" % draft.name) @@ -126,6 +148,10 @@ def test_search(self): self.assertEqual(r.status_code, 200) self.assertContains(r, draft.title) + r = self.client.get(base_url + "?activedrafts=on&by=group&group=%s" % draft.group.acronym.swapcase()) + self.assertEqual(r.status_code, 200) + self.assertContains(r, draft.title) + # find by area r = self.client.get(base_url + "?activedrafts=on&by=area&area=%s" % draft.group.parent_id) self.assertEqual(r.status_code, 200) @@ -146,6 +172,23 @@ def test_search(self): self.assertEqual(r.status_code, 200) self.assertContains(r, draft.title) + def test_search_became_rfc(self): + draft = WgDraftFactory() + rfc = WgRfcFactory() + draft.set_state(State.objects.get(type="draft", slug="rfc")) + draft.relateddocument_set.create(relationship_id="became_rfc", target=rfc) + base_url = urlreverse('ietf.doc.views_search.search') + + # find by RFC + r = self.client.get(base_url + f"?rfcs=on&name={rfc.name}") + self.assertEqual(r.status_code, 200) + self.assertContains(r, rfc.title) + + # find by draft + r = self.client.get(base_url + f"?activedrafts=on&rfcs=on&name={draft.name}") + self.assertEqual(r.status_code, 200) + self.assertContains(r, rfc.title) + def test_search_for_name(self): draft = WgDraftFactory(name='draft-ietf-mars-test',group=GroupFactory(acronym='mars',parent=Group.objects.get(acronym='farfut')),authors=[PersonFactory()],ad=PersonFactory()) draft.set_state(State.objects.get(used=True, type="draft-iesg", slug="pub-req")) @@ -167,16 +210,39 @@ def test_search_for_name(self): self.assertEqual(r.status_code, 302) self.assertEqual(urlparse(r["Location"]).path, urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=draft.name))) + # mixed-up case exact match + r = self.client.get(urlreverse('ietf.doc.views_search.search_for_name', kwargs=dict(name=draft.name.swapcase()))) + self.assertEqual(r.status_code, 302) + self.assertEqual(urlparse(r["Location"]).path, urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=draft.name))) + # prefix match r = self.client.get(urlreverse('ietf.doc.views_search.search_for_name', kwargs=dict(name="-".join(draft.name.split("-")[:-1])))) self.assertEqual(r.status_code, 302) self.assertEqual(urlparse(r["Location"]).path, urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=draft.name))) + # mixed-up case prefix match + r = self.client.get( + urlreverse( + 'ietf.doc.views_search.search_for_name', + kwargs=dict(name="-".join(draft.name.swapcase().split("-")[:-1])), + )) + self.assertEqual(r.status_code, 302) + self.assertEqual(urlparse(r["Location"]).path, urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=draft.name))) + # non-prefix match r = self.client.get(urlreverse('ietf.doc.views_search.search_for_name', kwargs=dict(name="-".join(draft.name.split("-")[1:])))) self.assertEqual(r.status_code, 302) self.assertEqual(urlparse(r["Location"]).path, urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=draft.name))) + # mixed-up case non-prefix match + r = self.client.get( + urlreverse( + 'ietf.doc.views_search.search_for_name', + kwargs=dict(name="-".join(draft.name.swapcase().split("-")[1:])), + )) + self.assertEqual(r.status_code, 302) + self.assertEqual(urlparse(r["Location"]).path, urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=draft.name))) + # other doctypes than drafts doc = Document.objects.get(name='charter-ietf-mars') r = self.client.get(urlreverse('ietf.doc.views_search.search_for_name', kwargs=dict(name='charter-ietf-ma'))) @@ -230,6 +296,17 @@ def test_search_for_name(self): parsed = urlparse(r["Location"]) self.assertEqual(parsed.path, urlreverse('ietf.doc.views_search.search')) self.assertEqual(parse_qs(parsed.query)["name"][0], "draft-ietf-doesnotexist-42") + + def test_search_rfc(self): + rfc = WgRfcFactory(name="rfc0000") + + # search for existing RFC should redirect directly to the RFC page + r = self.client.get(urlreverse('ietf.doc.views_search.search_for_name', kwargs=dict(name=rfc.name))) + self.assertRedirects(r, f'/doc/{rfc.name}/', status_code=302, target_status_code=200) + + # search for existing RFC with revision number should redirect to the RFC page + r = self.client.get(urlreverse('ietf.doc.views_search.search_for_name', kwargs=dict(name=rfc.name + "-99")), follow=True) + self.assertRedirects(r, f'/doc/{rfc.name}/', status_code=302, target_status_code=200) def test_frontpage(self): r = self.client.get("/") @@ -237,52 +314,68 @@ def test_frontpage(self): self.assertContains(r, "Document Search") def test_ad_workload(self): - Role.objects.filter(name_id='ad').delete() - ad = RoleFactory(name_id='ad',group__type_id='area',group__state_id='active',person__name='Example Areadirector').person - doc_type_names = ['bofreq', 'charter', 'conflrev', 'draft', 'statchg'] - expected = defaultdict(lambda :0) - for doc_type_name in doc_type_names: - if doc_type_name=='draft': - states = State.objects.filter(type='draft-iesg', used=True).values_list('slug', flat=True) - else: - states = State.objects.filter(type=doc_type_name, used=True).values_list('slug', flat=True) - - for state in states: - target_num = random.randint(0,2) + Role.objects.filter(name_id="ad").delete() + ad = RoleFactory( + name_id="ad", + group__type_id="area", + group__state_id="active", + person__name="Example Areadirector", + ).person + expected = defaultdict(lambda: 0) + for doc_type_slug in AD_WORKLOAD: + for state in AD_WORKLOAD[doc_type_slug]: + target_num = random.randint(0, 2) for _ in range(target_num): - if doc_type_name == 'draft': - doc = IndividualDraftFactory(ad=ad,states=[('draft-iesg', state),('draft','rfc' if state=='pub' else 'active')]) - elif doc_type_name == 'charter': - doc = CharterFactory(ad=ad, states=[(doc_type_name, state)]) - elif doc_type_name == 'bofreq': - # Note that the view currently doesn't handle bofreqs - doc = BofreqFactory(states=[(doc_type_name, state)], bofreqresponsibledocevent__responsible=[ad]) - elif doc_type_name == 'conflrev': - doc = ConflictReviewFactory(ad=ad, states=State.objects.filter(type_id=doc_type_name, slug=state)) - elif doc_type_name == 'statchg': - doc = StatusChangeFactory(ad=ad, states=State.objects.filter(type_id=doc_type_name, slug=state)) - else: - # Currently unreachable - doc = DocumentFactory(type_id=doc_type_name, ad=ad, states=[(doc_type_name, state)]) - - if not slugify(ad_dashboard_group_type(doc)) in ('document', 'none'): - expected[(slugify(ad_dashboard_group_type(doc)), slugify(ad.full_name_as_key()), slugify(shorten_group_name(ad_dashboard_group(doc))))] += 1 - - url = urlreverse('ietf.doc.views_search.ad_workload') + if ( + doc_type_slug == "draft" + or doc_type_slug == "rfc" + and state == "rfcqueue" + ): + IndividualDraftFactory( + ad=ad, + states=[ + ("draft-iesg", state), + ("draft", "rfc" if state == "pub" else "active"), + ], + ) + elif doc_type_slug == "rfc": + WgRfcFactory.create( + states=[("draft", "rfc"), ("draft-iesg", "pub")] + ) + + elif doc_type_slug == "charter": + CharterFactory(ad=ad, states=[(doc_type_slug, state)]) + elif doc_type_slug == "conflrev": + ConflictReviewFactory( + ad=ad, + states=State.objects.filter( + type_id=doc_type_slug, slug=state + ), + ) + elif doc_type_slug == "statchg": + StatusChangeFactory( + ad=ad, + states=State.objects.filter( + type_id=doc_type_slug, slug=state + ), + ) + self.client.login(username="ad", password="ad+password") + url = urlreverse("ietf.doc.views_search.ad_workload") r = self.client.get(url) self.assertEqual(r.status_code, 200) q = PyQuery(r.content) for group_type, ad, group in expected: - self.assertEqual(int(q(f'#{group_type}-{ad}-{group}').text()),expected[(group_type, ad, group)]) + self.assertEqual( + int(q(f"#{group_type}-{ad}-{group}").text()), + expected[(group_type, ad, group)], + ) def test_docs_for_ad(self): ad = RoleFactory(name_id='ad',group__type_id='area',group__state_id='active').person draft = IndividualDraftFactory(ad=ad) draft.action_holders.set([PersonFactory()]) draft.set_state(State.objects.get(type='draft-iesg', slug='lc')) - rfc = IndividualDraftFactory(ad=ad) - rfc.set_state(State.objects.get(type='draft', slug='rfc')) - DocAlias.objects.create(name='rfc6666').docs.add(rfc) + rfc = IndividualRfcFactory(ad=ad) conflrev = DocumentFactory(type_id='conflrev',ad=ad) conflrev.set_state(State.objects.get(type='conflrev', slug='iesgeval')) statchg = DocumentFactory(type_id='statchg',ad=ad) @@ -306,7 +399,7 @@ def test_docs_for_ad(self): self.assertEqual(r.status_code, 200) self.assertContains(r, draft.name) self.assertContains(r, escape(draft.action_holders.first().name)) - self.assertContains(r, rfc.canonical_name()) + self.assertContains(r, rfc.name) self.assertContains(r, conflrev.name) self.assertContains(r, statchg.name) self.assertContains(r, charter.name) @@ -314,6 +407,30 @@ def test_docs_for_ad(self): self.assertContains(r, discuss_other.doc.name) self.assertContains(r, block_other.doc.name) + def test_docs_for_iesg(self): + ad1 = RoleFactory(name_id='ad',group__type_id='area',group__state_id='active').person + ad2 = RoleFactory(name_id='ad',group__type_id='area',group__state_id='active').person + + draft = IndividualDraftFactory(ad=ad1) + draft.action_holders.set([PersonFactory()]) + draft.set_state(State.objects.get(type='draft-iesg', slug='lc')) + rfc = IndividualRfcFactory(ad=ad2) + conflrev = DocumentFactory(type_id='conflrev',ad=ad1) + conflrev.set_state(State.objects.get(type='conflrev', slug='iesgeval')) + statchg = DocumentFactory(type_id='statchg',ad=ad2) + statchg.set_state(State.objects.get(type='statchg', slug='iesgeval')) + charter = CharterFactory(name='charter-ietf-ames',ad=ad1) + charter.set_state(State.objects.get(type='charter', slug='iesgrev')) + + r = self.client.get(urlreverse('ietf.doc.views_search.docs_for_iesg')) + self.assertEqual(r.status_code, 200) + self.assertContains(r, draft.name) + self.assertContains(r, escape(draft.action_holders.first().name)) + self.assertNotContains(r, rfc.name) + self.assertContains(r, conflrev.name) + self.assertContains(r, statchg.name) + self.assertContains(r, charter.name) + def test_auth48_doc_for_ad(self): """Docs in AUTH48 state should have a decoration""" ad = RoleFactory(name_id='ad', group__type_id='area', group__state_id='active').person @@ -336,17 +453,6 @@ def test_drafts_in_last_call(self): self.assertContains(r, draft.title) self.assertContains(r, escape(draft.action_holders.first().name)) - def test_in_iesg_process(self): - doc_in_process = IndividualDraftFactory() - doc_in_process.action_holders.set([PersonFactory()]) - doc_in_process.set_state(State.objects.get(type='draft-iesg', slug='lc')) - doc_not_in_process = IndividualDraftFactory() - r = self.client.get(urlreverse('ietf.doc.views_search.drafts_in_iesg_process')) - self.assertEqual(r.status_code, 200) - self.assertContains(r, doc_in_process.title) - self.assertContains(r, escape(doc_in_process.action_holders.first().name)) - self.assertNotContains(r, doc_not_in_process.title) - def test_indexes(self): draft = IndividualDraftFactory() rfc = WgRfcFactory() @@ -354,16 +460,17 @@ def test_indexes(self): r = self.client.get(urlreverse('ietf.doc.views_search.index_all_drafts')) self.assertEqual(r.status_code, 200) self.assertContains(r, draft.name) - self.assertContains(r, rfc.canonical_name().upper()) + self.assertContains(r, rfc.name.upper()) r = self.client.get(urlreverse('ietf.doc.views_search.index_active_drafts')) self.assertEqual(r.status_code, 200) self.assertContains(r, draft.title) def test_ajax_search_docs(self): - draft = IndividualDraftFactory() + draft = IndividualDraftFactory(name="draft-ietf-rfc1234bis") + rfc = IndividualRfcFactory(rfc_number=1234) + bcp = IndividualRfcFactory(name="bcp12345", type_id="bcp") - # Document url = urlreverse('ietf.doc.views_search.ajax_select2_search_docs', kwargs={ "model_name": "document", "doc_type": "draft", @@ -373,18 +480,27 @@ def test_ajax_search_docs(self): data = r.json() self.assertEqual(data[0]["id"], draft.pk) - # DocAlias - doc_alias = draft.docalias.first() - url = urlreverse('ietf.doc.views_search.ajax_select2_search_docs', kwargs={ - "model_name": "docalias", - "doc_type": "draft", + "model_name": "document", + "doc_type": "rfc", }) + r = self.client.get(url, dict(q=rfc.name)) + self.assertEqual(r.status_code, 200) + data = r.json() + self.assertEqual(data[0]["id"], rfc.pk) - r = self.client.get(url, dict(q=doc_alias.name)) + url = urlreverse('ietf.doc.views_search.ajax_select2_search_docs', kwargs={ + "model_name": "document", + "doc_type": "all", + }) + r = self.client.get(url, dict(q="1234")) self.assertEqual(r.status_code, 200) data = r.json() - self.assertEqual(data[0]["id"], doc_alias.pk) + self.assertEqual(len(data), 3) + pks = set([data[i]["id"] for i in range(3)]) + self.assertEqual(pks, set([bcp.pk, rfc.pk, draft.pk])) + + def test_recent_drafts(self): # Three drafts to show with various warnings @@ -403,8 +519,8 @@ def test_recent_drafts(self): self.assertEqual(r.status_code, 200) q = PyQuery(r.content) self.assertEqual(len(q('td.doc')),3) - self.assertTrue(q('td.status span.bg-warning[title*="%s"]' % "for 15 days")) - self.assertTrue(q('td.status span.bg-danger[title*="%s"]' % "for 29 days")) + self.assertTrue(q('td.status span.text-bg-warning[title*="%s"]' % "for 15 days")) + self.assertTrue(q('td.status span.text-bg-danger[title*="%s"]' % "for 29 days")) for ah in [draft.action_holders.first() for draft in drafts]: self.assertContains(r, escape(ah.name)) @@ -586,23 +702,24 @@ def setUp(self): f.write(self.draft_text) def test_document_draft(self): - draft = WgDraftFactory(name='draft-ietf-mars-test',rev='01') + draft = WgDraftFactory(name='draft-ietf-mars-test',rev='01', create_revisions=range(0,2)) + HolderIprDisclosureFactory(docs=[draft]) # Docs for testing relationships. Does not test 'possibly-replaces'. The 'replaced_by' direction # is tested separately below. replaced = IndividualDraftFactory() - draft.relateddocument_set.create(relationship_id='replaces',source=draft,target=replaced.docalias.first()) + draft.relateddocument_set.create(relationship_id='replaces',source=draft,target=replaced) obsoleted = IndividualDraftFactory() - draft.relateddocument_set.create(relationship_id='obs',source=draft,target=obsoleted.docalias.first()) + draft.relateddocument_set.create(relationship_id='obs',source=draft,target=obsoleted) obsoleted_by = IndividualDraftFactory() - obsoleted_by.relateddocument_set.create(relationship_id='obs',source=obsoleted_by,target=draft.docalias.first()) + obsoleted_by.relateddocument_set.create(relationship_id='obs',source=obsoleted_by,target=draft) updated = IndividualDraftFactory() - draft.relateddocument_set.create(relationship_id='updates',source=draft,target=updated.docalias.first()) + draft.relateddocument_set.create(relationship_id='updates',source=draft,target=updated) updated_by = IndividualDraftFactory() - updated_by.relateddocument_set.create(relationship_id='updates',source=obsoleted_by,target=draft.docalias.first()) + updated_by.relateddocument_set.create(relationship_id='updates',source=obsoleted_by,target=draft) - external_resource = DocExtResourceFactory(doc=draft) + DocExtResourceFactory(doc=draft) # these tests aren't testing all attributes yet, feel free to # expand them @@ -613,69 +730,32 @@ def test_document_draft(self): if settings.USER_PREFERENCE_DEFAULTS['full_draft'] == 'off': self.assertContains(r, "Show full document") self.assertNotContains(r, "Deimos street") - self.assertContains(r, replaced.canonical_name()) + self.assertContains(r, replaced.name) self.assertContains(r, replaced.title) - # obs/updates not included until draft is RFC - self.assertNotContains(r, obsoleted.canonical_name()) - self.assertNotContains(r, obsoleted.title) - self.assertNotContains(r, obsoleted_by.canonical_name()) - self.assertNotContains(r, obsoleted_by.title) - self.assertNotContains(r, updated.canonical_name()) - self.assertNotContains(r, updated.title) - self.assertNotContains(r, updated_by.canonical_name()) - self.assertNotContains(r, updated_by.title) - self.assertContains(r, external_resource.value) r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=draft.name)) + "?include_text=0") self.assertEqual(r.status_code, 200) self.assertContains(r, "Active Internet-Draft") self.assertContains(r, "Show full document") self.assertNotContains(r, "Deimos street") - self.assertContains(r, replaced.canonical_name()) + self.assertContains(r, replaced.name) self.assertContains(r, replaced.title) - # obs/updates not included until draft is RFC - self.assertNotContains(r, obsoleted.canonical_name()) - self.assertNotContains(r, obsoleted.title) - self.assertNotContains(r, obsoleted_by.canonical_name()) - self.assertNotContains(r, obsoleted_by.title) - self.assertNotContains(r, updated.canonical_name()) - self.assertNotContains(r, updated.title) - self.assertNotContains(r, updated_by.canonical_name()) - self.assertNotContains(r, updated_by.title) r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=draft.name)) + "?include_text=foo") self.assertEqual(r.status_code, 200) self.assertContains(r, "Active Internet-Draft") self.assertNotContains(r, "Show full document") self.assertContains(r, "Deimos street") - self.assertContains(r, replaced.canonical_name()) + self.assertContains(r, replaced.name) self.assertContains(r, replaced.title) - # obs/updates not included until draft is RFC - self.assertNotContains(r, obsoleted.canonical_name()) - self.assertNotContains(r, obsoleted.title) - self.assertNotContains(r, obsoleted_by.canonical_name()) - self.assertNotContains(r, obsoleted_by.title) - self.assertNotContains(r, updated.canonical_name()) - self.assertNotContains(r, updated.title) - self.assertNotContains(r, updated_by.canonical_name()) - self.assertNotContains(r, updated_by.title) r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=draft.name)) + "?include_text=1") self.assertEqual(r.status_code, 200) self.assertContains(r, "Active Internet-Draft") self.assertNotContains(r, "Show full document") self.assertContains(r, "Deimos street") - self.assertContains(r, replaced.canonical_name()) + self.assertContains(r, replaced.name) self.assertContains(r, replaced.title) - # obs/updates not included until draft is RFC - self.assertNotContains(r, obsoleted.canonical_name()) - self.assertNotContains(r, obsoleted.title) - self.assertNotContains(r, obsoleted_by.canonical_name()) - self.assertNotContains(r, obsoleted_by.title) - self.assertNotContains(r, updated.canonical_name()) - self.assertNotContains(r, updated.title) - self.assertNotContains(r, updated_by.canonical_name()) - self.assertNotContains(r, updated_by.title) self.client.cookies = SimpleCookie({str('full_draft'): str('on')}) r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=draft.name))) @@ -683,17 +763,8 @@ def test_document_draft(self): self.assertContains(r, "Active Internet-Draft") self.assertNotContains(r, "Show full document") self.assertContains(r, "Deimos street") - self.assertContains(r, replaced.canonical_name()) + self.assertContains(r, replaced.name) self.assertContains(r, replaced.title) - # obs/updates not included until draft is RFC - self.assertNotContains(r, obsoleted.canonical_name()) - self.assertNotContains(r, obsoleted.title) - self.assertNotContains(r, obsoleted_by.canonical_name()) - self.assertNotContains(r, obsoleted_by.title) - self.assertNotContains(r, updated.canonical_name()) - self.assertNotContains(r, updated.title) - self.assertNotContains(r, updated_by.canonical_name()) - self.assertNotContains(r, updated_by.title) self.client.cookies = SimpleCookie({str('full_draft'): str('off')}) r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=draft.name))) @@ -701,17 +772,8 @@ def test_document_draft(self): self.assertContains(r, "Active Internet-Draft") self.assertContains(r, "Show full document") self.assertNotContains(r, "Deimos street") - self.assertContains(r, replaced.canonical_name()) + self.assertContains(r, replaced.name) self.assertContains(r, replaced.title) - # obs/updates not included until draft is RFC - self.assertNotContains(r, obsoleted.canonical_name()) - self.assertNotContains(r, obsoleted.title) - self.assertNotContains(r, obsoleted_by.canonical_name()) - self.assertNotContains(r, obsoleted_by.title) - self.assertNotContains(r, updated.canonical_name()) - self.assertNotContains(r, updated.title) - self.assertNotContains(r, updated_by.canonical_name()) - self.assertNotContains(r, updated_by.title) self.client.cookies = SimpleCookie({str('full_draft'): str('foo')}) r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=draft.name))) @@ -720,17 +782,8 @@ def test_document_draft(self): if settings.USER_PREFERENCE_DEFAULTS['full_draft'] == 'off': self.assertContains(r, "Show full document") self.assertNotContains(r, "Deimos street") - self.assertContains(r, replaced.canonical_name()) + self.assertContains(r, replaced.name) self.assertContains(r, replaced.title) - # obs/updates not included until draft is RFC - self.assertNotContains(r, obsoleted.canonical_name()) - self.assertNotContains(r, obsoleted.title) - self.assertNotContains(r, obsoleted_by.canonical_name()) - self.assertNotContains(r, obsoleted_by.title) - self.assertNotContains(r, updated.canonical_name()) - self.assertNotContains(r, updated.title) - self.assertNotContains(r, updated_by.canonical_name()) - self.assertNotContains(r, updated_by.title) r = self.client.get(urlreverse("ietf.doc.views_doc.document_html", kwargs=dict(name=draft.name))) self.assertEqual(r.status_code, 200) @@ -740,7 +793,7 @@ def test_document_draft(self): self.assertEqual(q('title').text(), 'draft-ietf-mars-test-01') self.assertEqual(len(q('.rfcmarkup pre')), 3) self.assertEqual(len(q('.rfcmarkup span.h1, .rfcmarkup h1')), 2) - self.assertEqual(len(q('.rfcmarkup a[href]')), 28) + self.assertEqual(len(q('.rfcmarkup a[href]')), 27) r = self.client.get(urlreverse("ietf.doc.views_doc.document_html", kwargs=dict(name=draft.name, rev=draft.rev))) self.assertEqual(r.status_code, 200) @@ -754,17 +807,18 @@ def test_document_draft(self): self.assertEqual(len(q('#sidebar option[value="draft-ietf-mars-test-00"][selected="selected"]')), 1) rfc = WgRfcFactory() + rfc.save_with_history([DocEventFactory(doc=rfc)]) (Path(settings.RFC_PATH) / rfc.get_base_name()).touch() - r = self.client.get(urlreverse("ietf.doc.views_doc.document_html", kwargs=dict(name=rfc.canonical_name()))) + r = self.client.get(urlreverse("ietf.doc.views_doc.document_html", kwargs=dict(name=rfc.name))) self.assertEqual(r.status_code, 200) q = PyQuery(r.content) - self.assertEqual(q('title').text(), f'RFC {rfc.rfc_number()} - {rfc.title}') + self.assertEqual(q('title').text(), f'RFC {rfc.rfc_number} - {rfc.title}') # synonyms for the rfc should be redirected to its canonical view - r = self.client.get(urlreverse("ietf.doc.views_doc.document_html", kwargs=dict(name=rfc.rfc_number()))) - self.assertRedirects(r, urlreverse("ietf.doc.views_doc.document_html", kwargs=dict(name=rfc.canonical_name()))) - r = self.client.get(urlreverse("ietf.doc.views_doc.document_html", kwargs=dict(name=f'RFC {rfc.rfc_number()}'))) - self.assertRedirects(r, urlreverse("ietf.doc.views_doc.document_html", kwargs=dict(name=rfc.canonical_name()))) + r = self.client.get(urlreverse("ietf.doc.views_doc.document_html", kwargs=dict(name=rfc.rfc_number))) + self.assertRedirects(r, urlreverse("ietf.doc.views_doc.document_html", kwargs=dict(name=rfc.name))) + r = self.client.get(urlreverse("ietf.doc.views_doc.document_html", kwargs=dict(name=f'RFC {rfc.rfc_number}'))) + self.assertRedirects(r, urlreverse("ietf.doc.views_doc.document_html", kwargs=dict(name=rfc.name))) # expired draft draft.set_state(State.objects.get(type="draft", slug="expired")) @@ -783,48 +837,55 @@ def test_document_draft(self): stream_id=draft.stream_id, group_id=draft.group_id, abstract=draft.abstract,stream=draft.stream, rev=draft.rev, pages=draft.pages, intended_std_level_id=draft.intended_std_level_id, shepherd_id=draft.shepherd_id, ad_id=draft.ad_id, expires=draft.expires, - notify=draft.notify, note=draft.note) + notify=draft.notify) rel = RelatedDocument.objects.create(source=replacement, - target=draft.docalias.get(name__startswith="draft"), + target=draft, relationship_id="replaces") r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=draft.name))) self.assertEqual(r.status_code, 200) self.assertContains(r, "Replaced Internet-Draft") - self.assertContains(r, replacement.canonical_name()) + self.assertContains(r, replacement.name) self.assertContains(r, replacement.title) rel.delete() # draft published as RFC draft.set_state(State.objects.get(type="draft", slug="rfc")) - draft.std_level_id = "bcp" - draft.save_with_history([DocEvent.objects.create(doc=draft, rev=draft.rev, type="published_rfc", by=Person.objects.get(name="(System)"))]) + draft.std_level_id = "ps" + rfc = WgRfcFactory(group=draft.group, name="rfc123456") + rfc.save_with_history([DocEvent.objects.create(doc=rfc, rev=None, type="published_rfc", by=Person.objects.get(name="(System)"))]) - rfc_alias = DocAlias.objects.create(name="rfc123456") - rfc_alias.docs.add(draft) - bcp_alias = DocAlias.objects.create(name="bcp123456") - bcp_alias.docs.add(draft) + draft.relateddocument_set.create(relationship_id="became_rfc", target=rfc) + + obsoleted = IndividualRfcFactory() + rfc.relateddocument_set.create(relationship_id='obs',target=obsoleted) + obsoleted_by = IndividualRfcFactory() + obsoleted_by.relateddocument_set.create(relationship_id='obs',target=rfc) + updated = IndividualRfcFactory() + rfc.relateddocument_set.create(relationship_id='updates',target=updated) + updated_by = IndividualRfcFactory() + updated_by.relateddocument_set.create(relationship_id='updates',target=rfc) + + r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=draft.name, rev=draft.rev))) + self.assertEqual(r.status_code, 200) + self.assertContains(r, "This is an older version of an Internet-Draft that was ultimately published as") r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=draft.name))) self.assertEqual(r.status_code, 302) - r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=bcp_alias.name))) - self.assertEqual(r.status_code, 302) - r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=rfc_alias.name))) + r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=rfc.name))) self.assertEqual(r.status_code, 200) self.assertContains(r, "RFC 123456") self.assertContains(r, draft.name) - self.assertContains(r, replaced.canonical_name()) - self.assertContains(r, replaced.title) # obs/updates included with RFC - self.assertContains(r, obsoleted.canonical_name()) + self.assertContains(r, obsoleted.name) self.assertContains(r, obsoleted.title) - self.assertContains(r, obsoleted_by.canonical_name()) + self.assertContains(r, obsoleted_by.name) self.assertContains(r, obsoleted_by.title) - self.assertContains(r, updated.canonical_name()) + self.assertContains(r, updated.name) self.assertContains(r, updated.title) - self.assertContains(r, updated_by.canonical_name()) + self.assertContains(r, updated_by.name) self.assertContains(r, updated_by.title) # naked RFC - also weird that we test a PS from the ISE @@ -857,7 +918,7 @@ def test_draft_status_changes(self): draft = WgRfcFactory() status_change_doc = StatusChangeFactory( group=draft.group, - changes_status_of=[('tops', draft.docalias.first())], + changes_status_of=[('tops', draft)], ) status_change_url = urlreverse( 'ietf.doc.views_doc.document_main', @@ -865,7 +926,7 @@ def test_draft_status_changes(self): ) proposed_status_change_doc = StatusChangeFactory( group=draft.group, - changes_status_of=[('tobcp', draft.docalias.first())], + changes_status_of=[('tobcp', draft)], states=[State.objects.get(slug='needshep', type='statchg')], ) proposed_status_change_url = urlreverse( @@ -876,7 +937,7 @@ def test_draft_status_changes(self): r = self.client.get( urlreverse( 'ietf.doc.views_doc.document_main', - kwargs={'name': draft.canonical_name()}, + kwargs={'name': draft.name}, ) ) self.assertEqual(r.status_code, 200) @@ -922,7 +983,7 @@ def test_edit_authors_permissions(self): # Relevant users not authorized to edit authors unauthorized_usernames = [ 'plain', - *[author.user.username for author in draft.authors()], + *[author.user.username for author in draft.author_persons()], draft.group.get_chair().person.user.username, 'ad' ] @@ -937,7 +998,7 @@ def test_edit_authors_permissions(self): self.client.logout() # Try to add an author via POST - still only the secretary should be able to do this. - orig_authors = draft.authors() + orig_authors = draft.author_persons() post_data = self.make_edit_authors_post_data( basis='permission test', authors=draft.documentauthor_set.all(), @@ -955,12 +1016,12 @@ def test_edit_authors_permissions(self): for username in unauthorized_usernames: login_testing_unauthorized(self, username, url, method='post', request_kwargs=dict(data=post_data)) draft = Document.objects.get(pk=draft.pk) - self.assertEqual(draft.authors(), orig_authors) # ensure draft author list was not modified + self.assertEqual(draft.author_persons(), orig_authors) # ensure draft author list was not modified login_testing_unauthorized(self, 'secretary', url, method='post', request_kwargs=dict(data=post_data)) r = self.client.post(url, post_data) self.assertEqual(r.status_code, 302) draft = Document.objects.get(pk=draft.pk) - self.assertEqual(draft.authors(), orig_authors + [new_auth_person]) + self.assertEqual(draft.author_persons(), orig_authors + [new_auth_person]) def make_edit_authors_post_data(self, basis, authors): """Helper to generate edit_authors POST data for a set of authors""" @@ -1308,8 +1369,8 @@ def test_edit_authors_edit_fields(self): basis=change_reason ) - old_address = draft.authors()[0].email() - new_email = EmailFactory(person=draft.authors()[0], address=f'changed-{old_address}') + old_address = draft.author_persons()[0].email() + new_email = EmailFactory(person=draft.author_persons()[0], address=f'changed-{old_address}') post_data['author-0-email'] = new_email.address post_data['author-1-affiliation'] = 'University of Nowhere' post_data['author-2-country'] = 'Chile' @@ -1342,17 +1403,17 @@ def test_edit_authors_edit_fields(self): country_event = change_events.filter(desc__icontains='changed country').first() self.assertIsNotNone(email_event) - self.assertIn(draft.authors()[0].name, email_event.desc) + self.assertIn(draft.author_persons()[0].name, email_event.desc) self.assertIn(before[0]['email'], email_event.desc) self.assertIn(after[0]['email'], email_event.desc) self.assertIsNotNone(affiliation_event) - self.assertIn(draft.authors()[1].name, affiliation_event.desc) + self.assertIn(draft.author_persons()[1].name, affiliation_event.desc) self.assertIn(before[1]['affiliation'], affiliation_event.desc) self.assertIn(after[1]['affiliation'], affiliation_event.desc) self.assertIsNotNone(country_event) - self.assertIn(draft.authors()[2].name, country_event.desc) + self.assertIn(draft.author_persons()[2].name, country_event.desc) self.assertIn(before[2]['country'], country_event.desc) self.assertIn(after[2]['country'], country_event.desc) @@ -1410,6 +1471,14 @@ def test_document_draft_action_holders_buttons(self, mock_method): """Buttons for action holders should be shown when AD or secretary""" draft = WgDraftFactory() draft.action_holders.set([PersonFactory()]) + other_group = GroupFactory(type_id=draft.group.type_id) + + # create a test RoleName and put it in the docman_roles for the document group + RoleName.objects.create(slug="wrangler", name="Wrangler", used=True) + draft.group.features.docman_roles.append("wrangler") + draft.group.features.save() + wrangler = RoleFactory(group=draft.group, name_id="wrangler").person + wrangler_of_other_group = RoleFactory(group=other_group, name_id="wrangler").person url = urlreverse('ietf.doc.views_doc.document_main', kwargs=dict(name=draft.name)) edit_ah_url = urlreverse('ietf.doc.views_doc.edit_action_holders', kwargs=dict(name=draft.name)) @@ -1442,6 +1511,8 @@ def _run_test(username=None, expect_buttons=False): _run_test(None, False) _run_test('plain', False) + _run_test(wrangler_of_other_group.user.username, False) + _run_test(wrangler.user.username, True) _run_test('ad', True) _run_test('secretary', True) @@ -1456,11 +1527,11 @@ def test_draft_group_link(self): self.assertEqual(r.status_code, 200) self.assert_correct_wg_group_link(r, group) - rfc = WgRfcFactory(name='draft-rfc-document-%s' % group_type_id, group=group) + rfc = WgRfcFactory(group=group) + draft = WgDraftFactory(group=group) + draft.relateddocument_set.create(relationship_id="became_rfc", target=rfc) DocEventFactory.create(doc=rfc, type='published_rfc', time=event_datetime) - # get the rfc name to avoid a redirect - rfc_name = rfc.docalias.filter(name__startswith='rfc').first().name - r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=rfc_name))) + r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=rfc.name))) self.assertEqual(r.status_code, 200) self.assert_correct_wg_group_link(r, group) @@ -1471,14 +1542,33 @@ def test_draft_group_link(self): self.assertEqual(r.status_code, 200) self.assert_correct_non_wg_group_link(r, group) - rfc = WgRfcFactory(name='draft-rfc-document-%s' % group_type_id, group=group) + rfc = WgRfcFactory(group=group) + draft = WgDraftFactory(name='draft-rfc-document-%s'% group_type_id, group=group) + draft.relateddocument_set.create(relationship_id="became_rfc", target=rfc) DocEventFactory.create(doc=rfc, type='published_rfc', time=event_datetime) - # get the rfc name to avoid a redirect - rfc_name = rfc.docalias.filter(name__startswith='rfc').first().name - r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=rfc_name))) + r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=rfc.name))) self.assertEqual(r.status_code, 200) self.assert_correct_non_wg_group_link(r, group) + def test_document_email_authors_button(self): + # rfc not from draft + rfc = WgRfcFactory() + DocEventFactory.create(doc=rfc, type='published_rfc') + url = urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=rfc.name)) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q('a:contains("Email authors")')), 0, 'Did not expect "Email authors" button') + + # rfc from draft + draft = WgDraftFactory(group=rfc.group) + draft.relateddocument_set.create(relationship_id="became_rfc", target=rfc) + draft.set_state(State.objects.get(used=True, type="draft", slug="rfc")) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q('a:contains("Email authors")')), 1, 'Expected "Email authors" button') + def test_document_primary_and_history_views(self): IndividualDraftFactory(name='draft-imaginary-independent-submission') ConflictReviewFactory(name='conflict-review-imaginary-irtf-submission') @@ -1576,13 +1666,17 @@ def test_status_change(self): statchg = StatusChangeFactory() r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=statchg.name))) self.assertEqual(r.status_code, 200) - r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=statchg.relateddocument_set.first().target.document))) - self.assertEqual(r.status_code, 302) + r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=statchg.relateddocument_set.first().target))) + self.assertEqual(r.status_code, 200) def test_document_charter(self): CharterFactory(name='charter-ietf-mars') r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name="charter-ietf-mars"))) self.assertEqual(r.status_code, 200) + + def test_incorrect_rfc_url(self): + r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name="rfc8989", rev="00"))) + self.assertEqual(r.status_code, 404) def test_document_conflict_review(self): ConflictReviewFactory(name='conflict-review-imaginary-irtf-submission') @@ -1620,6 +1714,17 @@ def test_document_material(self): r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=doc.name))) self.assertEqual(r.status_code, 200) + self.assertNotContains(r, "The session for this document was cancelled.") + + SchedulingEvent.objects.create( + session=session, + status_id='canceled', + by = Person.objects.get(user__username="marschairman"), + ) + + r = self.client.get(urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=doc.name))) + self.assertEqual(r.status_code, 200) + self.assertContains(r, "The session for this document was cancelled.") def test_document_ballot(self): doc = IndividualDraftFactory() @@ -1737,38 +1842,88 @@ def test_document_ballot_needed_positions(self): self.assertNotContains(r, 'more YES or NO') # status change - DocAlias.objects.create(name='rfc9998').docs.add(IndividualDraftFactory()) - DocAlias.objects.create(name='rfc9999').docs.add(IndividualDraftFactory()) + Document.objects.create(name='rfc9998') + Document.objects.create(name='rfc9999') doc = DocumentFactory(type_id='statchg',name='status-change-imaginary-mid-review') iesgeval_pk = str(State.objects.get(slug='iesgeval',type__slug='statchg').pk) empty_outbox() self.client.login(username='ad', password='ad+password') r = self.client.post(urlreverse('ietf.doc.views_status_change.change_state',kwargs=dict(name=doc.name)),dict(new_state=iesgeval_pk)) self.assertEqual(r.status_code, 302) - r = self.client.get(r._headers["location"][1]) + r = self.client.get(r.headers["location"]) self.assertContains(r, ">IESG Evaluation<") self.assertEqual(len(outbox), 2) self.assertIn('iesg-secretary',outbox[0]['To']) self.assertIn('drafts-eval',outbox[1]['To']) - doc.relateddocument_set.create(target=DocAlias.objects.get(name='rfc9998'),relationship_id='tohist') + doc.relateddocument_set.create(target=Document.objects.get(name='rfc9998'),relationship_id='tohist') r = self.client.get(urlreverse("ietf.doc.views_doc.document_ballot", kwargs=dict(name=doc.name))) self.assertNotContains(r, 'Needs a YES') self.assertNotContains(r, 'more YES or NO') - doc.relateddocument_set.create(target=DocAlias.objects.get(name='rfc9999'),relationship_id='tois') + doc.relateddocument_set.create(target=Document.objects.get(name='rfc9999'),relationship_id='tois') r = self.client.get(urlreverse("ietf.doc.views_doc.document_ballot", kwargs=dict(name=doc.name))) self.assertContains(r, 'more YES or NO') def test_document_json(self): doc = IndividualDraftFactory() - + author = DocumentAuthorFactory(document=doc) + r = self.client.get(urlreverse("ietf.doc.views_doc.document_json", kwargs=dict(name=doc.name))) self.assertEqual(r.status_code, 200) data = r.json() - self.assertEqual(doc.name, data['name']) - self.assertEqual(doc.pages,data['pages']) + self.assertEqual(data["name"], doc.name) + self.assertEqual(data["pages"], doc.pages) + self.assertEqual( + data["authors"], + [ + { + "name": author.person.name, + "email": author.email.address, + "affiliation": author.affiliation, + } + ] + ) + + def test_document_json_rfc(self): + doc = IndividualRfcFactory() + old_style_author = DocumentAuthorFactory(document=doc) + url = urlreverse("ietf.doc.views_doc.document_json", kwargs=dict(name=doc.name)) + + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + data = r.json() + self.assertEqual(data["name"], doc.name) + self.assertEqual(data["pages"], doc.pages) + self.assertEqual( + data["authors"], + [ + { + "name": old_style_author.person.name, + "email": old_style_author.email.address, + "affiliation": old_style_author.affiliation, + } + ] + ) + + new_style_author = RfcAuthorFactory(document=doc) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + data = r.json() + self.assertEqual(data["name"], doc.name) + self.assertEqual(data["pages"], doc.pages) + self.assertEqual( + data["authors"], + [ + { + "name": new_style_author.titlepage_name, + "email": new_style_author.email.address, + "affiliation": new_style_author.affiliation, + } + ] + ) + def test_writeup(self): doc = IndividualDraftFactory(states = [('draft','active'),('draft-iesg','iesg-eva')],) @@ -1803,6 +1958,18 @@ def test_writeup(self): self.assertContains(r, notes.text) self.assertContains(r, rfced_note.text) + def test_diff_revisions(self): + ind_doc = IndividualDraftFactory(create_revisions=range(2)) + wg_doc = WgDraftFactory( + relations=[("replaces", ind_doc)], create_revisions=range(2) + ) + diff_revisions = get_diff_revisions(HttpRequest(), wg_doc.name, wg_doc) + self.assertEqual(len(diff_revisions), 4) + self.assertEqual( + [t[3] for t in diff_revisions], + [f"{n}-{v:02d}" for n in [wg_doc.name, ind_doc.name] for v in [1, 0]], + ) + def test_history(self): doc = IndividualDraftFactory() @@ -1819,15 +1986,14 @@ def test_history(self): self.assertContains(r, e.desc) def test_history_bis_00(self): - rfcname='rfc9090' - rfc = WgRfcFactory(alias2=rfcname) - bis_draft = WgDraftFactory(name='draft-ietf-{}-{}bis'.format(rfc.group.acronym,rfcname)) + rfc = WgRfcFactory(rfc_number=9090) + bis_draft = WgDraftFactory(name='draft-ietf-{}-{}bis'.format(rfc.group.acronym,rfc.name)) url = urlreverse('ietf.doc.views_doc.document_history', kwargs=dict(name=bis_draft.name)) r = self.client.get(url) self.assertEqual(r.status_code, 200) q = PyQuery(unicontent(r)) - attr1='value="{}"'.format(rfcname) + attr1='value="{}"'.format(rfc.name) self.assertEqual(len(q('option['+attr1+'][selected="selected"]')), 1) @@ -1877,11 +2043,31 @@ def test_last_call_feed(self): self.assertContains(r, doc.name) def test_rfc_feed(self): - WgRfcFactory() + rfc = WgRfcFactory(rfc_number=9000) + DocEventFactory(doc=rfc, type="published_rfc") r = self.client.get("/feed/rfc/") self.assertTrue(r.status_code, 200) + q = PyQuery(r.content[39:]) # Strip off the xml declaration + self.assertEqual(len(q("item")), 1) + item = q("item")[0] + media_content = item.findall("{http://search.yahoo.com/mrss/}content") + self.assertEqual(len(media_content),4) + types = set([m.attrib["type"] for m in media_content]) + self.assertEqual(types, set(["application/rfc+xml", "text/plain", "text/html", "application/pdf"])) + rfcs_2016 = WgRfcFactory.create_batch(3) # rfc numbers will be well below v3 + for rfc in rfcs_2016: + e = DocEventFactory(doc=rfc, type="published_rfc") + e.time = e.time.replace(year=2016) + e.save() r = self.client.get("/feed/rfc/2016") self.assertTrue(r.status_code, 200) + q = PyQuery(r.content[39:]) + self.assertEqual(len(q("item")), 3) + item = q("item")[0] + media_content = item.findall("{http://search.yahoo.com/mrss/}content") + self.assertEqual(len(media_content), 3) + types = set([m.attrib["type"] for m in media_content]) + self.assertEqual(types, set(["text/plain", "text/html", "application/pdf"])) def test_state_help(self): url = urlreverse('ietf.doc.views_help.state_help', kwargs=dict(type="draft-iesg")) @@ -1914,70 +2100,91 @@ def _parse_bibtex_response(self, response) -> dict: @override_settings(RFC_EDITOR_INFO_BASE_URL='https://www.rfc-editor.ietf.org/info/') def test_document_bibtex(self): + + for factory in [CharterFactory, BcpFactory, StatusChangeFactory, ConflictReviewFactory]: # Should be extended to all other doc types + doc = factory() + url = urlreverse("ietf.doc.views_doc.document_bibtex", kwargs=dict(name=doc.name)) + r = self.client.get(url) + self.assertEqual(r.status_code, 404) rfc = WgRfcFactory.create( - #other_aliases = ['rfc6020',], - states = [('draft','rfc'),('draft-iesg','pub')], - std_level_id = 'ps', - time = datetime.datetime(2010, 10, 10, tzinfo=ZoneInfo(settings.TIME_ZONE)), - ) - num = rfc.rfc_number() + time=datetime.datetime(2010, 10, 10, tzinfo=ZoneInfo(settings.TIME_ZONE)) + ) + num = rfc.rfc_number DocEventFactory.create( doc=rfc, - type='published_rfc', + type="published_rfc", time=datetime.datetime(2010, 10, 10, tzinfo=RPC_TZINFO), ) # - url = urlreverse('ietf.doc.views_doc.document_bibtex', kwargs=dict(name=rfc.name)) + url = urlreverse("ietf.doc.views_doc.document_bibtex", kwargs=dict(name=rfc.name)) r = self.client.get(url) - entry = self._parse_bibtex_response(r)["rfc%s"%num] - self.assertEqual(entry['series'], 'Request for Comments') - self.assertEqual(entry['number'], num) - self.assertEqual(entry['doi'], '10.17487/RFC%s'%num) - self.assertEqual(entry['year'], '2010') - self.assertEqual(entry['month'].lower()[0:3], 'oct') - self.assertEqual(entry['url'], f'https://www.rfc-editor.ietf.org/info/rfc{num}') + entry = self._parse_bibtex_response(r)["rfc%s" % num] + self.assertEqual(entry["series"], "Request for Comments") + self.assertEqual(int(entry["number"]), num) + self.assertEqual(entry["doi"], "10.17487/RFC%s" % num) + self.assertEqual(entry["year"], "2010") + self.assertEqual(entry["month"].lower()[0:3], "oct") + self.assertEqual(entry["url"], f"https://www.rfc-editor.ietf.org/info/rfc{num}") # - self.assertNotIn('day', entry) - + self.assertNotIn("day", entry) + + # test for incorrect case - revision for RFC + rfc = WgRfcFactory(name="rfc0000") + url = urlreverse( + "ietf.doc.views_doc.document_bibtex", kwargs=dict(name=rfc.name, rev="00") + ) + r = self.client.get(url) + self.assertEqual(r.status_code, 404) + april1 = IndividualRfcFactory.create( - stream_id = 'ise', - states = [('draft','rfc'),('draft-iesg','pub')], - std_level_id = 'inf', - time = datetime.datetime(1990, 4, 1, tzinfo=ZoneInfo(settings.TIME_ZONE)), - ) - num = april1.rfc_number() + stream_id="ise", + std_level_id="inf", + time=datetime.datetime(1990, 4, 1, tzinfo=ZoneInfo(settings.TIME_ZONE)), + ) + num = april1.rfc_number DocEventFactory.create( doc=april1, - type='published_rfc', + type="published_rfc", time=datetime.datetime(1990, 4, 1, tzinfo=RPC_TZINFO), ) # - url = urlreverse('ietf.doc.views_doc.document_bibtex', kwargs=dict(name=april1.name)) + url = urlreverse( + "ietf.doc.views_doc.document_bibtex", kwargs=dict(name=april1.name) + ) r = self.client.get(url) - self.assertEqual(r.get('Content-Type'), 'text/plain; charset=utf-8') - entry = self._parse_bibtex_response(r)["rfc%s"%num] - self.assertEqual(entry['series'], 'Request for Comments') - self.assertEqual(entry['number'], num) - self.assertEqual(entry['doi'], '10.17487/RFC%s'%num) - self.assertEqual(entry['year'], '1990') - self.assertEqual(entry['month'].lower()[0:3], 'apr') - self.assertEqual(entry['day'], '1') - self.assertEqual(entry['url'], f'https://www.rfc-editor.ietf.org/info/rfc{num}') - + self.assertEqual(r.get("Content-Type"), "text/plain; charset=utf-8") + entry = self._parse_bibtex_response(r)["rfc%s" % num] + self.assertEqual(entry["series"], "Request for Comments") + self.assertEqual(int(entry["number"]), num) + self.assertEqual(entry["doi"], "10.17487/RFC%s" % num) + self.assertEqual(entry["year"], "1990") + self.assertEqual(entry["month"].lower()[0:3], "apr") + self.assertEqual(entry["day"], "1") + self.assertEqual(entry["url"], f"https://www.rfc-editor.ietf.org/info/rfc{num}") + draft = IndividualDraftFactory.create() - docname = '%s-%s' % (draft.name, draft.rev) - bibname = docname[6:] # drop the 'draft-' prefix - url = urlreverse('ietf.doc.views_doc.document_bibtex', kwargs=dict(name=draft.name)) + docname = "%s-%s" % (draft.name, draft.rev) + bibname = docname[6:] # drop the 'draft-' prefix + url = urlreverse("ietf.doc.views_doc.document_bibtex", kwargs=dict(name=draft.name)) r = self.client.get(url) entry = self._parse_bibtex_response(r)[bibname] - self.assertEqual(entry['note'], 'Work in Progress') - self.assertEqual(entry['number'], docname) - self.assertEqual(entry['year'], str(draft.pub_date().year)) - self.assertEqual(entry['month'].lower()[0:3], draft.pub_date().strftime('%b').lower()) - self.assertEqual(entry['day'], str(draft.pub_date().day)) - self.assertEqual(entry['url'], settings.IDTRACKER_BASE_URL + urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=draft.name, rev=draft.rev))) + self.assertEqual(entry["note"], "Work in Progress") + self.assertEqual(entry["number"], docname) + self.assertEqual(entry["year"], str(draft.pub_date().year)) + self.assertEqual( + entry["month"].lower()[0:3], draft.pub_date().strftime("%b").lower() + ) + self.assertEqual(entry["day"], str(draft.pub_date().day)) + self.assertEqual( + entry["url"], + settings.IDTRACKER_BASE_URL + + urlreverse( + "ietf.doc.views_doc.document_main", + kwargs=dict(name=draft.name, rev=draft.rev), + ), + ) # - self.assertNotIn('doi', entry) + self.assertNotIn("doi", entry) def test_document_bibxml(self): draft = IndividualDraftFactory.create() @@ -2008,20 +2215,19 @@ def test_trailing_hypen_digit_name_bibxml(self): class AddCommentTestCase(TestCase): def test_add_comment(self): - draft = WgDraftFactory(name='draft-ietf-mars-test',group__acronym='mars') - url = urlreverse('ietf.doc.views_doc.add_comment', kwargs=dict(name=draft.name)) + draft = WgDraftFactory(name="draft-ietf-mars-test", group__acronym="mars") + url = urlreverse("ietf.doc.views_doc.add_comment", kwargs=dict(name=draft.name)) login_testing_unauthorized(self, "secretary", url) # normal get r = self.client.get(url) self.assertEqual(r.status_code, 200) q = PyQuery(unicontent(r)) - self.assertEqual(len(q('form textarea[name=comment]')), 1) + self.assertEqual(len(q("form textarea[name=comment]")), 1) - # request resurrect events_before = draft.docevent_set.count() mailbox_before = len(outbox) - + r = self.client.post(url, dict(comment="This is a test.")) self.assertEqual(r.status_code, 302) @@ -2029,9 +2235,9 @@ def test_add_comment(self): self.assertEqual("This is a test.", draft.latest_event().desc) self.assertEqual("added_comment", draft.latest_event().type) self.assertEqual(len(outbox), mailbox_before + 1) - self.assertIn("Comment added", outbox[-1]['Subject']) - self.assertIn(draft.name, outbox[-1]['Subject']) - self.assertIn('draft-ietf-mars-test@', outbox[-1]['To']) + self.assertIn("Comment added", outbox[-1]["Subject"]) + self.assertIn(draft.name, outbox[-1]["Subject"]) + self.assertIn("draft-ietf-mars-test@", outbox[-1]["To"]) # Make sure we can also do it as IANA self.client.login(username="iana", password="iana+password") @@ -2040,7 +2246,22 @@ def test_add_comment(self): r = self.client.get(url) self.assertEqual(r.status_code, 200) q = PyQuery(unicontent(r)) - self.assertEqual(len(q('form textarea[name=comment]')), 1) + self.assertEqual(len(q("form textarea[name=comment]")), 1) + + empty_outbox() + rfc = WgRfcFactory() + self.client.login(username="rfc", password="rfc+password") + url = urlreverse("ietf.doc.views_doc.add_comment", kwargs=dict(name=rfc.name)) + r = self.client.post( + url, dict(comment="This is an RFC Editor comment on an RFC.") + ) + self.assertEqual(r.status_code, 302) + + self.assertEqual( + "This is an RFC Editor comment on an RFC.", rfc.latest_event().desc + ) + self.assertEqual(len(outbox), 1) + self.assertIn("This is an RFC Editor comment on an RFC.", get_payload_text(outbox[0])) class TemplateTagTest(TestCase): @@ -2054,7 +2275,7 @@ class ReferencesTest(TestCase): def test_references(self): doc1 = WgDraftFactory(name='draft-ietf-mars-test') - doc2 = IndividualDraftFactory(name='draft-imaginary-independent-submission').docalias.first() + doc2 = IndividualDraftFactory(name='draft-imaginary-independent-submission') RelatedDocument.objects.get_or_create(source=doc1,target=doc2,relationship=DocRelationshipName.objects.get(slug='refnorm')) url = urlreverse('ietf.doc.views_doc.document_references', kwargs=dict(name=doc1.name)) r = self.client.get(url) @@ -2066,124 +2287,169 @@ def test_references(self): self.assertContains(r, doc1.name) class GenerateDraftAliasesTests(TestCase): - def setUp(self): - super().setUp() - self.doc_aliases_file = NamedTemporaryFile(delete=False, mode='w+') - self.doc_aliases_file.close() - self.doc_virtual_file = NamedTemporaryFile(delete=False, mode='w+') - self.doc_virtual_file.close() - self.saved_draft_aliases_path = settings.DRAFT_ALIASES_PATH - self.saved_draft_virtual_path = settings.DRAFT_VIRTUAL_PATH - settings.DRAFT_ALIASES_PATH = self.doc_aliases_file.name - settings.DRAFT_VIRTUAL_PATH = self.doc_virtual_file.name - - def tearDown(self): - settings.DRAFT_ALIASES_PATH = self.saved_draft_aliases_path - settings.DRAFT_VIRTUAL_PATH = self.saved_draft_virtual_path - os.unlink(self.doc_aliases_file.name) - os.unlink(self.doc_virtual_file.name) - super().tearDown() - - def testManagementCommand(self): - a_month_ago = (timezone.now() - datetime.timedelta(30)).astimezone(RPC_TZINFO) - a_month_ago = a_month_ago.replace(hour=0, minute=0, second=0, microsecond=0) - ad = RoleFactory(name_id='ad', group__type_id='area', group__state_id='active').person - shepherd = PersonFactory() - author1 = PersonFactory() - author2 = PersonFactory() - author3 = PersonFactory() - author4 = PersonFactory() - author5 = PersonFactory() - author6 = PersonFactory() - mars = GroupFactory(type_id='wg', acronym='mars') - marschairman = PersonFactory(user__username='marschairman') - mars.role_set.create(name_id='chair', person=marschairman, email=marschairman.email()) - doc1 = IndividualDraftFactory(authors=[author1], shepherd=shepherd.email(), ad=ad) - doc2 = WgDraftFactory(name='draft-ietf-mars-test', group__acronym='mars', authors=[author2], ad=ad) - doc3 = WgRfcFactory.create(name='draft-ietf-mars-finished', group__acronym='mars', authors=[author3], ad=ad, std_level_id='ps', states=[('draft','rfc'),('draft-iesg','pub')], time=a_month_ago) - DocEventFactory.create(doc=doc3, type='published_rfc', time=a_month_ago) - doc4 = WgRfcFactory.create(authors=[author4,author5], ad=ad, std_level_id='ps', states=[('draft','rfc'),('draft-iesg','pub')], time=datetime.datetime(2010,10,10, tzinfo=ZoneInfo(settings.TIME_ZONE))) - DocEventFactory.create(doc=doc4, type='published_rfc', time=datetime.datetime(2010, 10, 10, tzinfo=RPC_TZINFO)) - doc5 = IndividualDraftFactory(authors=[author6]) - - args = [ ] - kwargs = { } - out = io.StringIO() - call_command("generate_draft_aliases", *args, **kwargs, stdout=out, stderr=out) - self.assertFalse(out.getvalue()) - - with open(settings.DRAFT_ALIASES_PATH) as afile: - acontent = afile.read() - self.assertTrue(all([x in acontent for x in [ - 'xfilter-' + doc1.name, - 'xfilter-' + doc1.name + '.ad', - 'xfilter-' + doc1.name + '.authors', - 'xfilter-' + doc1.name + '.shepherd', - 'xfilter-' + doc1.name + '.all', - 'xfilter-' + doc2.name, - 'xfilter-' + doc2.name + '.ad', - 'xfilter-' + doc2.name + '.authors', - 'xfilter-' + doc2.name + '.chairs', - 'xfilter-' + doc2.name + '.all', - 'xfilter-' + doc3.name, - 'xfilter-' + doc3.name + '.ad', - 'xfilter-' + doc3.name + '.authors', - 'xfilter-' + doc3.name + '.chairs', - 'xfilter-' + doc5.name, - 'xfilter-' + doc5.name + '.authors', - 'xfilter-' + doc5.name + '.all', - ]])) - self.assertFalse(all([x in acontent for x in [ - 'xfilter-' + doc1.name + '.chairs', - 'xfilter-' + doc2.name + '.shepherd', - 'xfilter-' + doc3.name + '.shepherd', - 'xfilter-' + doc4.name, - 'xfilter-' + doc5.name + '.shepherd', - 'xfilter-' + doc5.name + '.ad', - ]])) - - with open(settings.DRAFT_VIRTUAL_PATH) as vfile: - vcontent = vfile.read() - self.assertTrue(all([x in vcontent for x in [ - ad.email_address(), - shepherd.email_address(), - marschairman.email_address(), - author1.email_address(), - author2.email_address(), - author3.email_address(), - author6.email_address(), - ]])) - self.assertFalse(all([x in vcontent for x in [ - author4.email_address(), - author5.email_address(), - ]])) - self.assertTrue(all([x in vcontent for x in [ - 'xfilter-' + doc1.name, - 'xfilter-' + doc1.name + '.ad', - 'xfilter-' + doc1.name + '.authors', - 'xfilter-' + doc1.name + '.shepherd', - 'xfilter-' + doc1.name + '.all', - 'xfilter-' + doc2.name, - 'xfilter-' + doc2.name + '.ad', - 'xfilter-' + doc2.name + '.authors', - 'xfilter-' + doc2.name + '.chairs', - 'xfilter-' + doc2.name + '.all', - 'xfilter-' + doc3.name, - 'xfilter-' + doc3.name + '.ad', - 'xfilter-' + doc3.name + '.authors', - 'xfilter-' + doc3.name + '.chairs', - 'xfilter-' + doc5.name, - 'xfilter-' + doc5.name + '.authors', - 'xfilter-' + doc5.name + '.all', - ]])) - self.assertFalse(all([x in vcontent for x in [ - 'xfilter-' + doc1.name + '.chairs', - 'xfilter-' + doc2.name + '.shepherd', - 'xfilter-' + doc3.name + '.shepherd', - 'xfilter-' + doc4.name, - 'xfilter-' + doc5.name + '.shepherd', - 'xfilter-' + doc5.name + '.ad', - ]])) + @override_settings(TOOLS_SERVER="tools.example.org", DRAFT_ALIAS_DOMAIN="draft.example.org") + def test_generator_class(self): + """The DraftAliasGenerator should generate the same lists as the old mgmt cmd""" + a_month_ago = (timezone.now() - datetime.timedelta(30)).astimezone(RPC_TZINFO) + a_month_ago = a_month_ago.replace(hour=0, minute=0, second=0, microsecond=0) + ad = RoleFactory( + name_id="ad", group__type_id="area", group__state_id="active" + ).person + shepherd = PersonFactory() + author1 = PersonFactory() + author2 = PersonFactory() + author3 = PersonFactory() + author4 = PersonFactory() + author5 = PersonFactory() + author6 = PersonFactory() + mars = GroupFactory(type_id="wg", acronym="mars") + marschairman = PersonFactory(user__username="marschairman") + mars.role_set.create( + name_id="chair", person=marschairman, email=marschairman.email() + ) + doc1 = IndividualDraftFactory(authors=[author1], shepherd=shepherd.email(), ad=ad) + doc2 = WgDraftFactory( + name="draft-ietf-mars-test", group__acronym="mars", authors=[author2], ad=ad + ) + doc2.notify = f"{doc2.name}.ad@draft.example.org" + doc2.save() + doc3 = WgDraftFactory.create( + name="draft-ietf-mars-finished", + group__acronym="mars", + authors=[author3], + ad=ad, + std_level_id="ps", + states=[("draft", "rfc"), ("draft-iesg", "pub")], + time=a_month_ago, + ) + rfc3 = WgRfcFactory() + DocEventFactory.create(doc=rfc3, type="published_rfc", time=a_month_ago) + doc3.relateddocument_set.create(relationship_id="became_rfc", target=rfc3) + doc4 = WgDraftFactory.create( + authors=[author4, author5], + ad=ad, + std_level_id="ps", + states=[("draft", "rfc"), ("draft-iesg", "pub")], + time=datetime.datetime(2010, 10, 10, tzinfo=ZoneInfo(settings.TIME_ZONE)), + ) + rfc4 = WgRfcFactory() + DocEventFactory.create( + doc=rfc4, + type="published_rfc", + time=datetime.datetime(2010, 10, 10, tzinfo=RPC_TZINFO), + ) + doc4.relateddocument_set.create(relationship_id="became_rfc", target=rfc4) + doc5 = IndividualDraftFactory(authors=[author6]) + + output = [(alias, alist) for alias, alist in DraftAliasGenerator()] + alias_dict = dict(output) + self.assertEqual(len(alias_dict), len(output)) # no duplicate aliases + expected_dict = { + doc1.name: [author1.email_address()], + doc1.name + ".ad": [ad.email_address()], + doc1.name + ".authors": [author1.email_address()], + doc1.name + ".shepherd": [shepherd.email_address()], + doc1.name + + ".all": [ + author1.email_address(), + ad.email_address(), + shepherd.email_address(), + ], + doc2.name: [author2.email_address()], + doc2.name + ".ad": [ad.email_address()], + doc2.name + ".authors": [author2.email_address()], + doc2.name + ".chairs": [marschairman.email_address()], + doc2.name + ".notify": [ad.email_address()], + doc2.name + + ".all": [ + author2.email_address(), + ad.email_address(), + marschairman.email_address(), + ], + doc3.name: [author3.email_address()], + doc3.name + ".ad": [ad.email_address()], + doc3.name + ".authors": [author3.email_address()], + doc3.name + ".chairs": [marschairman.email_address()], + doc3.name + + ".all": [ + author3.email_address(), + ad.email_address(), + marschairman.email_address(), + ], + doc5.name: [author6.email_address()], + doc5.name + ".authors": [author6.email_address()], + doc5.name + ".all": [author6.email_address()], + } + # Sort lists for comparison + self.assertEqual( + {k: sorted(v) for k, v in alias_dict.items()}, + {k: sorted(v) for k, v in expected_dict.items()}, + ) + + # check single name + output = [(alias, alist) for alias, alist in DraftAliasGenerator(Document.objects.filter(name=doc1.name))] + alias_dict = dict(output) + self.assertEqual(len(alias_dict), len(output)) # no duplicate aliases + expected_dict = { + doc1.name: [author1.email_address()], + doc1.name + ".ad": [ad.email_address()], + doc1.name + ".authors": [author1.email_address()], + doc1.name + ".shepherd": [shepherd.email_address()], + doc1.name + + ".all": [ + author1.email_address(), + ad.email_address(), + shepherd.email_address(), + ], + } + # Sort lists for comparison + self.assertEqual( + {k: sorted(v) for k, v in alias_dict.items()}, + {k: sorted(v) for k, v in expected_dict.items()}, + ) + + @override_settings(TOOLS_SERVER="tools.example.org", DRAFT_ALIAS_DOMAIN="draft.example.org") + def test_get_draft_notify_emails(self): + ad = PersonFactory() + shepherd = PersonFactory() + author = PersonFactory() + doc = DocumentFactory(authors=[author], shepherd=shepherd.email(), ad=ad) + generator = DraftAliasGenerator() + + doc.notify = f"{doc.name}@draft.example.org" + doc.save() + self.assertCountEqual(generator.get_draft_notify_emails(doc), [author.email_address()]) + + doc.notify = f"{doc.name}.ad@draft.example.org" + doc.save() + self.assertCountEqual(generator.get_draft_notify_emails(doc), [ad.email_address()]) + + doc.notify = f"{doc.name}.shepherd@draft.example.org" + doc.save() + self.assertCountEqual(generator.get_draft_notify_emails(doc), [shepherd.email_address()]) + + doc.notify = f"{doc.name}.all@draft.example.org" + doc.save() + self.assertCountEqual( + generator.get_draft_notify_emails(doc), + [ad.email_address(), author.email_address(), shepherd.email_address()] + ) + + doc.notify = f"{doc.name}.notify@draft.example.org" + doc.save() + self.assertCountEqual(generator.get_draft_notify_emails(doc), []) + + doc.notify = f"{doc.name}.ad@somewhere.example.com" + doc.save() + self.assertCountEqual(generator.get_draft_notify_emails(doc), [f"{doc.name}.ad@somewhere.example.com"]) + + doc.notify = f"somebody@example.com, nobody@example.com, {doc.name}.ad@tools.example.org" + doc.save() + self.assertCountEqual( + generator.get_draft_notify_emails(doc), + ["somebody@example.com", "nobody@example.com", ad.email_address()] + ) + class EmailAliasesTests(TestCase): @@ -2192,37 +2458,20 @@ def setUp(self): WgDraftFactory(name='draft-ietf-mars-test',group__acronym='mars') WgDraftFactory(name='draft-ietf-ames-test',group__acronym='ames') RoleFactory(group__type_id='review', group__acronym='yangdoctors', name_id='secr') - self.doc_alias_file = NamedTemporaryFile(delete=False, mode='w+') - self.doc_alias_file.write("""# Generated by hand at 2015-02-12_16:26:45 -virtual.ietf.org anything -draft-ietf-mars-test@ietf.org xfilter-draft-ietf-mars-test -expand-draft-ietf-mars-test@virtual.ietf.org mars-author@example.com, mars-collaborator@example.com -draft-ietf-mars-test.authors@ietf.org xfilter-draft-ietf-mars-test.authors -expand-draft-ietf-mars-test.authors@virtual.ietf.org mars-author@example.mars, mars-collaborator@example.mars -draft-ietf-mars-test.chairs@ietf.org xfilter-draft-ietf-mars-test.chairs -expand-draft-ietf-mars-test.chairs@virtual.ietf.org mars-chair@example.mars -draft-ietf-mars-test.all@ietf.org xfilter-draft-ietf-mars-test.all -expand-draft-ietf-mars-test.all@virtual.ietf.org mars-author@example.mars, mars-collaborator@example.mars, mars-chair@example.mars -draft-ietf-ames-test@ietf.org xfilter-draft-ietf-ames-test -expand-draft-ietf-ames-test@virtual.ietf.org ames-author@example.com, ames-collaborator@example.com -draft-ietf-ames-test.authors@ietf.org xfilter-draft-ietf-ames-test.authors -expand-draft-ietf-ames-test.authors@virtual.ietf.org ames-author@example.ames, ames-collaborator@example.ames -draft-ietf-ames-test.chairs@ietf.org xfilter-draft-ietf-ames-test.chairs -expand-draft-ietf-ames-test.chairs@virtual.ietf.org ames-chair@example.ames -draft-ietf-ames-test.all@ietf.org xfilter-draft-ietf-ames-test.all -expand-draft-ietf-ames-test.all@virtual.ietf.org ames-author@example.ames, ames-collaborator@example.ames, ames-chair@example.ames - -""") - self.doc_alias_file.close() - self.saved_draft_virtual_path = settings.DRAFT_VIRTUAL_PATH - settings.DRAFT_VIRTUAL_PATH = self.doc_alias_file.name - - def tearDown(self): - settings.DRAFT_VIRTUAL_PATH = self.saved_draft_virtual_path - os.unlink(self.doc_alias_file.name) - super().tearDown() - - def testAliases(self): + + + @mock.patch("ietf.doc.views_doc.get_doc_email_aliases") + def testAliases(self, mock_get_aliases): + mock_get_aliases.return_value = [ + {"doc_name": "draft-ietf-mars-test", "alias_type": "", "expansion": "mars-author@example.mars, mars-collaborator@example.mars"}, + {"doc_name": "draft-ietf-mars-test", "alias_type": ".authors", "expansion": "mars-author@example.mars, mars-collaborator@example.mars"}, + {"doc_name": "draft-ietf-mars-test", "alias_type": ".chairs", "expansion": "mars-chair@example.mars"}, + {"doc_name": "draft-ietf-mars-test", "alias_type": ".all", "expansion": "mars-author@example.mars, mars-collaborator@example.mars, mars-chair@example.mars"}, + {"doc_name": "draft-ietf-ames-test", "alias_type": "", "expansion": "ames-author@example.ames, ames-collaborator@example.ames"}, + {"doc_name": "draft-ietf-ames-test", "alias_type": ".authors", "expansion": "ames-author@example.ames, ames-collaborator@example.ames"}, + {"doc_name": "draft-ietf-ames-test", "alias_type": ".chairs", "expansion": "ames-chair@example.ames"}, + {"doc_name": "draft-ietf-ames-test", "alias_type": ".all", "expansion": "ames-author@example.ames, ames-collaborator@example.ames, ames-chair@example.ames"}, + ] PersonFactory(user__username='plain') url = urlreverse('ietf.doc.urls.redirect.document_email', kwargs=dict(name="draft-ietf-mars-test")) r = self.client.get(url) @@ -2232,16 +2481,70 @@ def testAliases(self): login_testing_unauthorized(self, "plain", url) r = self.client.get(url) self.assertEqual(r.status_code, 200) + self.assertEqual(mock_get_aliases.call_args, mock.call()) self.assertTrue(all([x in unicontent(r) for x in ['mars-test@','mars-test.authors@','mars-test.chairs@']])) self.assertTrue(all([x in unicontent(r) for x in ['ames-test@','ames-test.authors@','ames-test.chairs@']])) - def testExpansions(self): + + @mock.patch("ietf.doc.views_doc.get_doc_email_aliases") + def testExpansions(self, mock_get_aliases): + mock_get_aliases.return_value = [ + {"doc_name": "draft-ietf-mars-test", "alias_type": "", "expansion": "mars-author@example.mars, mars-collaborator@example.mars"}, + {"doc_name": "draft-ietf-mars-test", "alias_type": ".authors", "expansion": "mars-author@example.mars, mars-collaborator@example.mars"}, + {"doc_name": "draft-ietf-mars-test", "alias_type": ".chairs", "expansion": "mars-chair@example.mars"}, + {"doc_name": "draft-ietf-mars-test", "alias_type": ".all", "expansion": "mars-author@example.mars, mars-collaborator@example.mars, mars-chair@example.mars"}, + ] url = urlreverse('ietf.doc.views_doc.document_email', kwargs=dict(name="draft-ietf-mars-test")) r = self.client.get(url) + self.assertEqual(mock_get_aliases.call_args, mock.call("draft-ietf-mars-test")) self.assertEqual(r.status_code, 200) self.assertContains(r, 'draft-ietf-mars-test.all@ietf.org') self.assertContains(r, 'iesg_ballot_saved') + + @mock.patch("ietf.doc.utils.DraftAliasGenerator") + def test_get_doc_email_aliases(self, mock_alias_gen_cls): + mock_alias_gen_cls.return_value = [ + ("draft-something-or-other.some-type", ["somebody@example.com"]), + ("draft-something-or-other", ["somebody@example.com"]), + ("draft-nothing-at-all", ["nobody@example.com"]), + ("draft-nothing-at-all.some-type", ["nobody@example.com"]), + ] + # order is important in the response - should be sorted by doc name and otherwise left + # in order + self.assertEqual( + get_doc_email_aliases(), + [ + { + "doc_name": "draft-nothing-at-all", + "alias_type": "", + "expansion": "nobody@example.com", + }, + { + "doc_name": "draft-nothing-at-all", + "alias_type": ".some-type", + "expansion": "nobody@example.com", + }, + { + "doc_name": "draft-something-or-other", + "alias_type": ".some-type", + "expansion": "somebody@example.com", + }, + { + "doc_name": "draft-something-or-other", + "alias_type": "", + "expansion": "somebody@example.com", + }, + ], + ) + self.assertEqual(mock_alias_gen_cls.call_args, mock.call(None)) + # Repeat with a name, no need to re-test that the alias list is actually passed through, just + # check that the DraftAliasGenerator is called correctly + draft = WgDraftFactory() + get_doc_email_aliases(draft.name) + self.assertQuerySetEqual(mock_alias_gen_cls.call_args[0][0], Document.objects.filter(pk=draft.pk)) + + class DocumentMeetingTests(TestCase): def setUp(self): @@ -2264,8 +2567,8 @@ def setUp(self): def test_view_document_meetings(self): doc = IndividualDraftFactory.create() - doc.sessionpresentation_set.create(session=self.inprog,rev=None) - doc.sessionpresentation_set.create(session=self.interim,rev=None) + doc.presentations.create(session=self.inprog,rev=None) + doc.presentations.create(session=self.interim,rev=None) url = urlreverse('ietf.doc.views_doc.all_presentations', kwargs=dict(name=doc.name)) response = self.client.get(url) @@ -2276,8 +2579,8 @@ def test_view_document_meetings(self): self.assertFalse(q('#addsessionsbutton')) self.assertFalse(q("a.btn:contains('Remove document')")) - doc.sessionpresentation_set.create(session=self.past_cutoff,rev=None) - doc.sessionpresentation_set.create(session=self.past,rev=None) + doc.presentations.create(session=self.past_cutoff,rev=None) + doc.presentations.create(session=self.past,rev=None) self.client.login(username="secretary", password="secretary+password") response = self.client.get(url) @@ -2310,41 +2613,72 @@ def test_view_document_meetings(self): self.assertFalse(q("#futuremeets a.btn:contains('Remove document')")) self.assertFalse(q("#pastmeets a.btn:contains('Remove document')")) - def test_edit_document_session(self): + @override_settings(MEETECHO_API_CONFIG="fake settings") + @mock.patch("ietf.doc.views_doc.SlidesManager") + def test_edit_document_session(self, mock_slides_manager_cls): doc = IndividualDraftFactory.create() - sp = doc.sessionpresentation_set.create(session=self.future,rev=None) + sp = doc.presentations.create(session=self.future,rev=None) url = urlreverse('ietf.doc.views_doc.edit_sessionpresentation',kwargs=dict(name='no-such-doc',session_id=sp.session_id)) response = self.client.get(url) self.assertEqual(response.status_code, 404) + self.assertFalse(mock_slides_manager_cls.called) url = urlreverse('ietf.doc.views_doc.edit_sessionpresentation',kwargs=dict(name=doc.name,session_id=0)) response = self.client.get(url) self.assertEqual(response.status_code, 404) + self.assertFalse(mock_slides_manager_cls.called) url = urlreverse('ietf.doc.views_doc.edit_sessionpresentation',kwargs=dict(name=doc.name,session_id=sp.session_id)) response = self.client.get(url) self.assertEqual(response.status_code, 404) + self.assertFalse(mock_slides_manager_cls.called) self.client.login(username=self.other_chair.user.username,password='%s+password'%self.other_chair.user.username) response = self.client.get(url) self.assertEqual(response.status_code, 404) - + self.assertFalse(mock_slides_manager_cls.called) + self.client.login(username=self.group_chair.user.username,password='%s+password'%self.group_chair.user.username) response = self.client.get(url) self.assertEqual(response.status_code, 200) q = PyQuery(response.content) self.assertEqual(2,len(q('select#id_version option'))) + self.assertFalse(mock_slides_manager_cls.called) + # edit draft self.assertEqual(1,doc.docevent_set.count()) response = self.client.post(url,{'version':'00','save':''}) self.assertEqual(response.status_code, 302) - self.assertEqual(doc.sessionpresentation_set.get(pk=sp.pk).rev,'00') + self.assertEqual(doc.presentations.get(pk=sp.pk).rev,'00') self.assertEqual(2,doc.docevent_set.count()) + self.assertFalse(mock_slides_manager_cls.called) + + # editing slides should call Meetecho API + slides = SessionPresentationFactory( + session=self.future, + document__type_id="slides", + document__rev="00", + rev=None, + order=1, + ).document + url = urlreverse( + "ietf.doc.views_doc.edit_sessionpresentation", + kwargs={"name": slides.name, "session_id": self.future.pk}, + ) + response = self.client.post(url, {"version": "00", "save": ""}) + self.assertEqual(response.status_code, 302) + self.assertEqual(mock_slides_manager_cls.call_count, 1) + self.assertEqual(mock_slides_manager_cls.call_args, mock.call(api_config="fake settings")) + self.assertEqual(mock_slides_manager_cls.return_value.send_update.call_count, 1) + self.assertEqual( + mock_slides_manager_cls.return_value.send_update.call_args, + mock.call(self.future), + ) def test_edit_document_session_after_proceedings_closed(self): doc = IndividualDraftFactory.create() - sp = doc.sessionpresentation_set.create(session=self.past_cutoff,rev=None) + sp = doc.presentations.create(session=self.past_cutoff,rev=None) url = urlreverse('ietf.doc.views_doc.edit_sessionpresentation',kwargs=dict(name=doc.name,session_id=sp.session_id)) self.client.login(username=self.group_chair.user.username,password='%s+password'%self.group_chair.user.username) @@ -2357,39 +2691,64 @@ def test_edit_document_session_after_proceedings_closed(self): q=PyQuery(response.content) self.assertEqual(1,len(q(".alert-warning:contains('may affect published proceedings')"))) - def test_remove_document_session(self): + @override_settings(MEETECHO_API_CONFIG="fake settings") + @mock.patch("ietf.doc.views_doc.SlidesManager") + def test_remove_document_session(self, mock_slides_manager_cls): doc = IndividualDraftFactory.create() - sp = doc.sessionpresentation_set.create(session=self.future,rev=None) + sp = doc.presentations.create(session=self.future,rev=None) url = urlreverse('ietf.doc.views_doc.remove_sessionpresentation',kwargs=dict(name='no-such-doc',session_id=sp.session_id)) response = self.client.get(url) self.assertEqual(response.status_code, 404) + self.assertFalse(mock_slides_manager_cls.called) url = urlreverse('ietf.doc.views_doc.remove_sessionpresentation',kwargs=dict(name=doc.name,session_id=0)) response = self.client.get(url) self.assertEqual(response.status_code, 404) + self.assertFalse(mock_slides_manager_cls.called) url = urlreverse('ietf.doc.views_doc.remove_sessionpresentation',kwargs=dict(name=doc.name,session_id=sp.session_id)) response = self.client.get(url) self.assertEqual(response.status_code, 404) + self.assertFalse(mock_slides_manager_cls.called) self.client.login(username=self.other_chair.user.username,password='%s+password'%self.other_chair.user.username) response = self.client.get(url) self.assertEqual(response.status_code, 404) - + self.assertFalse(mock_slides_manager_cls.called) + self.client.login(username=self.group_chair.user.username,password='%s+password'%self.group_chair.user.username) response = self.client.get(url) self.assertEqual(response.status_code, 200) + self.assertFalse(mock_slides_manager_cls.called) + # removing a draft self.assertEqual(1,doc.docevent_set.count()) response = self.client.post(url,{'remove_session':''}) self.assertEqual(response.status_code, 302) - self.assertFalse(doc.sessionpresentation_set.filter(pk=sp.pk).exists()) + self.assertFalse(doc.presentations.filter(pk=sp.pk).exists()) self.assertEqual(2,doc.docevent_set.count()) + self.assertFalse(mock_slides_manager_cls.called) + + # removing slides should call Meetecho API + slides = SessionPresentationFactory(session=self.future, document__type_id="slides", order=1).document + url = urlreverse( + "ietf.doc.views_doc.remove_sessionpresentation", + kwargs={"name": slides.name, "session_id": self.future.pk}, + ) + response = self.client.post(url, {"remove_session": ""}) + self.assertEqual(response.status_code, 302) + self.assertEqual(mock_slides_manager_cls.call_count, 1) + self.assertEqual(mock_slides_manager_cls.call_args, mock.call(api_config="fake settings")) + self.assertEqual(mock_slides_manager_cls.return_value.delete.call_count, 1) + self.assertEqual( + mock_slides_manager_cls.return_value.delete.call_args, + mock.call(self.future, slides), + ) def test_remove_document_session_after_proceedings_closed(self): doc = IndividualDraftFactory.create() - sp = doc.sessionpresentation_set.create(session=self.past_cutoff,rev=None) + sp = doc.presentations.create(session=self.past_cutoff,rev=None) url = urlreverse('ietf.doc.views_doc.remove_sessionpresentation',kwargs=dict(name=doc.name,session_id=sp.session_id)) self.client.login(username=self.group_chair.user.username,password='%s+password'%self.group_chair.user.username) @@ -2402,28 +2761,49 @@ def test_remove_document_session_after_proceedings_closed(self): q=PyQuery(response.content) self.assertEqual(1,len(q(".alert-warning:contains('may affect published proceedings')"))) - def test_add_document_session(self): + @override_settings(MEETECHO_API_CONFIG="fake settings") + @mock.patch("ietf.doc.views_doc.SlidesManager") + def test_add_document_session(self, mock_slides_manager_cls): doc = IndividualDraftFactory.create() url = urlreverse('ietf.doc.views_doc.add_sessionpresentation',kwargs=dict(name=doc.name)) login_testing_unauthorized(self,self.group_chair.user.username,url) response = self.client.get(url) self.assertEqual(response.status_code,200) - + self.assertFalse(mock_slides_manager_cls.called) + response = self.client.post(url,{'session':0,'version':'current'}) self.assertEqual(response.status_code,200) q=PyQuery(response.content) self.assertTrue(q('.form-select.is-invalid')) + self.assertFalse(mock_slides_manager_cls.called) response = self.client.post(url,{'session':self.future.pk,'version':'bogus version'}) self.assertEqual(response.status_code,200) q=PyQuery(response.content) self.assertTrue(q('.form-select.is-invalid')) + self.assertFalse(mock_slides_manager_cls.called) + # adding a draft self.assertEqual(1,doc.docevent_set.count()) response = self.client.post(url,{'session':self.future.pk,'version':'current'}) self.assertEqual(response.status_code,302) self.assertEqual(2,doc.docevent_set.count()) + self.assertEqual(doc.presentations.get(session__pk=self.future.pk).order, 0) + self.assertFalse(mock_slides_manager_cls.called) + + # adding slides should set order / call Meetecho API + slides = DocumentFactory(type_id="slides") + url = urlreverse("ietf.doc.views_doc.add_sessionpresentation", kwargs=dict(name=slides.name)) + response = self.client.post(url, {"session": self.future.pk, "version": "current"}) + self.assertEqual(response.status_code,302) + self.assertEqual(slides.presentations.get(session__pk=self.future.pk).order, 1) + self.assertEqual(mock_slides_manager_cls.call_args, mock.call(api_config="fake settings")) + self.assertEqual(mock_slides_manager_cls.return_value.add.call_count, 1) + self.assertEqual( + mock_slides_manager_cls.return_value.add.call_args, + mock.call(self.future, slides, order=1), + ) def test_get_related_meeting(self): """Should be able to retrieve related meeting""" @@ -2456,60 +2836,6 @@ def test_get_related_meeting(self): self.assertIsNone(doc.get_related_meeting(), f'{doc.type.slug} should not be related to meeting') class ChartTests(ResourceTestCaseMixin, TestCase): - def test_search_chart_conf(self): - doc = IndividualDraftFactory() - - conf_url = urlreverse('ietf.doc.views_stats.chart_conf_newrevisiondocevent') - - # No qurey arguments; expect an empty json object - r = self.client.get(conf_url) - self.assertValidJSONResponse(r) - self.assertEqual(unicontent(r), '{}') - - # No match - r = self.client.get(conf_url + '?activedrafts=on&name=thisisnotadocumentname') - self.assertValidJSONResponse(r) - d = r.json() - self.assertEqual(d['chart']['type'], settings.CHART_TYPE_COLUMN_OPTIONS['chart']['type']) - - r = self.client.get(conf_url + '?activedrafts=on&name=%s'%doc.name[6:12]) - self.assertValidJSONResponse(r) - d = r.json() - self.assertEqual(d['chart']['type'], settings.CHART_TYPE_COLUMN_OPTIONS['chart']['type']) - self.assertEqual(len(d['series'][0]['data']), 0) - - def test_search_chart_data(self): - doc = IndividualDraftFactory() - - data_url = urlreverse('ietf.doc.views_stats.chart_data_newrevisiondocevent') - - # No qurey arguments; expect an empty json list - r = self.client.get(data_url) - self.assertValidJSONResponse(r) - self.assertEqual(unicontent(r), '[]') - - # No match - r = self.client.get(data_url + '?activedrafts=on&name=thisisnotadocumentname') - self.assertValidJSONResponse(r) - d = r.json() - self.assertEqual(unicontent(r), '[]') - - r = self.client.get(data_url + '?activedrafts=on&name=%s'%doc.name[6:12]) - self.assertValidJSONResponse(r) - d = r.json() - self.assertEqual(len(d), 1) - self.assertEqual(len(d[0]), 2) - - def test_search_chart(self): - doc = IndividualDraftFactory() - - chart_url = urlreverse('ietf.doc.views_stats.chart_newrevisiondocevent') - r = self.client.get(chart_url) - self.assertEqual(r.status_code, 200) - - r = self.client.get(chart_url + '?activedrafts=on&name=%s'%doc.name[6:12]) - self.assertEqual(r.status_code, 200) - def test_personal_chart(self): person = PersonFactory.create() IndividualDraftFactory.create( @@ -2522,7 +2848,7 @@ def test_personal_chart(self): self.assertValidJSONResponse(r) d = r.json() self.assertEqual(d['chart']['type'], settings.CHART_TYPE_COLUMN_OPTIONS['chart']['type']) - self.assertEqual("New draft revisions over time for %s" % person.name, d['title']['text']) + self.assertEqual("New Internet-Draft revisions over time for %s" % person.name, d['title']['text']) data_url = urlreverse('ietf.doc.views_stats.chart_data_person_drafts', kwargs=dict(id=person.id)) @@ -2531,6 +2857,7 @@ def test_personal_chart(self): d = r.json() self.assertEqual(len(d), 1) self.assertEqual(len(d[0]), 2) + self.assertEqual(d[0][1], 1) page_url = urlreverse('ietf.person.views.profile', kwargs=dict(email_or_name=person.name)) r = self.client.get(page_url) @@ -2615,247 +2942,63 @@ def test_markdown_and_text(self): class Idnits2SupportTests(TestCase): settings_temp_path_overrides = TestCase.settings_temp_path_overrides + ['DERIVED_DIR'] - def test_obsoleted(self): - rfc = WgRfcFactory(alias2__name='rfc1001') - WgRfcFactory(alias2__name='rfc1003',relations=[('obs',rfc)]) - rfc = WgRfcFactory(alias2__name='rfc1005') - WgRfcFactory(alias2__name='rfc1007',relations=[('obs',rfc)]) + def test_generate_idnits2_rfcs_obsoleted(self): + rfc = WgRfcFactory(rfc_number=1001) + WgRfcFactory(rfc_number=1003,relations=[('obs',rfc)]) + rfc = WgRfcFactory(rfc_number=1005) + WgRfcFactory(rfc_number=1007,relations=[('obs',rfc)]) + blob = generate_idnits2_rfcs_obsoleted() + self.assertEqual(blob, b'1001 1003\n1005 1007\n'.decode("utf8")) + def test_obsoleted(self): url = urlreverse('ietf.doc.views_doc.idnits2_rfcs_obsoleted') r = self.client.get(url) self.assertEqual(r.status_code, 404) - call_command('generate_idnits2_rfcs_obsoleted') + # value written is arbitrary, expect it to be passed through + (Path(settings.DERIVED_DIR) / "idnits2-rfcs-obsoleted").write_bytes(b'1001 1003\n1005 1007\n') url = urlreverse('ietf.doc.views_doc.idnits2_rfcs_obsoleted') r = self.client.get(url) self.assertEqual(r.status_code, 200) self.assertEqual(r.content, b'1001 1003\n1005 1007\n') - def test_rfc_status(self): + def test_generate_idnits2_rfc_status(self): for slug in ('bcp', 'ds', 'exp', 'hist', 'inf', 'std', 'ps', 'unkn'): WgRfcFactory(std_level_id=slug) + blob = generate_idnits2_rfc_status().replace("\n", "") + self.assertEqual(blob[6312-1], "O") + + def test_rfc_status(self): url = urlreverse('ietf.doc.views_doc.idnits2_rfc_status') r = self.client.get(url) self.assertEqual(r.status_code,404) - call_command('generate_idnits2_rfc_status') + # value written is arbitrary, expect it to be passed through + (Path(settings.DERIVED_DIR) / "idnits2-rfc-status").write_bytes(b'1001 1003\n1005 1007\n') r = self.client.get(url) self.assertEqual(r.status_code,200) - blob = unicontent(r).replace('\n','') - self.assertEqual(blob[6312-1],'O') + self.assertEqual(r.content, b'1001 1003\n1005 1007\n') def test_idnits2_state(self): rfc = WgRfcFactory() - url = urlreverse('ietf.doc.views_doc.idnits2_state', kwargs=dict(name=rfc.canonical_name())) + draft = WgDraftFactory() + draft.relateddocument_set.create(relationship_id="became_rfc", target=rfc) + url = urlreverse('ietf.doc.views_doc.idnits2_state', kwargs=dict(name=rfc.name)) r = self.client.get(url) self.assertEqual(r.status_code, 200) self.assertContains(r,'rfcnum') draft = WgDraftFactory() - url = urlreverse('ietf.doc.views_doc.idnits2_state', kwargs=dict(name=draft.canonical_name())) + url = urlreverse('ietf.doc.views_doc.idnits2_state', kwargs=dict(name=draft.name)) r = self.client.get(url) self.assertEqual(r.status_code, 200) self.assertNotContains(r,'rfcnum') self.assertContains(r,'Unknown') draft = WgDraftFactory(intended_std_level_id='ps') - url = urlreverse('ietf.doc.views_doc.idnits2_state', kwargs=dict(name=draft.canonical_name())) + url = urlreverse('ietf.doc.views_doc.idnits2_state', kwargs=dict(name=draft.name)) r = self.client.get(url) self.assertEqual(r.status_code, 200) self.assertContains(r,'Proposed') -class RfcdiffSupportTests(TestCase): - - def setUp(self): - super().setUp() - self.target_view = 'ietf.doc.views_doc.rfcdiff_latest_json' - self._last_rfc_num = 8000 - - def getJson(self, view_args): - url = urlreverse(self.target_view, kwargs=view_args) - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - return r.json() - - def next_rfc_number(self): - self._last_rfc_num += 1 - return self._last_rfc_num - - def do_draft_test(self, name): - draft = IndividualDraftFactory(name=name, rev='00', create_revisions=range(0,13)) - draft = reload_db_objects(draft) - - received = self.getJson(dict(name=draft.name)) - self.assertEqual( - received, - dict( - name=draft.name, - rev=draft.rev, - content_url=draft.get_href(), - previous=f'{draft.name}-{(int(draft.rev)-1):02d}' - ), - 'Incorrect JSON when draft revision not specified', - ) - - received = self.getJson(dict(name=draft.name, rev=draft.rev)) - self.assertEqual( - received, - dict( - name=draft.name, - rev=draft.rev, - content_url=draft.get_href(), - previous=f'{draft.name}-{(int(draft.rev)-1):02d}' - ), - 'Incorrect JSON when latest revision specified', - ) - - received = self.getJson(dict(name=draft.name, rev='10')) - self.assertEqual( - received, - dict( - name=draft.name, - rev='10', - content_url=draft.history_set.get(rev='10').get_href(), - previous=f'{draft.name}-09' - ), - 'Incorrect JSON when historical revision specified', - ) - - received = self.getJson(dict(name=draft.name, rev='00')) - self.assertNotIn('previous', received, 'Rev 00 has no previous name when not replacing a draft') - - replaced = IndividualDraftFactory() - RelatedDocument.objects.create(relationship_id='replaces',source=draft,target=replaced.docalias.first()) - received = self.getJson(dict(name=draft.name, rev='00')) - self.assertEqual(received['previous'], f'{replaced.name}-{replaced.rev}', - 'Rev 00 has a previous name when replacing a draft') - - def test_draft(self): - # test with typical, straightforward names - self.do_draft_test(name='draft-somebody-did-a-thing') - # try with different potentially problematic names - self.do_draft_test(name='draft-someone-did-something-01-02') - self.do_draft_test(name='draft-someone-did-something-else-02') - self.do_draft_test(name='draft-someone-did-something-02-weird-01') - - def do_draft_with_broken_history_test(self, name): - draft = IndividualDraftFactory(name=name, rev='10') - received = self.getJson(dict(name=draft.name,rev='09')) - self.assertEqual(received['rev'],'09') - self.assertEqual(received['previous'], f'{draft.name}-08') - self.assertTrue('warning' in received) - - def test_draft_with_broken_history(self): - # test with typical, straightforward names - self.do_draft_with_broken_history_test(name='draft-somebody-did-something') - # try with different potentially problematic names - self.do_draft_with_broken_history_test(name='draft-someone-did-something-01-02') - self.do_draft_with_broken_history_test(name='draft-someone-did-something-else-02') - self.do_draft_with_broken_history_test(name='draft-someone-did-something-02-weird-03') - - def do_rfc_test(self, draft_name): - draft = WgDraftFactory(name=draft_name, create_revisions=range(0,2)) - draft.docalias.create(name=f'rfc{self.next_rfc_number():04}') - draft.set_state(State.objects.get(type_id='draft',slug='rfc')) - draft.set_state(State.objects.get(type_id='draft-iesg', slug='pub')) - draft = reload_db_objects(draft) - rfc = draft - - number = rfc.rfc_number() - received = self.getJson(dict(name=number)) - self.assertEqual( - received, - dict( - content_url=rfc.get_href(), - name=rfc.canonical_name(), - previous=f'{draft.name}-{draft.rev}', - ), - 'Can look up an RFC by number', - ) - - num_received = received - received = self.getJson(dict(name=rfc.canonical_name())) - self.assertEqual(num_received, received, 'RFC by canonical name gives same result as by number') - - received = self.getJson(dict(name=f'RfC {number}')) - self.assertEqual(num_received, received, 'RFC with unusual spacing/caps gives same result as by number') - - received = self.getJson(dict(name=draft.name)) - self.assertEqual(num_received, received, 'RFC by draft name and no rev gives same result as by number') - - received = self.getJson(dict(name=draft.name, rev='01')) - self.assertEqual( - received, - dict( - content_url=draft.history_set.get(rev='01').get_href(), - name=draft.name, - rev='01', - previous=f'{draft.name}-00', - ), - 'RFC by draft name with rev should give draft name, not canonical name' - ) - - def test_rfc(self): - # simple draft name - self.do_rfc_test(draft_name='draft-test-ar-ef-see') - # tricky draft names - self.do_rfc_test(draft_name='draft-whatever-02') - self.do_rfc_test(draft_name='draft-test-me-03-04') - - def test_rfc_with_tombstone(self): - draft = WgDraftFactory(create_revisions=range(0,2)) - draft.docalias.create(name='rfc3261') # See views_doc.HAS_TOMBSTONE - draft.set_state(State.objects.get(type_id='draft',slug='rfc')) - draft.set_state(State.objects.get(type_id='draft-iesg', slug='pub')) - draft = reload_db_objects(draft) - rfc = draft - - # Some old rfcs had tombstones that shouldn't be used for comparisons - received = self.getJson(dict(name=rfc.canonical_name())) - self.assertTrue(received['previous'].endswith('00')) - - def do_rfc_with_broken_history_test(self, draft_name): - draft = WgDraftFactory(rev='10', name=draft_name) - draft.docalias.create(name=f'rfc{self.next_rfc_number():04}') - draft.set_state(State.objects.get(type_id='draft',slug='rfc')) - draft.set_state(State.objects.get(type_id='draft-iesg', slug='pub')) - draft = reload_db_objects(draft) - rfc = draft - - received = self.getJson(dict(name=draft.name)) - self.assertEqual( - received, - dict( - content_url=rfc.get_href(), - name=rfc.canonical_name(), - previous=f'{draft.name}-10', - ), - 'RFC by draft name without rev should return canonical RFC name and no rev', - ) - - received = self.getJson(dict(name=draft.name, rev='10')) - self.assertEqual(received['name'], draft.name, 'RFC by draft name with rev should return draft name') - self.assertEqual(received['rev'], '10', 'Requested rev should be returned') - self.assertEqual(received['previous'], f'{draft.name}-09', 'Previous rev is one less than requested') - self.assertIn(f'{draft.name}-10', received['content_url'], 'Returned URL should include requested rev') - self.assertNotIn('warning', received, 'No warning when we have the rev requested') - - received = self.getJson(dict(name=f'{draft.name}-09')) - self.assertEqual(received['name'], draft.name, 'RFC by draft name with rev should return draft name') - self.assertEqual(received['rev'], '09', 'Requested rev should be returned') - self.assertEqual(received['previous'], f'{draft.name}-08', 'Previous rev is one less than requested') - self.assertIn(f'{draft.name}-09', received['content_url'], 'Returned URL should include requested rev') - self.assertEqual( - received['warning'], - 'History for this version not found - these results are speculation', - 'Warning should be issued when requested rev is not found' - ) - - def test_rfc_with_broken_history(self): - # simple draft name - self.do_rfc_with_broken_history_test(draft_name='draft-some-draft') - # tricky draft names - self.do_rfc_with_broken_history_test(draft_name='draft-gizmo-01') - self.do_rfc_with_broken_history_test(draft_name='draft-oh-boy-what-a-draft-02-03') - class RawIdTests(TestCase): @@ -2896,16 +3039,12 @@ def test_raw_id(self): self.should_succeed(dict(name=draft.name, rev='00',ext='txt')) self.should_404(dict(name=draft.name, rev='00',ext='html')) - def test_raw_id_rfc(self): - rfc = WgRfcFactory() - dir = settings.INTERNET_ALL_DRAFTS_ARCHIVE_DIR - (Path(dir) / f'{rfc.name}-{rfc.rev}.txt').touch() - self.should_succeed(dict(name=rfc.name)) - self.should_404(dict(name=rfc.canonical_name())) + # test_raw_id_rfc intentionally removed + # an rfc is no longer a pseudo-version of a draft. def test_non_draft(self): - charter = CharterFactory() - self.should_404(dict(name=charter.name)) + for doc in [CharterFactory(), WgRfcFactory()]: + self.should_404(dict(name=doc.name)) class PdfizedTests(TestCase): @@ -2917,31 +3056,47 @@ def should_succeed(self, argdict): url = urlreverse(self.view, kwargs=argdict) r = self.client.get(url) self.assertEqual(r.status_code,200) - self.assertEqual(r.get('Content-Type'),'application/pdf;charset=utf-8') + self.assertEqual(r.get('Content-Type'),'application/pdf') def should_404(self, argdict): url = urlreverse(self.view, kwargs=argdict) r = self.client.get(url) self.assertEqual(r.status_code, 404) + # This takes a _long_ time (32s on a 2022 m1 macbook pro) - is it worth what it covers? def test_pdfized(self): - rfc = WgRfcFactory(create_revisions=range(0,2)) + rfc = WgRfcFactory() + draft = WgDraftFactory(create_revisions=range(0,2)) + draft.relateddocument_set.create(relationship_id="became_rfc", target=rfc) dir = settings.RFC_PATH - with (Path(dir) / f'{rfc.canonical_name()}.txt').open('w') as f: + with (Path(dir) / f'{rfc.name}.txt').open('w') as f: f.write('text content') dir = settings.INTERNET_ALL_DRAFTS_ARCHIVE_DIR for r in range(0,2): - with (Path(dir) / f'{rfc.name}-{r:02d}.txt').open('w') as f: + with (Path(dir) / f'{draft.name}-{r:02d}.txt').open('w') as f: f.write('text content') - self.should_succeed(dict(name=rfc.canonical_name())) + self.assertTrue( + login_testing_unauthorized( + self, + PersonFactory().user.username, + urlreverse(self.view, kwargs={"name": draft.name}), + ) + ) self.should_succeed(dict(name=rfc.name)) + self.should_succeed(dict(name=draft.name)) for r in range(0,2): - self.should_succeed(dict(name=rfc.name,rev=f'{r:02d}')) + self.should_succeed(dict(name=draft.name,rev=f'{r:02d}')) for ext in ('pdf','txt','html','anythingatall'): - self.should_succeed(dict(name=rfc.name,rev=f'{r:02d}',ext=ext)) - self.should_404(dict(name=rfc.name,rev='02')) + self.should_succeed(dict(name=draft.name,rev=f'{r:02d}',ext=ext)) + self.should_404(dict(name=draft.name,rev='02')) + + with mock.patch('ietf.doc.models.DocumentInfo.pdfized', side_effect=URLFetchingError): + url = urlreverse(self.view, kwargs=dict(name=rfc.name)) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + self.assertContains(r, "Error while rendering PDF") class NotifyValidationTests(TestCase): def test_notify_validation(self): @@ -2982,3 +3137,425 @@ def test_notify_validation(self): self.assertFalse(f.is_valid()) self.assertTrue("Invalid addresses" in f.errors["notify"][0]) self.assertTrue("Duplicate addresses" in f.errors["notify"][0]) + +class CanRequestConflictReviewTests(TestCase): + def test_gets_request_conflict_review_action_button(self): + ise_draft = IndividualDraftFactory(stream_id="ise") + irtf_draft = RgDraftFactory() + + # This is blunt, trading off precision for time. A more thorough test would ensure + # that the text is in a button and that the correct link is absent/present as well. + + target_string = "Begin IETF conflict review" + + url = urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=irtf_draft.name)) + r = self.client.get(url) + self.assertNotContains(r, target_string) + self.client.login(username="secretary", password="secretary+password") + r = self.client.get(url) + self.assertContains(r, target_string) + self.client.logout() + self.client.login(username="irtf-chair", password="irtf-chair+password") + r = self.client.get(url) + self.assertContains(r, target_string) + self.client.logout() + self.client.login(username="ise-chair", password="ise-chair+password") + r = self.client.get(url) + self.assertNotContains(r, target_string) + self.client.logout() + + url = urlreverse("ietf.doc.views_doc.document_main", kwargs=dict(name=ise_draft.name)) + r = self.client.get(url) + self.assertNotContains(r, target_string) + self.client.login(username="secretary", password="secretary+password") + r = self.client.get(url) + self.assertContains(r, target_string) + self.client.logout() + self.client.login(username="irtf-chair", password="irtf-chair+password") + r = self.client.get(url) + self.assertNotContains(r, target_string) + self.client.logout() + self.client.login(username="ise-chair", password="ise-chair+password") + r = self.client.get(url) + self.assertContains(r, target_string) + +class DocInfoMethodsTests(TestCase): + + def test_became_rfc(self): + draft = WgDraftFactory() + rfc = WgRfcFactory() + draft.relateddocument_set.create(relationship_id="became_rfc",target=rfc) + self.assertEqual(draft.became_rfc(), rfc) + self.assertEqual(rfc.came_from_draft(), draft) + + charter = CharterFactory() + self.assertIsNone(charter.became_rfc()) + self.assertIsNone(charter.came_from_draft()) + + def test_revisions(self): + draft = WgDraftFactory(rev="09",create_revisions=range(0,10)) + self.assertEqual(draft.revisions_by_dochistory(),[f"{i:02d}" for i in range(0,10)]) + self.assertEqual(draft.revisions_by_newrevisionevent(),[f"{i:02d}" for i in range(0,10)]) + rfc = WgRfcFactory() + self.assertEqual(rfc.revisions_by_newrevisionevent(),[]) + self.assertEqual(rfc.revisions_by_dochistory(),[]) + + draft.history_set.filter(rev__lt="08").delete() + draft.docevent_set.filter(newrevisiondocevent__rev="05").delete() + self.assertEqual(draft.revisions_by_dochistory(),[f"{i:02d}" for i in range(8,10)]) + self.assertEqual(draft.revisions_by_newrevisionevent(),[f"{i:02d}" for i in [*range(0,5), *range(6,10)]]) + + def test_referenced_by_rfcs(self): + # n.b., no significance to the ref* values in this test + referring_draft = WgDraftFactory() + (rfc, referring_rfc) = WgRfcFactory.create_batch(2) + rfc.targets_related.create(relationship_id="refnorm", source=referring_draft) + rfc.targets_related.create(relationship_id="refnorm", source=referring_rfc) + self.assertCountEqual( + rfc.referenced_by_rfcs(), + rfc.targets_related.filter(source=referring_rfc), + ) + + def test_referenced_by_rfcs_as_rfc_or_draft(self): + # n.b., no significance to the ref* values in this test + draft = WgDraftFactory() + rfc = WgRfcFactory() + draft.relateddocument_set.create(relationship_id="became_rfc", target=rfc) + + # Draft referring to the rfc and the draft - should not be reported at all + draft_referring_to_both = WgDraftFactory() + draft_referring_to_both.relateddocument_set.create(relationship_id="refnorm", target=draft) + draft_referring_to_both.relateddocument_set.create(relationship_id="refnorm", target=rfc) + + # RFC referring only to the draft - should be reported for either the draft or the rfc + rfc_referring_to_draft = WgRfcFactory() + rfc_referring_to_draft.relateddocument_set.create(relationship_id="refinfo", target=draft) + + # RFC referring only to the rfc - should be reported only for the rfc + rfc_referring_to_rfc = WgRfcFactory() + rfc_referring_to_rfc.relateddocument_set.create(relationship_id="refinfo", target=rfc) + + # RFC referring only to the rfc - should be reported only for the rfc + rfc_referring_to_rfc = WgRfcFactory() + rfc_referring_to_rfc.relateddocument_set.create(relationship_id="refinfo", target=rfc) + + # RFC referring to the rfc and the draft - should be reported for both + rfc_referring_to_both = WgRfcFactory() + rfc_referring_to_both.relateddocument_set.create(relationship_id="refnorm", target=draft) + rfc_referring_to_both.relateddocument_set.create(relationship_id="refnorm", target=rfc) + + self.assertCountEqual( + draft.referenced_by_rfcs_as_rfc_or_draft(), + draft.targets_related.filter(source__type="rfc"), + ) + + self.assertCountEqual( + rfc.referenced_by_rfcs_as_rfc_or_draft(), + draft.targets_related.filter(source__type="rfc") | rfc.targets_related.filter(source__type="rfc"), + ) + +class StateIndexTests(TestCase): + + def test_state_index(self): + url = urlreverse('ietf.doc.views_help.state_index') + r = self.client.get(url) + q = PyQuery(r.content) + content = [ e.text for e in q('#content table td a ') ] + names = StateType.objects.values_list('slug', flat=True) + # The following doesn't cover all doc types, only a selection + for name in names: + if not '-' in name: + self.assertIn(name, content) + +class InvestigateTests(TestCase): + settings_temp_path_overrides = TestCase.settings_temp_path_overrides + [ + "AGENDA_PATH", + # "INTERNET_DRAFT_PATH", + # "INTERNET_DRAFT_ARCHIVE_DIR", + # "INTERNET_ALL_DRAFTS_ARCHIVE_DIR", + ] + + def setUp(self): + super().setUp() + # Contort the draft archive dir temporary replacement + # to match the "collections" concept + archive_tmp_dir = Path(settings.INTERNET_DRAFT_ARCHIVE_DIR) + new_archive_dir = archive_tmp_dir / "draft-archive" + new_archive_dir.mkdir() + settings.INTERNET_DRAFT_ARCHIVE_DIR = str(new_archive_dir) + donated_personal_copy_dir = archive_tmp_dir / "donated-personal-copy" + donated_personal_copy_dir.mkdir() + meeting_dir = Path(settings.AGENDA_PATH) / "666" + meeting_dir.mkdir() + all_archive_dir = Path(settings.INTERNET_ALL_DRAFTS_ARCHIVE_DIR) + repository_dir = Path(settings.INTERNET_DRAFT_PATH) + + for path in [repository_dir, all_archive_dir]: + (path / "draft-this-is-active-00.txt").touch() + for path in [new_archive_dir, all_archive_dir]: + (path / "draft-old-but-can-authenticate-00.txt").touch() + (path / "draft-has-mixed-provenance-01.txt").touch() + for path in [donated_personal_copy_dir, all_archive_dir]: + (path / "draft-donated-from-a-personal-collection-00.txt").touch() + (path / "draft-has-mixed-provenance-00.txt").touch() + (path / "draft-has-mixed-provenance-00.txt.Z").touch() + (all_archive_dir / "draft-this-should-not-be-possible-00.txt").touch() + (meeting_dir / "draft-this-predates-the-archive-00.txt").touch() + + def test_investigate_fragment(self): + + result = investigate_fragment("this-is-active") + self.assertEqual(len(result["can_verify"]), 1) + self.assertEqual(len(result["unverifiable_collections"]), 0) + self.assertEqual(len(result["unexpected"]), 0) + self.assertEqual( + list(result["can_verify"])[0].name, "draft-this-is-active-00.txt" + ) + + result = investigate_fragment("old-but-can") + self.assertEqual(len(result["can_verify"]), 1) + self.assertEqual(len(result["unverifiable_collections"]), 0) + self.assertEqual(len(result["unexpected"]), 0) + self.assertEqual( + list(result["can_verify"])[0].name, "draft-old-but-can-authenticate-00.txt" + ) + + result = investigate_fragment("predates") + self.assertEqual(len(result["can_verify"]), 1) + self.assertEqual(len(result["unverifiable_collections"]), 0) + self.assertEqual(len(result["unexpected"]), 0) + self.assertEqual( + list(result["can_verify"])[0].name, "draft-this-predates-the-archive-00.txt" + ) + + result = investigate_fragment("personal-collection") + self.assertEqual(len(result["can_verify"]), 0) + self.assertEqual(len(result["unverifiable_collections"]), 1) + self.assertEqual(len(result["unexpected"]), 0) + self.assertEqual( + list(result["unverifiable_collections"])[0].name, + "draft-donated-from-a-personal-collection-00.txt", + ) + + result = investigate_fragment("mixed-provenance") + self.assertEqual(len(result["can_verify"]), 1) + self.assertEqual(len(result["unverifiable_collections"]), 2) + self.assertEqual(len(result["unexpected"]), 0) + self.assertEqual( + list(result["can_verify"])[0].name, "draft-has-mixed-provenance-01.txt" + ) + self.assertEqual( + set([p.name for p in result["unverifiable_collections"]]), + set( + [ + "draft-has-mixed-provenance-00.txt", + "draft-has-mixed-provenance-00.txt.Z", + ] + ), + ) + + result = investigate_fragment("not-be-possible") + self.assertEqual(len(result["can_verify"]), 0) + self.assertEqual(len(result["unverifiable_collections"]), 0) + self.assertEqual(len(result["unexpected"]), 1) + self.assertEqual( + list(result["unexpected"])[0].name, + "draft-this-should-not-be-possible-00.txt", + ) + + @mock.patch("ietf.doc.utils.caches") + def test_investigate_fragment_cache(self, mock_caches): + """investigate_fragment should cache its result""" + mock_default_cache = mock_caches["default"] + mock_default_cache.get.return_value = None # disable cache + result = investigate_fragment("this-is-active") + self.assertEqual(len(result["can_verify"]), 1) + self.assertEqual(len(result["unverifiable_collections"]), 0) + self.assertEqual(len(result["unexpected"]), 0) + self.assertEqual( + list(result["can_verify"])[0].name, "draft-this-is-active-00.txt" + ) + self.assertTrue(mock_default_cache.get.called) + self.assertTrue(mock_default_cache.set.called) + expected_key = f"investigate_fragment:{sha384(b'this-is-active').hexdigest()}" + self.assertEqual(mock_default_cache.set.call_args.kwargs["key"], expected_key) + cached_value = mock_default_cache.set.call_args.kwargs["value"] # hang on to this + mock_default_cache.reset_mock() + + # Check that a cached value is used + mock_default_cache.get.return_value = cached_value + with mock.patch("ietf.doc.utils.Path") as mock_path: + result = investigate_fragment("this-is-active") + # Check that we got the same results + self.assertEqual(len(result["can_verify"]), 1) + self.assertEqual(len(result["unverifiable_collections"]), 0) + self.assertEqual(len(result["unexpected"]), 0) + self.assertEqual( + list(result["can_verify"])[0].name, "draft-this-is-active-00.txt" + ) + # And that we used the cache + self.assertFalse(mock_path.called) # a proxy for "did the method do any real work" + self.assertTrue(mock_default_cache.get.called) + self.assertEqual(mock_default_cache.get.call_args, mock.call(expected_key)) + + def test_investigate_get(self): + """GET with no querystring should retrieve the investigate UI""" + url = urlreverse("ietf.doc.views_doc.investigate") + login_testing_unauthorized(self, "secretary", url) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q("form#investigate")), 1) + self.assertEqual(len(q("div#results")), 0) + + @mock.patch("ietf.doc.views_doc.AsyncResult") + def test_investgate_get_task_id(self, mock_asyncresult): + """GET with querystring should lookup task status""" + url = urlreverse("ietf.doc.views_doc.investigate") + login_testing_unauthorized(self, "secretary", url) + mock_asyncresult.return_value.ready.return_value = True + r = self.client.get(url + "?id=a-task-id") + self.assertEqual(r.status_code, 200) + self.assertEqual(r.json(), {"status": "ready"}) + self.assertTrue(mock_asyncresult.called) + self.assertEqual(mock_asyncresult.call_args, mock.call("a-task-id")) + mock_asyncresult.reset_mock() + + mock_asyncresult.return_value.ready.return_value = False + r = self.client.get(url + "?id=a-task-id") + self.assertEqual(r.status_code, 200) + self.assertEqual(r.json(), {"status": "notready"}) + self.assertTrue(mock_asyncresult.called) + self.assertEqual(mock_asyncresult.call_args, mock.call("a-task-id")) + + @mock.patch("ietf.doc.views_doc.investigate_fragment_task") + def test_investigate_post(self, mock_investigate_fragment_task): + """POST with a name_fragment and no task_id should start a celery task""" + url = urlreverse("ietf.doc.views_doc.investigate") + login_testing_unauthorized(self, "secretary", url) + + # test some invalid cases + r = self.client.post(url, {"name_fragment": "short"}) # limit is >= 8 characters + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q("#id_name_fragment.is-invalid")), 1) + self.assertFalse(mock_investigate_fragment_task.delay.called) + for char in ["*", "%", "/", "\\"]: + r = self.client.post(url, {"name_fragment": f"bad{char}character"}) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q("#id_name_fragment.is-invalid")), 1) + self.assertFalse(mock_investigate_fragment_task.delay.called) + + # now a valid one + mock_investigate_fragment_task.delay.return_value.id = "a-task-id" + r = self.client.post(url, {"name_fragment": "this-is-a-valid-fragment"}) + self.assertEqual(r.status_code, 200) + self.assertTrue(mock_investigate_fragment_task.delay.called) + self.assertEqual(mock_investigate_fragment_task.delay.call_args, mock.call("this-is-a-valid-fragment")) + self.assertEqual(r.json(), {"id": "a-task-id"}) + + @mock.patch("ietf.doc.views_doc.AsyncResult") + def test_investigate_post_task_id(self, mock_asyncresult): + """POST with name_fragment and task_id should retrieve results""" + url = urlreverse("ietf.doc.views_doc.investigate") + login_testing_unauthorized(self, "secretary", url) + + # First, test a non-successful result - this could be a failure or non-existent task id + mock_result = mock_asyncresult.return_value + mock_result.successful.return_value = False + r = self.client.post(url, {"name_fragment": "some-fragment", "task_id": "a-task-id"}) + self.assertContains(r, "The investigation task failed.", status_code=200) + self.assertTrue(mock_asyncresult.called) + self.assertEqual(mock_asyncresult.call_args, mock.call("a-task-id")) + self.assertFalse(mock_result.get.called) + mock_asyncresult.reset_mock() + q = PyQuery(r.content) + self.assertEqual(q("#id_name_fragment").val(), "some-fragment") + self.assertEqual(q("#id_task_id").val(), "a-task-id") + + # now the various successful result mixes + mock_result = mock_asyncresult.return_value + mock_result.successful.return_value = True + mock_result.get.return_value = { + "name_fragment": "different-fragment", + "results": { + "can_verify": set(), + "unverifiable_collections": set(), + "unexpected": set(), + } + } + r = self.client.post(url, {"name_fragment": "some-fragment", "task_id": "a-task-id"}) + self.assertEqual(r.status_code, 200) + self.assertTrue(mock_asyncresult.called) + self.assertEqual(mock_asyncresult.call_args, mock.call("a-task-id")) + mock_asyncresult.reset_mock() + q = PyQuery(r.content) + self.assertEqual(q("#id_name_fragment").val(), "different-fragment", "name_fragment should be reset") + self.assertEqual(q("#id_task_id").val(), "", "task_id should be cleared") + self.assertEqual(len(q("div#results")), 1) + self.assertEqual(len(q("table#authenticated")), 0) + self.assertEqual(len(q("table#unverifiable")), 0) + self.assertEqual(len(q("table#unexpected")), 0) + + # This file was created in setUp. It allows the view to render properly + # but its location / content don't matter for this test otherwise. + a_file_that_exists = Path(settings.INTERNET_DRAFT_PATH) / "draft-this-is-active-00.txt" + + mock_result.get.return_value = { + "name_fragment": "different-fragment", + "results": { + "can_verify": {a_file_that_exists}, + "unverifiable_collections": {a_file_that_exists}, + "unexpected": set(), + } + } + r = self.client.post(url, {"name_fragment": "some-fragment", "task_id": "a-task-id"}) + self.assertEqual(r.status_code, 200) + self.assertTrue(mock_asyncresult.called) + self.assertEqual(mock_asyncresult.call_args, mock.call("a-task-id")) + mock_asyncresult.reset_mock() + q = PyQuery(r.content) + self.assertEqual(q("#id_name_fragment").val(), "different-fragment", "name_fragment should be reset") + self.assertEqual(q("#id_task_id").val(), "", "task_id should be cleared") + self.assertEqual(len(q("div#results")), 1) + self.assertEqual(len(q("table#authenticated")), 1) + self.assertEqual(len(q("table#unverifiable")), 1) + self.assertEqual(len(q("table#unexpected")), 0) + + mock_result.get.return_value = { + "name_fragment": "different-fragment", + "results": { + "can_verify": set(), + "unverifiable_collections": set(), + "unexpected": {a_file_that_exists}, + } + } + r = self.client.post(url, {"name_fragment": "some-fragment", "task_id": "a-task-id"}) + self.assertEqual(r.status_code, 200) + self.assertTrue(mock_asyncresult.called) + self.assertEqual(mock_asyncresult.call_args, mock.call("a-task-id")) + mock_asyncresult.reset_mock() + q = PyQuery(r.content) + self.assertEqual(q("#id_name_fragment").val(), "different-fragment", "name_fragment should be reset") + self.assertEqual(q("#id_task_id").val(), "", "task_id should be cleared") + self.assertEqual(len(q("div#results")), 1) + self.assertEqual(len(q("table#authenticated")), 0) + self.assertEqual(len(q("table#unverifiable")), 0) + self.assertEqual(len(q("table#unexpected")), 1) + + +class LogIOErrorTests(TestCase): + + def test_doc_text_io_error(self): + + d = IndividualDraftFactory() + + with mock.patch("ietf.doc.models.Path") as path_cls_mock: + with mock.patch("ietf.doc.models.log.log") as log_mock: + path_cls_mock.return_value.exists.return_value = True + path_cls_mock.return_value.open.return_value.__enter__.return_value.read.side_effect = IOError("Bad things happened") + text = d.text() + self.assertIsNone(text) + self.assertTrue(log_mock.called) + self.assertIn("Bad things happened", log_mock.call_args[0][0]) diff --git a/ietf/doc/tests_ballot.py b/ietf/doc/tests_ballot.py index 4de4de08ba..8420e411e2 100644 --- a/ietf/doc/tests_ballot.py +++ b/ietf/doc/tests_ballot.py @@ -3,7 +3,7 @@ import datetime -import mock +from unittest import mock from pyquery import PyQuery @@ -17,27 +17,37 @@ from ietf.doc.models import (Document, State, DocEvent, BallotPositionDocEvent, LastCallDocEvent, WriteupDocEvent, TelechatDocEvent) from ietf.doc.factories import (DocumentFactory, IndividualDraftFactory, IndividualRfcFactory, WgDraftFactory, - BallotPositionDocEventFactory, BallotDocEventFactory) + BallotPositionDocEventFactory, BallotDocEventFactory, IRSGBallotDocEventFactory, RgDraftFactory) +from ietf.doc.templatetags.ietf_filters import can_defer from ietf.doc.utils import create_ballot_if_not_open +from ietf.doc.views_ballot import parse_ballot_edit_return_point from ietf.doc.views_doc import document_ballot_content from ietf.group.models import Group, Role from ietf.group.factories import GroupFactory, RoleFactory, ReviewTeamFactory from ietf.ipr.factories import HolderIprDisclosureFactory -from ietf.name.models import BallotPositionName from ietf.iesg.models import TelechatDate -from ietf.person.models import Person, PersonalApiKey -from ietf.person.factories import PersonFactory +from ietf.person.models import Person +from ietf.person.factories import PersonFactory, PersonalApiKeyFactory from ietf.person.utils import get_active_ads from ietf.utils.test_utils import TestCase, login_testing_unauthorized from ietf.utils.mail import outbox, empty_outbox, get_payload_text from ietf.utils.text import unwrap -from ietf.utils.timezone import date_today +from ietf.utils.timezone import date_today, datetime_today class EditPositionTests(TestCase): + + # N.B. This test needs to be rewritten to exercise all types of ballots (iesg, irsg, rsab) + # and test against the output of the mailtriggers instead of looking for hardcoded values + # in the To and CC results. See #7864 def test_edit_position(self): ad = Person.objects.get(user__username="ad") - draft = IndividualDraftFactory(ad=ad,stream_id='ietf') + draft = WgDraftFactory( + ad=ad, + stream_id="ietf", + notify="somebody@example.com", + group__acronym="mars", + ) ballot = create_ballot_if_not_open(None, draft, ad, 'approve') url = urlreverse('ietf.doc.views_ballot.edit_position', kwargs=dict(name=draft.name, ballot_id=ballot.pk)) @@ -53,11 +63,20 @@ def test_edit_position(self): self.assertEqual(len(q('form textarea[name=comment]')), 1) # vote + empty_outbox() events_before = draft.docevent_set.count() - - r = self.client.post(url, dict(position="discuss", - discuss=" This is a discussion test. \n ", - comment=" This is a test. \n ")) + + r = self.client.post( + url, + dict( + position="discuss", + discuss=" This is a discussion test. \n ", + comment=" This is a test. \n ", + additional_cc="test298347@example.com", + cc_choices=["doc_notify", "doc_group_chairs"], + send_mail=1, + ), + ) self.assertEqual(r.status_code, 302) pos = draft.latest_event(BallotPositionDocEvent, balloter=ad) @@ -68,6 +87,22 @@ def test_edit_position(self): self.assertTrue(pos.comment_time != None) self.assertTrue("New position" in pos.desc) self.assertEqual(draft.docevent_set.count(), events_before + 3) + self.assertEqual(len(outbox),1) + m = outbox[0] + self.assertTrue("COMMENT" in m['Subject']) + self.assertTrue("DISCUSS" in m['Subject']) + self.assertTrue(draft.name in m['Subject']) + self.assertTrue("This is a discussion test." in str(m)) + self.assertTrue("This is a test" in str(m)) + self.assertTrue("iesg@" in m['To']) + # cc_choice doc_group_chairs + self.assertTrue("mars-chairs@" in m['Cc']) + # cc_choice doc_notify + self.assertTrue("somebody@example.com" in m['Cc']) + # cc_choice doc_group_email_list was not selected + self.assertFalse(draft.group.list_email in m['Cc']) + # extra-cc + self.assertTrue("test298347@example.com" in m['Cc']) # recast vote events_before = draft.docevent_set.count() @@ -109,7 +144,7 @@ def test_api_set_position(self): create_ballot_if_not_open(None, draft, ad, 'approve') ad.user.last_login = timezone.now() ad.user.save() - apikey = PersonalApiKey.objects.create(endpoint=url, person=ad) + apikey = PersonalApiKeyFactory(endpoint=url, person=ad) # vote events_before = draft.docevent_set.count() @@ -228,61 +263,6 @@ def test_cannot_edit_position_as_pre_ad(self): r = self.client.post(url, dict(position="discuss", discuss="Test discuss text")) self.assertEqual(r.status_code, 403) - def test_send_ballot_comment(self): - ad = Person.objects.get(user__username="ad") - draft = WgDraftFactory(ad=ad,group__acronym='mars') - draft.notify = "somebody@example.com" - draft.save_with_history([DocEvent.objects.create(doc=draft, rev=draft.rev, type="changed_document", by=Person.objects.get(user__username="secretary"), desc="Test")]) - - ballot = create_ballot_if_not_open(None, draft, ad, 'approve') - - BallotPositionDocEvent.objects.create( - doc=draft, rev=draft.rev, type="changed_ballot_position", - by=ad, balloter=ad, ballot=ballot, pos=BallotPositionName.objects.get(slug="discuss"), - discuss="This draft seems to be lacking a clearer title?", - discuss_time=timezone.now(), - comment="Test!", - comment_time=timezone.now()) - - url = urlreverse('ietf.doc.views_ballot.send_ballot_comment', kwargs=dict(name=draft.name, - ballot_id=ballot.pk)) - login_testing_unauthorized(self, "ad", url) - - # normal get - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - q = PyQuery(r.content) - self.assertTrue(len(q('form input[name="extra_cc"]')) > 0) - - # send - mailbox_before = len(outbox) - - r = self.client.post(url, dict(extra_cc="test298347@example.com", cc_choices=['doc_notify','doc_group_chairs'])) - self.assertEqual(r.status_code, 302) - - self.assertEqual(len(outbox), mailbox_before + 1) - m = outbox[-1] - self.assertTrue("COMMENT" in m['Subject']) - self.assertTrue("DISCUSS" in m['Subject']) - self.assertTrue(draft.name in m['Subject']) - self.assertTrue("clearer title" in str(m)) - self.assertTrue("Test!" in str(m)) - self.assertTrue("iesg@" in m['To']) - # cc_choice doc_group_chairs - self.assertTrue("mars-chairs@" in m['Cc']) - # cc_choice doc_notify - self.assertTrue("somebody@example.com" in m['Cc']) - # cc_choice doc_group_email_list was not selected - self.assertFalse(draft.group.list_email in m['Cc']) - # extra-cc - self.assertTrue("test298347@example.com" in m['Cc']) - - r = self.client.post(url, dict(cc="")) - self.assertEqual(r.status_code, 302) - self.assertEqual(len(outbox), mailbox_before + 2) - m = outbox[-1] - self.assertTrue("iesg@" in m['To']) - self.assertFalse(m['Cc'] and draft.group.list_email in m['Cc']) class BallotWriteupsTests(TestCase): @@ -355,7 +335,7 @@ def test_request_last_call(self): self.assertTrue('aread@' in outbox[-1]['Cc']) def test_edit_ballot_writeup(self): - draft = IndividualDraftFactory(states=[('draft','active'),('draft-iesg','iesg-eva')]) + draft = IndividualDraftFactory(states=[('draft','active'),('draft-iesg','iesg-eva')], stream_id='ietf') url = urlreverse('ietf.doc.views_ballot.ballot_writeupnotes', kwargs=dict(name=draft.name)) login_testing_unauthorized(self, "secretary", url) @@ -385,8 +365,25 @@ def test_edit_ballot_writeup(self): self.assertTrue("This is a simple test" in d.latest_event(WriteupDocEvent, type="changed_ballot_writeup_text").text) self.assertTrue('iesg-eva' == d.get_state_slug('draft-iesg')) + def test_edit_ballot_writeup_unauthorized_stream(self): + # Test that accessing a document from unauthorized (irtf) stream returns a 404 error + draft = RgDraftFactory() + url = urlreverse('ietf.doc.views_ballot.ballot_writeupnotes', kwargs=dict(name=draft.name)) + login_testing_unauthorized(self, "ad", url) + + r = self.client.get(url) + self.assertEqual(r.status_code, 404) + + def test_edit_ballot_writeup_invalid_name(self): + # Test that accessing a non-existent document returns a 404 error + url = urlreverse('ietf.doc.views_ballot.ballot_writeupnotes', kwargs=dict(name="invalid_name")) + login_testing_unauthorized(self, "ad", url) + + r = self.client.get(url) + self.assertEqual(r.status_code, 404) + def test_edit_ballot_writeup_already_approved(self): - draft = IndividualDraftFactory(states=[('draft','active'),('draft-iesg','approved')]) + draft = IndividualDraftFactory(states=[('draft','active'),('draft-iesg','approved')], stream_id='ietf') url = urlreverse('ietf.doc.views_ballot.ballot_writeupnotes', kwargs=dict(name=draft.name)) login_testing_unauthorized(self, "secretary", url) @@ -460,7 +457,7 @@ def test_edit_ballot_rfceditornote(self): def test_issue_ballot(self): ad = Person.objects.get(user__username="ad") for case in ('none','past','future'): - draft = IndividualDraftFactory(ad=ad) + draft = IndividualDraftFactory(ad=ad, stream_id='ietf') if case in ('past','future'): LastCallDocEvent.objects.create( by=Person.objects.get(name='(System)'), @@ -499,7 +496,7 @@ def test_issue_ballot(self): def test_issue_ballot_auto_state_change(self): ad = Person.objects.get(user__username="ad") - draft = IndividualDraftFactory(ad=ad, states=[('draft','active'),('draft-iesg','writeupw')]) + draft = IndividualDraftFactory(ad=ad, states=[('draft','active'),('draft-iesg','writeupw')], stream_id='ietf') url = urlreverse('ietf.doc.views_ballot.ballot_writeupnotes', kwargs=dict(name=draft.name)) login_testing_unauthorized(self, "secretary", url) @@ -523,11 +520,12 @@ def test_issue_ballot_auto_state_change(self): def test_issue_ballot_warn_if_early(self): ad = Person.objects.get(user__username="ad") - draft = IndividualDraftFactory(ad=ad, states=[('draft','active'),('draft-iesg','lc')]) + draft = IndividualDraftFactory(ad=ad, states=[('draft','active'),('draft-iesg','lc')], stream_id='ietf') url = urlreverse('ietf.doc.views_ballot.ballot_writeupnotes', kwargs=dict(name=draft.name)) login_testing_unauthorized(self, "secretary", url) # expect warning about issuing a ballot before IETF Last Call is done + # No last call has yet been issued r = self.client.get(url) self.assertEqual(r.status_code, 200) q = PyQuery(r.content) @@ -535,6 +533,38 @@ def test_issue_ballot_warn_if_early(self): self.assertTrue(q('[class=text-danger]:contains("not completed IETF Last Call")')) self.assertTrue(q('[type=submit]:contains("Save")')) + # Last call exists but hasn't expired + LastCallDocEvent.objects.create( + doc=draft, + expires=datetime_today()+datetime.timedelta(days=14), + by=Person.objects.get(name="(System)") + ) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertTrue(q('[class=text-danger]:contains("not completed IETF Last Call")')) + + # Last call exists and has expired + LastCallDocEvent.objects.filter(doc=draft).update(expires=datetime_today()-datetime.timedelta(days=2)) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertFalse(q('[class=text-danger]:contains("not completed IETF Last Call")')) + + for state_slug in ["lc", "ad-eval"]: + draft.set_state(State.objects.get(type="draft-iesg",slug=state_slug)) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertTrue(q('[class=text-danger]:contains("It would be unexpected to issue a ballot while in this state.")')) + + draft.set_state(State.objects.get(type="draft-iesg",slug="writeupw")) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertFalse(q('[class=text-danger]:contains("It would be unexpected to issue a ballot while in this state.")')) + + def test_edit_approval_text(self): ad = Person.objects.get(user__username="ad") draft = WgDraftFactory(ad=ad,states=[('draft','active'),('draft-iesg','iesg-eva')],intended_std_level_id='ps',group__parent=Group.objects.get(acronym='farfut')) @@ -772,7 +802,7 @@ def test_clear_ballot(self): ballot = create_ballot_if_not_open(None, draft, ad, 'approve') old_ballot_id = ballot.id draft.set_state(State.objects.get(used=True, type="draft-iesg", slug="iesg-eva")) - url = urlreverse('ietf.doc.views_ballot.clear_ballot', kwargs=dict(name=draft.name,ballot_type_slug=draft.ballot_open('approve').ballot_type.slug)) + url = urlreverse('ietf.doc.views_ballot.clear_ballot', kwargs=dict(name=draft.name,ballot_type_slug="approve")) login_testing_unauthorized(self, "secretary", url) r = self.client.get(url) self.assertEqual(r.status_code, 200) @@ -782,6 +812,11 @@ def test_clear_ballot(self): self.assertIsNotNone(ballot) self.assertEqual(ballot.ballotpositiondocevent_set.count(),0) self.assertNotEqual(old_ballot_id, ballot.id) + # It's not valid to clear a ballot of a type where there's no matching state + url = urlreverse('ietf.doc.views_ballot.clear_ballot', kwargs=dict(name=draft.name,ballot_type_slug="statchg")) + r = self.client.post(url,{}) + self.assertEqual(r.status_code, 404) + def test_ballot_downref_approve(self): ad = Person.objects.get(name="Areað Irector") @@ -802,8 +837,8 @@ def test_ballot_downref_approve(self): desc='Last call announcement was changed', text='this is simple last call text.' ) rfc = IndividualRfcFactory.create( + name = "rfc6666", stream_id='ise', - other_aliases=['rfc6666',], states=[('draft','rfc'),('draft-iesg','pub')], std_level_id='inf', ) @@ -820,7 +855,7 @@ def test_ballot_downref_approve(self): self.assertContains(r, "No downward references for") # Add a downref, the page should ask if it should be added to the registry - rel = draft.relateddocument_set.create(target=rfc.docalias.get(name='rfc6666'),relationship_id='refnorm') + rel = draft.relateddocument_set.create(target=rfc, relationship_id='refnorm') d = [rdoc for rdoc in draft.relateddocument_set.all() if rel.is_approved_downref()] original_len = len(d) r = self.client.get(url) @@ -1069,6 +1104,35 @@ def setUp(self): DocumentFactory(type_id='statchg',name='status-change-imaginary-mid-review',states=[('statchg','iesgeval')]) DocumentFactory(type_id='conflrev',name='conflict-review-imaginary-irtf-submission',states=[('conflrev','iesgeval')]) +class IetfFiltersTests(TestCase): + def test_can_defer(self): + secretariat = Person.objects.get(user__username="secretary").user + ad = Person.objects.get(user__username="ad").user + irtf_chair = Person.objects.get(user__username="irtf-chair").user + rsab_chair = Person.objects.get(user__username="rsab-chair").user + irsg_member = RoleFactory(group__type_id="rg", name_id="chair").person.user + rsab_member = RoleFactory(group=Group.objects.get(acronym="rsab"), name_id="member").person.user + nobody = PersonFactory().user + + users = set([secretariat, ad, irtf_chair, rsab_chair, irsg_member, rsab_member, nobody]) + + iesg_ballot = BallotDocEventFactory(doc__stream_id='ietf') + self.assertTrue(can_defer(secretariat, iesg_ballot.doc)) + self.assertTrue(can_defer(ad, iesg_ballot.doc)) + for user in users - set([secretariat, ad]): + self.assertFalse(can_defer(user, iesg_ballot.doc)) + + irsg_ballot = IRSGBallotDocEventFactory(doc__stream_id='irtf') + for user in users: + self.assertFalse(can_defer(user, irsg_ballot.doc)) + + rsab_ballot = BallotDocEventFactory(ballot_type__slug='rsab-approve', doc__stream_id='editorial') + for user in users: + self.assertFalse(can_defer(user, rsab_ballot.doc)) + + def test_can_clear_ballot(self): + pass # Right now, can_clear_ballot is implemented by can_defer + class RegenerateLastCallTestCase(TestCase): def test_regenerate_last_call(self): @@ -1091,13 +1155,13 @@ def test_regenerate_last_call(self): self.assertFalse("contains these normative down" in lc_text) rfc = IndividualRfcFactory.create( + rfc_number=6666, stream_id='ise', - other_aliases=['rfc6666',], states=[('draft','rfc'),('draft-iesg','pub')], std_level_id='inf', ) - draft.relateddocument_set.create(target=rfc.docalias.get(name='rfc6666'),relationship_id='refnorm') + draft.relateddocument_set.create(target=rfc,relationship_id='refnorm') r = self.client.post(url, dict(regenerate_last_call_text="1")) self.assertEqual(r.status_code, 200) @@ -1107,7 +1171,7 @@ def test_regenerate_last_call(self): self.assertTrue("rfc6666" in lc_text) self.assertTrue("Independent Submission" in lc_text) - draft.relateddocument_set.create(target=rfc.docalias.get(name='rfc6666'), relationship_id='downref-approval') + draft.relateddocument_set.create(target=rfc, relationship_id='downref-approval') r = self.client.post(url, dict(regenerate_last_call_text="1")) self.assertEqual(r.status_code, 200) @@ -1201,8 +1265,9 @@ def _assertBallotMessage(self, q, balloter, expected): heading = q(f'div.h5[id$="_{slugify(balloter.plain_name())}"]') self.assertEqual(len(heading), 1) # is followed by a panel with the message of interest, so use next() + next = heading.next() self.assertEqual( - len(heading.next().find( + len(next.find( f'*[title="{expected}"]' )), 1, @@ -1379,6 +1444,31 @@ def test_document_ballot_content_without_send_email_values(self): ballot_id=ballot.pk, ) q = PyQuery(content) - self._assertBallotMessage(q, balloters[0], 'No email send requests for this discuss') - self._assertBallotMessage(q, balloters[1], 'No ballot position send log available') + self._assertBallotMessage(q, balloters[0], 'No discuss send log available') + self._assertBallotMessage(q, balloters[1], 'No comment send log available') self._assertBallotMessage(q, old_balloter, 'No ballot position send log available') + +class ReturnToUrlTests(TestCase): + def test_invalid_return_to_url(self): + with self.assertRaises(ValueError): + parse_ballot_edit_return_point('/', 'draft-ietf-opsawg-ipfix-tcpo-v6eh', '998718') + + with self.assertRaises(ValueError): + parse_ballot_edit_return_point('/a-route-that-does-not-exist/', 'draft-ietf-opsawg-ipfix-tcpo-v6eh', '998718') + + with self.assertRaises(ValueError): + parse_ballot_edit_return_point('https://example.com/phishing', 'draft-ietf-opsawg-ipfix-tcpo-v6eh', '998718') + + def test_valid_default_return_to_url(self): + self.assertEqual(parse_ballot_edit_return_point( + None, + 'draft-ietf-opsawg-ipfix-tcpo-v6eh', + '998718' + ), '/doc/draft-ietf-opsawg-ipfix-tcpo-v6eh/ballot/998718/') + + def test_valid_return_to_url(self): + self.assertEqual(parse_ballot_edit_return_point( + '/doc/draft-ietf-opsawg-ipfix-tcpo-v6eh/ballot/998718/', + 'draft-ietf-opsawg-ipfix-tcpo-v6eh', + '998718' + ), '/doc/draft-ietf-opsawg-ipfix-tcpo-v6eh/ballot/998718/') diff --git a/ietf/doc/tests_bofreq.py b/ietf/doc/tests_bofreq.py index 9925ec3d17..6b142149be 100644 --- a/ietf/doc/tests_bofreq.py +++ b/ietf/doc/tests_bofreq.py @@ -16,9 +16,10 @@ from django.template.loader import render_to_string from django.utils import timezone +from ietf.doc.storage_utils import retrieve_str from ietf.group.factories import RoleFactory from ietf.doc.factories import BofreqFactory, NewRevisionDocEventFactory -from ietf.doc.models import State, Document, DocAlias, NewRevisionDocEvent +from ietf.doc.models import State, Document, NewRevisionDocEvent from ietf.doc.utils_bofreq import bofreq_editors, bofreq_responsible from ietf.ietfauth.utils import has_role from ietf.person.factories import PersonFactory @@ -32,7 +33,7 @@ class BofreqTests(TestCase): settings_temp_path_overrides = TestCase.settings_temp_path_overrides + ['BOFREQ_PATH'] def write_bofreq_file(self, bofreq): - fname = Path(settings.BOFREQ_PATH) / ("%s-%s.md" % (bofreq.canonical_name(), bofreq.rev)) + fname = Path(settings.BOFREQ_PATH) / ("%s-%s.md" % (bofreq.name, bofreq.rev)) with fname.open("w") as f: f.write(f"""# This is a test bofreq. Version: {bofreq.rev} @@ -54,8 +55,8 @@ def test_show_bof_requests(self): self.assertEqual(r.status_code, 200) q = PyQuery(r.content) for state in states: - self.assertEqual(len(q(f'#bofreqs-{state.slug}')), 1) - self.assertEqual(len(q(f'#bofreqs-{state.slug} tbody tr')), 3) + self.assertEqual(len(q(f'#bofreqs-{state.slug}')), 1 if state.slug!="spam" else 0) + self.assertEqual(len(q(f'#bofreqs-{state.slug} tbody tr')), 3 if state.slug!="spam" else 0) self.assertFalse(q('#start_button')) PersonFactory(user__username='nobody') self.client.login(username='nobody', password='nobody+password') @@ -63,6 +64,13 @@ def test_show_bof_requests(self): self.assertEqual(r.status_code, 200) q = PyQuery(r.content) self.assertTrue(q('#start_button')) + self.client.logout() + self.client.login(username='secretary', password='secretary+password') + r = self.client.get(url) + q = PyQuery(r.content) + for state in states: + self.assertEqual(len(q(f'#bofreqs-{state.slug}')), 1) + self.assertEqual(len(q(f'#bofreqs-{state.slug} tbody tr')), 3) def test_bofreq_main_page(self): @@ -299,17 +307,20 @@ def test_submit(self): url = urlreverse('ietf.doc.views_bofreq.submit', kwargs=dict(name=doc.name)) rev = doc.rev + doc_time = doc.time r = self.client.post(url,{'bofreq_submission':'enter','bofreq_content':'# oiwefrase'}) self.assertEqual(r.status_code, 302) doc = reload_db_objects(doc) - self.assertEqual(rev, doc.rev) + self.assertEqual(doc.rev, rev) + self.assertEqual(doc.time, doc_time) nobody = PersonFactory() self.client.login(username=nobody.user.username, password=nobody.user.username+'+password') r = self.client.post(url,{'bofreq_submission':'enter','bofreq_content':'# oiwefrase'}) self.assertEqual(r.status_code, 403) doc = reload_db_objects(doc) - self.assertEqual(rev, doc.rev) + self.assertEqual(doc.rev, rev) + self.assertEqual(doc.time, doc_time) self.client.logout() editor = bofreq_editors(doc).first() @@ -320,22 +331,29 @@ def test_submit(self): file = NamedTemporaryFile(delete=False,mode="w+",encoding='utf-8') file.write(f'# {username}') file.close() - for postdict in [ - {'bofreq_submission':'enter','bofreq_content':f'# {username}'}, - {'bofreq_submission':'upload','bofreq_file':open(file.name,'rb')}, - ]: - docevent_count = doc.docevent_set.count() - empty_outbox() - r = self.client.post(url, postdict) - self.assertEqual(r.status_code, 302) - doc = reload_db_objects(doc) - self.assertEqual('%02d'%(int(rev)+1) ,doc.rev) - self.assertEqual(f'# {username}', doc.text()) - self.assertEqual(docevent_count+1, doc.docevent_set.count()) - self.assertEqual(1, len(outbox)) - rev = doc.rev + try: + with open(file.name, 'rb') as bofreq_fd: + for postdict in [ + {'bofreq_submission':'enter','bofreq_content':f'# {username}'}, + {'bofreq_submission':'upload','bofreq_file':bofreq_fd}, + ]: + docevent_count = doc.docevent_set.count() + empty_outbox() + r = self.client.post(url, postdict) + self.assertEqual(r.status_code, 302) + doc = reload_db_objects(doc) + self.assertEqual(doc.rev, '%02d'%(int(rev)+1)) + self.assertGreater(doc.time, doc_time) + self.assertEqual(doc.text(), f'# {username}') + self.assertEqual(retrieve_str('bofreq', doc.get_base_name()), f'# {username}') + self.assertEqual(doc.docevent_set.count(), docevent_count+1) + self.assertEqual(len(outbox), 1) + rev = doc.rev + doc_time = doc.time + finally: + os.unlink(file.name) + self.client.logout() - os.unlink(file.name) def test_start_new_bofreq(self): url = urlreverse('ietf.doc.views_bofreq.new_bof_request') @@ -350,25 +368,28 @@ def test_start_new_bofreq(self): file = NamedTemporaryFile(delete=False,mode="w+",encoding='utf-8') file.write('some stuff') file.close() - for postdict in [ - dict(title='title one', bofreq_submission='enter', bofreq_content='some stuff'), - dict(title='title two', bofreq_submission='upload', bofreq_file=open(file.name,'rb')), - ]: - empty_outbox() - r = self.client.post(url, postdict) - self.assertEqual(r.status_code,302) - name = f"bofreq-{xslugify(nobody.last_name())[:64]}-{postdict['title']}".replace(' ','-') - bofreq = Document.objects.filter(name=name,type_id='bofreq').first() - self.assertIsNotNone(bofreq) - self.assertIsNotNone(DocAlias.objects.filter(name=name).first()) - self.assertEqual(bofreq.title, postdict['title']) - self.assertEqual(bofreq.rev, '00') - self.assertEqual(bofreq.get_state_slug(), 'proposed') - self.assertEqual(list(bofreq_editors(bofreq)), [nobody]) - self.assertEqual(bofreq.latest_event(NewRevisionDocEvent).rev, '00') - self.assertEqual(bofreq.text_or_error(), 'some stuff') - self.assertEqual(len(outbox),1) - os.unlink(file.name) + try: + with open(file.name,'rb') as bofreq_fd: + for postdict in [ + dict(title='title one', bofreq_submission='enter', bofreq_content='some stuff'), + dict(title='title two', bofreq_submission='upload', bofreq_file=bofreq_fd), + ]: + empty_outbox() + r = self.client.post(url, postdict) + self.assertEqual(r.status_code,302) + name = f"bofreq-{xslugify(nobody.last_name())[:64]}-{postdict['title']}".replace(' ','-') + bofreq = Document.objects.filter(name=name,type_id='bofreq').first() + self.assertIsNotNone(bofreq) + self.assertEqual(bofreq.title, postdict['title']) + self.assertEqual(bofreq.rev, '00') + self.assertEqual(bofreq.get_state_slug(), 'proposed') + self.assertEqual(list(bofreq_editors(bofreq)), [nobody]) + self.assertEqual(bofreq.latest_event(NewRevisionDocEvent).rev, '00') + self.assertEqual(bofreq.text_or_error(), 'some stuff') + self.assertEqual(retrieve_str('bofreq',bofreq.get_base_name()), 'some stuff') + self.assertEqual(len(outbox),1) + finally: + os.unlink(file.name) existing_bofreq = BofreqFactory(requester_lastname=nobody.last_name()) for postdict in [ dict(title='', bofreq_submission='enter', bofreq_content='some stuff'), diff --git a/ietf/doc/tests_charter.py b/ietf/doc/tests_charter.py index f65cf14e08..62e49559e2 100644 --- a/ietf/doc/tests_charter.py +++ b/ietf/doc/tests_charter.py @@ -1,5 +1,5 @@ # -*- coding: utf-8 -*- -# Copyright The IETF Trust 2011-2020, All Rights Reserved +# Copyright The IETF Trust 2011-2023, All Rights Reserved import datetime @@ -16,6 +16,7 @@ from ietf.doc.factories import CharterFactory, NewRevisionDocEventFactory, TelechatDocEventFactory from ietf.doc.models import ( Document, State, BallotDocEvent, BallotType, NewRevisionDocEvent, TelechatDocEvent, WriteupDocEvent ) +from ietf.doc.storage_utils import retrieve_str from ietf.doc.utils_charter import ( next_revision, default_review_text, default_action_text, charter_name_for_group ) from ietf.doc.utils import close_open_ballots @@ -87,11 +88,12 @@ def test_view_revisions(self): class EditCharterTests(TestCase): settings_temp_path_overrides = TestCase.settings_temp_path_overrides + ['CHARTER_PATH'] + def setUp(self): + super().setUp() + (Path(settings.FTP_DIR)/"charter").mkdir() + def write_charter_file(self, charter): - with (Path(settings.CHARTER_PATH) / - ("%s-%s.txt" % (charter.canonical_name(), charter.rev)) - ).open("w") as f: - f.write("This is a charter.") + (Path(settings.CHARTER_PATH) / f"{charter.name}-{charter.rev}.txt").write_text("This is a charter.") def test_startstop_process(self): CharterFactory(group__acronym='mars') @@ -509,8 +511,21 @@ def test_submit_charter(self): self.assertEqual(charter.rev, next_revision(prev_rev)) self.assertTrue("new_revision" in charter.latest_event().type) - with (Path(settings.CHARTER_PATH) / (charter.canonical_name() + "-" + charter.rev + ".txt")).open(encoding='utf-8') as f: - self.assertEqual(f.read(), "Windows line\nMac line\nUnix line\n" + utf_8_snippet.decode('utf-8')) + charter_path = Path(settings.CHARTER_PATH) / (charter.name + "-" + charter.rev + ".txt") + file_contents = (charter_path).read_text("utf-8") + self.assertEqual( + file_contents, + "Windows line\nMac line\nUnix line\n" + utf_8_snippet.decode("utf-8"), + ) + ftp_charter_path = Path(settings.FTP_DIR) / "charter" / charter_path.name + self.assertTrue(ftp_charter_path.exists()) + self.assertTrue(charter_path.samefile(ftp_charter_path)) + blobstore_contents = retrieve_str("charter", charter.get_base_name()) + self.assertEqual( + blobstore_contents, + "Windows line\nMac line\nUnix line\n" + utf_8_snippet.decode("utf-8"), + ) + def test_submit_initial_charter(self): group = GroupFactory(type_id='wg',acronym='mars',list_email='mars-wg@ietf.org') @@ -538,6 +553,24 @@ def test_submit_initial_charter(self): group = Group.objects.get(pk=group.pk) self.assertEqual(group.charter, charter) + def test_submit_charter_with_invalid_name(self): + self.client.login(username="secretary", password="secretary+password") + ietf_group = GroupFactory(type_id="wg") + for bad_name in ("charter-irtf-{}", "charter-randomjunk-{}", "charter-ietf-thisisnotagroup"): + url = urlreverse("ietf.doc.views_charter.submit", kwargs={"name": bad_name.format(ietf_group.acronym)}) + r = self.client.get(url) + self.assertEqual(r.status_code, 404, f"GET of charter named {bad_name} should 404") + r = self.client.post(url, {}) + self.assertEqual(r.status_code, 404, f"POST of charter named {bad_name} should 404") + + irtf_group = GroupFactory(type_id="rg") + for bad_name in ("charter-ietf-{}", "charter-whatisthis-{}", "charter-irtf-thisisnotagroup"): + url = urlreverse("ietf.doc.views_charter.submit", kwargs={"name": bad_name.format(irtf_group.acronym)}) + r = self.client.get(url) + self.assertEqual(r.status_code, 404, f"GET of charter named {bad_name} should 404") + r = self.client.post(url, {}) + self.assertEqual(r.status_code, 404, f"POST of charter named {bad_name} should 404") + def test_edit_review_announcement_text(self): area = GroupFactory(type_id='area') RoleFactory(name_id='ad',group=area,person=Person.objects.get(user__username='ad')) @@ -788,9 +821,11 @@ def test_approve(self): self.assertTrue(not charter.ballot_open("approve")) self.assertEqual(charter.rev, "01") - self.assertTrue( - (Path(settings.CHARTER_PATH) / ("charter-ietf-%s-%s.txt" % (group.acronym, charter.rev))).exists() - ) + charter_path = Path(settings.CHARTER_PATH) / ("charter-ietf-%s-%s.txt" % (group.acronym, charter.rev)) + charter_ftp_path = Path(settings.FTP_DIR) / "charter" / charter_path.name + self.assertTrue(charter_path.exists()) + self.assertTrue(charter_ftp_path.exists()) + self.assertTrue(charter_path.samefile(charter_ftp_path)) self.assertEqual(len(outbox), 2) # @@ -817,6 +852,19 @@ def test_approve(self): self.assertEqual(group.groupmilestone_set.filter(state="active", desc=m1.desc).count(), 1) self.assertEqual(group.groupmilestone_set.filter(state="active", desc=m4.desc).count(), 1) + def test_approve_irtf(self): + charter = CharterFactory(group__type_id='rg') + url = urlreverse('ietf.doc.views_charter.approve', kwargs=dict(name=charter.name)) + login_testing_unauthorized(self, "secretary", url) + empty_outbox() + r = self.client.post(url, dict()) + self.assertEqual(r.status_code, 302) + self.assertEqual(len(outbox), 2) + self.assertTrue("IRTF" in outbox[1]['From']) + self.assertTrue("irtf-announce" in outbox[1]['To']) + self.assertTrue(charter.group.acronym in outbox[1]['Cc']) + self.assertTrue("RG Action" in outbox[1]['Subject']) + def test_charter_with_milestones(self): charter = CharterFactory() diff --git a/ietf/doc/tests_conflict_review.py b/ietf/doc/tests_conflict_review.py index 6fa1a4d9b6..791db17f5a 100644 --- a/ietf/doc/tests_conflict_review.py +++ b/ietf/doc/tests_conflict_review.py @@ -1,9 +1,10 @@ -# Copyright The IETF Trust 2012-2020, All Rights Reserved +# Copyright The IETF Trust 2012-2023, All Rights Reserved # -*- coding: utf-8 -*- import io import os +from pathlib import Path from pyquery import PyQuery from textwrap import wrap @@ -13,8 +14,9 @@ import debug # pyflakes:ignore -from ietf.doc.factories import IndividualDraftFactory, ConflictReviewFactory -from ietf.doc.models import Document, DocEvent, NewRevisionDocEvent, BallotPositionDocEvent, TelechatDocEvent, State +from ietf.doc.factories import IndividualDraftFactory, ConflictReviewFactory, RgDraftFactory +from ietf.doc.models import Document, DocEvent, NewRevisionDocEvent, BallotPositionDocEvent, TelechatDocEvent, State, DocTagName +from ietf.doc.storage_utils import retrieve_str from ietf.doc.utils import create_ballot_if_not_open from ietf.doc.views_conflict_review import default_approval_text from ietf.group.models import Person @@ -70,12 +72,12 @@ def test_start_review_as_secretary(self): self.assertEqual(review_doc.ad.name,'Areað Irector') self.assertEqual(review_doc.notify,'ipu@ietf.org') doc = Document.objects.get(name='draft-imaginary-independent-submission') - self.assertTrue(doc in [x.target.document for x in review_doc.relateddocument_set.filter(relationship__slug='conflrev')]) + self.assertTrue(doc in [x.target for x in review_doc.relateddocument_set.filter(relationship__slug='conflrev')]) self.assertTrue(review_doc.latest_event(DocEvent,type="added_comment").desc.startswith("IETF conflict review requested")) self.assertTrue(doc.latest_event(DocEvent,type="added_comment").desc.startswith("IETF conflict review initiated")) self.assertTrue('Conflict Review requested' in outbox[-1]['Subject']) - + # verify you can't start a review when a review is already in progress r = self.client.post(url,dict(ad="Areað Irector",create_in_state="Needs Shepherd",notify='ipu@ietf.org')) self.assertEqual(r.status_code, 404) @@ -119,7 +121,7 @@ def test_start_review_as_stream_owner(self): self.assertEqual(review_doc.ad.name,'Ietf Chair') self.assertEqual(review_doc.notify,'ipu@ietf.org') doc = Document.objects.get(name='draft-imaginary-independent-submission') - self.assertTrue(doc in [x.target.document for x in review_doc.relateddocument_set.filter(relationship__slug='conflrev')]) + self.assertTrue(doc in [x.target for x in review_doc.relateddocument_set.filter(relationship__slug='conflrev')]) self.assertEqual(len(outbox), messages_before + 2) @@ -168,6 +170,21 @@ def test_change_state(self): self.assertTrue(review_doc.active_ballot()) self.assertEqual(review_doc.latest_event(BallotPositionDocEvent, type="changed_ballot_position").pos_id,'yes') + # try to change to an AD-forbidden state + appr_noprob_sent_pk = str(State.objects.get(used=True, slug='appr-noprob-sent',type__slug='conflrev').pk) + r = self.client.post(url,dict(review_state=appr_noprob_sent_pk,comment='xyzzy')) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertTrue(q('form .invalid-feedback')) + + # try again as secretariat + self.client.logout() + login_testing_unauthorized(self, 'secretary', url) + r = self.client.post(url,dict(review_state=appr_noprob_sent_pk,comment='xyzzy')) + self.assertEqual(r.status_code, 302) + review_doc = Document.objects.get(name='conflict-review-imaginary-irtf-submission') + self.assertEqual(review_doc.get_state('conflrev').slug, 'appr-noprob-sent') + def test_edit_notices(self): doc = Document.objects.get(name='conflict-review-imaginary-irtf-submission') @@ -372,7 +389,7 @@ def setUp(self): class ConflictReviewSubmitTests(TestCase): - settings_temp_path_overrides = TestCase.settings_temp_path_overrides + ['CONFLICT_REVIEW_PATH'] + settings_temp_path_overrides = TestCase.settings_temp_path_overrides + ['CONFLICT_REVIEW_PATH','FTP_PATH'] def test_initial_submission(self): doc = Document.objects.get(name='conflict-review-imaginary-irtf-submission') url = urlreverse('ietf.doc.views_conflict_review.submit',kwargs=dict(name=doc.name)) @@ -388,9 +405,15 @@ def test_initial_submission(self): # Right now, nothing to test - we let people put whatever the web browser will let them put into that textbox # sane post using textbox - path = os.path.join(settings.CONFLICT_REVIEW_PATH, '%s-%s.txt' % (doc.canonical_name(), doc.rev)) + basename = f"{doc.name}-{doc.rev}.txt" + path = Path(settings.CONFLICT_REVIEW_PATH) / basename + ftp_dir = Path(settings.FTP_DIR) / "conflict-reviews" + if not ftp_dir.exists(): + ftp_dir.mkdir() + ftp_path = ftp_dir / basename self.assertEqual(doc.rev,'00') - self.assertFalse(os.path.exists(path)) + self.assertFalse(path.exists()) + self.assertFalse(ftp_path.exists()) r = self.client.post(url,dict(content="Some initial review text\n",submit_response="1")) self.assertEqual(r.status_code,302) doc = Document.objects.get(name='conflict-review-imaginary-irtf-submission') @@ -398,7 +421,9 @@ def test_initial_submission(self): with io.open(path) as f: self.assertEqual(f.read(),"Some initial review text\n") f.close() + self.assertTrue(ftp_path.exists()) self.assertTrue( "submission-00" in doc.latest_event(NewRevisionDocEvent).desc) + self.assertEqual(retrieve_str("conflrev",basename), "Some initial review text\n") def test_subsequent_submission(self): doc = Document.objects.get(name='conflict-review-imaginary-irtf-submission') @@ -408,7 +433,7 @@ def test_subsequent_submission(self): # A little additional setup # doc.rev is u'00' per the test setup - double-checking that here - if it fails, the breakage is in setUp self.assertEqual(doc.rev,'00') - path = os.path.join(settings.CONFLICT_REVIEW_PATH, '%s-%s.txt' % (doc.canonical_name(), doc.rev)) + path = os.path.join(settings.CONFLICT_REVIEW_PATH, '%s-%s.txt' % (doc.name, doc.rev)) with io.open(path,'w') as f: f.write('This is the old proposal.') f.close() @@ -435,7 +460,7 @@ def test_subsequent_submission(self): self.assertEqual(r.status_code, 302) doc = Document.objects.get(name='conflict-review-imaginary-irtf-submission') self.assertEqual(doc.rev,'01') - path = os.path.join(settings.CONFLICT_REVIEW_PATH, '%s-%s.txt' % (doc.canonical_name(), doc.rev)) + path = os.path.join(settings.CONFLICT_REVIEW_PATH, '%s-%s.txt' % (doc.name, doc.rev)) with io.open(path) as f: self.assertEqual(f.read(),"This is a new proposal.") f.close() @@ -450,3 +475,89 @@ def test_subsequent_submission(self): def setUp(self): super().setUp() ConflictReviewFactory(name='conflict-review-imaginary-irtf-submission',review_of=IndividualDraftFactory(name='draft-imaginary-irtf-submission',stream_id='irtf'),notify='notifyme@example.net') + +class ConflictReviewStreamStateTests(TestCase): + + def start_review(self, stream, role, kwargs=None): + doc = RgDraftFactory() if stream=='irtf' else IndividualDraftFactory(stream=StreamName.objects.get(slug='ise')) + url = urlreverse('ietf.doc.views_conflict_review.start_review', kwargs=dict(name=doc.name)) + login_testing_unauthorized(self, role, url) + r = self.client.post(url, kwargs) + self.assertEqual(r.status_code, 302) + self.assertEqual(doc.get_state('draft-stream-'+stream).slug, 'iesg-rev') + + def test_start_irtf_review_as_secretary(self): + ad_strpk = str(Person.objects.get(name='Areað Irector').pk) + state_strpk = str(State.objects.get(used=True, slug='needshep', type__slug='conflrev').pk) + self.start_review('irtf', 'secretary', kwargs=dict(ad=ad_strpk, create_in_state=state_strpk)) + + def test_start_ise_review_as_secretary(self): + ad_strpk = str(Person.objects.get(name='Areað Irector').pk) + state_strpk = str(State.objects.get(used=True, slug='needshep', type__slug='conflrev').pk) + self.start_review('ise', 'secretary', kwargs=dict(ad=ad_strpk, create_in_state=state_strpk)) + + def test_start_irtf_review_as_stream_owner(self): + self.start_review('irtf', 'irtf-chair') + + def test_start_ise_review_as_stream_owner(self): + self.start_review('ise', 'ise-chair') + + def close_review(self, close_type, stream, role): + doc = RgDraftFactory() if stream=='irtf' else IndividualDraftFactory(stream=StreamName.objects.get(slug='ise')) + review = ConflictReviewFactory(review_of=doc) + url = urlreverse('ietf.doc.views_conflict_review.change_state', kwargs=dict(name=review.name)) + login_testing_unauthorized(self, role, url) + strpk = str(State.objects.get(used=True, slug=close_type, type__slug='conflrev').pk) + r = self.client.post(url, dict(review_state=strpk)) + self.assertEqual(r.status_code, 302) + self.assertEqual(doc.get_state('draft-stream-'+stream).slug, 'chair-w' if stream=='irtf' else 'ise-rev') + self.assertIn(DocTagName.objects.get(pk='iesg-com'), doc.tags.all()) + + def test_close_irtf_review_reqnopub_as_secretary(self): + self.close_review('appr-reqnopub-sent', 'irtf', 'secretary') + + def test_close_ise_review_reqnopub_as_secretary(self): + self.close_review('appr-reqnopub-sent', 'ise', 'secretary') + + def test_close_irtf_review_noprob_as_secretary(self): + self.close_review('appr-noprob-sent', 'irtf', 'secretary') + + def test_close_ise_review_noprob_as_secretary(self): + self.close_review('appr-noprob-sent', 'ise', 'secretary') + + def test_close_irtf_review_withdraw_as_secretary(self): + self.close_review('withdraw', 'irtf', 'secretary') + + def test_close_ise_review_withdraw_as_secretary(self): + self.close_review('withdraw', 'ise', 'secretary') + + def test_close_irtf_review_dead_as_secretary(self): + self.close_review('dead', 'irtf', 'secretary') + + def test_close_ise_review_dead_as_secretary(self): + self.close_review('dead', 'ise', 'secretary') + + def test_close_irtf_review_withdraw_as_ad(self): + self.close_review('withdraw', 'irtf', 'ad') + + def test_close_ise_review_withdraw_as_ad(self): + self.close_review('withdraw', 'ise', 'ad') + + def test_close_irtf_review_dead_as_ad(self): + self.close_review('dead', 'irtf', 'ad') + + def test_close_ise_review_dead_as_ad(self): + self.close_review('dead', 'ise', 'ad') + + def test_approve_review(self): + doc = RgDraftFactory() + review = ConflictReviewFactory(review_of=doc) + review.set_state(State.objects.get(used=True, slug='appr-noprob-pend', type='conflrev')) + + url = urlreverse('ietf.doc.views_conflict_review.approve_conflict_review', kwargs=dict(name=review.name)) + login_testing_unauthorized(self, 'secretary', url) + + r = self.client.post(url, dict(announcement_text=default_approval_text(review))) + self.assertEqual(r.status_code, 302) + self.assertEqual(doc.get_state('draft-stream-irtf').slug, 'chair-w') + self.assertIn(DocTagName.objects.get(pk='iesg-com'), doc.tags.all()) diff --git a/ietf/doc/tests_downref.py b/ietf/doc/tests_downref.py index dae65cb07d..0222ad7942 100644 --- a/ietf/doc/tests_downref.py +++ b/ietf/doc/tests_downref.py @@ -19,12 +19,9 @@ def setUp(self): super().setUp() PersonFactory(name='Plain Man',user__username='plain') self.draft = WgDraftFactory(name='draft-ietf-mars-test') - self.draftalias = self.draft.docalias.get(name='draft-ietf-mars-test') self.doc = WgDraftFactory(name='draft-ietf-mars-approved-document',states=[('draft-iesg','rfcqueue')]) - self.docalias = self.doc.docalias.get(name='draft-ietf-mars-approved-document') - self.rfc = WgRfcFactory(alias2__name='rfc9998') - self.rfcalias = self.rfc.docalias.get(name='rfc9998') - RelatedDocument.objects.create(source=self.doc, target=self.rfcalias, relationship_id='downref-approval') + self.rfc = WgRfcFactory(rfc_number=9998) + RelatedDocument.objects.create(source=self.doc, target=self.rfc, relationship_id='downref-approval') def test_downref_registry(self): url = urlreverse('ietf.doc.views_downref.downref_registry') @@ -64,44 +61,44 @@ def test_downref_registry_add(self): self.assertContains(r, 'Save downref') # error - already in the downref registry - r = self.client.post(url, dict(rfc=self.rfcalias.pk, drafts=(self.doc.pk, ))) + r = self.client.post(url, dict(rfc=self.rfc.pk, drafts=(self.doc.pk, ))) self.assertContains(r, 'Downref is already in the registry') # error - source is not in an approved state r = self.client.get(url) self.assertEqual(r.status_code, 200) - r = self.client.post(url, dict(rfc=self.rfcalias.pk, drafts=(self.draft.pk, ))) + r = self.client.post(url, dict(rfc=self.rfc.pk, drafts=(self.draft.pk, ))) self.assertContains(r, 'Draft is not yet approved') # error - the target is not a normative reference of the source self.draft.set_state(State.objects.get(used=True, type="draft-iesg", slug="pub")) r = self.client.get(url) self.assertEqual(r.status_code, 200) - r = self.client.post(url, dict(rfc=self.rfcalias.pk, drafts=(self.draft.pk, ))) + r = self.client.post(url, dict(rfc=self.rfc.pk, drafts=(self.draft.pk, ))) self.assertContains(r, 'There does not seem to be a normative reference to RFC') self.assertContains(r, 'Save downref anyway') # normal - approve the document so the downref is now okay - RelatedDocument.objects.create(source=self.draft, target=self.rfcalias, relationship_id='refnorm') + RelatedDocument.objects.create(source=self.draft, target=self.rfc, relationship_id='refnorm') draft_de_count_before = self.draft.docevent_set.count() rfc_de_count_before = self.rfc.docevent_set.count() r = self.client.get(url) self.assertEqual(r.status_code, 200) - r = self.client.post(url, dict(rfc=self.rfcalias.pk, drafts=(self.draft.pk, ))) + r = self.client.post(url, dict(rfc=self.rfc.pk, drafts=(self.draft.pk, ))) self.assertEqual(r.status_code, 302) newurl = urlreverse('ietf.doc.views_downref.downref_registry') r = self.client.get(newurl) self.assertContains(r, 'tr>th:first-child").text() + self.assertNotIn("IESG", top_level_metadata_headings) + self.assertNotIn("IANA", top_level_metadata_headings) + +class IetfGroupActionHelperTests(TestCase): + def test_manage_adoption_routing(self): + draft = IndividualDraftFactory() + nobody = PersonFactory() + rgchair = RoleFactory(group__type_id="rg", name_id="chair").person + wgchair = RoleFactory(group__type_id="wg", name_id="chair").person + multichair = RoleFactory(group__type_id="rg", name_id="chair").person + RoleFactory(group__type_id="wg", person=multichair, name_id="chair") + ad = RoleFactory(group__type_id="area", name_id="ad").person + secretary = Role.objects.filter( + name_id="secr", group__acronym="secretariat" + ).first() + self.assertIsNotNone(secretary) + secretary = secretary.person + self.assertFalse( + has_role(rgchair.user, ["Secretariat", "Area Director", "WG Chair"]) + ) + url = urlreverse( + "ietf.doc.views_doc.document_main", kwargs={"name": draft.name} + ) + ask_about_ietf_link = urlreverse( + "ietf.doc.views_draft.ask_about_ietf_adoption_call", + kwargs={"name": draft.name}, + ) + non_ietf_adoption_link = urlreverse( + "ietf.doc.views_draft.adopt_draft", kwargs={"name": draft.name} + ) + for person in (None, nobody, rgchair, wgchair, multichair, ad, secretary): + if person is not None: + self.client.login( + username=person.user.username, + password=f"{person.user.username}+password", + ) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + has_ask_about_ietf_link = len(q(f'a[href="{ask_about_ietf_link}"]')) != 0 + has_non_ietf_adoption_link = ( + len(q(f'a[href="{non_ietf_adoption_link}"]')) != 0 + ) + ask_about_r = self.client.get(ask_about_ietf_link) + ask_about_link_return_code = ask_about_r.status_code + if person == rgchair: + self.assertFalse(has_ask_about_ietf_link) + self.assertTrue(has_non_ietf_adoption_link) + self.assertEqual(ask_about_link_return_code, 403) + elif person in (ad, nobody, None): + self.assertFalse(has_ask_about_ietf_link) + self.assertFalse(has_non_ietf_adoption_link) + self.assertEqual( + ask_about_link_return_code, 302 if person is None else 403 + ) + else: + self.assertTrue(has_ask_about_ietf_link) + self.assertFalse(has_non_ietf_adoption_link) + self.assertEqual(ask_about_link_return_code, 200) + self.client.logout() + + def test_ask_about_ietf_adoption_call(self): + # Basic permission tests above + doc = IndividualDraftFactory() + self.assertEqual(doc.docevent_set.count(), 1) + chair_role = RoleFactory(group__type_id="wg", name_id="chair") + chair = chair_role.person + group = chair_role.group + othergroup = GroupFactory(type_id="wg") + url = urlreverse( + "ietf.doc.views_draft.ask_about_ietf_adoption_call", + kwargs={"name": doc.name}, + ) + login_testing_unauthorized(self, chair.user.username, url) + r = self.client.post(url, {"group": othergroup.pk}) + self.assertEqual(r.status_code, 200) + r = self.client.post(url, {"group": group.pk}) + self.assertEqual(r.status_code, 302) + + def test_offer_wg_action_helpers(self): + def _assert_view_presents_buttons(testcase, response, expected): + q = PyQuery(response.content) + for id, expect in expected: + button = q(f"#{id}") + testcase.assertEqual( + len(button) != 0, + expect + ) + + # View rejects access + came_from_draft = WgDraftFactory(states=[("draft","rfc")]) + rfc = WgRfcFactory(group=came_from_draft.group) + came_from_draft.relateddocument_set.create(relationship_id="became_rfc",target=rfc) + rfc_chair = RoleFactory(name_id="chair", group=rfc.group).person + url = urlreverse("ietf.doc.views_draft.offer_wg_action_helpers", kwargs=dict(name=came_from_draft.name)) + login_testing_unauthorized(self, rfc_chair.user.username, url) + r = self.client.get(url) + self.assertEqual(r.status_code, 404) + self.client.logout() + rg_draft = RgDraftFactory() + rg_chair = RoleFactory(group=rg_draft.group, name_id="chair").person + url = urlreverse("ietf.doc.views_draft.offer_wg_action_helpers", kwargs=dict(name=rg_draft.name)) + login_testing_unauthorized(self, rg_chair.user.username, url) + r = self.client.get(url) + self.assertEqual(r.status_code,404) + self.client.logout() + + # View offers access + draft = WgDraftFactory() + chair = RoleFactory(group=draft.group, name_id="chair").person + url = urlreverse("ietf.doc.views_draft.offer_wg_action_helpers", kwargs=dict(name=draft.name)) + login_testing_unauthorized(self, chair.user.username, url) + r = self.client.get(url) + self.assertEqual(r.status_code,200) + _assert_view_presents_buttons( + self, + r, + [ + ("id_wgadopt_button", False), + ("id_wglc_button", True), + ("id_pubreq_button", True), + ], + ) + draft.set_state(State.objects.get(type_id="draft-stream-ietf", slug="wg-cand")) + r = self.client.get(url) + self.assertEqual(r.status_code,200) + _assert_view_presents_buttons( + self, + r, + [ + ("id_wgadopt_button", True), + ("id_wglc_button", False), + ("id_pubreq_button", False), + ], + ) + draft.set_state(State.objects.get(type_id="draft-stream-ietf", slug="wg-lc")) + StateDocEventFactory( + doc=draft, + state_type_id="draft-stream-ietf", + state=("draft-stream-ietf", "wg-lc"), + ) + self.assertEqual(draft.docevent_set.count(), 2) + r = self.client.get(url) + self.assertEqual(r.status_code,200) + _assert_view_presents_buttons( + self, + r, + [ + ("id_wgadopt_button", False), + ("id_wglc_button", False), + ("id_pubreq_button", True), + ], + ) + draft.set_state(State.objects.get(type_id="draft-stream-ietf",slug="chair-w")) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + _assert_view_presents_buttons( + self, + r, + [ + ("id_wgadopt_button", False), + ("id_wglc_button", True), + ("id_pubreq_button", True), + ], + ) + self.assertContains(response=r,text="Issue Another Working Group Last Call", status_code=200) + other_draft = WgDraftFactory() + self.client.logout() + url = urlreverse("ietf.doc.views_draft.offer_wg_action_helpers", kwargs=dict(name=other_draft.name)) + login_testing_unauthorized(self, "secretary", url) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + _assert_view_presents_buttons( + self, + r, + [ + ("id_wgadopt_button", False), + ("id_wglc_button", True), + ("id_pubreq_button", True), + ], + ) + self.assertContains( + response=r, text="Issue Working Group Last Call", status_code=200 + ) + +class BallotEmailAjaxTests(TestCase): + def test_ajax_build_position_email(self): + def _post_json(self, url, json_to_post): + r = self.client.post( + url, json.dumps(json_to_post), content_type="application/json" + ) + self.assertEqual(r.status_code, 200) + return json.loads(r.content) + + doc = WgDraftFactory() + ad = RoleFactory( + name_id="ad", group=doc.group, person__name="Some Areadirector" + ).person + url = urlreverse("ietf.doc.views_ballot.ajax_build_position_email") + login_testing_unauthorized(self, "secretary", url) + r = self.client.get(url) + self.assertEqual(r.status_code, 405) + response = _post_json(self, url, {}) + self.assertFalse(response["success"]) + self.assertEqual(response["errors"], ["post_data not provided"]) + response = _post_json(self, url, {"dictis": "not empty"}) + self.assertFalse(response["success"]) + self.assertEqual(response["errors"], ["post_data not provided"]) + response = _post_json(self, url, {"post_data": {}}) + self.assertFalse(response["success"]) + self.assertEqual(len(response["errors"]), 7) + response = _post_json( + self, + url, + { + "post_data": { + "discuss": "aaaaaa", + "comment": "bbbbbb", + "position": "discuss", + "balloter": Person.objects.aggregate(maxpk=Max("pk") + 1)["maxpk"], + "docname": "this-draft-does-not-exist", + "cc_choices": ["doc_group_mail_list"], + "additional_cc": "foo@example.com", + } + }, + ) + self.assertFalse(response["success"]) + self.assertEqual( + response["errors"], + ["No person found matching balloter", "No document found matching docname"], + ) + response = _post_json( + self, + url, + { + "post_data": { + "discuss": "aaaaaa", + "comment": "bbbbbb", + "position": "discuss", + "balloter": ad.pk, + "docname": doc.name, + "cc_choices": ["doc_group_mail_list"], + "additional_cc": "foo@example.com", + } + }, + ) + self.assertTrue(response["success"]) + for snippet in [ + "aaaaaa", + "bbbbbb", + "DISCUSS", + ad.plain_name(), + doc.name, + doc.group.list_email, + "foo@example.com", + ]: + self.assertIn(snippet, response["text"]) + diff --git a/ietf/doc/tests_irsg_ballot.py b/ietf/doc/tests_irsg_ballot.py index da1b48fc6c..d96cf9dbef 100644 --- a/ietf/doc/tests_irsg_ballot.py +++ b/ietf/doc/tests_irsg_ballot.py @@ -288,7 +288,7 @@ def test_edit_ballot_position_permissions(self): def test_iesg_ballot_no_irsg_actions(self): ad = Person.objects.get(user__username="ad") - wg_draft = IndividualDraftFactory(ad=ad) + wg_draft = IndividualDraftFactory(ad=ad, stream_id='ietf') irsgmember = get_active_irsg()[0] url = urlreverse('ietf.doc.views_ballot.ballot_writeupnotes', kwargs=dict(name=wg_draft.name)) @@ -355,28 +355,35 @@ def test_issue_ballot(self): def test_take_and_email_position(self): draft = RgDraftFactory() ballot = IRSGBallotDocEventFactory(doc=draft) - url = urlreverse('ietf.doc.views_ballot.edit_position', kwargs=dict(name=draft.name, ballot_id=ballot.pk)) + self.balloter + url = ( + urlreverse( + "ietf.doc.views_ballot.edit_position", + kwargs=dict(name=draft.name, ballot_id=ballot.pk), + ) + + self.balloter + ) empty_outbox() login_testing_unauthorized(self, self.username, url) r = self.client.get(url) self.assertEqual(r.status_code, 200) - r = self.client.post(url, dict(position='yes', comment='oib239sb', send_mail='Save and send email')) + empty_outbox() + r = self.client.post( + url, + dict( + position="yes", + comment="oib239sb", + send_mail="Save and send email", + cc_choices=["doc_authors", "doc_group_chairs", "doc_group_mail_list"], + ), + ) self.assertEqual(r.status_code, 302) e = draft.latest_event(BallotPositionDocEvent) - self.assertEqual(e.pos.slug,'yes') - self.assertEqual(e.comment, 'oib239sb') - - url = urlreverse('ietf.doc.views_ballot.send_ballot_comment', kwargs=dict(name=draft.name, ballot_id=ballot.pk)) + self.balloter - - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - - r = self.client.post(url, dict(cc_choices=['doc_authors','doc_group_chairs','doc_group_mail_list'], body="Stuff")) - self.assertEqual(r.status_code, 302) - self.assertEqual(len(outbox),1) - self.assertNotIn('discuss-criteria', get_payload_text(outbox[0])) + self.assertEqual(e.pos.slug, "yes") + self.assertEqual(e.comment, "oib239sb") + self.assertEqual(len(outbox), 1) + self.assertNotIn("discuss-criteria", get_payload_text(outbox[0])) def test_close_ballot(self): draft = RgDraftFactory() @@ -446,7 +453,7 @@ def setUp(self): def test_cant_issue_irsg_ballot(self): draft = RgDraftFactory() due = datetime_today(DEADLINE_TZINFO) + datetime.timedelta(days=14) - url = urlreverse('ietf.doc.views_ballot.close_irsg_ballot', kwargs=dict(name=draft.name)) + url = urlreverse('ietf.doc.views_ballot.issue_irsg_ballot', kwargs=dict(name=draft.name)) self.client.login(username = self.username, password = self.username+'+password') r = self.client.get(url) @@ -482,27 +489,31 @@ def test_cant_take_position_on_iesg_ballot(self): def test_take_and_email_position(self): draft = RgDraftFactory() ballot = IRSGBallotDocEventFactory(doc=draft) - url = urlreverse('ietf.doc.views_ballot.edit_position', kwargs=dict(name=draft.name, ballot_id=ballot.pk)) + url = urlreverse( + "ietf.doc.views_ballot.edit_position", + kwargs=dict(name=draft.name, ballot_id=ballot.pk), + ) empty_outbox() login_testing_unauthorized(self, self.username, url) r = self.client.get(url) self.assertEqual(r.status_code, 200) - r = self.client.post(url, dict(position='yes', comment='oib239sb', send_mail='Save and send email')) + r = self.client.post( + url, + dict( + position="yes", + comment="oib239sb", + send_mail="Save and send email", + cc_choices=["doc_authors", "doc_group_chairs", "doc_group_mail_list"], + ), + ) self.assertEqual(r.status_code, 302) e = draft.latest_event(BallotPositionDocEvent) - self.assertEqual(e.pos.slug,'yes') - self.assertEqual(e.comment, 'oib239sb') - - url = urlreverse('ietf.doc.views_ballot.send_ballot_comment', kwargs=dict(name=draft.name, ballot_id=ballot.pk)) - - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - - r = self.client.post(url, dict(cc_choices=['doc_authors','doc_group_chairs','doc_group_mail_list'], body="Stuff")) + self.assertEqual(e.pos.slug, "yes") + self.assertEqual(e.comment, "oib239sb") self.assertEqual(r.status_code, 302) - self.assertEqual(len(outbox),1) + self.assertEqual(len(outbox), 1) class IESGMemberTests(TestCase): diff --git a/ietf/doc/tests_js.py b/ietf/doc/tests_js.py index 02daaae904..9a5aad13b9 100644 --- a/ietf/doc/tests_js.py +++ b/ietf/doc/tests_js.py @@ -41,7 +41,7 @@ def _fill_in_author_form(form_elt, name, email, affiliation, country): (By.CSS_SELECTOR, result_selector), name )) - input.send_keys('\n') # select the object + self.driver.find_element(By.CSS_SELECTOR, result_selector).click() # After the author is selected, the email select options will be populated. # Wait for that, then click on the option corresponding to the requested email. @@ -92,12 +92,8 @@ def _read_author_form(form_elt): self.assertEqual(len(author_forms), 1) # get the "add author" button so we can add blank author forms - add_author_button = self.driver.find_element(By.ID, 'add-author-button') for index, auth in enumerate(authors): - self.driver.execute_script("arguments[0].scrollIntoView();", add_author_button) # FIXME: no idea why this fails: - # self.scroll_to_element(add_author_button) # Can only click if it's in view! - self.driver.execute_script("arguments[0].click();", add_author_button) # FIXME: no idea why this fails: - # add_author_button.click() # Create a new form. Automatically scrolls to it. + self.scroll_and_click((By.ID, 'add-author-button')) # Create new form. Automatically scrolls to it. author_forms = authors_list.find_elements(By.CLASS_NAME, 'author-panel') authors_added = index + 1 self.assertEqual(len(author_forms), authors_added + 1) # Started with 1 author, hence +1 @@ -118,10 +114,9 @@ def _read_author_form(form_elt): # Must provide a "basis" (change reason) self.driver.find_element(By.ID, 'id_basis').send_keys('change testing') # Now click the 'submit' button and check that the update was accepted. - submit_button = self.driver.find_element(By.CSS_SELECTOR, 'button[type="submit"]') - self.driver.execute_script("arguments[0].click();", submit_button) # FIXME: no idea why this fails: - # self.scroll_to_element(submit_button) - # submit_button.click() + submit_button = self.driver.find_element(By.CSS_SELECTOR, '#content button[type="submit"]') + self.scroll_to_element(submit_button) + submit_button.click() # Wait for redirect to the document_main view self.wait.until( expected_conditions.url_to_be( @@ -132,4 +127,4 @@ def _read_author_form(form_elt): self.assertEqual( list(draft.documentauthor_set.values_list('person', flat=True)), [first_auth.person.pk] + [auth.pk for auth in authors] - ) \ No newline at end of file + ) diff --git a/ietf/doc/tests_material.py b/ietf/doc/tests_material.py index 05bbc2078b..04779bdaf1 100644 --- a/ietf/doc/tests_material.py +++ b/ietf/doc/tests_material.py @@ -6,19 +6,22 @@ import shutil import io +from unittest.mock import call, patch from pathlib import Path from pyquery import PyQuery import debug # pyflakes:ignore from django.conf import settings +from django.test import override_settings from django.urls import reverse as urlreverse from django.utils import timezone -from ietf.doc.models import Document, State, DocAlias, NewRevisionDocEvent +from ietf.doc.models import Document, State, NewRevisionDocEvent +from ietf.doc.storage_utils import retrieve_str from ietf.group.factories import RoleFactory from ietf.group.models import Group -from ietf.meeting.factories import MeetingFactory, SessionFactory +from ietf.meeting.factories import MeetingFactory, SessionFactory, SessionPresentationFactory from ietf.meeting.models import Meeting, SessionPresentation, SchedulingEvent from ietf.name.models import SessionStatusName from ietf.person.models import Person @@ -26,7 +29,7 @@ class GroupMaterialTests(TestCase): - settings_temp_path_overrides = TestCase.settings_temp_path_overrides + ['AGENDA_PATH'] + settings_temp_path_overrides = TestCase.settings_temp_path_overrides + ['AGENDA_PATH', 'FTP_DIR'] def setUp(self): super().setUp() self.materials_dir = self.tempdir("materials") @@ -35,6 +38,10 @@ def setUp(self): self.slides_dir.mkdir() self.saved_document_path_pattern = settings.DOCUMENT_PATH_PATTERN settings.DOCUMENT_PATH_PATTERN = self.materials_dir + "/{doc.type_id}/" + self.assertTrue(Path(settings.FTP_DIR).exists()) + ftp_slides_dir = Path(settings.FTP_DIR) / "slides" + if not ftp_slides_dir.exists(): + ftp_slides_dir.mkdir() self.meeting_slides_dir = Path(settings.AGENDA_PATH) / "42" / "slides" if not self.meeting_slides_dir.exists(): @@ -54,7 +61,6 @@ def create_slides(self): doc = Document.objects.create(name="slides-testteam-test-file", rev="01", type_id="slides", group=group) doc.set_state(State.objects.get(type="slides", slug="active")) doc.set_state(State.objects.get(type="reuse_policy", slug="multiple")) - DocAlias.objects.create(name=doc.name).docs.add(doc) NewRevisionDocEvent.objects.create(doc=doc,by=Person.objects.get(name="(System)"),rev='00',type='new_revision',desc='New revision available') NewRevisionDocEvent.objects.create(doc=doc,by=Person.objects.get(name="(System)"),rev='01',type='new_revision',desc='New revision available') @@ -111,8 +117,16 @@ def test_upload_slides(self): self.assertEqual(doc.title, "Test File - with fancy title") self.assertEqual(doc.get_state_slug(), "active") - with io.open(os.path.join(self.materials_dir, "slides", doc.name + "-" + doc.rev + ".pdf")) as f: + basename=f"{doc.name}-{doc.rev}.pdf" + filepath=Path(self.materials_dir) / "slides" / basename + with filepath.open() as f: self.assertEqual(f.read(), content) + ftp_filepath=Path(settings.FTP_DIR) / "slides" / basename + with ftp_filepath.open() as f: + self.assertEqual(f.read(), content) + # This test is very sloppy wrt the actual file content. + # Working with/around that for the moment. + self.assertEqual(retrieve_str("slides", basename), content) # check that posting same name is prevented test_file.seek(0) @@ -136,19 +150,47 @@ def test_change_state(self): doc = Document.objects.get(name=doc.name) self.assertEqual(doc.get_state_slug(), "deleted") - def test_edit_title(self): + @override_settings(MEETECHO_API_CONFIG="fake settings") + @patch("ietf.doc.views_material.SlidesManager") + def test_edit_title(self, mock_slides_manager_cls): doc = self.create_slides() url = urlreverse('ietf.doc.views_material.edit_material', kwargs=dict(name=doc.name, action="title")) login_testing_unauthorized(self, "secretary", url) + self.assertFalse(mock_slides_manager_cls.called) # post r = self.client.post(url, dict(title="New title")) self.assertEqual(r.status_code, 302) doc = Document.objects.get(name=doc.name) self.assertEqual(doc.title, "New title") + self.assertFalse(mock_slides_manager_cls.return_value.send_update.called) + + # assign to a session to see that it now sends updates to Meetecho + session = SessionPresentationFactory(session__group=doc.group, document=doc).session + + # Grab the title on the slides when the API call was made (to be sure it's not before it was updated) + titles_sent = [] + mock_slides_manager_cls.return_value.send_update.side_effect = lambda sess: titles_sent.extend( + list(sess.presentations.values_list("document__title", flat=True)) + ) + + r = self.client.post(url, dict(title="Newer title")) + self.assertEqual(r.status_code, 302) + doc = Document.objects.get(name=doc.name) + self.assertEqual(doc.title, "Newer title") + self.assertTrue(mock_slides_manager_cls.called) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertEqual(mock_slides_manager_cls.return_value.send_update.call_count, 1) + self.assertEqual( + mock_slides_manager_cls.return_value.send_update.call_args, + call(session), + ) + self.assertEqual(titles_sent, ["Newer title"]) - def test_revise(self): + @override_settings(MEETECHO_API_CONFIG="fake settings") + @patch("ietf.doc.views_material.SlidesManager") + def test_revise(self, mock_slides_manager_cls): doc = self.create_slides() session = SessionFactory( @@ -166,11 +208,18 @@ def test_revise(self): url = urlreverse('ietf.doc.views_material.edit_material', kwargs=dict(name=doc.name, action="revise")) login_testing_unauthorized(self, "secretary", url) + self.assertFalse(mock_slides_manager_cls.called) content = "some text" test_file = io.StringIO(content) test_file.name = "unnamed.txt" + # Grab the title on the slides when the API call was made (to be sure it's not before it was updated) + titles_sent = [] + mock_slides_manager_cls.return_value.send_update.side_effect = lambda sess: titles_sent.extend( + list(sess.presentations.values_list("document__title", flat=True)) + ) + # post r = self.client.post(url, dict(title="New title", abstract="New abstract", @@ -181,7 +230,17 @@ def test_revise(self): self.assertEqual(doc.rev, "02") self.assertEqual(doc.title, "New title") self.assertEqual(doc.get_state_slug(), "active") + self.assertTrue(mock_slides_manager_cls.called) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertEqual(mock_slides_manager_cls.return_value.send_update.call_count, 1) + self.assertEqual( + mock_slides_manager_cls.return_value.send_update.call_args, + call(session), + ) + self.assertEqual(titles_sent, ["New title"]) with io.open(os.path.join(doc.get_file_path(), doc.name + "-" + doc.rev + ".txt")) as f: self.assertEqual(f.read(), content) + self.assertEqual(retrieve_str("slides", f"{doc.name}-{doc.rev}.txt"), content) + diff --git a/ietf/doc/tests_models.py b/ietf/doc/tests_models.py new file mode 100644 index 0000000000..d835f646fb --- /dev/null +++ b/ietf/doc/tests_models.py @@ -0,0 +1,113 @@ +# Copyright The IETF Trust 2016-2023, All Rights Reserved +# -*- coding: utf-8 -*- + +import itertools + +from ietf.doc.factories import WgRfcFactory +from ietf.doc.models import RelatedDocument +from ietf.utils.test_utils import TestCase + + +class RelatedDocumentTests(TestCase): + def test_is_downref(self): + rfcs = [ + WgRfcFactory(std_level_id=lvl) + for lvl in ["inf", "exp", "bcp", "ps", "ds", "std", "unkn"] + ] + + result_matrix = { + # source + "inf": { + "inf": None, # target + "exp": None, # target + "bcp": None, # target + "ps": None, # target + "ds": None, # target + "std": None, # target + "unkn": None, # target + }, + # source + "exp": { + "inf": None, # target + "exp": None, # target + "bcp": None, # target + "ps": None, # target + "ds": None, # target + "std": None, # target + "unkn": None, # target + }, + # source + "bcp": { + "inf": "Downref", # target + "exp": "Downref", # target + "bcp": None, # target + "ps": None, # target + "ds": None, # target + "std": None, # target + "unkn": "Possible Downref", # target + }, + # source + "ps": { + "inf": "Downref", # target + "exp": "Downref", # target + "bcp": None, # target + "ps": None, # target + "ds": None, # target + "std": None, # target + "unkn": "Possible Downref", # target + }, + # source + "ds": { + "inf": "Downref", # target + "exp": "Downref", # target + "bcp": None, # target + "ps": "Downref", # target + "ds": None, # target + "std": None, # target + "unkn": "Possible Downref", # target + }, + # source + "std": { + "inf": "Downref", # target + "exp": "Downref", # target + "bcp": None, # target + "ps": "Downref", # target + "ds": "Downref", # target + "std": None, # target + "unkn": "Possible Downref", # target + }, + # source + "unkn": { + "inf": None, # target + "exp": None, # target + "bcp": None, # target + "ps": "Possible Downref", # target + "ds": "Possible Downref", # target + "std": None, # target + "unkn": "Possible Downref", # target + }, + } + + for rel in ["refnorm", "refinfo", "refunk", "refold"]: + for source, target in itertools.product(rfcs, rfcs): + ref = RelatedDocument.objects.create( + source=source, + target=target, + relationship_id=rel, + ) + + result = ref.is_downref() + + desired_result = ( + result_matrix[source.std_level_id][target.std_level_id] + if ref.relationship.slug in ["refnorm", "refunk"] + else None + ) + if ( + ref.relationship.slug == "refunk" + and desired_result is not None + and not desired_result.startswith("Possible") + ): + desired_result = f"Possible {desired_result}" + + self.assertEqual(desired_result, result) diff --git a/ietf/doc/tests_notprepped.py b/ietf/doc/tests_notprepped.py new file mode 100644 index 0000000000..f417aa7931 --- /dev/null +++ b/ietf/doc/tests_notprepped.py @@ -0,0 +1,122 @@ +# Copyright The IETF Trust 2026, All Rights Reserved + +from django.conf import settings +from django.utils import timezone +from django.urls import reverse as urlreverse + +from pyquery import PyQuery + +from ietf.doc.factories import WgRfcFactory +from ietf.doc.models import StoredObject +from ietf.doc.storage_utils import store_bytes +from ietf.utils.test_utils import TestCase + + +class NotpreppedRfcXmlTests(TestCase): + def test_editor_source_button_visibility(self): + pre_v3 = WgRfcFactory(rfc_number=settings.FIRST_V3_RFC - 1) + first_v3 = WgRfcFactory(rfc_number=settings.FIRST_V3_RFC) + post_v3 = WgRfcFactory(rfc_number=settings.FIRST_V3_RFC + 1) + + for rfc, expect_button in [(pre_v3, False), (first_v3, True), (post_v3, True)]: + r = self.client.get( + urlreverse( + "ietf.doc.views_doc.document_main", kwargs=dict(name=rfc.name) + ) + ) + self.assertEqual(r.status_code, 200) + buttons = PyQuery(r.content)('a.btn:contains("Get editor source")') + if expect_button: + self.assertEqual(len(buttons), 1, msg=f"rfc_number={rfc.rfc_number}") + expected_href = urlreverse( + "ietf.doc.views_doc.rfcxml_notprepped_wrapper", + kwargs=dict(number=rfc.rfc_number), + ) + self.assertEqual( + buttons.attr("href"), + expected_href, + msg=f"rfc_number={rfc.rfc_number}", + ) + else: + self.assertEqual(len(buttons), 0, msg=f"rfc_number={rfc.rfc_number}") + + def test_rfcxml_notprepped(self): + number = settings.FIRST_V3_RFC + stored_name = f"notprepped/rfc{number}.notprepped.xml" + url = f"/doc/rfc{number}/notprepped/" + + # 404 for pre-v3 RFC numbers (no document needed) + r = self.client.get(f"/doc/rfc{number - 1}/notprepped/") + self.assertEqual(r.status_code, 404) + + # 404 when no RFC document exists in the database + r = self.client.get(url) + self.assertEqual(r.status_code, 404) + + # 404 when RFC document exists but has no StoredObject + WgRfcFactory(rfc_number=number) + r = self.client.get(url) + self.assertEqual(r.status_code, 404) + + # 404 when StoredObject exists but backing storage is missing (FileNotFoundError) + now = timezone.now() + StoredObject.objects.create( + store="rfc", + name=stored_name, + sha384="a" * 96, + len=0, + store_created=now, + created=now, + modified=now, + ) + r = self.client.get(url) + self.assertEqual(r.status_code, 404) + + # 200 with correct content-type, attachment disposition, and body when object is fully stored + xml_content = b"test" + store_bytes("rfc", stored_name, xml_content, allow_overwrite=True) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + self.assertEqual(r["Content-Type"], "application/xml") + self.assertEqual( + r["Content-Disposition"], + f'attachment; filename="rfc{number}.notprepped.xml"', + ) + self.assertEqual(b"".join(r.streaming_content), xml_content) + + def test_rfcxml_notprepped_wrapper(self): + number = settings.FIRST_V3_RFC + + # 404 for pre-v3 RFC numbers (no document needed) + r = self.client.get( + urlreverse( + "ietf.doc.views_doc.rfcxml_notprepped_wrapper", + kwargs=dict(number=number - 1), + ) + ) + self.assertEqual(r.status_code, 404) + + # 404 when no RFC document exists in the database + r = self.client.get( + urlreverse( + "ietf.doc.views_doc.rfcxml_notprepped_wrapper", + kwargs=dict(number=number), + ) + ) + self.assertEqual(r.status_code, 404) + + # 200 with rendered template when RFC document exists + rfc = WgRfcFactory(rfc_number=number) + r = self.client.get( + urlreverse( + "ietf.doc.views_doc.rfcxml_notprepped_wrapper", + kwargs=dict(number=number), + ) + ) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertIn(str(rfc.rfc_number), q("h1").text()) + download_url = urlreverse( + "ietf.doc.views_doc.rfcxml_notprepped", kwargs=dict(number=number) + ) + self.assertEqual(len(q(f'a.btn[href="{download_url}"]')), 1) diff --git a/ietf/doc/tests_review.py b/ietf/doc/tests_review.py index f4c814005a..82d1b5c232 100644 --- a/ietf/doc/tests_review.py +++ b/ietf/doc/tests_review.py @@ -1,13 +1,14 @@ -# Copyright The IETF Trust 2016-2020, All Rights Reserved +# Copyright The IETF Trust 2016-2023, All Rights Reserved # -*- coding: utf-8 -*- -import datetime, os, shutil +from pathlib import Path +import datetime import io -import tarfile, tempfile, mailbox -import email.mime.multipart, email.mime.text, email.utils +import os +import shutil -from mock import patch +from unittest.mock import patch, Mock from requests import Response from django.apps import apps @@ -19,11 +20,12 @@ import debug # pyflakes:ignore +from ietf.doc.storage_utils import retrieve_str import ietf.review.mailarch from ietf.doc.factories import ( NewRevisionDocEventFactory, IndividualDraftFactory, WgDraftFactory, WgRfcFactory, ReviewFactory, DocumentFactory) -from ietf.doc.models import ( Document, DocumentAuthor, RelatedDocument, DocEvent, ReviewRequestDocEvent, +from ietf.doc.models import ( DocumentAuthor, RelatedDocument, DocEvent, ReviewRequestDocEvent, ReviewAssignmentDocEvent, ) from ietf.group.factories import RoleFactory, ReviewTeamFactory from ietf.group.models import Group @@ -47,6 +49,7 @@ def setUp(self): self.review_dir = self.tempdir('review') self.old_document_path_pattern = settings.DOCUMENT_PATH_PATTERN settings.DOCUMENT_PATH_PATTERN = self.review_dir + "/{doc.type_id}/" + (Path(settings.FTP_DIR) / "review").mkdir() self.review_subdir = os.path.join(self.review_dir, "review") if not os.path.exists(self.review_subdir): @@ -57,6 +60,17 @@ def tearDown(self): settings.DOCUMENT_PATH_PATTERN = self.old_document_path_pattern super().tearDown() + def verify_review_files_were_written(self, assignment, expected_content = "This is a review\nwith two lines"): + review_file = Path(self.review_subdir) / f"{assignment.review.name}.txt" + content = review_file.read_text() + self.assertEqual(content, expected_content) + self.assertEqual( + retrieve_str("review", review_file.name), + expected_content + ) + review_ftp_file = Path(settings.FTP_DIR) / "review" / review_file.name + self.assertTrue(review_file.samefile(review_ftp_file)) + def test_request_review(self): doc = WgDraftFactory(group__acronym='mars',rev='01') NewRevisionDocEventFactory(doc=doc,rev='01') @@ -137,10 +151,18 @@ def test_request_review_of_rfc(self): url = urlreverse('ietf.doc.views_review.request_review', kwargs={ "name": doc.name }) login_testing_unauthorized(self, "ad", url) - # get should fail + # get should fail - all non draft types 404 + r = self.client.get(url) + self.assertEqual(r.status_code, 404) + + # Can only request reviews on active draft documents + doc = WgDraftFactory(states=[("draft","rfc")]) + url = urlreverse('ietf.doc.views_review.request_review', kwargs={ "name": doc.name }) r = self.client.get(url) self.assertEqual(r.status_code, 403) + + def test_doc_page(self): doc = WgDraftFactory(group__acronym='mars',rev='01') @@ -153,8 +175,8 @@ def test_doc_page(self): # check we can fish it out old_doc = WgDraftFactory(name="draft-foo-mars-test") older_doc = WgDraftFactory(name="draft-older") - RelatedDocument.objects.create(source=old_doc, target=older_doc.docalias.first(), relationship_id='replaces') - RelatedDocument.objects.create(source=doc, target=old_doc.docalias.first(), relationship_id='replaces') + RelatedDocument.objects.create(source=old_doc, target=older_doc, relationship_id='replaces') + RelatedDocument.objects.create(source=doc, target=old_doc, relationship_id='replaces') review_req.doc = older_doc review_req.save() @@ -355,6 +377,42 @@ def test_assign_reviewer(self): request_events = review_req.reviewrequestdocevent_set.all() self.assertEqual(request_events.count(), 0) + def test_assign_reviewer_after_reject(self): + doc = WgDraftFactory() + review_team = ReviewTeamFactory() + rev_role = RoleFactory(group=review_team,person__user__username='reviewer',person__user__email='reviewer@example.com',name_id='reviewer') + reviewer_email = Email.objects.get(person__user__username="reviewer") + RoleFactory(group=review_team,person__user__username='reviewsecretary',name_id='secr') + review_req = ReviewRequestFactory(team=review_team,doc=doc) + ReviewAssignmentFactory(review_request=review_req, state_id='rejected', reviewer=rev_role.person.email_set.first()) + + url = urlreverse('ietf.doc.views_review.assign_reviewer', kwargs={ "name": doc.name, "request_id": review_req.pk }) + login_testing_unauthorized(self, "reviewsecretary", url) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + reviewer_label = q("option[value=\"{}\"]".format(reviewer_email.address)).text().lower() + self.assertIn("rejected review of document before", reviewer_label) + + def test_assign_reviewer_after_withdraw(self): + doc = WgDraftFactory() + review_team = ReviewTeamFactory() + rev_role = RoleFactory(group=review_team,person__user__username='reviewer',person__user__email='reviewer@example.com',name_id='reviewer') + RoleFactory(group=review_team,person__user__username='reviewsecretary',name_id='secr') + review_req = ReviewRequestFactory(team=review_team,doc=doc) + reviewer = rev_role.person.email_set.first() + ReviewAssignmentFactory(review_request=review_req, state_id='withdrawn', reviewer=reviewer) + req_url = urlreverse('ietf.doc.views_review.review_request', kwargs={ "name": doc.name, "request_id": review_req.pk }) + assign_url = urlreverse('ietf.doc.views_review.assign_reviewer', kwargs={ "name": doc.name, "request_id": review_req.pk }) + + login_testing_unauthorized(self, "reviewsecretary", assign_url) + r = self.client.post(assign_url, { "action": "assign", "reviewer": reviewer.pk }) + self.assertRedirects(r, req_url) + review_req = reload_db_objects(review_req) + assignment = review_req.reviewassignment_set.last() + self.assertEqual(assignment.state, ReviewAssignmentStateName.objects.get(slug='assigned')) + self.assertEqual(review_req.state, ReviewRequestStateName.objects.get(slug='assigned')) + def test_previously_reviewed_replaced_doc(self): review_team = ReviewTeamFactory(acronym="reviewteam", name="Review Team", type_id="review", list_email="reviewteam@ietf.org", parent=Group.objects.get(acronym="farfut")) rev_role = RoleFactory(group=review_team,person__user__username='reviewer',person__user__email='reviewer@example.com',person__name='Some Reviewer',name_id='reviewer') @@ -418,9 +476,50 @@ def test_reject_reviewer_assignment(self): r = self.client.get(req_url) self.assertEqual(r.status_code, 200) self.assertContains(r, reject_url) + + # anonymous user should not be able to reject self.client.logout() + r = self.client.post(reject_url, { "action": "reject", "message_to_secretary": "Test message" }) + self.assertEqual(r.status_code, 302) # forwards to login page + assignment = reload_db_objects(assignment) + self.assertEqual(assignment.state_id, "accepted") - # get reject page + # unrelated person should not be able to reject + other_person = PersonFactory() + login_testing_unauthorized(self, other_person.user.username, reject_url) + r = self.client.post(reject_url, { "action": "reject", "message_to_secretary": "Test message" }) + self.assertEqual(r.status_code, 403) + assignment = reload_db_objects(assignment) + self.assertEqual(assignment.state_id, "accepted") + + # Check that user can reject it + login_testing_unauthorized(self, assignment.reviewer.person.user.username, reject_url) + r = self.client.get(reject_url) + self.assertEqual(r.status_code, 200) + self.assertContains(r, escape(assignment.reviewer.person.name)) + self.assertNotContains(r, 'can not be rejected') + self.assertContains(r, '
MODIFY ...;' in order to fix foregn keys - # to the altered field, but as it seems does _not_ fix up m2m - # intermediary tables in an equivalent manner, so here we remove and - # then recreate the m2m tables so they will have the appropriate field - # types. - - operations = [ - migrations.RemoveField( - model_name='groupmilestone', - name='docs', - ), - migrations.RemoveField( - model_name='groupmilestonehistory', - name='docs', - ), - migrations.AddField( - model_name='groupmilestone', - name='docs', - field=models.ManyToManyField(to='doc.Document'), - ), - migrations.AddField( - model_name='groupmilestonehistory', - name='docs', - field=models.ManyToManyField(to='doc.Document'), - ), - ] diff --git a/ietf/group/migrations/0015_2_add_docs_m2m_table.py b/ietf/group/migrations/0015_2_add_docs_m2m_table.py deleted file mode 100644 index e7c0e3bfbf..0000000000 --- a/ietf/group/migrations/0015_2_add_docs_m2m_table.py +++ /dev/null @@ -1,41 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-22 08:00 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0015_1_del_docs_m2m_table'), - ] - - # The implementation of AlterField in Django 1.11 applies - # 'ALTER TABLE
MODIFY ...;' in order to fix foregn keys - # to the altered field, but as it seems does _not_ fix up m2m - # intermediary tables in an equivalent manner, so here we remove and - # then recreate the m2m tables so they will have the appropriate field - # types. - - operations = [ - migrations.RemoveField( - model_name='groupmilestone', - name='docs', - ), - migrations.RemoveField( - model_name='groupmilestonehistory', - name='docs', - ), - migrations.AddField( - model_name='groupmilestone', - name='docs', - field=models.ManyToManyField(blank=True, to='doc.Document'), - ), - migrations.AddField( - model_name='groupmilestonehistory', - name='docs', - field=models.ManyToManyField(blank=True, to='doc.Document'), - ), - ] diff --git a/ietf/group/migrations/0016_copy_docs_m2m_table.py b/ietf/group/migrations/0016_copy_docs_m2m_table.py deleted file mode 100644 index 2871c34a95..0000000000 --- a/ietf/group/migrations/0016_copy_docs_m2m_table.py +++ /dev/null @@ -1,58 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-27 05:57 - - -import sys, time - -from tqdm import tqdm - -from django.db import migrations - - -def forward(apps, schema_editor): - - GroupMilestone = apps.get_model('group', 'GroupMilestone') - GroupMilestoneDocs = apps.get_model('group', 'GroupMilestoneDocs') - GroupMilestoneHistory = apps.get_model('group', 'GroupMilestoneHistory') - GroupMilestoneHistoryDocs = apps.get_model('group', 'GroupMilestoneHistoryDocs') - - # Document id fixup ------------------------------------------------------------ - - sys.stderr.write('\n') - - sys.stderr.write(' %s.%s:\n' % (GroupMilestone.__name__, 'docs')) - for m in tqdm(GroupMilestone.objects.all()): - m.docs.set([ d.document for d in GroupMilestoneDocs.objects.filter(groupmilestone=m) ]) - - sys.stderr.write(' %s.%s:\n' % (GroupMilestoneHistory.__name__, 'docs')) - for m in tqdm(GroupMilestoneHistory.objects.all()): - m.docs.set([ d.document for d in GroupMilestoneHistoryDocs.objects.filter(groupmilestonehistory=m) ]) - - -def reverse(apps, schema_editor): - pass - -def timestamp(apps, schema_editor): - sys.stderr.write('\n %s' % time.strftime('%Y-%m-%d %H:%M:%S')) - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0015_2_add_docs_m2m_table'), - ] - - operations = [ - #migrations.RunPython(forward, reverse), - migrations.RunPython(timestamp, timestamp), - migrations.RunSQL( - "INSERT INTO group_groupmilestone_docs SELECT * FROM group_groupmilestonedocs;", - "" - ), - migrations.RunPython(timestamp, timestamp), - migrations.RunSQL( - "INSERT INTO group_groupmilestonehistory_docs SELECT * FROM group_groupmilestonehistorydocs;", - "" - ), - migrations.RunPython(timestamp, timestamp), - ] diff --git a/ietf/group/migrations/0017_remove_docs2_m2m.py b/ietf/group/migrations/0017_remove_docs2_m2m.py deleted file mode 100644 index 00a44303db..0000000000 --- a/ietf/group/migrations/0017_remove_docs2_m2m.py +++ /dev/null @@ -1,46 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-30 03:23 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0016_copy_docs_m2m_table'), - ] - - operations = [ - migrations.RemoveField( - model_name='groupmilestonedocs', - name='document', - ), - migrations.RemoveField( - model_name='groupmilestonedocs', - name='groupmilestone', - ), - migrations.RemoveField( - model_name='groupmilestonehistorydocs', - name='document', - ), - migrations.RemoveField( - model_name='groupmilestonehistorydocs', - name='groupmilestonehistory', - ), - migrations.RemoveField( - model_name='groupmilestone', - name='docs2', - ), - migrations.RemoveField( - model_name='groupmilestonehistory', - name='docs2', - ), - migrations.DeleteModel( - name='GroupMilestoneDocs', - ), - migrations.DeleteModel( - name='GroupMilestoneHistoryDocs', - ), - ] diff --git a/ietf/group/migrations/0018_remove_old_document_field.py b/ietf/group/migrations/0018_remove_old_document_field.py deleted file mode 100644 index 5a47e545c5..0000000000 --- a/ietf/group/migrations/0018_remove_old_document_field.py +++ /dev/null @@ -1,20 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-25 06:51 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0017_remove_docs2_m2m'), - ] - - operations = [ - migrations.RemoveField( - model_name='group', - name='charter', - ), - ] diff --git a/ietf/group/migrations/0019_rename_field_document2.py b/ietf/group/migrations/0019_rename_field_document2.py deleted file mode 100644 index 438c669d26..0000000000 --- a/ietf/group/migrations/0019_rename_field_document2.py +++ /dev/null @@ -1,22 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-25 06:52 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0019_rename_field_document2'), - ('group', '0018_remove_old_document_field'), - ] - - operations = [ - migrations.RenameField( - model_name='group', - old_name='charter2', - new_name='charter', - ), - ] diff --git a/ietf/group/migrations/0020_add_uses_milestone_dates.py b/ietf/group/migrations/0020_add_uses_milestone_dates.py deleted file mode 100644 index a14b49fbec..0000000000 --- a/ietf/group/migrations/0020_add_uses_milestone_dates.py +++ /dev/null @@ -1,26 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.25 on 2019-10-30 11:41 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0019_rename_field_document2'), - ] - - operations = [ - migrations.AddField( - model_name='group', - name='uses_milestone_dates', - field=models.BooleanField(default=False), - ), - migrations.AddField( - model_name='grouphistory', - name='uses_milestone_dates', - field=models.BooleanField(default=False), - ), - ] diff --git a/ietf/group/migrations/0021_add_order_to_milestones.py b/ietf/group/migrations/0021_add_order_to_milestones.py deleted file mode 100644 index 22f40af6c3..0000000000 --- a/ietf/group/migrations/0021_add_order_to_milestones.py +++ /dev/null @@ -1,44 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.25 on 2019-10-30 13:37 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0020_add_uses_milestone_dates'), - ] - - operations = [ - migrations.AlterModelOptions( - name='groupmilestone', - options={'ordering': ['order', 'due', 'id']}, - ), - migrations.AlterModelOptions( - name='groupmilestonehistory', - options={'ordering': ['order', 'due', 'id']}, - ), - migrations.AddField( - model_name='groupmilestone', - name='order', - field=models.IntegerField(blank=True, null=True), - ), - migrations.AddField( - model_name='groupmilestonehistory', - name='order', - field=models.IntegerField(blank=True, null=True), - ), - migrations.AlterField( - model_name='groupmilestone', - name='due', - field=models.DateField(blank=True, null=True), - ), - migrations.AlterField( - model_name='groupmilestonehistory', - name='due', - field=models.DateField(blank=True, null=True), - ), - ] diff --git a/ietf/group/migrations/0022_populate_uses_milestone_dates.py b/ietf/group/migrations/0022_populate_uses_milestone_dates.py deleted file mode 100644 index a7b7d3eaa7..0000000000 --- a/ietf/group/migrations/0022_populate_uses_milestone_dates.py +++ /dev/null @@ -1,30 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.25 on 2019-10-30 11:42 - - -from django.db import migrations - -def forward(apps, schema_editor): - Group = apps.get_model('group','Group') - GroupHistory = apps.get_model('group','GroupHistory') - - Group.objects.filter(type__features__has_milestones=True).update(uses_milestone_dates=True) - GroupHistory.objects.filter(type__features__has_milestones=True).update(uses_milestone_dates=True) - -def reverse(apps, schema_editor): - Group = apps.get_model('group','Group') - GroupHistory = apps.get_model('group','GroupHistory') - - Group.objects.filter(type__features__has_milestones=True).update(uses_milestone_dates=False) - GroupHistory.objects.filter(type__features__has_milestones=True).update(uses_milestone_dates=False) - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0021_add_order_to_milestones'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/group/migrations/0023_use_milestone_dates_default_to_true.py b/ietf/group/migrations/0023_use_milestone_dates_default_to_true.py deleted file mode 100644 index 440fce54a5..0000000000 --- a/ietf/group/migrations/0023_use_milestone_dates_default_to_true.py +++ /dev/null @@ -1,26 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright The IETF Trust 2020, All Rights Reserved -# Generated by Django 1.11.28 on 2020-02-11 07:47 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0022_populate_uses_milestone_dates'), - ] - - operations = [ - migrations.AlterField( - model_name='group', - name='uses_milestone_dates', - field=models.BooleanField(default=True), - ), - migrations.AlterField( - model_name='grouphistory', - name='uses_milestone_dates', - field=models.BooleanField(default=True), - ), - ] diff --git a/ietf/group/migrations/0024_add_groupman_authroles.py b/ietf/group/migrations/0024_add_groupman_authroles.py deleted file mode 100644 index 5c64618054..0000000000 --- a/ietf/group/migrations/0024_add_groupman_authroles.py +++ /dev/null @@ -1,26 +0,0 @@ -# -*- coding: utf-8 -*- -# Generated by Django 1.11.29 on 2020-05-04 13:10 -from __future__ import unicode_literals - -from django.db import migrations -import jsonfield.fields - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0023_use_milestone_dates_default_to_true'), - ] - - operations = [ - migrations.AddField( - model_name='groupfeatures', - name='groupman_authroles', - field=jsonfield.fields.JSONField(default=['Secretariat'], max_length=128), - ), - migrations.AddField( - model_name='historicalgroupfeatures', - name='groupman_authroles', - field=jsonfield.fields.JSONField(default=['Secretariat'], max_length=128), - ), - ] diff --git a/ietf/group/migrations/0025_populate_groupman_authroles.py b/ietf/group/migrations/0025_populate_groupman_authroles.py deleted file mode 100644 index d024c8bd2d..0000000000 --- a/ietf/group/migrations/0025_populate_groupman_authroles.py +++ /dev/null @@ -1,47 +0,0 @@ -# -*- coding: utf-8 -*- -# Generated by Django 1.11.29 on 2020-05-01 12:54 -from __future__ import unicode_literals - -from django.db import migrations - -authroles_map = { - 'adhoc': ['Secretariat'], - 'admin': ['Secretariat'], - 'ag': ['Secretariat', 'Area Director'], - 'area': ['Secretariat'], - 'dir': ['Secretariat'], - 'iab': ['Secretariat'], - 'iana': ['Secretariat'], - 'iesg': ['Secretariat'], - 'ietf': ['Secretariat'], - 'individ': [], - 'irtf': ['Secretariat'], - 'ise': ['Secretariat'], - 'isoc': ['Secretariat'], - 'nomcom': ['Secretariat'], - 'program': ['Secretariat', 'IAB'], - 'review': ['Secretariat'], - 'rfcedtyp': ['Secretariat'], - 'rg': ['Secretariat', 'IRTF Chair'], - 'sdo': ['Secretariat'], - 'team': ['Secretariat'], - 'wg': ['Secretariat', 'Area Director'], -} - -def forward(apps, schema_editor): - GroupFeatures = apps.get_model('group', 'GroupFeatures') - for type_id, authroles in authroles_map.items(): - GroupFeatures.objects.filter(type_id=type_id).update(groupman_authroles=authroles) - -def reverse(apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0024_add_groupman_authroles'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/group/migrations/0026_programs_meet.py b/ietf/group/migrations/0026_programs_meet.py deleted file mode 100644 index 4fb6fda06b..0000000000 --- a/ietf/group/migrations/0026_programs_meet.py +++ /dev/null @@ -1,36 +0,0 @@ -# -*- coding: utf-8 -*- -# Generated by Django 1.11.29 on 2020-05-01 12:54 -from __future__ import unicode_literals - -from django.db import migrations - - -def forward(apps, schema_editor): - GroupFeatures = apps.get_model('group', 'GroupFeatures') - program = GroupFeatures.objects.get(type_id='program') - program.has_meetings = True - program.matman_roles = ['lead', 'chair', 'secr'] - program.docman_roles = ['lead', 'chair', 'secr'] - program.groupman_roles = ['lead', 'chair', 'secr'] - program.role_order = ['lead', 'chair', 'secr'] - program.save() - -def reverse(apps, schema_editor): - GroupFeatures = apps.get_model('group', 'GroupFeatures') - program = GroupFeatures.objects.get(type_id='program') - program.has_meetings = False - program.matman_roles = ['lead', 'secr'] - program.docman_roles = ['lead', 'secr'] - program.groupman_roles = ['lead', 'secr'] - program.role_order = ['lead', 'secr'] - program.save() - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0025_populate_groupman_authroles'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/group/migrations/0027_programs_have_parents.py b/ietf/group/migrations/0027_programs_have_parents.py deleted file mode 100644 index d05d020092..0000000000 --- a/ietf/group/migrations/0027_programs_have_parents.py +++ /dev/null @@ -1,23 +0,0 @@ -# -*- coding: utf-8 -*- -# Generated by Django 1.11.29 on 2020-05-08 09:02 -from __future__ import unicode_literals - -from django.db import migrations - -def forward(apps, schema_editor): - Group = apps.get_model('group','Group') - iab = Group.objects.get(acronym='iab') - Group.objects.filter(type_id='program').update(parent=iab) - -def reverse(apps, schema_editor): - pass # No point in removing the parents - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0026_programs_meet'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/group/migrations/0028_add_robots.py b/ietf/group/migrations/0028_add_robots.py deleted file mode 100644 index ab4c8749bc..0000000000 --- a/ietf/group/migrations/0028_add_robots.py +++ /dev/null @@ -1,72 +0,0 @@ -# -*- coding: utf-8 -*- -# Generated by Django 1.11.29 on 2020-05-01 12:54 -from __future__ import unicode_literals - -from django.db import migrations -from django.utils.text import slugify - -robots = [ - ('Mail Archive', '/api/v2/person/person', 'secretariat', 'mailarchive@ietf.org'), - ('Registration System', '/api/notify/meeting/registration', 'secretariat', 'registration@ietf.org'), -] - -def forward(apps, schema_editor): - RoleName = apps.get_model('name', 'RoleName') - Role = apps.get_model('group', 'Role') - Group = apps.get_model('group', 'Group') - User = apps.get_model('auth', 'User') - Person = apps.get_model('person', 'Person') - Email = apps.get_model('person', 'Email') - PersonalApiKey = apps.get_model('person', 'PersonalApiKey') - # - rname, __ = RoleName.objects.get_or_create(slug='robot') - # - for (name, endpoint, acronym, address) in robots: - first_name, last_name = name.rsplit(None, 1) - user, created = User.objects.get_or_create(username=slugify(name)) - if created: - user.first_name=first_name - user.last_name=last_name - user.is_staff=True - user.is_active=True - user.save() - # - person, created = Person.objects.get_or_create(name=name) - if created: - person.user = user - person.ascii = name - person.consent = True - person.save() - else: - assert person.user == user - # - email, created = Email.objects.get_or_create(address=address) - if created: - email.origin = 'registration' - email.person = person - email.active = True - email.save() - else: - assert email.person == person - # - group = Group.objects.get(acronym=acronym) - role, created = Role.objects.get_or_create(person=person, name=rname, group=group, email=email) - # - key, created = PersonalApiKey.objects.get_or_create(person=person, endpoint=endpoint, valid=True) - -def reverse(apps, schema_editor): - Person = apps.get_model('person', 'Person') - for (name, endpoint, acronym, address) in robots: - deleted = Person.objects.filter(name=name).delete() - print('deleted: %s' % deleted) - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0027_programs_have_parents'), - ('name', '0012_role_name_robots'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/group/migrations/0029_add_used_roles_and_default_used_roles.py b/ietf/group/migrations/0029_add_used_roles_and_default_used_roles.py deleted file mode 100644 index 6f4a562c61..0000000000 --- a/ietf/group/migrations/0029_add_used_roles_and_default_used_roles.py +++ /dev/null @@ -1,34 +0,0 @@ -# Generated by Django 2.0.13 on 2020-05-22 12:00 - -from django.db import migrations -import jsonfield.fields - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0028_add_robots'), - ] - - operations = [ - migrations.AddField( - model_name='group', - name='used_roles', - field=jsonfield.fields.JSONField(default=[], help_text="Leave an empty list to get the group_type's default used roles", max_length=128), - ), - migrations.AddField( - model_name='groupfeatures', - name='default_used_roles', - field=jsonfield.fields.JSONField(default=[], max_length=128), - ), - migrations.AddField( - model_name='grouphistory', - name='used_roles', - field=jsonfield.fields.JSONField(default=[], help_text="Leave an empty list to get the group_type's default used roles", max_length=128), - ), - migrations.AddField( - model_name='historicalgroupfeatures', - name='default_used_roles', - field=jsonfield.fields.JSONField(default=[], max_length=128), - ), - ] diff --git a/ietf/group/migrations/0030_populate_default_used_roles.py b/ietf/group/migrations/0030_populate_default_used_roles.py deleted file mode 100644 index e6ba3a0fc1..0000000000 --- a/ietf/group/migrations/0030_populate_default_used_roles.py +++ /dev/null @@ -1,46 +0,0 @@ -# Generated by Django 2.0.13 on 2020-05-22 11:41 - -from django.db import migrations - -grouptype_defaults = { - 'adhoc': ['matman', 'ad', 'chair', 'lead'], - 'admin': ['member', 'chair'], - 'ag': ['ad', 'chair', 'secr'], - 'area': ['ad'], - 'dir': ['ad', 'chair', 'reviewer', 'secr'], - 'review': ['ad', 'chair', 'reviewer', 'secr'], - 'iab': ['chair'], - 'iana': ['auth'], - 'iesg': [], - 'ietf': ['ad', 'member', 'comdir', 'delegate', 'execdir', 'recman', 'secr', 'trac-editor', 'trac-admin', 'chair'], - 'individ': ['ad'], - 'irtf': ['member', 'atlarge', 'chair'], - 'ise': ['chair'], - 'isoc': ['chair', 'ceo'], - 'nomcom': ['member', 'advisor', 'liaison', 'chair', 'techadv'], - 'program': ['member', 'chair', 'lead'], - 'rfcedtyp': ['auth', 'chair'], - 'rg': ['chair', 'techadv', 'secr', 'delegate'], - 'sdo': ['liaiman', 'ceo', 'coord', 'auth', 'chair'], - 'team': ['ad', 'member', 'delegate', 'secr', 'liaison', 'atlarge', 'chair', 'matman', 'techadv'], - 'wg': ['ad', 'editor', 'delegate', 'secr', 'chair', 'matman', 'techadv'], -} - -def forward(apps, schema_editor): - GroupFeatures = apps.get_model('group','GroupFeatures') - for type_id, roles in grouptype_defaults.items(): - GroupFeatures.objects.filter(type_id=type_id).update(default_used_roles=roles) - -def reverse(apps, schema_editor): - pass # intentional - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0029_add_used_roles_and_default_used_roles'), - ('stats', '0003_meetingregistration_attended'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/group/migrations/0031_allow_blank_used_roles.py b/ietf/group/migrations/0031_allow_blank_used_roles.py deleted file mode 100644 index 331727f6f3..0000000000 --- a/ietf/group/migrations/0031_allow_blank_used_roles.py +++ /dev/null @@ -1,24 +0,0 @@ -# Generated by Django 2.0.13 on 2020-06-15 11:10 - -from django.db import migrations -import jsonfield.fields - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0030_populate_default_used_roles'), - ] - - operations = [ - migrations.AlterField( - model_name='group', - name='used_roles', - field=jsonfield.fields.JSONField(blank=True, default=[], help_text="Leave an empty list to get the group_type's default used roles", max_length=128), - ), - migrations.AlterField( - model_name='grouphistory', - name='used_roles', - field=jsonfield.fields.JSONField(blank=True, default=[], help_text="Leave an empty list to get the group_type's default used roles", max_length=128), - ), - ] diff --git a/ietf/group/migrations/0032_add_meeting_seen_as_area.py b/ietf/group/migrations/0032_add_meeting_seen_as_area.py deleted file mode 100644 index 9be9ebe0fb..0000000000 --- a/ietf/group/migrations/0032_add_meeting_seen_as_area.py +++ /dev/null @@ -1,37 +0,0 @@ -# Copyright The IETF Trust 2020', 'All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.27 on 2020-02-12 07:11 -from __future__ import unicode_literals - -from django.db import migrations, models - - -def forward(apps, schema_editor): - Group = apps.get_model('group', 'Group') - initial_area_groups = ['dispatch', 'gendispatch', 'intarea', 'opsarea', 'opsawg', 'rtgarea', 'rtgwg', 'saag', 'secdispatch', 'tsvarea', 'irtfopen'] - Group.objects.filter(acronym__in=initial_area_groups).update(meeting_seen_as_area=True) - - -def reverse(apps, schema_editor): - pass - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0031_allow_blank_used_roles'), - ] - - operations = [ - migrations.AddField( - model_name='group', - name='meeting_seen_as_area', - field=models.BooleanField(default=False, help_text='For meeting scheduling, should be considered an area meeting, even if the type is WG'), - ), - migrations.AddField( - model_name='grouphistory', - name='meeting_seen_as_area', - field=models.BooleanField(default=False, help_text='For meeting scheduling, should be considered an area meeting, even if the type is WG'), - ), - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/group/migrations/0033_extres.py b/ietf/group/migrations/0033_extres.py deleted file mode 100644 index 2e3d037a2a..0000000000 --- a/ietf/group/migrations/0033_extres.py +++ /dev/null @@ -1,28 +0,0 @@ -# -*- coding: utf-8 -*- -# Generated by Django 1.11.29 on 2020-04-15 10:20 -from __future__ import unicode_literals - -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0014_extres'), - ('group', '0032_add_meeting_seen_as_area'), - ] - - operations = [ - migrations.CreateModel( - name='GroupExtResource', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('display_name', models.CharField(blank=True, default='', max_length=255)), - ('value', models.CharField(max_length=2083)), - ('group', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='group.Group')), - ('name', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ExtResourceName')), - ], - ), - ] diff --git a/ietf/group/migrations/0034_populate_groupextresources.py b/ietf/group/migrations/0034_populate_groupextresources.py deleted file mode 100644 index a8c0e1b989..0000000000 --- a/ietf/group/migrations/0034_populate_groupextresources.py +++ /dev/null @@ -1,120 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.29 on 2020-03-19 13:06 -from __future__ import unicode_literals - -import re - -import debug # pyflakes:ignore - -from collections import OrderedDict, Counter -from io import StringIO - -from django.db import migrations -from django.core.exceptions import ValidationError - - -from ietf.utils.validators import validate_external_resource_value - -name_map = { - "Issue.*": "tracker", - ".*FAQ.*": "faq", - ".*Area Web Page": "webpage", - ".*Wiki": "wiki", - "Home Page": "webpage", - "Slack.*": "slack", - "Additional .* Web Page": "webpage", - "Additional .* Page": "webpage", - "Yang catalog entry.*": "yc_entry", - "Yang impact analysis.*": "yc_impact", - "GitHub": "github_repo", - "Github page": "github_repo", - "GitHub repo.*": "github_repo", - "Github repository.*": "github_repo", - "GitHub org.*": "github_org", - "GitHub User.*": "github_username", - "GitLab User": "gitlab_username", - "GitLab User Name": "gitlab_username", -} - -url_map = OrderedDict({ - "https?://github\\.com": "github_repo", - "https?://trac\\.ietf\\.org/.*/wiki": "wiki", - "ietf\\.org.*/trac/wiki": "wiki", - "trac.*wiki": "wiki", - "www\\.ietf\\.org/mailman" : "mailing_list", - "www\\.ietf\\.org/mail-archive" : "mailing_list_archive", - "ietf\\.org/logs": "jabber_log", - "ietf\\.org/jabber/logs": "jabber_log", - "xmpp:.*?join": "jabber_room", - "https?://.*": "webpage" -}) - -def forward(apps, schema_editor): - GroupExtResource = apps.get_model('group', 'GroupExtResource') - ExtResourceName = apps.get_model('name', 'ExtResourceName') - GroupUrl = apps.get_model('group', 'GroupUrl') - - stats = Counter() - stats_file = StringIO() - - for group_url in GroupUrl.objects.all(): - group_url.url = group_url.url.strip() - match_found = False - for regext,slug in name_map.items(): - if re.fullmatch(regext, group_url.name): - match_found = True - stats['mapped'] += 1 - name = ExtResourceName.objects.get(slug=slug) - try: - validate_external_resource_value(name, group_url.url) - GroupExtResource.objects.create(group=group_url.group, name_id=slug, value=group_url.url, display_name=group_url.name) - except ValidationError as e: # pyflakes:ignore - print("Failed validation:", group_url.url, e, file=stats_file) - stats['failed_validation'] +=1 - break - if not match_found: - for regext, slug in url_map.items(): - if re.search(regext, group_url.url): - match_found = True - if slug: - stats['mapped'] +=1 - name = ExtResourceName.objects.get(slug=slug) - # Munge the URL if it's the first github repo match - # Remove "/tree/master" substring if it exists - # Remove trailing "/issues" substring if it exists - # Remove "/blob/master/.*" pattern if present - if regext == "https?://github\\.com": - group_url.url = group_url.url.replace("/tree/master","") - group_url.url = re.sub('/issues$', '', group_url.url) - group_url.url = re.sub('/blob/master.*$', '', group_url.url) - try: - validate_external_resource_value(name, group_url.url) - GroupExtResource.objects.create(group=group_url.group, name=name, value=group_url.url, display_name=group_url.name) - except ValidationError as e: # pyflakes:ignore - print("Failed validation:", group_url.url, e, file=stats_file) - stats['failed_validation'] +=1 - else: - stats['ignored'] +=1 - break - if not match_found: - print("Not Mapped:",group_url.group.acronym, group_url.name, group_url.url, file=stats_file) - stats['not_mapped'] += 1 - print('') - print(stats_file.getvalue()) - print(stats) - -def reverse(apps, schema_editor): - GroupExtResource = apps.get_model('group', 'GroupExtResource') - GroupExtResource.objects.all().delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0033_extres'), - ('name', '0015_populate_extres'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/group/migrations/0035_add_research_area_groups.py b/ietf/group/migrations/0035_add_research_area_groups.py deleted file mode 100644 index 212ef30f7d..0000000000 --- a/ietf/group/migrations/0035_add_research_area_groups.py +++ /dev/null @@ -1,56 +0,0 @@ -# Generated by Django 2.2.14 on 2020-07-28 09:30 - -from django.db import migrations - -def forward(apps, schema_editor): - GroupFeatures = apps.get_model('group','GroupFeatures') - GroupFeatures.objects.create( - about_page = 'ietf.group.views.group_about', - acts_like_wg = True, - admin_roles = ['chair'], - agenda_type_id = 'ietf', - create_wiki = True, - custom_group_roles = True, - customize_workflow = False, - default_tab = 'ietf.group.views.group_about', - default_used_roles = ['chair', 'secr'], - docman_roles = ['chair', 'delegate', 'secr'], - groupman_authroles = ['Secretariat', 'IRTF Chair'], - groupman_roles = ['chair', 'delegate'], - has_chartering_process = False, - has_default_jabber = False, - has_documents = True, - has_meetings = True, - has_milestones = False, - has_nonsession_materials = False, - has_reviews = False, - has_session_materials = True, - is_schedulable = True, - material_types = ['slides'], - matman_roles = ['chair', 'delegate', 'secr'], - req_subm_approval = True, - role_order = ['chair', 'secr'], - show_on_agenda = True, - type_id = 'rag', - ) - - Group = apps.get_model('group','Group') - Group.objects.filter(type_id='ag',parent__acronym='irtf').update(type_id='rag') - -def reverse(apps, schema_editor): - Group = apps.get_model('group','Group') - Group.objects.filter(type_id='rag',parent__acronym='irtf').update(type_id='ag') - - GroupFeatures = apps.get_model('group','GroupFeatures') - GroupFeatures.objects.filter(type_id='rag').delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0034_populate_groupextresources'), - ('name', '0016_add_research_area_groups'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/group/migrations/0036_orgs_vs_repos.py b/ietf/group/migrations/0036_orgs_vs_repos.py deleted file mode 100644 index b764b86f44..0000000000 --- a/ietf/group/migrations/0036_orgs_vs_repos.py +++ /dev/null @@ -1,34 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved - -from urllib.parse import urlparse -from django.db import migrations - -def categorize(url): - # This will categorize a few urls pointing into files in a repo as a repo, but that's better than calling them an org - element_count = len(urlparse(url).path.strip('/').split('/')) - if element_count < 1: - print("Bad github resource:",url) - return 'github_org' if element_count == 1 else 'github_repo' - -def forward(apps, schema_editor): - GroupExtResource = apps.get_model('group','GroupExtResource') - - for resource in GroupExtResource.objects.filter(name_id__in=('github_org','github_repo',)): - category = categorize(resource.value) - if resource.name_id != category: - resource.name_id = category - resource.save() - -def reverse(apps, schema_editor): - # Intentionally don't try to return to former worse state - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0035_add_research_area_groups'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/group/migrations/0037_initial_yc_roles.py b/ietf/group/migrations/0037_initial_yc_roles.py deleted file mode 100644 index 20eb471d60..0000000000 --- a/ietf/group/migrations/0037_initial_yc_roles.py +++ /dev/null @@ -1,70 +0,0 @@ -# Copyright The IETF Trust 2020 All Rights Reserved - -from django.db import migrations - -def forward(apps, schema_editor): - RoleName = apps.get_model('name','RoleName') - Group = apps.get_model('group','Group') - Role = apps.get_model('group','Role') - Person = apps.get_model('person','Person') - - RoleName.objects.create( - slug = 'yc_operator', - name = 'YangCatalog Operator', - desc = 'Can grant user api rights and browse the YangCatalog directory structure', - ) - - ycsupport = Group.objects.create( - acronym='ycsupport', - name="YangCatalog Support", - state_id='active', - type_id='team', - parent = Group.objects.get(acronym='ops'), - description = "Team for supporting YangCatalog.org operations", - ) - - RoleName.objects.create( - slug = 'yc_admin', - name = 'YangCatalog Administrator', - desc = 'Can operate the YangCatalog, change its configuration, and edit its data', - ) - - for name,role_name_id in ( - ('Robert Sparks','yc_operator'), - ('Benoit Claise','yc_operator'), - ('Eric Vyncke','yc_operator'), - ('Miroslav Kovac','yc_admin'), - ('Slavomir Mazur','yc_admin'), - ): - person = Person.objects.get(name=name) - email = person.email_set.filter(primary=True).first() - if not email: - email = person.email_set.filter(active=True).order_by("-time").first() - Role.objects.create( - name_id = role_name_id, - group = ycsupport, - person = person, - email = email, - ) - -def reverse(apps, schema_editor): - RoleName = apps.get_model('name','RoleName') - Group = apps.get_model('group','Group') - Role = apps.get_model('group','Role') - - Role.objects.filter(name_id__in = ( 'yc_operator' , 'yc_admin' )).delete() - Group.objects.filter(acronym='ycsupport').delete() - RoleName.objects.filter(slug__in=( 'yc_operator' , 'yc_admin' )).delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0036_orgs_vs_repos'), - ('name', '0020_add_rescheduled_session_name'), - ('person','0016_auto_20200807_0750'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/group/migrations/0038_auto_20201109_0439.py b/ietf/group/migrations/0038_auto_20201109_0439.py deleted file mode 100644 index 26bfd03eda..0000000000 --- a/ietf/group/migrations/0038_auto_20201109_0439.py +++ /dev/null @@ -1,17 +0,0 @@ -# Generated by Django 2.2.17 on 2020-11-09 04:39 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0037_initial_yc_roles'), - ] - - operations = [ - migrations.AddIndex( - model_name='groupevent', - index=models.Index(fields=['-time', '-id'], name='group_group_time_ee7c7c_idx'), - ), - ] diff --git a/ietf/group/migrations/0039_remove_historicalgroupfeatures.py b/ietf/group/migrations/0039_remove_historicalgroupfeatures.py deleted file mode 100644 index 669cde8750..0000000000 --- a/ietf/group/migrations/0039_remove_historicalgroupfeatures.py +++ /dev/null @@ -1,48 +0,0 @@ -# Generated by Django 2.2.17 on 2020-12-18 08:54 - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0038_auto_20201109_0439'), - ] - - operations = [ - migrations.DeleteModel( - name='HistoricalGroupFeatures', - ), - ] - -# In case these are ever needed again, here's what had been captured. -# Note that many of the values are not well formed as JSONFields and need to be manually corrected if they are brought back. -# -# -- Table structure for table `group_historicalgroupfeatures` -# DROP TABLE IF EXISTS `group_historicalgroupfeatures`; -# CREATE TABLE `group_historicalgroupfeatures` ( -# KEY `group_historicalgroupfeatures_agenda_type_id_089e752b` (`agenda_type_id`), -# KEY `group_historicalgroupfeatures_history_user_id_0d1368d2` (`history_user_id`), -# KEY `group_historicalgroupfeatures_type_id_4ed21f10` (`type_id`) -# -- Dumping data for table `group_historicalgroupfeatures` -# LOCK TABLES `group_historicalgroupfeatures` WRITE; -# /*!40000 ALTER TABLE `group_historicalgroupfeatures` DISABLE KEYS */; -# INSERT INTO `group_historicalgroupfeatures` (`has_milestones`, `has_chartering_process`, `has_documents`, `has_nonsession_materials`, `has_meetings`, `has_reviews`, `has_default_jabber`, `customize_workflow`, `about_page`, `default_tab`, `material_types`, `admin_roles`, `history_id`, `history_change_reason`, `history_date`, `history_type`, `agenda_type_id`, `history_user_id`, `type_id`, `acts_like_wg`, `create_wiki`, `custom_group_roles`, `has_session_materials`, `is_schedulable`, `role_order`, `show_on_agenda`, `req_subm_approval`, `matman_roles`, `docman_roles`, `groupman_roles`, `groupman_authroles`, `default_used_roles`) VALUES (0,0,0,0,1,0,0,0,'ietf.group.views.group_about','ietf.group.views.group_about','slides','chair,lead',1,NULL,'2018-07-15 08:23:57.183851','~','ietf',433,'ietf',0,0,0,0,0,'chair,secr,member',0,0,'ad,chair,delegate,secr','[\"ad\",\"chair\",\"delegate\",\"secr\"]','[\"ad\",\"chair\"]','[\"Secretariat\"]','[]'); -# INSERT INTO `group_historicalgroupfeatures` (`has_milestones`, `has_chartering_process`, `has_documents`, `has_nonsession_materials`, `has_meetings`, `has_reviews`, `has_default_jabber`, `customize_workflow`, `about_page`, `default_tab`, `material_types`, `admin_roles`, `history_id`, `history_change_reason`, `history_date`, `history_type`, `agenda_type_id`, `history_user_id`, `type_id`, `acts_like_wg`, `create_wiki`, `custom_group_roles`, `has_session_materials`, `is_schedulable`, `role_order`, `show_on_agenda`, `req_subm_approval`, `matman_roles`, `docman_roles`, `groupman_roles`, `groupman_authroles`, `default_used_roles`) VALUES (0,0,0,0,0,0,0,0,'ietf.group.views.group_about','ietf.group.views.group_about','slides','ad',2,NULL,'2018-07-19 13:07:33.440449','~','ietf',433,'area',0,0,0,0,0,'chair,secr,member',0,0,'ad,chair,delegate,secr','[\"ad\",\"chair\",\"delegate\",\"secr\"]','[\"ad\",\"chair\"]','[\"Secretariat\"]','[]'); -# INSERT INTO `group_historicalgroupfeatures` (`has_milestones`, `has_chartering_process`, `has_documents`, `has_nonsession_materials`, `has_meetings`, `has_reviews`, `has_default_jabber`, `customize_workflow`, `about_page`, `default_tab`, `material_types`, `admin_roles`, `history_id`, `history_change_reason`, `history_date`, `history_type`, `agenda_type_id`, `history_user_id`, `type_id`, `acts_like_wg`, `create_wiki`, `custom_group_roles`, `has_session_materials`, `is_schedulable`, `role_order`, `show_on_agenda`, `req_subm_approval`, `matman_roles`, `docman_roles`, `groupman_roles`, `groupman_authroles`, `default_used_roles`) VALUES (0,0,1,0,1,0,0,0,'ietf.group.views.group_about','ietf.group.views.group_about','[\"slides\"]','[\"chair\"]',3,NULL,'2019-02-04 07:41:05.566267','~','ietf',433,'ag',1,1,1,1,1,'[\"chair\",\"secr\"]',1,1,'[\"ad\",\"chair\",\"delegate\",\"secr\"]','[\"chair\",\"delegate\",\"secr\"]','[\"ad\",\"chair\",\"delegate\"]','[\"Secretariat\"]','[]'); -# INSERT INTO `group_historicalgroupfeatures` (`has_milestones`, `has_chartering_process`, `has_documents`, `has_nonsession_materials`, `has_meetings`, `has_reviews`, `has_default_jabber`, `customize_workflow`, `about_page`, `default_tab`, `material_types`, `admin_roles`, `history_id`, `history_change_reason`, `history_date`, `history_type`, `agenda_type_id`, `history_user_id`, `type_id`, `acts_like_wg`, `create_wiki`, `custom_group_roles`, `has_session_materials`, `is_schedulable`, `role_order`, `show_on_agenda`, `req_subm_approval`, `matman_roles`, `docman_roles`, `groupman_roles`, `groupman_authroles`, `default_used_roles`) VALUES (0,0,0,0,1,0,1,0,'ietf.group.views.group_about','ietf.group.views.group_about','\"[\\\"slides\\\"]\"','\"[\\\"chair\\\"]\"',4,NULL,'2019-03-07 14:32:17.595737','~','ietf',433,'adhoc',0,1,0,1,1,'\"[\\\"chair\\\",\\\"delegate\\\",\\\"matman\\\"]\"',1,1,'\"[\\\"chair\\\",\\\"delegate\\\",\\\"matman\\\"]\"','\"[\\\"chair\\\"]\"','\"[\\\"chair\\\",\\\"delegate\\\"]\"','[\"Secretariat\"]','[]'); -# INSERT INTO `group_historicalgroupfeatures` (`has_milestones`, `has_chartering_process`, `has_documents`, `has_nonsession_materials`, `has_meetings`, `has_reviews`, `has_default_jabber`, `customize_workflow`, `about_page`, `default_tab`, `material_types`, `admin_roles`, `history_id`, `history_change_reason`, `history_date`, `history_type`, `agenda_type_id`, `history_user_id`, `type_id`, `acts_like_wg`, `create_wiki`, `custom_group_roles`, `has_session_materials`, `is_schedulable`, `role_order`, `show_on_agenda`, `req_subm_approval`, `matman_roles`, `docman_roles`, `groupman_roles`, `groupman_authroles`, `default_used_roles`) VALUES (0,0,0,0,1,0,1,0,'ietf.group.views.group_about','ietf.group.views.group_about','\"[\\\"slides\\\"]\"','\"[\\\"chair\\\"]\"',5,NULL,'2019-03-07 15:13:37.631043','~','ietf',433,'adhoc',0,1,0,1,1,'\"[\\\"chair\\\",\\\"lead\\\",\\\"delegate\\\",\\\"matman\\\"]\"',1,1,'\"[\\\"chair\\\",\\\"lead\\\",\\\"delegate\\\",\\\"matman\\\"]\"','\"[\\\"chair\\\"]\"','\"[\\\"chair\\\",\\\"lead\\\",\\\"delegate\\\"]\"','[\"Secretariat\"]','[]'); -# INSERT INTO `group_historicalgroupfeatures` (`has_milestones`, `has_chartering_process`, `has_documents`, `has_nonsession_materials`, `has_meetings`, `has_reviews`, `has_default_jabber`, `customize_workflow`, `about_page`, `default_tab`, `material_types`, `admin_roles`, `history_id`, `history_change_reason`, `history_date`, `history_type`, `agenda_type_id`, `history_user_id`, `type_id`, `acts_like_wg`, `create_wiki`, `custom_group_roles`, `has_session_materials`, `is_schedulable`, `role_order`, `show_on_agenda`, `req_subm_approval`, `matman_roles`, `docman_roles`, `groupman_roles`, `groupman_authroles`, `default_used_roles`) VALUES (0,0,0,1,1,0,0,0,'ietf.group.views.group_about','ietf.group.views.group_about','[\"slides\"]','[\"chair\"]',6,NULL,'2019-03-13 11:02:32.696034','~','ietf',433,'team',0,1,1,0,0,'[\"chair\",\"member\",\"matman\"]',0,0,'[\"chair\",\"matman\"]','[\"chair\"]','[\"chair\"]','[\"Secretariat\"]','[]'); -# INSERT INTO `group_historicalgroupfeatures` (`has_milestones`, `has_chartering_process`, `has_documents`, `has_nonsession_materials`, `has_meetings`, `has_reviews`, `has_default_jabber`, `customize_workflow`, `about_page`, `default_tab`, `material_types`, `admin_roles`, `history_id`, `history_change_reason`, `history_date`, `history_type`, `agenda_type_id`, `history_user_id`, `type_id`, `acts_like_wg`, `create_wiki`, `custom_group_roles`, `has_session_materials`, `is_schedulable`, `role_order`, `show_on_agenda`, `req_subm_approval`, `matman_roles`, `docman_roles`, `groupman_roles`, `groupman_authroles`, `default_used_roles`) VALUES (0,0,0,1,1,0,0,0,'ietf.group.views.group_about','ietf.group.views.group_about','[\"slides\"]','[\"chair\"]',7,NULL,'2019-03-13 13:59:38.964013','~','ietf',433,'team',0,1,1,0,0,'[\"chair\",\"member\",\"matman\"]',0,0,'[\"chair\",\"matman\"]','[\"chair\"]','[\"chair\"]','[\"Secretariat\"]','[]'); -# INSERT INTO `group_historicalgroupfeatures` (`has_milestones`, `has_chartering_process`, `has_documents`, `has_nonsession_materials`, `has_meetings`, `has_reviews`, `has_default_jabber`, `customize_workflow`, `about_page`, `default_tab`, `material_types`, `admin_roles`, `history_id`, `history_change_reason`, `history_date`, `history_type`, `agenda_type_id`, `history_user_id`, `type_id`, `acts_like_wg`, `create_wiki`, `custom_group_roles`, `has_session_materials`, `is_schedulable`, `role_order`, `show_on_agenda`, `req_subm_approval`, `matman_roles`, `docman_roles`, `groupman_roles`, `groupman_authroles`, `default_used_roles`) VALUES (0,0,0,0,1,0,1,0,'ietf.group.views.group_about','ietf.group.views.group_about','[\"slides\"]','[\"chair\"]',8,NULL,'2019-03-13 14:02:03.061530','~','ietf',433,'adhoc',0,1,0,1,1,'[\"chair\",\"lead\",\"delegate\",\"matman\"]',1,1,'[\"chair\",\"lead\",\"delegate\",\"matman\"]','[\"chair\"]','[\"chair\",\"lead\",\"delegate\"]','[\"Secretariat\"]','[]'); -# INSERT INTO `group_historicalgroupfeatures` (`has_milestones`, `has_chartering_process`, `has_documents`, `has_nonsession_materials`, `has_meetings`, `has_reviews`, `has_default_jabber`, `customize_workflow`, `about_page`, `default_tab`, `material_types`, `admin_roles`, `history_id`, `history_change_reason`, `history_date`, `history_type`, `agenda_type_id`, `history_user_id`, `type_id`, `acts_like_wg`, `create_wiki`, `custom_group_roles`, `has_session_materials`, `is_schedulable`, `role_order`, `show_on_agenda`, `req_subm_approval`, `matman_roles`, `docman_roles`, `groupman_roles`, `groupman_authroles`, `default_used_roles`) VALUES (0,0,0,0,0,0,0,0,'ietf.group.views.group_about','ietf.group.views.group_about','\"[]\"','[\"chair\"]',9,NULL,'2019-03-13 14:04:12.810180','~','ad',433,'iesg',0,0,1,0,0,'[\"chair\",\"delegate\",\"member\"]',0,1,'[\"chair\",\"delegate\",\"member\"]','[\"chair\"]','[\"chair\",\"delegate\"]','[\"Secretariat\"]','[]'); -# INSERT INTO `group_historicalgroupfeatures` (`has_milestones`, `has_chartering_process`, `has_documents`, `has_nonsession_materials`, `has_meetings`, `has_reviews`, `has_default_jabber`, `customize_workflow`, `about_page`, `default_tab`, `material_types`, `admin_roles`, `history_id`, `history_change_reason`, `history_date`, `history_type`, `agenda_type_id`, `history_user_id`, `type_id`, `acts_like_wg`, `create_wiki`, `custom_group_roles`, `has_session_materials`, `is_schedulable`, `role_order`, `show_on_agenda`, `req_subm_approval`, `matman_roles`, `docman_roles`, `groupman_roles`, `groupman_authroles`, `default_used_roles`) VALUES (0,0,0,0,0,0,0,0,'ietf.group.views.group_about','ietf.group.views.group_about','[\"slides\"]','[\"chair\",\"lead\"]',10,NULL,'2019-03-13 14:05:47.617726','~','ad',433,'ise',0,0,1,0,0,'[\"chair\",\"delegate\"]',0,1,'[\"chair\",\"delegate\"]','[\"chair\"]','[\"chair\",\"delegate\"]','[\"Secretariat\"]','[]'); -# INSERT INTO `group_historicalgroupfeatures` (`has_milestones`, `has_chartering_process`, `has_documents`, `has_nonsession_materials`, `has_meetings`, `has_reviews`, `has_default_jabber`, `customize_workflow`, `about_page`, `default_tab`, `material_types`, `admin_roles`, `history_id`, `history_change_reason`, `history_date`, `history_type`, `agenda_type_id`, `history_user_id`, `type_id`, `acts_like_wg`, `create_wiki`, `custom_group_roles`, `has_session_materials`, `is_schedulable`, `role_order`, `show_on_agenda`, `req_subm_approval`, `matman_roles`, `docman_roles`, `groupman_roles`, `groupman_authroles`, `default_used_roles`) VALUES (1,1,1,0,1,0,1,1,'ietf.group.views.group_about','ietf.group.views.group_documents','[\"slides\"]','[\"chair\"]',11,NULL,'2019-04-23 04:11:30.770056','~','ietf',433,'rg',1,1,0,1,1,'[\"chair\",\"delegate\",\"secr\"]',1,1,'[\"chair\",\"delegate\",\"secr\"]','[\"chair\",\"delegate\",\"secr\"]','[\"chair\",\"delegate\"]','[\"Secretariat\"]','[]'); -# INSERT INTO `group_historicalgroupfeatures` (`has_milestones`, `has_chartering_process`, `has_documents`, `has_nonsession_materials`, `has_meetings`, `has_reviews`, `has_default_jabber`, `customize_workflow`, `about_page`, `default_tab`, `material_types`, `admin_roles`, `history_id`, `history_change_reason`, `history_date`, `history_type`, `agenda_type_id`, `history_user_id`, `type_id`, `acts_like_wg`, `create_wiki`, `custom_group_roles`, `has_session_materials`, `is_schedulable`, `role_order`, `show_on_agenda`, `req_subm_approval`, `matman_roles`, `docman_roles`, `groupman_roles`, `groupman_authroles`, `default_used_roles`) VALUES (1,0,1,0,0,0,0,0,'ietf.group.views.group_about','ietf.group.views.group_about','[\"slides\"]','[\"lead\"]',12,NULL,'2019-04-24 04:03:41.967314','~','ad',433,'program',0,0,1,0,0,'[\"lead\",\"secr\"]',0,0,'[\"lead\",\"secr\"]','[\"lead\",\"secr\"]','[\"lead\",\"secr\"]','[\"Secretariat\"]','[]'); -# INSERT INTO `group_historicalgroupfeatures` (`has_milestones`, `has_chartering_process`, `has_documents`, `has_nonsession_materials`, `has_meetings`, `has_reviews`, `has_default_jabber`, `customize_workflow`, `about_page`, `default_tab`, `material_types`, `admin_roles`, `history_id`, `history_change_reason`, `history_date`, `history_type`, `agenda_type_id`, `history_user_id`, `type_id`, `acts_like_wg`, `create_wiki`, `custom_group_roles`, `has_session_materials`, `is_schedulable`, `role_order`, `show_on_agenda`, `req_subm_approval`, `matman_roles`, `docman_roles`, `groupman_roles`, `groupman_authroles`, `default_used_roles`) VALUES (0,0,0,0,0,0,0,0,'ietf.group.views.group_about','ietf.group.views.group_about','[\"slides\"]','[\"chair\",\"advisor\"]',13,NULL,'2019-04-29 04:44:26.522936','~','side',433,'nomcom',0,1,1,0,0,'[\"chair\",\"member\",\"advisor\"]',0,1,'[\"chair\"]','[\"chair\"]','[\"chair\",\"advisor\"]','[\"Secretariat\"]','[]'); -# INSERT INTO `group_historicalgroupfeatures` (`has_milestones`, `has_chartering_process`, `has_documents`, `has_nonsession_materials`, `has_meetings`, `has_reviews`, `has_default_jabber`, `customize_workflow`, `about_page`, `default_tab`, `material_types`, `admin_roles`, `history_id`, `history_change_reason`, `history_date`, `history_type`, `agenda_type_id`, `history_user_id`, `type_id`, `acts_like_wg`, `create_wiki`, `custom_group_roles`, `has_session_materials`, `is_schedulable`, `role_order`, `show_on_agenda`, `req_subm_approval`, `matman_roles`, `docman_roles`, `groupman_roles`, `groupman_authroles`, `default_used_roles`) VALUES (0,0,0,0,0,0,0,0,'ietf.group.views.group_about','ietf.group.views.group_about','[\"slides\"]','[\"chair\"]',14,NULL,'2019-06-26 13:28:11.695889','+','ietf',433,'admin',0,0,0,0,0,'[\"chair\"]',0,0,'[\"chair\"]','[\"chair\"]','[\"chair\"]','[\"Secretariat\"]','[]'); -# INSERT INTO `group_historicalgroupfeatures` (`has_milestones`, `has_chartering_process`, `has_documents`, `has_nonsession_materials`, `has_meetings`, `has_reviews`, `has_default_jabber`, `customize_workflow`, `about_page`, `default_tab`, `material_types`, `admin_roles`, `history_id`, `history_change_reason`, `history_date`, `history_type`, `agenda_type_id`, `history_user_id`, `type_id`, `acts_like_wg`, `create_wiki`, `custom_group_roles`, `has_session_materials`, `is_schedulable`, `role_order`, `show_on_agenda`, `req_subm_approval`, `matman_roles`, `docman_roles`, `groupman_roles`, `groupman_authroles`, `default_used_roles`) VALUES (0,0,0,0,0,0,0,0,'ietf.group.views.group_about','ietf.group.views.group_about','[\"slides\"]','[\"chair\"]',15,NULL,'2019-06-26 13:29:29.706999','+','ietf',433,'iana',0,0,0,0,0,'[\"chair\"]',0,0,'[\"chair\"]','[\"chair\"]','[\"chair\"]','[\"Secretariat\"]','[]'); -# INSERT INTO `group_historicalgroupfeatures` (`has_milestones`, `has_chartering_process`, `has_documents`, `has_nonsession_materials`, `has_meetings`, `has_reviews`, `has_default_jabber`, `customize_workflow`, `about_page`, `default_tab`, `material_types`, `admin_roles`, `history_id`, `history_change_reason`, `history_date`, `history_type`, `agenda_type_id`, `history_user_id`, `type_id`, `acts_like_wg`, `create_wiki`, `custom_group_roles`, `has_session_materials`, `is_schedulable`, `role_order`, `show_on_agenda`, `req_subm_approval`, `matman_roles`, `docman_roles`, `groupman_roles`, `groupman_authroles`, `default_used_roles`) VALUES (0,0,1,0,0,0,0,0,'ietf.group.views.group_about','ietf.group.views.group_about','[\"slides\"]','[\"chair\",\"lead\"]',16,NULL,'2019-07-17 04:20:03.049554','~','ad',433,'ise',0,0,1,0,0,'[\"chair\",\"delegate\"]',0,1,'[\"chair\",\"delegate\"]','[\"chair\"]','[\"chair\",\"delegate\"]','[\"Secretariat\"]','[]'); -# INSERT INTO `group_historicalgroupfeatures` (`has_milestones`, `has_chartering_process`, `has_documents`, `has_nonsession_materials`, `has_meetings`, `has_reviews`, `has_default_jabber`, `customize_workflow`, `about_page`, `default_tab`, `material_types`, `admin_roles`, `history_id`, `history_change_reason`, `history_date`, `history_type`, `agenda_type_id`, `history_user_id`, `type_id`, `acts_like_wg`, `create_wiki`, `custom_group_roles`, `has_session_materials`, `is_schedulable`, `role_order`, `show_on_agenda`, `req_subm_approval`, `matman_roles`, `docman_roles`, `groupman_roles`, `groupman_authroles`, `default_used_roles`) VALUES (0,0,0,0,1,1,0,0,'ietf.group.views.group_about','ietf.group.views.review_requests','[\n \"slides\"\n]','[\n \"chair\",\n \"secr\"\n]',17,NULL,'2020-11-23 10:19:39.119039','~','ietf',420,'review',0,1,1,0,0,'[\n \"chair\",\n \"secr\"\n]',0,1,'[\n \"ad\",\n \"secr\"\n]','[\n \"secr\"\n]','[\n \"ad\",\n \"secr\"\n]','[\n \"Secretariat\"\n]','[\n \"ad\",\n \"chair\",\n \"reviewer\",\n \"secr\"\n]'); -# /*!40000 ALTER TABLE `group_historicalgroupfeatures` ENABLE KEYS */; -# \ No newline at end of file diff --git a/ietf/group/migrations/0040_lengthen_used_roles_fields.py b/ietf/group/migrations/0040_lengthen_used_roles_fields.py deleted file mode 100644 index d33310495e..0000000000 --- a/ietf/group/migrations/0040_lengthen_used_roles_fields.py +++ /dev/null @@ -1,35 +0,0 @@ -# Generated by Django 2.2.17 on 2020-12-11 08:48 - -from django.db import migrations -import jsonfield.fields - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0039_remove_historicalgroupfeatures'), - ] - - operations = [ - migrations.AlterField( - model_name='group', - name='used_roles', - field=jsonfield.fields.JSONField(blank=True, default=[], help_text="Leave an empty list to get the group_type's default used roles", max_length=256), - ), - migrations.AlterField( - model_name='groupfeatures', - name='default_used_roles', - field=jsonfield.fields.JSONField(default=[], max_length=256), - ), - migrations.AlterField( - model_name='grouphistory', - name='used_roles', - field=jsonfield.fields.JSONField(blank=True, default=[], help_text="Leave an empty list to get the group_type's default used roles", max_length=256), - ), - # historicalgroupfeatures has been removed - # migrations.AlterField( - # model_name='historicalgroupfeatures', - # name='default_used_roles', - # field=jsonfield.fields.JSONField(default=[], max_length=256), - # ), - ] diff --git a/ietf/group/migrations/0041_create_liaison_contact_roles.py b/ietf/group/migrations/0041_create_liaison_contact_roles.py deleted file mode 100644 index 0194cc077f..0000000000 --- a/ietf/group/migrations/0041_create_liaison_contact_roles.py +++ /dev/null @@ -1,158 +0,0 @@ -# Generated by Django 2.2.17 on 2020-12-09 06:59 - -from django.db import migrations - -from ietf.person.name import plain_name -from ietf.utils.mail import parseaddr - - -def find_or_create_email(email_model, person_model, formatted_email, group): - """Look up an email address or create if needed - - Also creates a Person if the email does not have one. Created Email will have - the origin field set to the origin parameter to this method. - """ - name, address = parseaddr(formatted_email) - if not address: - raise ValueError('Could not parse email "%s"' % formatted_email) - email, _ = email_model.objects.get_or_create( - address=address, - defaults=dict(origin='liaison contact: ' + group.acronym) - ) - - if not email.person: - person = person_model.objects.create(name=name if name else address) - email.person = person - email.save() - - # Display an alert if the formatted address sent from the Role will differ - # from what was in the original contacts list - if not email.person.plain and email.person.name == email.address: - recreated_contact_email = email.address - else: - person_plain = email.person.plain if email.person.plain else plain_name(email.person.name) - recreated_contact_email = "%s <%s>" % (person_plain, email.address) - if recreated_contact_email != formatted_email: - print('>> Note: address "%s" is now "%s" (%s)' % ( - formatted_email, - recreated_contact_email, - group.acronym, - )) - return email - - -def forward(apps, schema_editor): - """Perform forward migration - - Creates liaison_contact and liaison_cc_contact Roles corresponding to existing - LiaisonStatementGroupContact instances. - """ - Group = apps.get_model('group', 'Group') - Role = apps.get_model('group', 'Role') - Email = apps.get_model('person', 'Email') - Person = apps.get_model('person', 'Person') - - RoleName = apps.get_model('name', 'RoleName') - contact_role_name = RoleName.objects.get(slug='liaison_contact') - cc_contact_role_name = RoleName.objects.get(slug='liaison_cc_contact') - - print() - LiaisonStatementGroupContacts = apps.get_model('liaisons', 'LiaisonStatementGroupContacts') - for lsgc in LiaisonStatementGroupContacts.objects.all(): - group = lsgc.group - for contact_email in lsgc.contacts.split(','): - if contact_email: - email = find_or_create_email(Email, Person, - contact_email.strip(), - group) - Role.objects.create( - group=group, - name=contact_role_name, - person=email.person, - email=email, - ) - - for contact_email in lsgc.cc_contacts.split(','): - if contact_email: - email = find_or_create_email(Email, Person, - contact_email.strip(), - group) - Role.objects.create( - group=group, - name=cc_contact_role_name, - person=email.person, - email=email, - ) - - # Now validate that we got them all. As much as possible, use independent code - # to avoid replicating any bugs from the original migration. - for group in Group.objects.all(): - lsgc = LiaisonStatementGroupContacts.objects.filter(group_id=group.pk).first() - - if not lsgc: - if group.role_set.filter(name__in=[contact_role_name, cc_contact_role_name]).exists(): - raise ValueError('%s group has contact roles after migration but had no LiaisonStatementGroupContacts' % ( - group.acronym, - )) - else: - contacts = group.role_set.filter(name=contact_role_name) - num_lsgc_contacts = len(lsgc.contacts.split(',')) if lsgc.contacts else 0 - if len(contacts) != num_lsgc_contacts: - raise ValueError( - '%s group has %d contact(s) but only %d address(es) in its LiaisonStatementGroupContacts (contact addresses = "%s", LSGC.contacts="%s")' % ( - group.acronym, len(contacts), num_lsgc_contacts, - '","'.join([c.email.address for c in contacts]), - lsgc.contacts, - ) - ) - for contact in contacts: - email = contact.email.address - if email.lower() not in lsgc.contacts.lower(): - raise ValueError( - '%s group has "%s" contact but not found in LiaisonStatementGroupContacts.contacts = "%s"' % ( - group.acronym, email, lsgc.contacts, - ) - ) - - cc_contacts = group.role_set.filter(name=cc_contact_role_name) - num_lsgc_cc_contacts = len(lsgc.cc_contacts.split(',')) if lsgc.cc_contacts else 0 - if len(cc_contacts) != num_lsgc_cc_contacts: - raise ValueError( - '%s group has %d CC contact(s) but %d address(es) in its LiaisonStatementGroupContacts (cc_contact addresses = "%s", LSGC.cc_contacts="%s")' % ( - group.acronym, len(cc_contacts), num_lsgc_cc_contacts, - '","'.join([c.email.address for c in cc_contacts]), - lsgc.cc_contacts, - ) - ) - for cc_contact in cc_contacts: - email = cc_contact.email.address - if email.lower() not in lsgc.cc_contacts.lower(): - raise ValueError( - '%s group has "%s" CC contact but not found in LiaisonStatementGroupContacts.cc_contacts = "%s"' % ( - group.acronym, email, lsgc.cc_contacts, - ) - ) - -def reverse(apps, schema_editor): - """Perform reverse migration - - Removes liaison_contact and liaison_cc_contact Roles. The forward migration creates missing - Email and Person instances, but these are not removed because it's difficult to do this - safely and correctly. - """ - Role = apps.get_model('group', 'Role') - Role.objects.filter( - name_id__in=['liaison_contact', 'liaison_cc_contact'] - ).delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0040_lengthen_used_roles_fields'), - ('name', '0022_add_liaison_contact_rolenames'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/group/migrations/0042_add_liaison_contact_roles_to_used_roles.py b/ietf/group/migrations/0042_add_liaison_contact_roles_to_used_roles.py deleted file mode 100644 index 2b16d7485b..0000000000 --- a/ietf/group/migrations/0042_add_liaison_contact_roles_to_used_roles.py +++ /dev/null @@ -1,63 +0,0 @@ -# Generated by Django 2.2.17 on 2020-12-11 08:52 - -from django.db import migrations - - -def forward(apps, schema_editor): - role_names_to_add = ['liaison_contact', 'liaison_cc_contact'] - - Group = apps.get_model('group', 'Group') - GroupFeatures = apps.get_model('group', 'GroupFeatures') - Role = apps.get_model('group', 'Role') - - # Add new liaison contact roles to default_used_fields for wg, sdo, and area groups - for group_type in ['wg', 'sdo', 'area']: - gf = GroupFeatures.objects.get(type_id=group_type) - for role_name in role_names_to_add: - if role_name not in gf.default_used_roles: - gf.default_used_roles.append(role_name) - gf.save() - - # Add new role names to any groups that both have liaison contacts - # and use a custom used_roles list. - for group in Group.objects.filter(type_id=group_type): - used_roles_is_set = len(group.used_roles) > 0 - has_contacts = Role.objects.filter(name_id__in=role_names_to_add).exists() - if used_roles_is_set and has_contacts: - for role_name in role_names_to_add: - if role_name not in group.used_roles: - print('>> Adding %s to used_roles for %s' % (role_name, group.acronym)) - group.used_roles.append(role_name) - group.save() - - -def reverse(apps, schema_editor): - role_names_to_remove = ['liaison_contact', 'liaison_cc_contact'] - - Group = apps.get_model('group', 'Group') - GroupFeatures = apps.get_model('group', 'GroupFeatures') - - for group in Group.objects.all(): - for role_name in role_names_to_remove: - if role_name in group.used_roles: - print('>> Removing %s from used_roles for %s' % (role_name, group.acronym)) - group.used_roles.remove(role_name) - group.save() - - for gf in GroupFeatures.objects.all(): - for role_name in role_names_to_remove: - if role_name in gf.default_used_roles: - print('>> Removing %s from default_used_roles for %s' % (role_name, gf.type_id)) - gf.default_used_roles.remove(role_name) - gf.save() - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0041_create_liaison_contact_roles'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/group/migrations/0043_add_groupfeatures_parent_type_fields.py b/ietf/group/migrations/0043_add_groupfeatures_parent_type_fields.py deleted file mode 100644 index 82cb0990f3..0000000000 --- a/ietf/group/migrations/0043_add_groupfeatures_parent_type_fields.py +++ /dev/null @@ -1,29 +0,0 @@ -# Generated by Django 2.2.19 on 2021-04-13 05:17 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0023_change_stream_descriptions'), - ('group', '0042_add_liaison_contact_roles_to_used_roles'), - ] - - operations = [ - migrations.AddField( - model_name='groupfeatures', - name='parent_types', - field=models.ManyToManyField(blank=True, help_text='Group types allowed as parent of this group type', related_name='child_features', to='name.GroupTypeName'), - ), - migrations.AddField( - model_name='groupfeatures', - name='need_parent', - field=models.BooleanField(default=False, help_text='Does this group type require a parent group?', verbose_name='Need Parent'), - ), - migrations.AddField( - model_name='groupfeatures', - name='default_parent', - field=models.CharField(blank=True, default='', help_text='Default parent group acronym for this group type', max_length=40, verbose_name='Default Parent'), - ), - ] diff --git a/ietf/group/migrations/0044_populate_groupfeatures_parent_type_fields.py b/ietf/group/migrations/0044_populate_groupfeatures_parent_type_fields.py deleted file mode 100644 index 855b47c90f..0000000000 --- a/ietf/group/migrations/0044_populate_groupfeatures_parent_type_fields.py +++ /dev/null @@ -1,92 +0,0 @@ -# Generated by Django 2.2.19 on 2021-04-13 09:17 - -from django.db import migrations - -def populate_parent_types(apps, schema_editor): - """Add default parent_types entries - - Data were determined from existing groups via this query: - {t.slug: list( - Group.objects.filter(type=t, parent__isnull=False).values_list('parent__type', flat=True).distinct() - ) for t in GroupTypeName.objects.all()} - """ - GroupFeatures = apps.get_model('group', 'GroupFeatures') - GroupTypeName = apps.get_model('name', 'GroupTypeName') - type_map = { - 'adhoc': ['ietf'], - 'admin': [], - 'ag': ['area', 'ietf'], - 'area': ['ietf'], - 'dir': ['area'], - 'iab': ['ietf'], - 'iana': [], - 'iesg': [], - 'ietf': ['ietf'], - 'individ': ['area'], - 'irtf': ['irtf'], - 'ise': [], - 'isoc': ['isoc'], - 'nomcom': ['area'], - 'program': ['ietf'], - 'rag': ['irtf'], - 'review': ['area'], - 'rfcedtyp': [], - 'rg': ['irtf'], - 'sdo': ['sdo', 'area'], - 'team': ['area'], - 'wg': ['area'] - } - for type_slug, parent_slugs in type_map.items(): - if len(parent_slugs) > 0: - features = GroupFeatures.objects.get(type__slug=type_slug) - features.parent_types.add(*GroupTypeName.objects.filter(slug__in=parent_slugs)) - - # validate - for gtn in GroupTypeName.objects.all(): - slugs_in_db = set(type.slug for type in gtn.features.parent_types.all()) - assert(slugs_in_db == set(type_map[gtn.slug])) - - -def set_need_parent_values(apps, schema_editor): - """Set need_parent values - - Data determined from existing groups using: - - GroupTypeName.objects.exclude(pk__in=Group.objects.filter(parent__isnull=True).values('type')) - - 'iesg' has been removed because there are no groups of this type, so no parent types have - been made available to it. - """ - GroupFeatures = apps.get_model('group', 'GroupFeatures') - - GroupFeatures.objects.filter( - type_id__in=('area', 'dir', 'individ', 'review', 'rg',) - ).update(need_parent=True) - - -def set_default_parents(apps, schema_editor): - GroupFeatures = apps.get_model('group', 'GroupFeatures') - - # rg-typed groups are children of the irtf group - rg_features = GroupFeatures.objects.filter(type_id='rg').first() - if rg_features: - rg_features.default_parent = 'irtf' - rg_features.save() - - -def empty_reverse(apps, schema_editor): - pass # nothing to do, field will be dropped - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0043_add_groupfeatures_parent_type_fields'), - ('person', '0019_auto_20210604_1443'), - ] - - operations = [ - migrations.RunPython(populate_parent_types, empty_reverse), - migrations.RunPython(set_need_parent_values, empty_reverse), - migrations.RunPython(set_default_parents, empty_reverse), - ] diff --git a/ietf/group/migrations/0045_iabasg.py b/ietf/group/migrations/0045_iabasg.py deleted file mode 100644 index 371b7c45b5..0000000000 --- a/ietf/group/migrations/0045_iabasg.py +++ /dev/null @@ -1,38 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -from django.db import migrations - -def forward(apps, schema_editor): - GroupFeatures = apps.get_model('group', 'GroupFeatures') - Group = apps.get_model('group', 'Group') - - # Copy program to iabasg - feat = GroupFeatures.objects.get(pk='program') - feat.pk = 'iabasg' - feat.save() - feat.parent_types.add('ietf') - - # List provided by Cindy on 30Aug2021 - Group.objects.filter(acronym__in=['iana-evolution','iproc','liaison-oversight','ietfiana','plenary-planning','rfcedprog']).update(type_id='iabasg') - - Group.objects.filter(acronym='model-t').update(parent=Group.objects.get(acronym='iab')) - -def reverse(apps, schema_editor): - GroupFeatures = apps.get_model('group', 'GroupFeatures') - Group = apps.get_model('group', 'Group') - Group.objects.filter(type_id='iabasg').update(type_id='program') - # Intentionally not removing the parent of model-t - GroupFeatures.objects.filter(pk='iabasg').delete() - - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0044_populate_groupfeatures_parent_type_fields'), - ('name', '0028_iabasg'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/group/migrations/0046_grouptypename_admin_to_adm.py b/ietf/group/migrations/0046_grouptypename_admin_to_adm.py deleted file mode 100644 index 73c2689938..0000000000 --- a/ietf/group/migrations/0046_grouptypename_admin_to_adm.py +++ /dev/null @@ -1,55 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -from django.db import migrations - -def forward(apps, schema_editor): - GroupTypeName = apps.get_model('name','GroupTypeName') - Group = apps.get_model('group', 'Group') - GroupHistory = apps.get_model('group', 'GroupHistory') - GroupFeatures = apps.get_model('group', 'GroupFeatures') - - a = GroupTypeName.objects.get(pk='admin') - a.pk='adm' - a.order=1 - a.save() - f = GroupFeatures.objects.get(pk='admin') - f.pk='adm' - f.save() - - Group.objects.filter(type_id='admin').update(type_id='adm') - GroupHistory.objects.filter(type_id='admin').update(type_id='adm') - - GroupFeatures.objects.filter(pk='admin').delete() - GroupTypeName.objects.filter(pk='admin').delete() - -def reverse(apps, schema_editor): - GroupTypeName = apps.get_model('name','GroupTypeName') - Group = apps.get_model('group', 'Group') - GroupHistory = apps.get_model('group','GroupHistory') - GroupFeatures = apps.get_model('group', 'GroupFeatures') - - a = GroupTypeName.objects.get(pk='adm') - a.pk='admin' - a.order=0 - a.save() - f = GroupFeatures.objects.get(pk='adm') - f.pk='admin' - f.save() - - Group.objects.filter(type_id='adm').update(type_id='admin') - GroupHistory.objects.filter(type_id='adm').update(type_id='admin') - - GroupFeatures.objects.filter(type_id='adm').delete() - GroupTypeName.objects.filter(pk='adm').delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0045_iabasg'), - ('name', '0028_iabasg'), - ] - - operations = [ - migrations.RunPython(forward,reverse) - ] diff --git a/ietf/group/migrations/0047_ietfllc.py b/ietf/group/migrations/0047_ietfllc.py deleted file mode 100644 index a946e2bf01..0000000000 --- a/ietf/group/migrations/0047_ietfllc.py +++ /dev/null @@ -1,88 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -from django.db import migrations - -def email(person): - e = person.email_set.filter(primary=True).first() - if not e: - e = person.email_set.filter(active=True).order_by("-time").first() - return e - -def forward(apps, schema_editor): - Group = apps.get_model('group', 'Group') - Person = apps.get_model('person', 'Person') - llc = Group.objects.create( - acronym='ietfadminllc', - name="IETF Administration LLC", - state_id='active', - type_id='adm', - description="The IETF Administration LLC (IETF LLC) provides the corporate legal home for the IETF, the Internet Architecture Board (IAB), and the Internet Research Task Force (IRTF). The Administration (https://www.ietf.org/about/administration/) section of the website has full details of the LLC and is where the various policies and reports produced by the LLC are published.", - ) - Group.objects.filter(acronym='llc-board').update(parent=llc, description="The IETF Administration LLC (IETF LLC) provides the corporate legal home for the IETF, the Internet Architecture Board (IAB), and the Internet Research Task Force (IRTF). The Administration (https://www.ietf.org/about/administration/) section of the website has full details of the LLC and is where the various policies and reports produced by the LLC are published.") - llc_staff= Group.objects.create( - acronym='llc-staff', - name="IETF LLC employees", - state_id='active', - type_id='adm', - parent=llc, - description="The IETF Administration LLC (IETF LLC) provides the corporate legal home for the IETF, the Internet Architecture Board (IAB), and the Internet Research Task Force (IRTF). The Administration (https://www.ietf.org/about/administration/) section of the website has full details of the LLC and is where the various policies and reports produced by the LLC are published.", - ) - legal = Group.objects.create( - acronym='legal-consult', - name="Legal consultation group", - state_id='active', - type_id='adm', - parent=llc, - description="The legal-consult list is a group of community participants who provide their views to the IETF Administration LLC in private on various legal matters. This was first established under the IAOC and has not been reviewed since. Legal advice is provided separately to the LLC by contracted external counsel.", - ) - - for email_addr in ('jay@ietf.org', 'ghwood@ietf.org', 'lbshaw@ietf.org', 'krathnayake@ietf.org'): - p = Person.objects.get(email__address=email_addr) - llc_staff.role_set.create(name_id='member',person=p,email=email(p)) - - for email_addr in ( - 'amorris@amsl.com', - 'brad@biddle.law', - 'David.Wilson@thompsonhine.com', - 'glenn.deen@nbcuni.com', - 'hall@isoc.org', - 'Jason_Livingood@comcast.com', - 'jay@ietf.org', - 'jmh@joelhalpern.com', - 'johnl@taugh.com', - 'kathleen.moriarty.ietf@gmail.com', - 'lars@eggert.org', - 'lflynn@amsl.com', - 'stewe@stewe.org', - 'vigdis@biddle.law', - 'wendy@seltzer.org', - ): - p = Person.objects.filter(email__address=email_addr).first() - if p: - legal.role_set.create(name_id='member', person=p, email=email(p)) - - -def reverse(apps, schema_editor): - Group = apps.get_model('group', 'Group') - Group.objects.filter(acronym='llc-board').update(parent=None) - Group.objects.filter(acronym__in=['llc_staff','legal-consult']).delete() - Group.objects.filter(acronym='ietfadminllc').delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0046_grouptypename_admin_to_adm'), - ('person', '0019_auto_20210604_1443'), - # The below are needed for reverse migrations to work - ('name','0028_iabasg'), - ('doc', '0043_bofreq_docevents'), - ('liaisons','0009_delete_liaisonstatementgroupcontacts_model'), - ('meeting', '0018_document_primary_key_cleanup'), - ('review', '0014_document_primary_key_cleanup'), - ('submit', '0008_submissionextresource'), - ] - - operations = [ - migrations.RunPython(forward,reverse) - ] diff --git a/ietf/group/migrations/0048_has_session_materials.py b/ietf/group/migrations/0048_has_session_materials.py deleted file mode 100644 index 3a081275d7..0000000000 --- a/ietf/group/migrations/0048_has_session_materials.py +++ /dev/null @@ -1,29 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -from django.db import migrations - -# Not adding team at this time - need to untangle the nonsession_materials mess first - -types_to_change = [ - 'program', - 'dir', - 'review', -] - -def forward(apps, schema_editor): - GroupFeatures = apps.get_model('group', 'GroupFeatures') - GroupFeatures.objects.filter(type__in=types_to_change).update(has_session_materials=True) - -def reverse(apps, schema_editor): - GroupFeatures = apps.get_model('group', 'GroupFeatures') - GroupFeatures.objects.filter(type__in=types_to_change).update(has_session_materials=False) - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0047_ietfllc'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/group/migrations/0049_auto_20211019_1136.py b/ietf/group/migrations/0049_auto_20211019_1136.py deleted file mode 100644 index b9464f8178..0000000000 --- a/ietf/group/migrations/0049_auto_20211019_1136.py +++ /dev/null @@ -1,54 +0,0 @@ -# Generated by Django 2.2.24 on 2021-10-19 11:36 - -from django.db import migrations -import ietf.utils.db - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0048_has_session_materials'), - ] - - operations = [ - migrations.AlterField( - model_name='groupfeatures', - name='admin_roles', - field=ietf.utils.db.IETFJSONField(default=['chair'], max_length=64), - ), - migrations.AlterField( - model_name='groupfeatures', - name='default_used_roles', - field=ietf.utils.db.IETFJSONField(default=[], max_length=256), - ), - migrations.AlterField( - model_name='groupfeatures', - name='docman_roles', - field=ietf.utils.db.IETFJSONField(default=['ad', 'chair', 'delegate', 'secr'], max_length=128), - ), - migrations.AlterField( - model_name='groupfeatures', - name='groupman_authroles', - field=ietf.utils.db.IETFJSONField(default=['Secretariat'], max_length=128), - ), - migrations.AlterField( - model_name='groupfeatures', - name='groupman_roles', - field=ietf.utils.db.IETFJSONField(default=['ad', 'chair'], max_length=128), - ), - migrations.AlterField( - model_name='groupfeatures', - name='material_types', - field=ietf.utils.db.IETFJSONField(default=['slides'], max_length=64), - ), - migrations.AlterField( - model_name='groupfeatures', - name='matman_roles', - field=ietf.utils.db.IETFJSONField(default=['ad', 'chair', 'delegate', 'secr'], max_length=128), - ), - migrations.AlterField( - model_name='groupfeatures', - name='role_order', - field=ietf.utils.db.IETFJSONField(default=['chair', 'secr', 'member'], help_text='The order in which roles are shown, for instance on photo pages. Enter valid JSON.', max_length=128), - ), - ] diff --git a/ietf/group/migrations/0050_groupfeatures_agenda_filter_type.py b/ietf/group/migrations/0050_groupfeatures_agenda_filter_type.py deleted file mode 100644 index e7cdc64dec..0000000000 --- a/ietf/group/migrations/0050_groupfeatures_agenda_filter_type.py +++ /dev/null @@ -1,20 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -from django.db import migrations, models -import django.db.models.deletion - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0033_populate_agendafiltertypename'), - ('group', '0049_auto_20211019_1136'), - ] - - operations = [ - migrations.AddField( - model_name='groupfeatures', - name='agenda_filter_type', - field=models.ForeignKey(default='none', on_delete=django.db.models.deletion.PROTECT, to='name.AgendaFilterTypeName'), - ), - ] diff --git a/ietf/group/migrations/0051_populate_groupfeatures_agenda_filter_type.py b/ietf/group/migrations/0051_populate_groupfeatures_agenda_filter_type.py deleted file mode 100644 index fa5025902b..0000000000 --- a/ietf/group/migrations/0051_populate_groupfeatures_agenda_filter_type.py +++ /dev/null @@ -1,35 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -from django.db import migrations - - -def forward(apps, schema_editor): - GroupFeatures = apps.get_model('group', 'GroupFeatures') - - # map AgendaFilterTypeName slug to group types - unlisted get 'none' - filter_types = dict( - # list previously hard coded in agenda view, plus 'review' - normal={'wg', 'ag', 'rg', 'rag', 'iab', 'program', 'review'}, - heading={'area', 'ietf', 'irtf'}, - special={'team', 'adhoc'}, - ) - - for ft, group_types in filter_types.items(): - for gf in GroupFeatures.objects.filter(type__slug__in=group_types): - gf.agenda_filter_type_id = ft - gf.save() - - -def reverse(apps, schema_editor): - pass # nothing to do, model will be deleted anyway - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0050_groupfeatures_agenda_filter_type'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/group/migrations/0052_groupfeatures_session_purposes.py b/ietf/group/migrations/0052_groupfeatures_session_purposes.py deleted file mode 100644 index 8a3668e7e1..0000000000 --- a/ietf/group/migrations/0052_groupfeatures_session_purposes.py +++ /dev/null @@ -1,24 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -# Generated by Django 2.2.24 on 2021-09-26 11:29 - -from django.db import migrations -import ietf.name.models -import ietf.utils.db -import ietf.utils.validators - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0051_populate_groupfeatures_agenda_filter_type'), - ('name', '0034_sessionpurposename'), - ] - - operations = [ - migrations.AddField( - model_name='groupfeatures', - name='session_purposes', - field=ietf.utils.db.IETFJSONField(default=[], help_text='Allowed session purposes for this group type', max_length=256, validators=[ietf.utils.validators.JSONForeignKeyListValidator(ietf.name.models.SessionPurposeName)]), - ), - ] diff --git a/ietf/group/migrations/0053_populate_groupfeatures_session_purposes.py b/ietf/group/migrations/0053_populate_groupfeatures_session_purposes.py deleted file mode 100644 index 642aa5f215..0000000000 --- a/ietf/group/migrations/0053_populate_groupfeatures_session_purposes.py +++ /dev/null @@ -1,64 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -# Generated by Django 2.2.24 on 2021-09-26 11:29 - -from django.db import migrations - - -default_purposes = dict( - adhoc=['presentation'], - adm=['closed_meeting', 'officehours'], - ag=['regular'], - area=['regular'], - dir=['open_meeting', 'presentation', 'regular', 'social', 'tutorial'], - iab=['closed_meeting', 'regular'], - iabasg=['closed_meeting', 'officehours', 'open_meeting'], - iana=['officehours'], - iesg=['closed_meeting', 'open_meeting'], - ietf=['admin', 'plenary', 'presentation', 'social'], - irtf=[], - ise=['officehours'], - isoc=['officehours', 'open_meeting', 'presentation'], - nomcom=['closed_meeting', 'officehours'], - program=['regular', 'tutorial'], - rag=['regular'], - review=['open_meeting', 'social'], - rfcedtyp=['officehours'], - rg=['regular'], - team=['coding', 'presentation', 'social', 'tutorial'], - wg=['regular'], -) - - -def forward(apps, schema_editor): - GroupFeatures = apps.get_model('group', 'GroupFeatures') - SessionPurposeName = apps.get_model('name', 'SessionPurposeName') - - # verify that we're not about to use an invalid purpose - for purposes in default_purposes.values(): - for purpose in purposes: - SessionPurposeName.objects.get(pk=purpose) # throws an exception unless exists - - for type_, purposes in default_purposes.items(): - GroupFeatures.objects.filter( - type=type_ - ).update( - session_purposes=purposes - ) - -def reverse(apps, schema_editor): - GroupFeatures = apps.get_model('group', 'GroupFeatures') - GroupFeatures.objects.update(session_purposes=[]) # clear back out to default - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0052_groupfeatures_session_purposes'), - ('name', '0035_populate_sessionpurposename'), - - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/group/migrations/0054_enable_delegation.py b/ietf/group/migrations/0054_enable_delegation.py deleted file mode 100644 index 9655f819e6..0000000000 --- a/ietf/group/migrations/0054_enable_delegation.py +++ /dev/null @@ -1,29 +0,0 @@ -# Copyright The IETF Trust 2022 All Rights Reserved - -from django.db import migrations - -def forward(apps, schema_editor): - GroupFeatures = apps.get_model('group','GroupFeatures') - for type_id in ('dir', 'iabasg', 'program', 'review', 'team'): - f = GroupFeatures.objects.get(type_id=type_id) - if 'delegate' not in f.groupman_roles: - f.groupman_roles.append('delegate') - f.save() - for type_id in ('adhoc', 'ag', 'iesg', 'irtf', 'ise', 'rag', 'dir', 'iabasg', 'program', 'review'): - f = GroupFeatures.objects.get(type_id=type_id) - if 'delegate' not in f.default_used_roles: - f.default_used_roles.append('delegate') - f.save() - -def reverse (apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0053_populate_groupfeatures_session_purposes'), - ] - - operations = [ - migrations.RunPython(forward,reverse), - ] diff --git a/ietf/group/migrations/0055_editorial_stream.py b/ietf/group/migrations/0055_editorial_stream.py deleted file mode 100644 index a7dc9ca9c3..0000000000 --- a/ietf/group/migrations/0055_editorial_stream.py +++ /dev/null @@ -1,38 +0,0 @@ -# Copyright The IETF Trust 2022 All Rights Reserved - -from django.db import migrations - -def forward(apps, schema_editor): - Group = apps.get_model('group', 'Group') - GroupFeatures = apps.get_model('group', 'GroupFeatures') - Group.objects.create( - acronym='editorial', - name='Editorial Stream', - state_id='active', - type_id='editorial', - parent=None, - ) - templ = GroupFeatures.objects.get(type='rfcedtyp') - templ.pk = None - templ.type_id='editorial' - templ.save() - - - -def reverse(apps, schema_editor): - Group = apps.get_model('group', 'Group') - GroupFeatures = apps.get_model('group', 'GroupFeatures') - GroupFeatures.objects.filter(type='editorial').delete() - Group.objects.filter(acronym='editorial').delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0054_enable_delegation'), - ('name', '0043_editorial_stream_grouptype'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/group/migrations/0056_dir_chair_groupman_role.py b/ietf/group/migrations/0056_dir_chair_groupman_role.py deleted file mode 100644 index b69eb03ae8..0000000000 --- a/ietf/group/migrations/0056_dir_chair_groupman_role.py +++ /dev/null @@ -1,30 +0,0 @@ -# Generated by Django 2.2.28 on 2022-06-14 13:14 - -from django.db import migrations - - -def forward(apps, schema_editor): - GroupFeatures = apps.get_model('group', 'GroupFeatures') - features = GroupFeatures.objects.get(type_id='dir') - if 'chair' not in features.groupman_roles: - features.groupman_roles.append('chair') - features.save() - - -def reverse(apps, schema_editor): - GroupFeatures = apps.get_model('group', 'GroupFeatures') - features = GroupFeatures.objects.get(type_id='dir') - if 'chair' in features.groupman_roles: - features.groupman_roles.remove('chair') - features.save() - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0055_editorial_stream'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/group/migrations/0057_nojabber_onlychat.py b/ietf/group/migrations/0057_nojabber_onlychat.py deleted file mode 100644 index c089a9761b..0000000000 --- a/ietf/group/migrations/0057_nojabber_onlychat.py +++ /dev/null @@ -1,19 +0,0 @@ -# Copyright The IETF Trust 2022, All Rights Reserved -# Generated by Django 2.2.28 on 2022-07-14 09:09 - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0056_dir_chair_groupman_role'), - ] - - operations = [ - migrations.RenameField( - model_name='groupfeatures', - old_name='has_default_jabber', - new_name='has_default_chat', - ), - ] diff --git a/ietf/group/migrations/0058_alter_has_default_chat.py b/ietf/group/migrations/0058_alter_has_default_chat.py deleted file mode 100644 index 8f0464764c..0000000000 --- a/ietf/group/migrations/0058_alter_has_default_chat.py +++ /dev/null @@ -1,19 +0,0 @@ -# Copyright The IETF Trust 2022, All Rights Reserved -# Generated by Django 2.2.28 on 2022-07-15 12:24 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0057_nojabber_onlychat'), - ] - - operations = [ - migrations.AlterField( - model_name='groupfeatures', - name='has_default_chat', - field=models.BooleanField(default=False, verbose_name='Chat'), - ), - ] diff --git a/ietf/group/migrations/0059_use_timezone_now_for_group_models.py b/ietf/group/migrations/0059_use_timezone_now_for_group_models.py deleted file mode 100644 index 24c083855b..0000000000 --- a/ietf/group/migrations/0059_use_timezone_now_for_group_models.py +++ /dev/null @@ -1,29 +0,0 @@ -# Generated by Django 2.2.28 on 2022-07-12 11:24 - -from django.db import migrations, models -import django.utils.timezone - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0058_alter_has_default_chat'), - ] - - operations = [ - migrations.AlterField( - model_name='group', - name='time', - field=models.DateTimeField(default=django.utils.timezone.now), - ), - migrations.AlterField( - model_name='groupevent', - name='time', - field=models.DateTimeField(default=django.utils.timezone.now, help_text='When the event happened'), - ), - migrations.AlterField( - model_name='grouphistory', - name='time', - field=models.DateTimeField(default=django.utils.timezone.now), - ), - ] diff --git a/ietf/group/milestones.py b/ietf/group/milestones.py index 43748a1f25..52f2eaebee 100644 --- a/ietf/group/milestones.py +++ b/ietf/group/milestones.py @@ -29,7 +29,7 @@ class MilestoneForm(forms.Form): desc = forms.CharField(max_length=500, label="Milestone", required=True) due = DatepickerDateField(date_format="MM yyyy", picker_settings={"min-view-mode": "months", "autoclose": "1", "view-mode": "years" }, required=True) order = forms.IntegerField(required=True, widget=forms.HiddenInput) - docs = SearchableDocumentsField(label="Drafts", required=False, help_text="Any drafts that the milestone concerns.") + docs = SearchableDocumentsField(label="Internet-Drafts", required=False, help_text="Any Internet-Drafts that the milestone concerns.") resolved_checkbox = forms.BooleanField(required=False, label="Resolved") resolved = forms.CharField(label="Resolved as", max_length=50, required=False) @@ -369,7 +369,7 @@ def save_milestone_form(f): email_milestones_changed(request, group, changes, states) if milestone_set == "charter": - return redirect('ietf.doc.views_doc.document_main', name=group.charter.canonical_name()) + return redirect('ietf.doc.views_doc.document_main', name=group.charter.name) else: return HttpResponseRedirect(group.about_url()) else: @@ -399,7 +399,7 @@ def save_milestone_form(f): can_change_uses_milestone_dates=can_change_uses_milestone_dates)) @login_required -def reset_charter_milestones(request, group_type, acronym): +def reset_charter_milestones(request, acronym, group_type=None): """Reset charter milestones to the currently in-use milestones.""" group = get_group_or_404(acronym, group_type) if not group.features.has_milestones: diff --git a/ietf/group/models.py b/ietf/group/models.py index 112522f2c9..a7e3c6616e 100644 --- a/ietf/group/models.py +++ b/ietf/group/models.py @@ -3,7 +3,6 @@ import email.utils -import jsonfield import os import re @@ -13,16 +12,19 @@ from django.db.models.deletion import CASCADE, PROTECT from django.dispatch import receiver from django.utils import timezone +from django.utils.text import slugify import debug # pyflakes:ignore from ietf.name.models import (GroupStateName, GroupTypeName, DocTagName, GroupMilestoneStateName, RoleName, - AgendaTypeName, AgendaFilterTypeName, ExtResourceName, SessionPurposeName) + AgendaTypeName, AgendaFilterTypeName, ExtResourceName, SessionPurposeName, + AppealArtifactTypeName ) from ietf.person.models import Email, Person -from ietf.utils.db import IETFJSONField +from ietf.utils.db import EmptyAwareJSONField from ietf.utils.mail import formataddr, send_mail_text from ietf.utils import log from ietf.utils.models import ForeignKey, OneToOneField +from ietf.utils.timezone import date_today from ietf.utils.validators import JSONForeignKeyListValidator @@ -43,7 +45,7 @@ class GroupInfo(models.Model): unused_states = models.ManyToManyField('doc.State', help_text="Document states that have been disabled for the group.", blank=True) unused_tags = models.ManyToManyField(DocTagName, help_text="Document tags that have been disabled for the group.", blank=True) - used_roles = jsonfield.JSONField(max_length=256, blank=True, default=[], help_text="Leave an empty list to get the group_type's default used roles") + used_roles = models.JSONField(max_length=256, blank=True, default=list, help_text="Leave an empty list to get the group_type's default used roles") uses_milestone_dates = models.BooleanField(default=True) @@ -109,6 +111,9 @@ def active_wgs(self): def closed_wgs(self): return self.wgs().exclude(state__in=Group.ACTIVE_STATE_IDS) + def areas(self): + return self.get_queryset().filter(type="area") + def with_meetings(self): return self.get_queryset().filter(type__features__has_meetings=True) @@ -232,6 +237,36 @@ def chat_archive_url(self): ) +# JSONFields need callable defaults that work with migrations to avoid sharing +# data structures between instances. These helpers provide that. +def default_material_types(): + return ["slides"] + + +def default_admin_roles(): + return ["chair"] + + +def default_docman_roles(): + return ["ad", "chair", "delegate", "secr"] + + +def default_groupman_roles(): + return ["ad", "chair"] + + +def default_groupman_authroles(): + return ["Secretariat"] + + +def default_matman_roles(): + return ["ad", "chair", "delegate", "secr"] + + +def default_role_order(): + return ["chair", "secr", "member"] + + class GroupFeatures(models.Model): type = OneToOneField(GroupTypeName, primary_key=True, null=False, related_name='features') #history = HistoricalRecords() @@ -265,16 +300,16 @@ class GroupFeatures(models.Model): agenda_type = models.ForeignKey(AgendaTypeName, null=True, default="ietf", on_delete=CASCADE) about_page = models.CharField(max_length=64, blank=False, default="ietf.group.views.group_about" ) default_tab = models.CharField(max_length=64, blank=False, default="ietf.group.views.group_about" ) - material_types = IETFJSONField(max_length=64, accepted_empty_values=[[], {}], blank=False, default=["slides"]) - default_used_roles = IETFJSONField(max_length=256, accepted_empty_values=[[], {}], blank=False, default=[]) - admin_roles = IETFJSONField(max_length=64, accepted_empty_values=[[], {}], blank=False, default=["chair"]) # Trac Admin - docman_roles = IETFJSONField(max_length=128, accepted_empty_values=[[], {}], blank=False, default=["ad","chair","delegate","secr"]) - groupman_roles = IETFJSONField(max_length=128, accepted_empty_values=[[], {}], blank=False, default=["ad","chair",]) - groupman_authroles = IETFJSONField(max_length=128, accepted_empty_values=[[], {}], blank=False, default=["Secretariat",]) - matman_roles = IETFJSONField(max_length=128, accepted_empty_values=[[], {}], blank=False, default=["ad","chair","delegate","secr"]) - role_order = IETFJSONField(max_length=128, accepted_empty_values=[[], {}], blank=False, default=["chair","secr","member"], - help_text="The order in which roles are shown, for instance on photo pages. Enter valid JSON.") - session_purposes = IETFJSONField(max_length=256, accepted_empty_values=[[], {}], blank=False, default=[], + material_types = EmptyAwareJSONField(max_length=64, accepted_empty_values=[[], {}], blank=False, default=default_material_types) + default_used_roles = EmptyAwareJSONField(max_length=256, accepted_empty_values=[[], {}], blank=False, default=list) + admin_roles = EmptyAwareJSONField(max_length=64, accepted_empty_values=[[], {}], blank=False, default=default_admin_roles) # Trac Admin + docman_roles = EmptyAwareJSONField(max_length=128, accepted_empty_values=[[], {}], blank=False, default=default_docman_roles) + groupman_roles = EmptyAwareJSONField(max_length=128, accepted_empty_values=[[], {}], blank=False, default=default_groupman_roles) + groupman_authroles = EmptyAwareJSONField(max_length=128, accepted_empty_values=[[], {}], blank=False, default=default_groupman_authroles) + matman_roles = EmptyAwareJSONField(max_length=128, accepted_empty_values=[[], {}], blank=False, default=default_matman_roles) + role_order = EmptyAwareJSONField(max_length=128, accepted_empty_values=[[], {}], blank=False, default=default_role_order, + help_text="The order in which roles are shown, for instance on photo pages. Enter valid JSON.") + session_purposes = EmptyAwareJSONField(max_length=256, accepted_empty_values=[[], {}], blank=False, default=list, help_text="Allowed session purposes for this group type", validators=[JSONForeignKeyListValidator(SessionPurposeName)]) @@ -409,6 +444,46 @@ def __str__(self): class Meta: verbose_name_plural = "role histories" +class Appeal(models.Model): + name = models.CharField(max_length=512) + group = models.ForeignKey(Group, on_delete=models.PROTECT) + date = models.DateField(default=date_today) + + class Meta: + ordering = ['-date', '-id'] + + def __str__(self): + return f"{self.date} - {self.name}" + +class AppealArtifact(models.Model): + appeal = ForeignKey(Appeal) + artifact_type = ForeignKey(AppealArtifactTypeName) + date = models.DateField(default=date_today) + title = models.CharField(max_length=256, blank=True, help_text="The artifact_type.name will be used if this field is blank") + order = models.IntegerField(default=0) + content_type = models.CharField(max_length=32) + # "Abusing" BinaryField (see the django docs) for the small number of + # these things we have on purpose. Later, any non-markdown content may + # move off into statics instead. + bits = models.BinaryField(editable=True) + + class Meta: + ordering = ['date', 'order', 'artifact_type__order'] + + def display_title(self): + if self.title != "": + return self.title + else: + return self.artifact_type.name + + def is_markdown(self): + return self.content_type == "text/markdown;charset=utf-8" + + def download_name(self): + return f"{self.date}-{slugify(self.display_title())}.{'md' if self.is_markdown() else 'pdf'}" + + def __str__(self): + return f"{self.date} {self.display_title()} : {self.appeal.name}" # --- Signal hooks for group models --- @@ -419,6 +494,8 @@ def notify_rfceditor_of_group_name_change(sender, instance=None, **kwargs): current = Group.objects.get(pk=instance.pk) except Group.DoesNotExist: return + if current.type_id == "sdo": + return addr = settings.RFC_EDITOR_GROUP_NOTIFICATION_EMAIL if addr and instance.name != current.name: msg = """ diff --git a/ietf/group/resources.py b/ietf/group/resources.py index 310fb5fb9b..698f50460a 100644 --- a/ietf/group/resources.py +++ b/ietf/group/resources.py @@ -13,7 +13,7 @@ from ietf.group.models import (Group, GroupStateTransitions, GroupMilestone, GroupHistory, # type: ignore GroupURL, Role, GroupEvent, RoleHistory, GroupMilestoneHistory, MilestoneGroupEvent, - ChangeStateGroupEvent, GroupFeatures, GroupExtResource) + ChangeStateGroupEvent, GroupFeatures, GroupExtResource, Appeal, AppealArtifact) from ietf.person.resources import PersonResource @@ -333,3 +333,42 @@ class Meta: "name": ALL_WITH_RELATIONS, } api.group.register(GroupExtResourceResource()) + + +class AppealResource(ModelResource): + group = ToOneField(GroupResource, 'group') + class Meta: + queryset = Appeal.objects.all() + serializer = api.Serializer() + cache = SimpleCache() + #resource_name = 'appeal' + ordering = ['id', ] + filtering = { + "id": ALL, + "name": ALL, + "date": ALL, + "group": ALL_WITH_RELATIONS, + } +api.group.register(AppealResource()) + +from ietf.name.resources import AppealArtifactTypeNameResource +class AppealArtifactResource(ModelResource): + appeal = ToOneField(AppealResource, 'appeal') + artifact_type = ToOneField(AppealArtifactTypeNameResource, 'artifact_type') + class Meta: + excludes= ("bits",) + queryset = AppealArtifact.objects.all() + serializer = api.Serializer() + cache = SimpleCache() + #resource_name = 'appealartifact' + ordering = [ "id", ] + filtering = { + "id": ALL, + "date": ALL, + "title": ALL, + "order": ALL, + "content_type": ALL, + "appeal": ALL_WITH_RELATIONS, + "artifact_type": ALL_WITH_RELATIONS, + } +api.group.register(AppealArtifactResource()) diff --git a/ietf/group/serializers.py b/ietf/group/serializers.py new file mode 100644 index 0000000000..e789ba46bf --- /dev/null +++ b/ietf/group/serializers.py @@ -0,0 +1,50 @@ +# Copyright The IETF Trust 2024-2026, All Rights Reserved +"""django-rest-framework serializers""" + +from drf_spectacular.utils import extend_schema_field +from rest_framework import serializers + +from ietf.person.models import Email +from .models import Group, Role + + +class GroupSerializer(serializers.ModelSerializer): + class Meta: + model = Group + fields = ["acronym", "name", "type", "list_email"] + + +class AreaDirectorSerializer(serializers.Serializer): + """Serialize an area director + + Works with Email or Role + """ + + name = serializers.SerializerMethodField() + email = serializers.SerializerMethodField() + + @extend_schema_field(serializers.CharField) + def get_name(self, instance: Email | Role): + person = getattr(instance, 'person', None) + return person.plain_name() if person else None + + @extend_schema_field(serializers.EmailField) + def get_email(self, instance: Email | Role): + if isinstance(instance, Role): + return instance.email.email_address() + return instance.email_address() + + +class AreaSerializer(serializers.ModelSerializer): + ads = serializers.SerializerMethodField() + + class Meta: + model = Group + fields = ["acronym", "name", "ads"] + + @extend_schema_field(AreaDirectorSerializer(many=True)) + def get_ads(self, area: Group): + return AreaDirectorSerializer( + area.ads if area.is_active else Role.objects.none(), + many=True, + ).data diff --git a/ietf/group/tasks.py b/ietf/group/tasks.py new file mode 100644 index 0000000000..ada83e80e2 --- /dev/null +++ b/ietf/group/tasks.py @@ -0,0 +1,232 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +# +# Celery task definitions +# +import shutil + +from celery import shared_task +from pathlib import Path + +from django.conf import settings +from django.template.loader import render_to_string +from django.utils import timezone + +from ietf.doc.storage_utils import store_file +from ietf.liaisons.models import LiaisonStatement +from ietf.utils import log +from ietf.utils.test_runner import disable_coverage + +from .models import Group, GroupHistory +from .utils import fill_in_charter_info, fill_in_wg_drafts, fill_in_wg_roles, save_group_in_history +from .views import extract_last_name, roles + + +@shared_task +def generate_wg_charters_files_task(): + areas = Group.objects.filter(type="area", state="active").order_by("name") + groups = ( + Group.objects.filter(type="wg", state="active") + .exclude(parent=None) + .order_by("acronym") + ) + for group in groups: + fill_in_charter_info(group) + fill_in_wg_roles(group) + fill_in_wg_drafts(group) + for area in areas: + area.groups = [g for g in groups if g.parent_id == area.pk] + charter_path = Path(settings.CHARTER_PATH) + charters_file = charter_path / "1wg-charters.txt" + charters_file.write_text( + render_to_string("group/1wg-charters.txt", {"areas": areas}), + encoding="utf8", + ) + charters_by_acronym_file = charter_path / "1wg-charters-by-acronym.txt" + charters_by_acronym_file.write_text( + render_to_string("group/1wg-charters-by-acronym.txt", {"groups": groups}), + encoding="utf8", + ) + + with charters_file.open("rb") as f: + store_file("indexes", "1wg-charters.txt", f, allow_overwrite=True) + with charters_by_acronym_file.open("rb") as f: + store_file("indexes", "1wg-charters-by-acronym.txt", f, allow_overwrite=True) + + charter_copy_dests = [ + getattr(settings, "CHARTER_COPY_PATH", None), + getattr(settings, "CHARTER_COPY_OTHER_PATH", None), + getattr(settings, "CHARTER_COPY_THIRD_PATH", None), + ] + for charter_copy_dest in charter_copy_dests: + if charter_copy_dest is not None: + if not Path(charter_copy_dest).is_dir(): + log.log( + f"Error copying 1wg-charter files to {charter_copy_dest}: it does not exist or is not a directory" + ) + else: + try: + shutil.copy2(charters_file, charter_copy_dest) + except IOError as err: + log.log(f"Error copying {charters_file} to {charter_copy_dest}: {err}") + try: + shutil.copy2(charters_by_acronym_file, charter_copy_dest) + except IOError as err: + log.log( + f"Error copying {charters_by_acronym_file} to {charter_copy_dest}: {err}" + ) + + +@shared_task +def generate_wg_summary_files_task(): + # Active WGs (all should have a parent, but filter to be sure) + groups = ( + Group.objects.filter(type="wg", state="active") + .exclude(parent=None) + .order_by("acronym") + ) + # Augment groups with chairs list + for group in groups: + group.chairs = sorted(roles(group, "chair"), key=extract_last_name) + + # Active areas with one or more active groups in them + areas = Group.objects.filter( + type="area", + state="active", + group__in=groups, + ).distinct().order_by("name") + # Augment areas with their groups + for area in areas: + area.groups = [g for g in groups if g.parent_id == area.pk] + summary_path = Path(settings.GROUP_SUMMARY_PATH) + summary_file = summary_path / "1wg-summary.txt" + summary_file.write_text( + render_to_string("group/1wg-summary.txt", {"areas": areas}), + encoding="utf8", + ) + summary_by_acronym_file = summary_path / "1wg-summary-by-acronym.txt" + summary_by_acronym_file.write_text( + render_to_string( + "group/1wg-summary-by-acronym.txt", + {"areas": areas, "groups": groups}, + ), + encoding="utf8", + ) + + with summary_file.open("rb") as f: + store_file("indexes", "1wg-summary.txt", f, allow_overwrite=True) + with summary_by_acronym_file.open("rb") as f: + store_file("indexes", "1wg-summary-by-acronym.txt", f, allow_overwrite=True) + +@shared_task +@disable_coverage() +def run_once_adjust_liaison_groups(): # pragma: no cover + log.log("Starting run_once_adjust_liaison_groups") + if all( + [ + Group.objects.filter( + acronym__in=[ + "3gpp-tsg-ct", + "3gpp-tsg-ran-wg1", + "3gpp-tsg-ran-wg4", + "3gpp-tsg-sa", + "3gpp-tsg-sa-wg5", + "3gpp-tsgct", # duplicates 3gpp-tsg-ct above already + "3gpp-tsgct-ct1", # will normalize all acronyms to hyphenated form + "3gpp-tsgct-ct3", # and consistently match the name + "3gpp-tsgct-ct4", # (particularly use of WG) + "3gpp-tsgran", + "3gpp-tsgran-ran2", + "3gpp-tsgsa", # duplicates 3gpp-tsg-sa above + "3gpp-tsgsa-sa2", # will normalize + "3gpp-tsgsa-sa3", + "3gpp-tsgsa-sa4", + "3gpp-tsgt-wg2", + ] + ).count() + == 16, + not Group.objects.filter( + acronym__in=[ + "3gpp-tsg-ran-wg3", + "3gpp-tsg-ct-wg1", + "3gpp-tsg-ct-wg3", + "3gpp-tsg-ct-wg4", + "3gpp-tsg-ran", + "3gpp-tsg-ran-wg2", + "3gpp-tsg-sa-wg2", + "3gpp-tsg-sa-wg3", + "3gpp-tsg-sa-wg4", + "3gpp-tsg-t-wg2", + ] + ).exists(), + Group.objects.filter(acronym="o3gpptsgran3").exists(), + not LiaisonStatement.objects.filter( + to_groups__acronym__in=["3gpp-tsgct", "3gpp-tsgsa"] + ).exists(), + not LiaisonStatement.objects.filter( + from_groups__acronym="3gpp-tsgct" + ).exists(), + LiaisonStatement.objects.filter(from_groups__acronym="3gpp-tsgsa").count() + == 1, + LiaisonStatement.objects.get(from_groups__acronym="3gpp-tsgsa").pk == 1448, + ] + ): + for old_acronym, new_acronym, new_name in ( + ("o3gpptsgran3", "3gpp-tsg-ran-wg3", "3GPP TSG RAN WG3"), + ("3gpp-tsgct-ct1", "3gpp-tsg-ct-wg1", "3GPP TSG CT WG1"), + ("3gpp-tsgct-ct3", "3gpp-tsg-ct-wg3", "3GPP TSG CT WG3"), + ("3gpp-tsgct-ct4", "3gpp-tsg-ct-wg4", "3GPP TSG CT WG4"), + ("3gpp-tsgran", "3gpp-tsg-ran", "3GPP TSG RAN"), + ("3gpp-tsgran-ran2", "3gpp-tsg-ran-wg2", "3GPP TSG RAN WG2"), + ("3gpp-tsgsa-sa2", "3gpp-tsg-sa-wg2", "3GPP TSG SA WG2"), + ("3gpp-tsgsa-sa3", "3gpp-tsg-sa-wg3", "3GPP TSG SA WG3"), + ("3gpp-tsgsa-sa4", "3gpp-tsg-sa-wg4", "3GPP TSG SA WG4"), + ("3gpp-tsgt-wg2", "3gpp-tsg-t-wg2", "3GPP TSG T WG2"), + ): + group = Group.objects.get(acronym=old_acronym) + save_group_in_history(group) + group.time = timezone.now() + group.acronym = new_acronym + group.name = new_name + if old_acronym.startswith("3gpp-tsgct-"): + group.parent = Group.objects.get(acronym="3gpp-tsg-ct") + elif old_acronym.startswith("3gpp-tsgsa-"): + group.parent = Group.objects.get(acronym="3gpp-tsg-sa") + group.save() + group.groupevent_set.create( + time=group.time, + by_id=1, # (System) + type="info_changed", + desc=f"acronym changed from {old_acronym} to {new_acronym}, name set to {new_name}", + ) + + for acronym, new_name in (("3gpp-tsg-ct", "3GPP TSG CT"),): + group = Group.objects.get(acronym=acronym) + save_group_in_history(group) + group.time = timezone.now() + group.name = new_name + group.save() + group.groupevent_set.create( + time=group.time, + by_id=1, # (System) + type="info_changed", + desc=f"name set to {new_name}", + ) + + ls = LiaisonStatement.objects.get(pk=1448) + ls.from_groups.remove(Group.objects.get(acronym="3gpp-tsgsa")) + ls.from_groups.add(Group.objects.get(acronym="3gpp-tsg-sa")) + + # Rewriting history to effectively merge the histories of the duplicate groups + GroupHistory.objects.filter(parent__acronym="3gpp-tsgsa").update( + parent=Group.objects.get(acronym="3gpp-tsg-sa") + ) + GroupHistory.objects.filter(parent__acronym="3gpp-tsgct").update( + parent=Group.objects.get(acronym="3gpp-tsg-ct") + ) + + deleted = Group.objects.filter( + acronym__in=["3gpp-tsgsa", "3gpp-tsgct"] + ).delete() + log.log(f"Deleted Groups: {deleted}") + else: + log.log("* Refusing to continue as preconditions have changed") diff --git a/ietf/group/templatetags/group_filters.py b/ietf/group/templatetags/group_filters.py index e7fb4a1819..bf2ad71949 100644 --- a/ietf/group/templatetags/group_filters.py +++ b/ietf/group/templatetags/group_filters.py @@ -2,7 +2,7 @@ import debug # pyflakes:ignore -from ietf.group.models import Group +from ietf.nomcom.models import NomCom register = template.Library() @@ -19,14 +19,15 @@ def active_nomcoms(user): if not (user and hasattr(user, "is_authenticated") and user.is_authenticated): return [] - groups = [] - - groups.extend(Group.objects.filter( - role__person__user=user, - type_id='nomcom', - state__slug='active').distinct().select_related("type")) - - return groups + return list( + NomCom.objects.filter( + group__role__person__user=user, + group__type_id='nomcom', # just in case... + group__state__slug='active', + ) + .distinct() + .order_by("group__acronym") + ) @register.inclusion_tag('person/person_link.html') def role_person_link(role, **kwargs): @@ -36,3 +37,10 @@ def role_person_link(role, **kwargs): plain_name = role.person.plain_name() email = role.email.address return {'name': name, 'plain_name': plain_name, 'email': email, 'title': title, 'class': cls} + +@register.filter +def name_with_conditional_acronym(group): + if group.type_id in ("sdo", "isoc", "individ", "nomcom", "ietf", "irtf", ): + return group.name + else: + return f"{group.name} ({group.acronym})" diff --git a/ietf/group/tests.py b/ietf/group/tests.py index 532f599a38..229744388c 100644 --- a/ietf/group/tests.py +++ b/ietf/group/tests.py @@ -1,15 +1,10 @@ # Copyright The IETF Trust 2013-2020, All Rights Reserved # -*- coding: utf-8 -*- -import io -import os import datetime import json +from unittest import mock -from tempfile import NamedTemporaryFile - -from django.core.management import call_command -from django.conf import settings from django.urls import reverse as urlreverse from django.db.models import Q from django.test import Client @@ -17,13 +12,20 @@ import debug # pyflakes:ignore -from ietf.doc.factories import DocumentFactory, WgDraftFactory -from ietf.doc.models import DocEvent, RelatedDocument +from ietf.doc.factories import DocumentFactory, WgDraftFactory, EditorialDraftFactory +from ietf.doc.models import DocEvent, RelatedDocument, Document from ietf.group.models import Role, Group -from ietf.group.utils import get_group_role_emails, get_child_group_role_emails, get_group_ad_emails +from ietf.group.utils import ( + get_group_role_emails, + get_child_group_role_emails, + get_group_ad_emails, + get_group_email_aliases, + GroupAliasGenerator, + role_holder_emails, +) from ietf.group.factories import GroupFactory, RoleFactory from ietf.person.factories import PersonFactory, EmailFactory -from ietf.person.models import Person +from ietf.person.models import Email, Person from ietf.utils.test_utils import login_testing_unauthorized, TestCase class StreamTests(TestCase): @@ -41,6 +43,11 @@ def test_stream_documents(self): self.assertEqual(r.status_code, 200) self.assertContains(r, draft.name) + EditorialDraftFactory() # Quick way to ensure RSWG exists. + r = self.client.get(urlreverse("ietf.group.views.stream_documents", kwargs=dict(acronym="editorial"))) + self.assertRedirects(r, expected_url=urlreverse('ietf.group.views.group_documents',kwargs={"acronym":"rswg"})) + + def test_stream_edit(self): EmailFactory(address="ad2@ietf.org") @@ -58,13 +65,86 @@ def test_stream_edit(self): self.assertTrue(Role.objects.filter(name="delegate", group__acronym=stream_acronym, email__address="ad2@ietf.org")) +class GroupLeadershipTests(TestCase): + def test_leadership_wg(self): + # setup various group states + bof_role = RoleFactory( + group__type_id="wg", group__state_id="bof", name_id="chair" + ) + proposed_role = RoleFactory( + group__type_id="wg", group__state_id="proposed", name_id="chair" + ) + active_role = RoleFactory( + group__type_id="wg", group__state_id="active", name_id="chair" + ) + conclude_role = RoleFactory( + group__type_id="wg", group__state_id="conclude", name_id="chair" + ) + url = urlreverse( + "ietf.group.views.group_leadership", kwargs={"group_type": "wg"} + ) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + self.assertContains(r, "Group Leadership") + self.assertContains(r, bof_role.person.last_name()) + self.assertContains(r, proposed_role.person.last_name()) + self.assertContains(r, active_role.person.last_name()) + self.assertNotContains(r, conclude_role.person.last_name()) + + def test_leadership_wg_csv(self): + url = urlreverse( + "ietf.group.views.group_leadership_csv", kwargs={"group_type": "wg"} + ) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + self.assertEqual(r["Content-Type"], "text/csv") + self.assertContains(r, "Chairman, Sops") + + def test_leadership_rg(self): + role = RoleFactory(group__type_id="rg", name_id="chair") + url = urlreverse( + "ietf.group.views.group_leadership", kwargs={"group_type": "rg"} + ) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + self.assertContains(r, "Group Leadership") + self.assertContains(r, role.person.last_name()) + self.assertNotContains(r, "Chairman, Sops") + + +class GroupStatsTests(TestCase): + def setUp(self): + super().setUp() + a = WgDraftFactory() + b = WgDraftFactory() + RelatedDocument.objects.create( + source=a, target=b, relationship_id="refnorm" + ) + + def test_group_stats(self): + client = Client(Accept="application/json") + url = urlreverse("ietf.group.views.group_stats_data") + r = client.get(url) + self.assertTrue(r.status_code == 200, "Failed to receive group stats") + self.assertGreater(len(r.content), 0, "Group stats have no content") + + try: + data = json.loads(r.content) + except Exception as e: + self.fail("JSON load failed: %s" % e) + + ids = [d["id"] for d in data] + for doc in Document.objects.all(): + self.assertIn(doc.name, ids) + + class GroupDocDependencyTests(TestCase): def setUp(self): super().setUp() a = WgDraftFactory() b = WgDraftFactory() RelatedDocument.objects.create( - source=a, target=b.docalias.first(), relationship_id="refnorm" + source=a, target=b, relationship_id="refnorm" ) def test_group_document_dependencies(self): @@ -97,25 +177,13 @@ def test_group_document_dependencies(self): class GenerateGroupAliasesTests(TestCase): - def setUp(self): - super().setUp() - self.doc_aliases_file = NamedTemporaryFile(delete=False, mode='w+') - self.doc_aliases_file.close() - self.doc_virtual_file = NamedTemporaryFile(delete=False, mode='w+') - self.doc_virtual_file.close() - self.saved_draft_aliases_path = settings.GROUP_ALIASES_PATH - self.saved_draft_virtual_path = settings.GROUP_VIRTUAL_PATH - settings.GROUP_ALIASES_PATH = self.doc_aliases_file.name - settings.GROUP_VIRTUAL_PATH = self.doc_virtual_file.name - - def tearDown(self): - settings.GROUP_ALIASES_PATH = self.saved_draft_aliases_path - settings.GROUP_VIRTUAL_PATH = self.saved_draft_virtual_path - os.unlink(self.doc_aliases_file.name) - os.unlink(self.doc_virtual_file.name) - super().tearDown() - - def testManagementCommand(self): + def test_generator_class(self): + """The GroupAliasGenerator should generate the same lists as the old mgmt cmd""" + # clean out test fixture group roles we don't need for this test + Role.objects.filter( + group__acronym__in=["farfut", "iab", "ietf", "irtf", "ise", "ops", "rsab", "rsoc", "sops"] + ).delete() + a_month_ago = timezone.now() - datetime.timedelta(30) a_decade_ago = timezone.now() - datetime.timedelta(3650) role1 = RoleFactory(name_id='ad', group__type_id='area', group__acronym='myth', group__state_id='active') @@ -132,12 +200,11 @@ def testManagementCommand(self): recent = GroupFactory(type_id='wg', acronym='recent', parent=area, state_id='conclude', time=a_month_ago) recentchair = PersonFactory(user__username='recentchair') recent.role_set.create(name_id='chair', person=recentchair, email=recentchair.email()) - wayold = GroupFactory(type_id='wg', acronym='recent', parent=area, state_id='conclude', time=a_decade_ago) + wayold = GroupFactory(type_id='wg', acronym='wayold', parent=area, state_id='conclude', time=a_decade_ago) wayoldchair = PersonFactory(user__username='wayoldchair') wayold.role_set.create(name_id='chair', person=wayoldchair, email=wayoldchair.email()) - role2 = RoleFactory(name_id='ad', group__type_id='area', group__acronym='done', group__state_id='conclude') - done = role2.group - done_ad = role2.person + # create a "done" group that should not be included anywhere + RoleFactory(name_id='ad', group__type_id='area', group__acronym='done', group__state_id='conclude') irtf = Group.objects.get(acronym='irtf') testrg = GroupFactory(type_id='rg', acronym='testrg', parent=irtf) testrgchair = PersonFactory(user__username='testrgchair') @@ -145,77 +212,88 @@ def testManagementCommand(self): testrag = GroupFactory(type_id='rg', acronym='testrag', parent=irtf) testragchair = PersonFactory(user__username='testragchair') testrag.role_set.create(name_id='chair', person=testragchair, email=testragchair.email()) - individual = PersonFactory() - - args = [ ] - kwargs = { } - out = io.StringIO() - call_command("generate_group_aliases", *args, **kwargs, stdout=out, stderr=out) - self.assertFalse(out.getvalue()) - - with open(settings.GROUP_ALIASES_PATH) as afile: - acontent = afile.read() - self.assertTrue('xfilter-' + area.acronym + '-ads' in acontent) - self.assertTrue('xfilter-' + area.acronym + '-chairs' in acontent) - self.assertTrue('xfilter-' + mars.acronym + '-ads' in acontent) - self.assertTrue('xfilter-' + mars.acronym + '-chairs' in acontent) - self.assertTrue('xfilter-' + ames.acronym + '-ads' in acontent) - self.assertTrue('xfilter-' + ames.acronym + '-chairs' in acontent) - self.assertTrue(all([x in acontent for x in [ - 'xfilter-' + area.acronym + '-ads', - 'xfilter-' + area.acronym + '-chairs', - 'xfilter-' + mars.acronym + '-ads', - 'xfilter-' + mars.acronym + '-chairs', - 'xfilter-' + ames.acronym + '-ads', - 'xfilter-' + ames.acronym + '-chairs', - 'xfilter-' + recent.acronym + '-ads', - 'xfilter-' + recent.acronym + '-chairs', - ]])) - self.assertFalse(all([x in acontent for x in [ - 'xfilter-' + done.acronym + '-ads', - 'xfilter-' + done.acronym + '-chairs', - 'xfilter-' + wayold.acronym + '-ads', - 'xfilter-' + wayold.acronym + '-chairs', - ]])) - - with open(settings.GROUP_VIRTUAL_PATH) as vfile: - vcontent = vfile.read() - self.assertTrue(all([x in vcontent for x in [ - ad.email_address(), - marschair.email_address(), - marssecr.email_address(), - ameschair.email_address(), - recentchair.email_address(), - testrgchair.email_address(), - testragchair.email_address(), - ]])) - self.assertFalse(all([x in vcontent for x in [ - done_ad.email_address(), - wayoldchair.email_address(), - individual.email_address(), - ]])) - self.assertTrue(all([x in vcontent for x in [ - 'xfilter-' + area.acronym + '-ads', - 'xfilter-' + area.acronym + '-chairs', - 'xfilter-' + mars.acronym + '-ads', - 'xfilter-' + mars.acronym + '-chairs', - 'xfilter-' + ames.acronym + '-ads', - 'xfilter-' + ames.acronym + '-chairs', - 'xfilter-' + recent.acronym + '-ads', - 'xfilter-' + recent.acronym + '-chairs', - 'xfilter-' + testrg.acronym + '-chairs', - 'xfilter-' + testrag.acronym + '-chairs', - testrg.acronym + '-chairs@ietf.org', - testrg.acronym + '-chairs@irtf.org', - testrag.acronym + '-chairs@ietf.org', - testrag.acronym + '-chairs@irtf.org', - ]])) - self.assertFalse(all([x in vcontent for x in [ - 'xfilter-' + done.acronym + '-ads', - 'xfilter-' + done.acronym + '-chairs', - 'xfilter-' + wayold.acronym + '-ads', - 'xfilter-' + wayold.acronym + '-chairs', - ]])) + + output = [(alias, (domains, alist)) for alias, domains, alist in GroupAliasGenerator()] + alias_dict = dict(output) + self.maxDiff = None + self.assertEqual(len(alias_dict), len(output)) # no duplicate aliases + expected_dict = { + area.acronym + "-ads": (["ietf"], [ad.email_address()]), + area.acronym + "-chairs": (["ietf"], [ad.email_address(), marschair.email_address(), marssecr.email_address(), ameschair.email_address()]), + mars.acronym + "-ads": (["ietf"], [ad.email_address()]), + mars.acronym + "-chairs": (["ietf"], [marschair.email_address(), marssecr.email_address()]), + ames.acronym + "-ads": (["ietf"], [ad.email_address()]), + ames.acronym + "-chairs": (["ietf"], [ameschair.email_address()]), + recent.acronym + "-ads": (["ietf"], [ad.email_address()]), + recent.acronym + "-chairs": (["ietf"], [recentchair.email_address()]), + testrg.acronym + "-chairs": (["ietf", "irtf"], [testrgchair.email_address()]), + testrag.acronym + "-chairs": (["ietf", "irtf"], [testragchair.email_address()]), + } + # Sort lists for comparison + self.assertEqual( + {k: (sorted(doms), sorted(addrs)) for k, (doms, addrs) in alias_dict.items()}, + {k: (sorted(doms), sorted(addrs)) for k, (doms, addrs) in expected_dict.items()}, + ) + + @mock.patch("ietf.group.utils.GroupAliasGenerator") + def test_get_group_email_aliases(self, mock_alias_gen_cls): + GroupFactory(name="agroup", type_id="rg") + GroupFactory(name="bgroup") + GroupFactory(name="cgroup", type_id="rg") + GroupFactory(name="dgroup") + + mock_alias_gen_cls.return_value = [ + ("bgroup-chairs", ["ietf"], ["c1@example.com", "c2@example.com"]), + ("agroup-ads", ["ietf", "irtf"], ["ad@example.com"]), + ("bgroup-ads", ["ietf"], ["ad@example.com"]), + ] + # order is important - should be by acronym, otherwise left in order returned by generator + self.assertEqual( + get_group_email_aliases(None, None), + [ + { + "acronym": "agroup", + "alias_type": "-ads", + "expansion": "ad@example.com", + }, + { + "acronym": "bgroup", + "alias_type": "-chairs", + "expansion": "c1@example.com, c2@example.com", + }, + { + "acronym": "bgroup", + "alias_type": "-ads", + "expansion": "ad@example.com", + }, + ], + ) + self.assertQuerySetEqual( + mock_alias_gen_cls.call_args[0][0], + Group.objects.all(), + ordered=False, + ) + + # test other parameter combinations but we already checked that the alias generator's + # output will be passed through, so don't re-test the processing + get_group_email_aliases("agroup", None) + self.assertQuerySetEqual( + mock_alias_gen_cls.call_args[0][0], + Group.objects.filter(acronym="agroup"), + ordered=False, + ) + get_group_email_aliases(None, "wg") + self.assertQuerySetEqual( + mock_alias_gen_cls.call_args[0][0], + Group.objects.filter(type_id="wg"), + ordered=False, + ) + get_group_email_aliases("agroup", "wg") + self.assertQuerySetEqual( + mock_alias_gen_cls.call_args[0][0], + Group.objects.none(), + ordered=False, + ) class GroupRoleEmailTests(TestCase): @@ -255,3 +333,41 @@ def test_group_ad_emails(self): self.assertGreater(len(emails), 0) for item in emails: self.assertIn('@', item) + + def test_role_holder_emails(self): + # The test fixtures create a bunch of addresses that pollute this test's results - disable them + Email.objects.update(active=False) + + role_holders = [ + RoleFactory(name_id="member", group__type_id=gt).person + for gt in [ + "ag", + "area", + "dir", + "iab", + "ietf", + "irtf", + "nomcom", + "rg", + "team", + "wg", + "rag", + ] + ] + # Expect an additional active email to be included + EmailFactory( + person=role_holders[0], + active=True, + ) + # Do not expect an inactive email to be included + EmailFactory( + person=role_holders[1], + active=False, + ) + # Do not expect address on a role-holder for a different group type + RoleFactory(name_id="member", group__type_id="adhoc") # arbitrary type not in the of-interest list + + self.assertCountEqual( + role_holder_emails(), + Email.objects.filter(active=True, person__in=role_holders), + ) diff --git a/ietf/group/tests_appeals.py b/ietf/group/tests_appeals.py new file mode 100644 index 0000000000..ae6c54bf47 --- /dev/null +++ b/ietf/group/tests_appeals.py @@ -0,0 +1,71 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +import debug # pyflakes: ignore +import datetime + +from pyquery import PyQuery + +from django.urls import reverse as urlreverse + +from ietf.utils.test_utils import login_testing_unauthorized, TestCase +from ietf.group.factories import ( + AppealFactory, + AppealArtifactFactory, +) +class AppealTests(TestCase): + + def test_download_name(self): + artifact = AppealArtifactFactory() + self.assertEqual(artifact.download_name(),f"{artifact.date}-appeal.md") + artifact = AppealArtifactFactory(content_type="application/pdf",artifact_type__slug="response") + self.assertEqual(artifact.download_name(),f"{artifact.date}-response.pdf") + + + def test_appeal_list_view(self): + appeal_date = datetime.date.today()-datetime.timedelta(days=14) + response_date = appeal_date+datetime.timedelta(days=8) + appeal = AppealFactory(name="A name to look for", date=appeal_date) + appeal_artifact = AppealArtifactFactory(appeal=appeal, artifact_type__slug="appeal", date=appeal_date) + response_artifact = AppealArtifactFactory(appeal=appeal, artifact_type__slug="response", content_type="application/pdf", date=response_date) + + url = urlreverse("ietf.group.views.appeals", kwargs=dict(acronym="iab")) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q("#appeals > tbody > tr")), 1) + self.assertEqual(q("#appeal-1-date").text(), f"{appeal_date}") + self.assertEqual(f"{appeal_artifact.display_title()} - {appeal_date}", q("#artifact-1-1").text()) + self.assertEqual(f"{response_artifact.display_title()} - {response_date}", q("#artifact-1-2").text()) + self.assertIsNone(q("#artifact-1-1").attr("download")) + self.assertEqual(q("#artifact-1-2").attr("download"), response_artifact.download_name()) + + def test_markdown_view(self): + artifact = AppealArtifactFactory() + url = urlreverse("ietf.group.views.appeal_artifact", kwargs=dict(acronym="iab", artifact_id=artifact.pk)) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(q("#content>p>strong").text(),"Markdown") + self.assertIsNone(q("#content a").attr("download")) + self.client.login(username='secretary', password='secretary+password') + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(q("#content a").attr("download"), artifact.download_name()) + + def test_markdown_download(self): + artifact = AppealArtifactFactory() + url = urlreverse("ietf.group.views.appeal_artifact_markdown", kwargs=dict(acronym="iab", artifact_id=artifact.pk)) + login_testing_unauthorized(self, "secretary", url) + r = self.client.get(url) + self.assertContains(r, "**Markdown**", status_code=200) + + def test_pdf_download(self): + artifact = AppealArtifactFactory(content_type="application/pdf") # The bits won't _really_ be pdf + url = urlreverse("ietf.group.views.appeal_artifact", kwargs=dict(acronym="iab", artifact_id=artifact.pk)) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.get("Content-Disposition"), f'attachment; filename="{artifact.download_name()}"') + self.assertEqual(r.get("Content-Type"), artifact.content_type) + self.assertEqual(r.content, artifact.bits.tobytes()) + diff --git a/ietf/group/tests_info.py b/ietf/group/tests_info.py index 98af69ba07..3f24e2e3d6 100644 --- a/ietf/group/tests_info.py +++ b/ietf/group/tests_info.py @@ -1,22 +1,23 @@ -# Copyright The IETF Trust 2009-2022, All Rights Reserved +# Copyright The IETF Trust 2009-2024, All Rights Reserved # -*- coding: utf-8 -*- -import os import calendar import datetime import io import bleach +from unittest import mock -from unittest.mock import patch +from unittest.mock import call, patch from pathlib import Path from pyquery import PyQuery -from tempfile import NamedTemporaryFile import debug # pyflakes:ignore from django.conf import settings +from django.http import Http404, HttpResponse from django.test import RequestFactory +from django.test.utils import override_settings from django.urls import reverse as urlreverse from django.urls import NoReverseMatch from django.utils import timezone @@ -26,14 +27,17 @@ from ietf.community.models import CommunityList from ietf.community.utils import reset_name_contains_index_for_rule -from ietf.doc.factories import WgDraftFactory, IndividualDraftFactory, CharterFactory, BallotDocEventFactory -from ietf.doc.models import Document, DocAlias, DocEvent, State +from ietf.doc.factories import WgDraftFactory, RgDraftFactory, IndividualDraftFactory, CharterFactory, BallotDocEventFactory +from ietf.doc.models import Document, DocEvent, State +from ietf.doc.storage_utils import retrieve_str from ietf.doc.utils_charter import charter_name_for_group from ietf.group.admin import GroupForm as AdminGroupForm from ietf.group.factories import (GroupFactory, RoleFactory, GroupEventFactory, DatedGroupMilestoneFactory, DatelessGroupMilestoneFactory) from ietf.group.forms import GroupForm from ietf.group.models import Group, GroupEvent, GroupMilestone, GroupStateTransitions, Role +from ietf.group.tasks import generate_wg_charters_files_task, generate_wg_summary_files_task +from ietf.group.views import response_from_file from ietf.group.utils import save_group_in_history, setup_default_community_list_for_group from ietf.meeting.factories import SessionFactory from ietf.name.models import DocTagName, GroupStateName, GroupTypeName, ExtResourceName, RoleName @@ -56,7 +60,13 @@ def pklist(docs): return [ str(doc.pk) for doc in docs.all() ] class GroupPagesTests(TestCase): - settings_temp_path_overrides = TestCase.settings_temp_path_overrides + ['CHARTER_PATH'] + settings_temp_path_overrides = TestCase.settings_temp_path_overrides + [ + "CHARTER_PATH", + "CHARTER_COPY_PATH", + "CHARTER_COPY_OTHER_PATH", # Note: not explicitly testing use of + "CHARTER_COPY_THIRD_PATH", # either of these settings + "GROUP_SUMMARY_PATH", + ] def test_active_groups(self): area = GroupFactory.create(type_id='area') @@ -71,7 +81,7 @@ def test_active_groups(self): self.assertContains(r, group.name) self.assertContains(r, escape(group.ad_role().person.name)) - for t in ('rg','area','ag', 'rag', 'dir','review','team','program','iabasg','adm','rfcedtyp'): + for t in ('rg','area','ag', 'rag', 'dir','review','team','program','iabasg','iabworkshop','adm','rfcedtyp'): # See issue 5120 g = GroupFactory.create(type_id=t,state_id='active') if t in ['dir','review']: g.parent = GroupFactory.create(type_id='area',state_id='active') @@ -80,6 +90,12 @@ def test_active_groups(self): r = self.client.get(url) self.assertEqual(r.status_code, 200) self.assertContains(r, g.acronym) + if t == "area": + q = PyQuery(r.content) + wg_url = urlreverse("ietf.group.views.active_groups", kwargs=dict(group_type="wg")) + href = f"{wg_url}#{g.acronym.upper()}" + self.assertEqual(q(f"h2#id-{g.acronym} a").attr("href"), href) + self.assertEqual(q(f'h2#id-{g.acronym} a[href="{href}"]').text(), f"({g.acronym.upper()})") url = urlreverse('ietf.group.views.active_groups', kwargs=dict()) r = self.client.get(url) @@ -87,7 +103,7 @@ def test_active_groups(self): self.assertContains(r, "Directorate") self.assertContains(r, "AG") - for slug in GroupTypeName.objects.exclude(slug__in=['wg','rg','ag','rag','area','dir','review','team','program','adhoc','ise','adm','iabasg','rfcedtyp']).values_list('slug',flat=True): + for slug in GroupTypeName.objects.exclude(slug__in=['wg','rg','ag','rag','area','dir','review','team','program','adhoc','ise','adm','iabasg','iabworkshop','rfcedtyp', 'edwg', 'edappr']).values_list('slug',flat=True): with self.assertRaises(NoReverseMatch): url=urlreverse('ietf.group.views.active_groups', kwargs=dict(group_type=slug)) @@ -110,48 +126,204 @@ def test_group_home(self): self.assertContains(r, draft.name) self.assertContains(r, draft.title) - def test_wg_summaries(self): - group = CharterFactory(group__type_id='wg',group__parent=GroupFactory(type_id='area')).group - RoleFactory(group=group,name_id='chair',person=PersonFactory()) - RoleFactory(group=group,name_id='ad',person=PersonFactory()) + def test_response_from_file(self): + # n.b., GROUP_SUMMARY_PATH is a temp dir that will be cleaned up automatically + fp = Path(settings.GROUP_SUMMARY_PATH) / "some-file.txt" + fp.write_text("This is a charters file with an é") + r = response_from_file(fp) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.headers["Content-Type"], "text/plain; charset=utf-8") + self.assertEqual(r.content.decode("utf8"), "This is a charters file with an é") + # now try with a nonexistent file + fp.unlink() + with self.assertRaises(Http404): + response_from_file(fp) + + @patch("ietf.group.views.response_from_file") + def test_wg_summary_area(self, mock): + r = self.client.get( + urlreverse("ietf.group.views.wg_summary_area", kwargs={"group_type": "rg"}) + ) # not wg + self.assertEqual(r.status_code, 404) + self.assertFalse(mock.called) + mock.return_value = HttpResponse("yay") + r = self.client.get( + urlreverse("ietf.group.views.wg_summary_area", kwargs={"group_type": "wg"}) + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.content.decode(), "yay") + self.assertEqual(mock.call_args, call(Path(settings.GROUP_SUMMARY_PATH) / "1wg-summary.txt")) + + @patch("ietf.group.views.response_from_file") + def test_wg_summary_acronym(self, mock): + r = self.client.get( + urlreverse( + "ietf.group.views.wg_summary_acronym", kwargs={"group_type": "rg"} + ) + ) # not wg + self.assertEqual(r.status_code, 404) + self.assertFalse(mock.called) + mock.return_value = HttpResponse("yay") + r = self.client.get( + urlreverse( + "ietf.group.views.wg_summary_acronym", kwargs={"group_type": "wg"} + ) + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.content.decode(), "yay") + self.assertEqual( + mock.call_args, call(Path(settings.GROUP_SUMMARY_PATH) / "1wg-summary-by-acronym.txt") + ) + @patch("ietf.group.views.response_from_file") + def test_wg_charters(self, mock): + r = self.client.get( + urlreverse("ietf.group.views.wg_charters", kwargs={"group_type": "rg"}) + ) # not wg + self.assertEqual(r.status_code, 404) + self.assertFalse(mock.called) + mock.return_value = HttpResponse("yay") + r = self.client.get( + urlreverse("ietf.group.views.wg_charters", kwargs={"group_type": "wg"}) + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.content.decode(), "yay") + self.assertEqual(mock.call_args, call(Path(settings.CHARTER_PATH) / "1wg-charters.txt")) + + @patch("ietf.group.views.response_from_file") + def test_wg_charters_by_acronym(self, mock): + r = self.client.get( + urlreverse( + "ietf.group.views.wg_charters_by_acronym", kwargs={"group_type": "rg"} + ) + ) # not wg + self.assertEqual(r.status_code, 404) + self.assertFalse(mock.called) + mock.return_value = HttpResponse("yay") + r = self.client.get( + urlreverse( + "ietf.group.views.wg_charters_by_acronym", kwargs={"group_type": "wg"} + ) + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.content.decode(), "yay") + self.assertEqual( + mock.call_args, call(Path(settings.CHARTER_PATH) / "1wg-charters-by-acronym.txt") + ) + + def test_generate_wg_charters_files_task(self): + group = CharterFactory( + group__type_id="wg", group__parent=GroupFactory(type_id="area") + ).group + RoleFactory(group=group, name_id="chair", person=PersonFactory()) + RoleFactory(group=group, name_id="ad", person=PersonFactory()) chair = Email.objects.filter(role__group=group, role__name="chair")[0] + ( + Path(settings.CHARTER_PATH) / f"{group.charter.name}-{group.charter.rev}.txt" + ).write_text("This is a charter.") - with (Path(settings.CHARTER_PATH) / ("%s-%s.txt" % (group.charter.canonical_name(), group.charter.rev))).open("w") as f: - f.write("This is a charter.") + generate_wg_charters_files_task() + wg_charters_contents = (Path(settings.CHARTER_PATH) / "1wg-charters.txt").read_text( + encoding="utf8" + ) + self.assertIn(group.acronym, wg_charters_contents) + self.assertIn(group.name, wg_charters_contents) + self.assertIn(group.ad_role().person.plain_name(), wg_charters_contents) + self.assertIn(chair.address, wg_charters_contents) + self.assertIn("This is a charter.", wg_charters_contents) + wg_charters_copy = ( + Path(settings.CHARTER_COPY_PATH) / "1wg-charters.txt" + ).read_text(encoding="utf8") + self.assertEqual(wg_charters_copy, wg_charters_contents) + + wg_charters_by_acronym_contents = ( + Path(settings.CHARTER_PATH) / "1wg-charters-by-acronym.txt" + ).read_text(encoding="utf8") + self.assertIn(group.acronym, wg_charters_by_acronym_contents) + self.assertIn(group.name, wg_charters_by_acronym_contents) + self.assertIn(group.ad_role().person.plain_name(), wg_charters_by_acronym_contents) + self.assertIn(chair.address, wg_charters_by_acronym_contents) + self.assertIn("This is a charter.", wg_charters_by_acronym_contents) + wg_charters_by_acronymcopy = ( + Path(settings.CHARTER_COPY_PATH) / "1wg-charters-by-acronym.txt" + ).read_text(encoding="utf8") + self.assertEqual(wg_charters_by_acronymcopy, wg_charters_by_acronym_contents) + + def test_generate_wg_charters_files_task_without_copy(self): + """Test disabling charter file copying + + Note that these tests mostly check that errors are not encountered. Because they unset + the CHARTER_COPY_PATH or set it to a non-directory destination, it's not clear where to + look to see whether the files were (incorrectly) copied somewhere. + """ + group = CharterFactory( + group__type_id="wg", group__parent=GroupFactory(type_id="area") + ).group + ( + Path(settings.CHARTER_PATH) / f"{group.charter.name}-{group.charter.rev}.txt" + ).write_text("This is a charter.") + + # No directory set + with override_settings(): + del settings.CHARTER_COPY_PATH + generate_wg_charters_files_task() + # n.b., CHARTER_COPY_PATH is set again outside the with block + self.assertTrue((Path(settings.CHARTER_PATH) / "1wg-charters.txt").exists()) + self.assertFalse((Path(settings.CHARTER_COPY_PATH) / "1wg-charters.txt").exists()) + self.assertTrue( + (Path(settings.CHARTER_PATH) / "1wg-charters-by-acronym.txt").exists() + ) + self.assertFalse( + (Path(settings.CHARTER_COPY_PATH) / "1wg-charters-by-acronym.txt").exists() + ) + (Path(settings.CHARTER_PATH) / "1wg-charters.txt").unlink() + (Path(settings.CHARTER_PATH) / "1wg-charters-by-acronym.txt").unlink() + + # Set to a file, not a directory + not_a_dir = Path(settings.CHARTER_COPY_PATH) / "not-a-dir.txt" + not_a_dir.write_text("Not a dir") + with override_settings(CHARTER_COPY_PATH=str(not_a_dir)): + generate_wg_charters_files_task() + # n.b., CHARTER_COPY_PATH is set again outside the with block + self.assertTrue((Path(settings.CHARTER_PATH) / "1wg-charters.txt").exists()) + self.assertFalse((Path(settings.CHARTER_COPY_PATH) / "1wg-charters.txt").exists()) + self.assertTrue( + (Path(settings.CHARTER_PATH) / "1wg-charters-by-acronym.txt").exists() + ) + self.assertFalse( + (Path(settings.CHARTER_COPY_PATH) / "1wg-charters-by-acronym.txt").exists() + ) + self.assertEqual(not_a_dir.read_text(), "Not a dir") - url = urlreverse('ietf.group.views.wg_summary_area', kwargs=dict(group_type="wg")) - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - self.assertContains(r, group.parent.name) - self.assertContains(r, group.acronym) - self.assertContains(r, group.name) - self.assertContains(r, chair.address) + def test_generate_wg_summary_files_task(self): + group = CharterFactory(group__type_id='wg',group__parent=GroupFactory(type_id='area')).group + RoleFactory(group=group,name_id='chair',person=PersonFactory()) + RoleFactory(group=group,name_id='ad',person=PersonFactory()) - url = urlreverse('ietf.group.views.wg_summary_acronym', kwargs=dict(group_type="wg")) - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - self.assertContains(r, group.acronym) - self.assertContains(r, group.name) - self.assertContains(r, chair.address) - - url = urlreverse('ietf.group.views.wg_charters', kwargs=dict(group_type="wg")) - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - self.assertContains(r, group.acronym) - self.assertContains(r, group.name) - self.assertContains(r, group.ad_role().person.plain_name()) - self.assertContains(r, chair.address) - self.assertContains(r, "This is a charter.") + chair = Email.objects.filter(role__group=group, role__name="chair")[0] - url = urlreverse('ietf.group.views.wg_charters_by_acronym', kwargs=dict(group_type="wg")) - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - self.assertContains(r, group.acronym) - self.assertContains(r, group.name) - self.assertContains(r, group.ad_role().person.plain_name()) - self.assertContains(r, chair.address) - self.assertContains(r, "This is a charter.") + generate_wg_summary_files_task() + + for summary_by_area_contents in [ + ( + Path(settings.GROUP_SUMMARY_PATH) / "1wg-summary.txt" + ).read_text(encoding="utf8"), + retrieve_str("indexes", "1wg-summary.txt") + ]: + self.assertIn(group.parent.name, summary_by_area_contents) + self.assertIn(group.acronym, summary_by_area_contents) + self.assertIn(group.name, summary_by_area_contents) + self.assertIn(chair.address, summary_by_area_contents) + + for summary_by_acronym_contents in [ + ( + Path(settings.GROUP_SUMMARY_PATH) / "1wg-summary-by-acronym.txt" + ).read_text(encoding="utf8"), + retrieve_str("indexes", "1wg-summary-by-acronym.txt") + ]: + self.assertIn(group.acronym, summary_by_acronym_contents) + self.assertIn(group.name, summary_by_acronym_contents) + self.assertIn(chair.address, summary_by_acronym_contents) def test_chartering_groups(self): group = CharterFactory(group__type_id='wg',group__parent=GroupFactory(type_id='area'),states=[('charter','intrev')]).group @@ -241,6 +413,7 @@ def test_group_documents(self): self.assertContains(r, draft3.name) for ah in draft3.action_holders.all(): self.assertContains(r, escape(ah.name)) + self.assertContains(r, "Active with the IESG Internet-Draft") # draft3 is pub-req hence should have such a divider self.assertContains(r, 'for 173 days', count=1) # the old_dah should be tagged self.assertContains(r, draft4.name) self.assertNotContains(r, draft5.name) @@ -253,6 +426,25 @@ def test_group_documents(self): q = PyQuery(r.content) self.assertTrue(any([draft2.name in x.attrib['href'] for x in q('table td a.track-untrack-doc')])) + # Let's also check the IRTF stream + rg = GroupFactory(type_id='rg') + setup_default_community_list_for_group(rg) + rgDraft = RgDraftFactory(group=rg) + rgDraft4 = RgDraftFactory(group=rg) + rgDraft4.set_state(State.objects.get(slug='irsg-w')) + rgDraft7 = RgDraftFactory(group=rg) + rgDraft7.set_state(State.objects.get(type='draft-stream-%s' % rgDraft7.stream_id, slug='dead')) + for url in group_urlreverse_list(rg, 'ietf.group.views.group_documents'): + with self.settings(DOC_ACTION_HOLDER_MAX_AGE_DAYS=20): + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + self.assertContains(r, rgDraft.name) + self.assertContains(r, rg.name) + self.assertContains(r, rg.acronym) + self.assertNotContains(r, draft3.name) # As draft3 is a WG draft, it should not be listed here + self.assertContains(r, rgDraft4.name) + self.assertNotContains(r, rgDraft7.name) + # test the txt version too while we're at it for url in group_urlreverse_list(group, 'ietf.group.views.group_documents_txt'): r = self.client.get(url) @@ -264,8 +456,9 @@ def test_group_charter(self): group = CharterFactory().group draft = WgDraftFactory(group=group) - with (Path(settings.CHARTER_PATH) / ("%s-%s.txt" % (group.charter.canonical_name(), group.charter.rev))).open("w") as f: - f.write("This is a charter.") + ( + Path(settings.CHARTER_PATH) / f"{group.charter.name}-{group.charter.rev}.txt" + ).write_text("This is a charter.") milestone = GroupMilestone.objects.create( group=group, @@ -283,12 +476,6 @@ def test_group_charter(self): self.assertContains(r, milestone.desc) self.assertContains(r, milestone.docs.all()[0].name) - def test_about_rendertest(self): - group = CharterFactory().group - url = urlreverse('ietf.group.views.group_about_rendertest', kwargs=dict(acronym=group.acronym)) - r = self.client.get(url) - self.assertEqual(r.status_code,200) - def test_group_about(self): @@ -356,6 +543,25 @@ def verify_can_edit_group(url, group, username): for username in list(set(interesting_users)-set(can_edit[group.type_id])): verify_cannot_edit_group(url, group, username) + def test_group_about_team_parent(self): + """Team about page should show parent when parent is not an area""" + GroupFactory(type_id='team', parent=GroupFactory(type_id='area', acronym='gen')) + GroupFactory(type_id='team', parent=GroupFactory(type_id='ietf', acronym='iab')) + GroupFactory(type_id='team', parent=None) + + for team in Group.objects.filter(type='team').select_related('parent'): + url = urlreverse('ietf.group.views.group_about', kwargs=dict(acronym=team.acronym)) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + if team.parent and team.parent.type_id != 'area': + self.assertContains(r, 'Parent') + self.assertContains(r, team.parent.acronym) + elif team.parent and team.parent.type_id == 'area': + self.assertContains(r, team.parent.name) + self.assertNotContains(r, '>Parent<') + else: + self.assertNotContains(r, '>Parent<') + def test_group_about_personnel(self): """Correct personnel should appear on the group About page""" group = GroupFactory() @@ -391,7 +597,6 @@ def test_materials(self): type_id="slides", ) doc.set_state(State.objects.get(type="slides", slug="active")) - DocAlias.objects.create(name=doc.name).docs.add(doc) for url in group_urlreverse_list(group, 'ietf.group.views.materials'): r = self.client.get(url) @@ -478,13 +683,13 @@ def test_nonactive_group_badges(self): r = self.client.get(url) self.assertEqual(r.status_code,200) q = PyQuery(r.content) - self.assertEqual(q('.badge.bg-warning').text(),"Concluded WG") + self.assertEqual(q('.badge.text-bg-warning').text(),"Concluded WG") replaced_group = GroupFactory(state_id='replaced') url = urlreverse("ietf.group.views.history",kwargs={'acronym':replaced_group.acronym}) r = self.client.get(url) self.assertEqual(r.status_code,200) q = PyQuery(r.content) - self.assertEqual(q('.badge.bg-warning').text(),"Replaced WG") + self.assertEqual(q('.badge.text-bg-warning').text(),"Replaced WG") class GroupEditTests(TestCase): @@ -616,7 +821,7 @@ def test_create_based_on_existing_bof(self): def test_create_non_chartered_includes_description(self): parent = GroupFactory(type_id='area') - group_type = GroupTypeName.objects.filter(used=True, features__has_chartering_process=False).first() + group_type = GroupTypeName.objects.filter(used=True, features__has_chartering_process=False, features__parent_types='area').first() self.assertIsNotNone(group_type) url = urlreverse('ietf.group.views.edit', kwargs=dict(group_type=group_type.slug, action="create")) login_testing_unauthorized(self, "secretary", url) @@ -674,8 +879,9 @@ def test_edit_info(self): self.assertTrue(len(q('form .is-invalid')) > 0) # edit info - with (Path(settings.CHARTER_PATH) / ("%s-%s.txt" % (group.charter.canonical_name(), group.charter.rev))).open("w") as f: - f.write("This is a charter.") + ( + Path(settings.CHARTER_PATH) / f"{group.charter.name}-{group.charter.rev}.txt" + ).write_text("This is a charter.") area = group.parent ad = Person.objects.get(name="Areað Irector") state = GroupStateName.objects.get(slug="bof") @@ -717,7 +923,9 @@ def test_edit_info(self): self.assertEqual(group.list_archive, "archive.mars") self.assertEqual(group.description, '') - self.assertTrue((Path(settings.CHARTER_PATH) / ("%s-%s.txt" % (group.charter.canonical_name(), group.charter.rev))).exists()) + self.assertTrue( + (Path(settings.CHARTER_PATH) / f"{group.charter.name}-{group.charter.rev}.txt").exists() + ) self.assertEqual(len(outbox), 2) self.assertTrue('Personnel change' in outbox[0]['Subject']) for prefix in ['ad1','ad2','aread','marschairman','marsdelegate']: @@ -952,10 +1160,88 @@ def test_edit_description_field(self): r = self.client.post(url, { 'description': 'Ignored description', }) - self.assertEqual(r.status_code, 302) + self.assertEqual(r.status_code, 403) group = Group.objects.get(pk=group.pk) # refresh self.assertEqual(group.description, 'Updated description') + def test_edit_parent(self): + group = GroupFactory.create(type_id='wg', parent=GroupFactory.create(type_id='area')) + chair = RoleFactory(group=group, name_id='chair').person + url = urlreverse('ietf.group.views.edit', kwargs=dict(group_type=group.type_id, acronym=group.acronym, action='edit')) + + # parent is not shown to group chair + login_testing_unauthorized(self, chair.user.username, url) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q('form select[name=parent]')), 0) + + # view ignores attempt to change parent + old_parent = group.parent + new_parent = GroupFactory(type_id='area') + self.assertNotEqual(new_parent.acronym, group.parent.acronym) + r = self.client.post(url, dict( + name=group.name, + acronym=group.acronym, + state=group.state_id, + parent=new_parent.pk)) + self.assertEqual(r.status_code, 302) + group = Group.objects.get(pk=group.pk) + self.assertNotEqual(group.parent, new_parent) + self.assertEqual(group.parent, old_parent) + + # parent is shown to AD and Secretariat + for priv_user in ('ad', 'secretary'): + self.client.logout() + login_testing_unauthorized(self, priv_user, url) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q('form select[name=parent]')), 1) + + new_parent = GroupFactory(type_id='area') + self.assertNotEqual(new_parent.acronym, group.parent.acronym) + r = self.client.post(url, dict( + name=group.name, + acronym=group.acronym, + state=group.state_id, + parent=new_parent.pk)) + self.assertEqual(r.status_code, 302) + group = Group.objects.get(pk=group.pk) + self.assertEqual(group.parent, new_parent) + + def test_edit_parent_field(self): + group = GroupFactory.create(type_id='wg', parent=GroupFactory.create(type_id='area')) + chair = RoleFactory(group=group, name_id='chair').person + url = urlreverse('ietf.group.views.edit', kwargs=dict(group_type=group.type_id, acronym=group.acronym, action='edit', field='parent')) + + # parent is not shown to group chair + login_testing_unauthorized(self, chair.user.username, url) + r = self.client.get(url) + self.assertEqual(r.status_code, 403) + + # chair is not allowed to change parent + new_parent = GroupFactory(type_id='area') + self.assertNotEqual(new_parent.acronym, group.parent.acronym) + r = self.client.post(url, dict(parent=new_parent.pk)) + self.assertEqual(r.status_code, 403) + + # parent is shown to AD and Secretariat + for priv_user in ('ad', 'secretary'): + self.client.logout() + login_testing_unauthorized(self, priv_user, url) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q('form select[name=parent]')), 1) + + new_parent = GroupFactory(type_id='area') + self.assertNotEqual(new_parent.acronym, group.parent.acronym) + r = self.client.post(url, dict(parent=new_parent.pk)) + self.assertEqual(r.status_code, 302) + group = Group.objects.get(pk=group.pk) + self.assertEqual(group.parent, new_parent) + def test_conclude(self): group = GroupFactory(acronym="mars") @@ -1193,7 +1479,7 @@ def create_test_milestones(self): RoleFactory(group=group,name_id='chair',person=PersonFactory(user__username='marschairman')) draft = WgDraftFactory(group=group) - m1 = GroupMilestone.objects.create(id=1, + m1 = GroupMilestone.objects.create( group=group, desc="Test 1", due=date_today(DEADLINE_TZINFO), @@ -1201,7 +1487,7 @@ def create_test_milestones(self): state_id="active") m1.docs.set([draft]) - m2 = GroupMilestone.objects.create(id=2, + m2 = GroupMilestone.objects.create( group=group, desc="Test 2", due=date_today(DEADLINE_TZINFO), @@ -1342,13 +1628,14 @@ def test_accept_milestone(self): events_before = group.groupevent_set.count() # add - r = self.client.post(url, { 'prefix': "m1", - 'm1-id': m1.id, - 'm1-desc': m1.desc, - 'm1-due': m1.due.strftime("%B %Y"), - 'm1-resolved': m1.resolved, - 'm1-docs': pklist(m1.docs), - 'm1-review': "accept", + mstr = f"m{m1.id}" + r = self.client.post(url, { 'prefix': mstr, + f'{mstr}-id': m1.id, + f'{mstr}-desc': m1.desc, + f'{mstr}-due': m1.due.strftime("%B %Y"), + f'{mstr}-resolved': m1.resolved, + f'{mstr}-docs': pklist(m1.docs), + f'{mstr}-review': "accept", 'action': "save", }) self.assertEqual(r.status_code, 302) @@ -1368,13 +1655,14 @@ def test_delete_milestone(self): events_before = group.groupevent_set.count() # delete - r = self.client.post(url, { 'prefix': "m1", - 'm1-id': m1.id, - 'm1-desc': m1.desc, - 'm1-due': m1.due.strftime("%B %Y"), - 'm1-resolved': "", - 'm1-docs': pklist(m1.docs), - 'm1-delete': "checked", + mstr = f"m{m1.id}" + r = self.client.post(url, { 'prefix': mstr, + f'{mstr}-id': m1.id, + f'{mstr}-desc': m1.desc, + f'{mstr}-due': m1.due.strftime("%B %Y"), + f'{mstr}-resolved': "", + f'{mstr}-docs': pklist(m1.docs), + f'{mstr}-delete': "checked", 'action': "save", }) self.assertEqual(r.status_code, 302) @@ -1397,13 +1685,14 @@ def test_edit_milestone(self): due = self.last_day_of_month(date_today(DEADLINE_TZINFO) + datetime.timedelta(days=365)) + mstr = f"m{m1.id}" # faulty post - r = self.client.post(url, { 'prefix': "m1", - 'm1-id': m1.id, - 'm1-desc': "", # no description - 'm1-due': due.strftime("%B %Y"), - 'm1-resolved': "", - 'm1-docs': doc_pks, + r = self.client.post(url, { 'prefix': mstr, + f'{mstr}-id': m1.id, + f'{mstr}-desc': "", # no description + f'{mstr}-due': due.strftime("%B %Y"), + f'{mstr}-resolved': "", + f'{mstr}-docs': doc_pks, 'action': "save", }) self.assertEqual(r.status_code, 200) @@ -1415,13 +1704,13 @@ def test_edit_milestone(self): # edit mailbox_before = len(outbox) - r = self.client.post(url, { 'prefix': "m1", - 'm1-id': m1.id, - 'm1-desc': "Test 2 - changed", - 'm1-due': due.strftime("%B %Y"), - 'm1-resolved': "Done", - 'm1-resolved_checkbox': "checked", - 'm1-docs': doc_pks, + r = self.client.post(url, { 'prefix': mstr, + f'{mstr}-id': m1.id, + f'{mstr}-desc': "Test 2 - changed", + f'{mstr}-due': due.strftime("%B %Y"), + f'{mstr}-resolved': "Done", + f'{mstr}-resolved_checkbox': "checked", + f'{mstr}-docs': doc_pks, 'action': "save", }) self.assertEqual(r.status_code, 302) @@ -1692,58 +1981,72 @@ def setUp(self): PersonFactory(user__username='plain') GroupFactory(acronym='mars',parent=GroupFactory(type_id='area')) GroupFactory(acronym='ames',parent=GroupFactory(type_id='area')) - self.group_alias_file = NamedTemporaryFile(delete=False) - self.group_alias_file.write(b"""# Generated by hand at 2015-02-12_16:30:52 -virtual.ietf.org anything -mars-ads@ietf.org xfilter-mars-ads -expand-mars-ads@virtual.ietf.org aread@example.org -mars-chairs@ietf.org xfilter-mars-chairs -expand-mars-chairs@virtual.ietf.org mars_chair@ietf.org -ames-ads@ietf.org xfilter-mars-ads -expand-ames-ads@virtual.ietf.org aread@example.org -ames-chairs@ietf.org xfilter-mars-chairs -expand-ames-chairs@virtual.ietf.org mars_chair@ietf.org -""") - self.group_alias_file.close() - self.saved_group_virtual_path = settings.GROUP_VIRTUAL_PATH - settings.GROUP_VIRTUAL_PATH = self.group_alias_file.name - - def tearDown(self): - settings.GROUP_VIRTUAL_PATH = self.saved_group_virtual_path - os.unlink(self.group_alias_file.name) - super().tearDown() - - def testAliases(self): + + @mock.patch("ietf.group.views.get_group_email_aliases") + def testAliases(self, mock_get_aliases): url = urlreverse('ietf.group.urls_info_details.redirect.email', kwargs=dict(acronym="mars")) r = self.client.get(url) self.assertEqual(r.status_code, 302) + mock_get_aliases.return_value = [ + {"acronym": "mars", "alias_type": "-ads", "expansion": "aread@example.org"}, + {"acronym": "mars", "alias_type": "-chairs", "expansion": "mars_chair@ietf.org"}, + ] for testdict in [dict(acronym="mars"),dict(acronym="mars",group_type="wg")]: url = urlreverse('ietf.group.urls_info_details.redirect.email', kwargs=testdict) r = self.client.get(url,follow=True) + self.assertEqual( + mock_get_aliases.call_args, + mock.call(testdict.get("acronym", None), testdict.get("group_type", None)), + ) self.assertTrue(all([x in unicontent(r) for x in ['mars-ads@','mars-chairs@']])) self.assertFalse(any([x in unicontent(r) for x in ['ames-ads@','ames-chairs@']])) url = urlreverse('ietf.group.views.email_aliases', kwargs=dict()) login_testing_unauthorized(self, "plain", url) + + mock_get_aliases.return_value = [ + {"acronym": "mars", "alias_type": "-ads", "expansion": "aread@example.org"}, + {"acronym": "mars", "alias_type": "-chairs", "expansion": "mars_chair@ietf.org"}, + {"acronym": "ames", "alias_type": "-ads", "expansion": "aread@example.org"}, + {"acronym": "ames", "alias_type": "-chairs", "expansion": "mars_chair@ietf.org"}, + ] r = self.client.get(url) self.assertTrue(r.status_code,200) + self.assertEqual(mock_get_aliases.call_args, mock.call(None, None)) self.assertTrue(all([x in unicontent(r) for x in ['mars-ads@','mars-chairs@','ames-ads@','ames-chairs@']])) url = urlreverse('ietf.group.views.email_aliases', kwargs=dict(group_type="wg")) + mock_get_aliases.return_value = [ + {"acronym": "mars", "alias_type": "-ads", "expansion": "aread@example.org"}, + {"acronym": "mars", "alias_type": "-chairs", "expansion": "mars_chair@ietf.org"}, + {"acronym": "ames", "alias_type": "-ads", "expansion": "aread@example.org"}, + {"acronym": "ames", "alias_type": "-chairs", "expansion": "mars_chair@ietf.org"}, + ] r = self.client.get(url) self.assertEqual(r.status_code,200) + self.assertEqual(mock_get_aliases.call_args, mock.call(None, "wg")) self.assertContains(r, 'mars-ads@') url = urlreverse('ietf.group.views.email_aliases', kwargs=dict(group_type="rg")) + mock_get_aliases.return_value = [] r = self.client.get(url) self.assertEqual(r.status_code,200) + self.assertEqual(mock_get_aliases.call_args, mock.call(None, "rg")) self.assertNotContains(r, 'mars-ads@') - def testExpansions(self): + @mock.patch("ietf.group.views.get_group_email_aliases") + def testExpansions(self, mock_get_aliases): + mock_get_aliases.return_value = [ + {"acronym": "mars", "alias_type": "-ads", "expansion": "aread@example.org"}, + {"acronym": "mars", "alias_type": "-chairs", "expansion": "mars_chair@ietf.org"}, + {"acronym": "ames", "alias_type": "-ads", "expansion": "aread@example.org"}, + {"acronym": "ames", "alias_type": "-chairs", "expansion": "mars_chair@ietf.org"}, + ] url = urlreverse('ietf.group.views.email', kwargs=dict(acronym="mars")) r = self.client.get(url) self.assertEqual(r.status_code,200) + self.assertEqual(mock_get_aliases.call_args, mock.call("mars", None)) self.assertContains(r, 'Email aliases') self.assertContains(r, 'mars-ads@ietf.org') self.assertContains(r, 'group_personnel_change') @@ -1826,7 +2129,7 @@ def ensure_cant_edit(group,user): self.assertEqual(response.status_code, 404) self.client.logout() - for type_id in GroupTypeName.objects.exclude(slug__in=('wg','rg','ag','rag','team')).values_list('slug',flat=True): + for type_id in GroupTypeName.objects.exclude(slug__in=('wg','rg','ag','rag','team','edwg')).values_list('slug',flat=True): group = GroupFactory.create(type_id=type_id) for user in (None,User.objects.get(username='secretary')): ensure_updates_dont_show(group,user) @@ -1968,8 +2271,17 @@ def test_admin_acronym_validation(self): self.assertTrue(form.is_valid()) form = AdminGroupForm({'acronym':'shouldfail-','name':'should fail','type':'wg','state':'active','used_roles':'[]','time':now}) self.assertIn('acronym',form.errors) + form = AdminGroupForm({'acronym':'shouldfail-','name':'should fail','type':'sdo','state':'active','used_roles':'[]','time':now}) + self.assertIn('acronym',form.errors) form = AdminGroupForm({'acronym':'-shouldfail','name':'should fail','type':'wg','state':'active','used_roles':'[]','time':now}) self.assertIn('acronym',form.errors) + form = AdminGroupForm({'acronym':'-shouldfail','name':'should fail','type':'sdo','state':'active','used_roles':'[]','time':now}) + self.assertIn('acronym',form.errors) + # SDO groups (and only SDO groups) can have a leading number + form = AdminGroupForm({'acronym':'3gpp-should-pass','name':'should pass','type':'sdo','state':'active','used_roles':'[]','time':now}) + self.assertTrue(form.is_valid()) + form = AdminGroupForm({'acronym':'123shouldfail','name':'should fail','type':'wg','state':'active','used_roles':'[]','time':now}) + self.assertIn('acronym',form.errors) wg = GroupFactory(acronym='bad-idea', type_id='wg') # There are some existing wg and programs with hyphens in their acronyms. form = AdminGroupForm({'acronym':wg.acronym,'name':wg.name,'type':wg.type_id,'state':wg.state_id,'used_roles':str(wg.used_roles),'time':now},instance=wg) diff --git a/ietf/group/tests_js.py b/ietf/group/tests_js.py index 1c7f9fc9a4..7c51c23319 100644 --- a/ietf/group/tests_js.py +++ b/ietf/group/tests_js.py @@ -100,8 +100,8 @@ def test_add_milestone(self): # fill in the edit milestone form desc_input.send_keys(description) - due_input.send_keys(due_date.strftime('%m %Y\n')) # \n closes the date selector self._search_draft_and_locate_result(draft_input, draft_search_string, draft).click() + due_input.send_keys(due_date.strftime('%m %Y')) self._click_milestone_submit_button('Review') result_row = self._assert_milestone_changed() @@ -165,7 +165,7 @@ def test_edit_milestone(self): # modify the fields new_due_date = (milestone.due + datetime.timedelta(days=31)).strftime('%m %Y') due_field.clear() - due_field.send_keys(new_due_date + '\n') + due_field.send_keys(new_due_date) self._search_draft_and_locate_result(draft_input, draft_search_string, draft).click() @@ -189,4 +189,4 @@ def test_edit_milestone(self): gms = self.group.groupmilestone_set.first() self.assertEqual(gms.desc, expected_desc) self.assertEqual(gms.due.strftime('%m %Y'), expected_due_date) - self.assertCountEqual(expected_docs, gms.docs.all()) \ No newline at end of file + self.assertCountEqual(expected_docs, gms.docs.all()) diff --git a/ietf/group/tests_review.py b/ietf/group/tests_review.py index 43101cd378..bb9b79a416 100644 --- a/ietf/group/tests_review.py +++ b/ietf/group/tests_review.py @@ -1,4 +1,4 @@ -# Copyright The IETF Trust 2016-2020, All Rights Reserved +# Copyright The IETF Trust 2016-2023, All Rights Reserved # -*- coding: utf-8 -*- @@ -41,7 +41,7 @@ def test_review_requests(self): r = self.client.get(url) self.assertEqual(r.status_code, 200) self.assertContains(r, review_req.doc.name) - self.assertContains(r, assignment.reviewer.person.name) + self.assertContains(r, escape(assignment.reviewer.person.name)) url = urlreverse(ietf.group.views.review_requests, kwargs={ 'acronym': group.acronym }) @@ -101,8 +101,20 @@ def test_suggested_review_requests(self): self.assertEqual(list(suggested_review_requests_for_team(team)), []) + # blocked by an already existing request (don't suggest it again) + review_req.state_id = "requested" + review_req.save() + self.assertEqual(list(suggested_review_requests_for_team(team)), []) + + # ... but not for a previous version + review_req.requested_rev = prev_rev + review_req.save() + self.assertEqual(len(suggested_review_requests_for_team(team)), 1) + + # blocked by completion review_req.state = ReviewRequestStateName.objects.get(slug="assigned") + review_req.requested_rev = "" review_req.save() assignment.state = ReviewAssignmentStateName.objects.get(slug="completed") assignment.reviewed_rev = review_req.doc.rev @@ -116,6 +128,7 @@ def test_suggested_review_requests(self): self.assertEqual(len(suggested_review_requests_for_team(team)), 1) + def test_suggested_review_requests_on_lc_and_telechat(self): review_req = ReviewRequestFactory(state_id='assigned') doc = review_req.doc @@ -183,7 +196,7 @@ def test_reviewer_overview(self): urlreverse(ietf.group.views.reviewer_overview, kwargs={ 'acronym': group.acronym, 'group_type': group.type_id })]: r = self.client.get(url) self.assertEqual(r.status_code, 200) - self.assertContains(r, reviewer.name) + self.assertContains(r, escape(reviewer.name)) self.assertContains(r, review_req1.doc.name) # without a login, reason for being unavailable should not be seen self.assertNotContains(r, "Availability") @@ -199,13 +212,13 @@ def test_reviewer_overview(self): r = self.client.get(url) self.assertEqual(r.status_code, 200) # review team members can see reason for being unavailable - self.assertContains(r, "Availability") + self.assertContains(r, "Available") self.client.login(username="secretary", password="secretary+password") r = self.client.get(url) self.assertEqual(r.status_code, 200) # secretariat can see reason for being unavailable - self.assertContains(r, "Availability") + self.assertContains(r, "Available") # add one closed review with no response and see it is visible review_req2 = ReviewRequestFactory(state_id='completed',team=team) @@ -223,7 +236,7 @@ def test_reviewer_overview(self): r = self.client.get(url) self.assertEqual(r.status_code, 200) # We should see the new document with status of no response - self.assertContains(r, "No Response") + self.assertContains(r, "No response") self.assertContains(r, review_req1.doc.name) self.assertContains(r, review_req2.doc.name) # None of the reviews should be completed this time, @@ -698,6 +711,22 @@ def test_change_review_secretary_settings(self): self.assertEqual(settings.max_items_to_show_in_reviewer_list, 10) self.assertEqual(settings.days_to_show_in_reviewer_list, 365) + def test_assign_reviewer_after_reject(self): + team = ReviewTeamFactory() + reviewer = RoleFactory(name_id='reviewer', group=team).person + ReviewerSettingsFactory(person=reviewer, team=team) + review_req = ReviewRequestFactory(team=team) + ReviewAssignmentFactory(review_request=review_req, state_id='rejected', reviewer=reviewer.email()) + + unassigned_url = urlreverse(ietf.group.views.manage_review_requests, kwargs={ 'acronym': team.acronym, 'group_type': team.type_id, "assignment_status": "unassigned" }) + login_testing_unauthorized(self, "secretary", unassigned_url) + + r = self.client.get(unassigned_url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + reviewer_label = q("option[value=\"{}\"]".format(reviewer.email())).text().lower() + self.assertIn("rejected review of document before", reviewer_label) + class BulkAssignmentTests(TestCase): @@ -731,7 +760,7 @@ def test_rotation_queue_update(self): r = self.client.post(unassigned_url, postdict) self.assertEqual(r.status_code,302) self.assertEqual(expected_ending_head_of_rotation, policy.default_reviewer_rotation_list()[0]) - self.assertMailboxContains(outbox, subject='Last Call assignment', text='Requested by', count=4) + self.assertMailboxContains(outbox, subject='Last Call', text='Requested by', count=4) class ResetNextReviewerInTeamTests(TestCase): @@ -786,3 +815,170 @@ def test_reset_next_reviewer(self): self.assertEqual(NextReviewerInTeam.objects.get(team=group).next_reviewer, reviewers[target_index].person) self.client.logout() target_index += 2 + +class RequestsHistoryTests(TestCase): + def test_requests_history_overview_page(self): + # Make assigned assignment + review_req = ReviewRequestFactory(state_id='assigned') + assignment = ReviewAssignmentFactory(review_request=review_req, + state_id='assigned', + reviewer=EmailFactory(), + assigned_on = review_req.time) + group = review_req.team + + for url in [urlreverse(ietf.group.views.review_requests_history, + kwargs={ 'acronym': group.acronym }), + urlreverse(ietf.group.views.review_requests_history, + kwargs={ 'acronym': group.acronym , + 'group_type': group.type_id}), + urlreverse(ietf.group.views.review_requests_history, + kwargs={ 'acronym': group.acronym }) + + '?since=3m', + urlreverse(ietf.group.views.review_requests_history, + kwargs={ 'acronym': group.acronym , + 'group_type': group.type_id }) + + '?since=3m']: + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + self.assertContains(r, review_req.doc.name) + self.assertContains(r, 'Assigned') + self.assertContains(r, escape(assignment.reviewer.person.name)) + + url = urlreverse(ietf.group.views.review_requests_history, + kwargs={ 'acronym': group.acronym }) + + assignment.state = ReviewAssignmentStateName.objects.get(slug="completed") + assignment.result = ReviewResultName.objects.get(slug="ready") + assignment.save() + + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + self.assertContains(r, review_req.doc.name) + self.assertContains(r, 'Assigned') + self.assertContains(r, 'Completed') + + def test_requests_history_filter_page(self): + # First assignment as assigned + review_req = ReviewRequestFactory(state_id = 'assigned', + doc = DocumentFactory()) + assignment = ReviewAssignmentFactory(review_request = review_req, + state_id = 'assigned', + reviewer = EmailFactory(), + assigned_on = review_req.time) + group = review_req.team + + # Second assignment in same group as accepted + review_req2 = ReviewRequestFactory(state_id = 'assigned', + team = review_req.team, + doc = DocumentFactory()) + assignment2 = ReviewAssignmentFactory(review_request = review_req2, + state_id='accepted', + reviewer = EmailFactory(), + assigned_on = review_req2.time) + + # Modify the assignment to be completed, and mark it ready + assignment2.state = ReviewAssignmentStateName.objects.get(slug="completed") + assignment2.result = ReviewResultName.objects.get(slug="ready") + assignment2.save() + + # Check that we have all information when we do not filter + url = urlreverse(ietf.group.views.review_requests_history, + kwargs={ 'acronym': group.acronym }) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + self.assertContains(r, review_req.doc.name) + self.assertContains(r, review_req2.doc.name) + self.assertContains(r, 'data-text="Assigned"') + self.assertContains(r, 'data-text="Accepted"') + self.assertContains(r, 'data-text="Completed"') + self.assertContains(r, 'data-text="Ready"') + self.assertContains(r, escape(assignment.reviewer.person.name)) + self.assertContains(r, escape(assignment2.reviewer.person.name)) + + # Check first reviewer history + for url in [urlreverse(ietf.group.views.review_requests_history, + kwargs={ 'acronym': group.acronym }) + + '?reviewer_email=' + str(assignment.reviewer), + urlreverse(ietf.group.views.review_requests_history, + kwargs={ 'acronym': group.acronym , + 'group_type': group.type_id}) + + '?reviewer_email=' + str(assignment.reviewer)]: + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + self.assertContains(r, review_req.doc.name) + self.assertNotContains(r, review_req2.doc.name) + self.assertContains(r, 'data-text="Assigned"') + self.assertNotContains(r, 'data-text="Accepted"') + self.assertNotContains(r, 'data-text="Completed"') + self.assertNotContains(r, 'data-text="Ready"') + self.assertContains(r, escape(assignment.reviewer.person.name)) + self.assertNotContains(r, escape(assignment2.reviewer.person.name)) + + # Check second reviewer history + for url in [urlreverse(ietf.group.views.review_requests_history, + kwargs={ 'acronym': group.acronym }) + + '?reviewer_email=' + str(assignment2.reviewer), + urlreverse(ietf.group.views.review_requests_history, + kwargs={ 'acronym': group.acronym , + 'group_type': group.type_id}) + + '?reviewer_email=' + str(assignment2.reviewer)]: + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + self.assertNotContains(r, review_req.doc.name) + self.assertContains(r, review_req2.doc.name) + self.assertNotContains(r, 'data-text="Assigned"') + self.assertContains(r, 'data-text="Accepted"') + self.assertContains(r, 'data-text="Completed"') + self.assertContains(r, 'data-text="Ready"') + self.assertNotContains(r, escape(assignment.reviewer.person.name)) + self.assertContains(r, escape(assignment2.reviewer.person.name)) + + # Check for reviewer that does not have anything + url = urlreverse(ietf.group.views.review_requests_history, + kwargs={ 'acronym': group.acronym }) + '?reviewer_email=nobody@nowhere.example.org' + + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + self.assertNotContains(r, review_req.doc.name) + self.assertNotContains(r, 'data-text="Assigned"') + self.assertNotContains(r, 'data-text="Accepted"') + self.assertNotContains(r, 'data-text="Completed"') + + def test_requests_history_invalid_filter_parameters(self): + # First assignment as assigned + review_req = ReviewRequestFactory(state_id="assigned", doc=DocumentFactory()) + group = review_req.team + url = urlreverse( + "ietf.group.views.review_requests_history", + kwargs={"acronym": group.acronym}, + ) + invalid_reviewer_emails = [ + "%00null@example.com", # urlencoded null character + "null@exa%00mple.com", # urlencoded null character + "\x00null@example.com", # literal null character + "null@ex\x00ample.com", # literal null character + ] + for invalid_email in invalid_reviewer_emails: + r = self.client.get( + url + f"?reviewer_email={invalid_email}" + ) + self.assertEqual( + r.status_code, + 400, + f"should return a 400 response for reviewer_email={repr(invalid_email)}" + ) + + invalid_since_choices = [ + "forever", # not an option + "all\x00", # literal null character + "a%00ll", # urlencoded null character + ] + for invalid_since in invalid_since_choices: + r = self.client.get( + url + f"?since={invalid_since}" + ) + self.assertEqual( + r.status_code, + 400, + f"should return a 400 response for since={repr(invalid_since)}" + ) diff --git a/ietf/group/tests_serializers.py b/ietf/group/tests_serializers.py new file mode 100644 index 0000000000..b584a17ae2 --- /dev/null +++ b/ietf/group/tests_serializers.py @@ -0,0 +1,96 @@ +# Copyright The IETF Trust 2026, All Rights Reserved +from ietf.group.factories import RoleFactory, GroupFactory +from ietf.group.serializers import ( + AreaDirectorSerializer, + AreaSerializer, + GroupSerializer, +) +from ietf.person.factories import EmailFactory +from ietf.utils.test_utils import TestCase + + +class GroupSerializerTests(TestCase): + def test_serializes(self): + wg = GroupFactory() + serialized = GroupSerializer(wg).data + self.assertEqual( + serialized, + { + "acronym": wg.acronym, + "name": wg.name, + "type": "wg", + "list_email": wg.list_email, + }, + ) + + +class AreaDirectorSerializerTests(TestCase): + def test_serializes_role(self): + """Should serialize a Role correctly""" + role = RoleFactory(group__type_id="area", name_id="ad") + serialized = AreaDirectorSerializer(role).data + self.assertEqual( + serialized, + {"email": role.email.email_address(), "name": role.person.plain_name()}, + ) + + def test_serializes_email(self): + """Should serialize an Email correctly""" + email = EmailFactory() + serialized = AreaDirectorSerializer(email).data + self.assertEqual( + serialized, + { + "email": email.email_address(), + "name": email.person.plain_name() if email.person else None, + }, + ) + + +class AreaSerializerTests(TestCase): + def test_serializes_active_area(self): + """Should serialize an active area correctly""" + area = GroupFactory(type_id="area", state_id="active") + serialized = AreaSerializer(area).data + self.assertEqual( + serialized, + { + "acronym": area.acronym, + "name": area.name, + "ads": [], + }, + ) + ad_roles = RoleFactory.create_batch(2, group=area, name_id="ad") + serialized = AreaSerializer(area).data + self.assertEqual(serialized["acronym"], area.acronym) + self.assertEqual(serialized["name"], area.name) + self.assertCountEqual( + serialized["ads"], + [ + {"email": ad.email.email_address(), "name": ad.person.plain_name()} + for ad in ad_roles + ], + ) + + def test_serializes_inactive_area(self): + """Should serialize an inactive area correctly""" + area = GroupFactory(type_id="area", state_id="conclude") + serialized = AreaSerializer(area).data + self.assertEqual( + serialized, + { + "acronym": area.acronym, + "name": area.name, + "ads": [], + }, + ) + RoleFactory.create_batch(2, group=area, name_id="ad") + serialized = AreaSerializer(area).data + self.assertEqual( + serialized, + { + "acronym": area.acronym, + "name": area.name, + "ads": [], + }, + ) diff --git a/ietf/group/urls.py b/ietf/group/urls.py index 46bde4ede3..8354aba063 100644 --- a/ietf/group/urls.py +++ b/ietf/group/urls.py @@ -1,7 +1,7 @@ -# Copyright The IETF Trust 2013-2020, All Rights Reserved +# Copyright The IETF Trust 2013-2023, All Rights Reserved from django.conf import settings -from django.conf.urls import include +from django.urls import include from django.views.generic import RedirectView from ietf.community import views as community_views @@ -20,11 +20,11 @@ url(r'^documents/subscription/$', community_views.subscription), url(r'^charter/$', views.group_about), url(r'^about/$', views.group_about), - url(r'^about/rendertest/$', views.group_about_rendertest), url(r'^about/status/$', views.group_about_status), url(r'^about/status/edit/$', views.group_about_status_edit), url(r'^about/status/meeting/(?P\d+)/$', views.group_about_status_meeting), url(r'^history/$',views.history), + url(r'^requestshistory/$',views.review_requests_history), url(r'^history/addcomment/$',views.add_comment), url(r'^email/$', views.email), url(r'^deps\.json$', views.dependencies), @@ -48,11 +48,20 @@ url(r'^secretarysettings/$', views.change_review_secretary_settings), url(r'^reset_next_reviewer/$', views.reset_next_reviewer), url(r'^email-aliases/$', RedirectView.as_view(pattern_name=views.email,permanent=False),name='ietf.group.urls_info_details.redirect.email'), + url(r'^statements/$', views.statements), + url(r'^appeals/$', views.appeals), + url(r'^appeals/artifact/(?P\d+)$', views.appeal_artifact), + url(r'^appeals/artifact/(?P\d+)/markdown$', views.appeal_artifact_markdown), + + ] group_urls = [ - url(r'^$', views.active_groups), + url(r'^$', views.active_groups), + url(r'^leadership/(?P(wg|rg))/$', views.group_leadership), + url(r'^leadership/(?P(wg|rg))/csv/$', views.group_leadership_csv), + url(r'^groupstats.json', views.group_stats_data, None, 'ietf.group.views.group_stats_data'), url(r'^groupmenu.json', views.group_menu_data, None, 'ietf.group.views.group_menu_data'), url(r'^chartering/$', views.chartering_groups), url(r'^chartering/create/(?P(wg|rg))/$', views.edit, {'action': "charter"}), diff --git a/ietf/group/utils.py b/ietf/group/utils.py index b701d6a7c7..6777ed1933 100644 --- a/ietf/group/utils.py +++ b/ietf/group/utils.py @@ -1,12 +1,13 @@ -# Copyright The IETF Trust 2012-2021, All Rights Reserved +# Copyright The IETF Trust 2012-2023, All Rights Reserved # -*- coding: utf-8 -*- +import datetime - -import io -import os +from itertools import chain +from pathlib import Path from django.db.models import Q from django.shortcuts import get_object_or_404 +from django.utils import timezone from django.utils.html import format_html from django.utils.safestring import mark_safe from django.urls import reverse as urlreverse @@ -15,13 +16,13 @@ from ietf.community.models import CommunityList, SearchRule from ietf.community.utils import reset_name_contains_index_for_rule, can_manage_community_list -from ietf.doc.models import Document, State +from ietf.doc.models import Document, State, RelatedDocument from ietf.group.models import Group, RoleHistory, Role, GroupFeatures, GroupEvent from ietf.ietfauth.utils import has_role from ietf.name.models import GroupTypeName, RoleName from ietf.person.models import Email from ietf.review.utils import can_manage_review_requests_for_team -from ietf.utils import log +from ietf.utils import log, markdown from ietf.utils.history import get_history_object_for, copy_many_to_many_for_history from ietf.doc.templatetags.ietf_filters import is_valid_url from functools import reduce @@ -55,15 +56,14 @@ def get_charter_text(group): if (h.rev > c.rev and not (c_appr and not h_appr)) or (h_appr and not c_appr): c = h - filename = os.path.join(c.get_file_path(), "%s-%s.txt" % (c.canonical_name(), c.rev)) + filename = Path(c.get_file_path()) / f"{c.name}-{c.rev}.txt" try: - with io.open(filename, 'rb') as f: - text = f.read() - try: - text = text.decode('utf8') - except UnicodeDecodeError: - text = text.decode('latin1') - return text + text = filename.read_bytes() + try: + text = text.decode('utf8') + except UnicodeDecodeError: + text = text.decode('latin1') + return text except IOError: return 'Error Loading Group Charter' @@ -154,17 +154,23 @@ def can_manage_materials(user, group): def can_manage_session_materials(user, group, session): return has_role(user, 'Secretariat') or (group.has_role(user, group.features.matman_roles) and not session.is_material_submission_cutoff()) -# Maybe this should be cached... def can_manage_some_groups(user): if not user.is_authenticated: return False + authroles = set( + chain.from_iterable( + GroupFeatures.objects.values_list("groupman_authroles", flat=True) + ) + ) + extra_role_qs = dict() for gf in GroupFeatures.objects.all(): - for authrole in gf.groupman_authroles: - if has_role(user, authrole): - return True - if Role.objects.filter(name__in=gf.groupman_roles, group__type_id=gf.type_id, person__user=user).exists(): - return True - return False + extra_role_qs[f"{gf.type_id} groupman roles"] = Q( + name__in=gf.groupman_roles, + group__type_id=gf.type_id, + group__state__in=["active", "bof", "proposed"], + ) + return has_role(user, authroles, extra_role_qs=extra_role_qs) + def can_provide_status_update(user, group): if not group.features.acts_like_wg: @@ -191,7 +197,7 @@ def setup_default_community_list_for_group(group): community_list=clist, rule_type="group_rfc", group=group, - state=State.objects.get(slug="rfc", type="draft"), + state=State.objects.get(slug="published", type="rfc"), ) SearchRule.objects.create( community_list=clist, @@ -230,9 +236,14 @@ def construct_group_menu_context(request, group, selected, group_type, others): import ietf.group.views entries.append(("Review requests", urlreverse(ietf.group.views.review_requests, kwargs=kwargs))) entries.append(("Reviewers", urlreverse(ietf.group.views.reviewer_overview, kwargs=kwargs))) - + entries.append(("Reviews History", urlreverse(ietf.group.views.review_requests_history, kwargs=kwargs))) if group.features.has_meetings: entries.append(("Meetings", urlreverse("ietf.group.views.meetings", kwargs=kwargs))) + if group.acronym in ["iesg"]: + entries.append(("Working Groups", urlreverse("ietf.iesg.views.working_groups"))) + if group.acronym in ["iab", "iesg"]: + entries.append(("Statements", urlreverse("ietf.group.views.statements", kwargs=kwargs))) + entries.append(("Appeals", urlreverse("ietf.group.views.appeals", kwargs=kwargs))) entries.append(("History", urlreverse("ietf.group.views.history", kwargs=kwargs))) entries.append(("Photos", urlreverse("ietf.group.views.group_photos", kwargs=kwargs))) entries.append(("Email expansions", urlreverse("ietf.group.views.email", kwargs=kwargs))) @@ -240,7 +251,6 @@ def construct_group_menu_context(request, group, selected, group_type, others): if is_valid_url(group.list_archive): entries.append((mark_safe("List archive »"), group.list_archive)) - # actions actions = [] @@ -352,3 +362,188 @@ def update_role_set(group, role_name, new_value, by): e.save() return added, removed + + +class GroupAliasGenerator: + days = 5 * 365 + active_states = ["active", "bof", "proposed"] + group_types = [ + "wg", + "rg", + "rag", + "dir", + "team", + "review", + "program", + "rfcedtyp", + "edappr", + "edwg", + ] # This should become groupfeature driven... + no_ad_group_types = ["rg", "rag", "team", "program", "rfcedtyp", "edappr", "edwg"] + + def __init__(self, group_queryset=None): + if group_queryset is None: + self.group_queryset = Group.objects.all() + else: + self.group_queryset = group_queryset + + def __iter__(self): + show_since = timezone.now() - datetime.timedelta(days=self.days) + + # Loop through each group type and build -ads and -chairs entries + for g in self.group_types: + domains = ["ietf"] + if g in ("rg", "rag"): + domains.append("irtf") + if g == "program": + domains.append("iab") + + entries = self.group_queryset.filter(type=g).all() + active_entries = entries.filter(state__in=self.active_states) + inactive_recent_entries = entries.exclude( + state__in=self.active_states + ).filter(time__gte=show_since) + interesting_entries = active_entries | inactive_recent_entries + + for e in interesting_entries.distinct().iterator(): + name = e.acronym + + # Research groups, teams, and programs do not have -ads lists + if not g in self.no_ad_group_types: + ad_emails = get_group_ad_emails(e) + if ad_emails: + yield name + "-ads", domains, list(ad_emails) + # All group types have -chairs lists + chair_emails = get_group_role_emails(e, ["chair", "secr"]) + if chair_emails: + yield name + "-chairs", domains, list(chair_emails) + + # The area lists include every chair in active working groups in the area + areas = self.group_queryset.filter(type="area").all() + active_areas = areas.filter(state__in=self.active_states) + for area in active_areas: + name = area.acronym + area_ad_emails = get_group_role_emails(area, ["pre-ad", "ad", "chair"]) + if area_ad_emails: + yield name + "-ads", ["ietf"], list(area_ad_emails) + chair_emails = get_child_group_role_emails(area, ["chair", "secr"]) | area_ad_emails + if chair_emails: + yield name + "-chairs", ["ietf"], list(chair_emails) + + # Other groups with chairs that require Internet-Draft submission approval + gtypes = GroupTypeName.objects.values_list("slug", flat=True) + special_groups = self.group_queryset.filter( + type__features__req_subm_approval=True, acronym__in=gtypes, state="active" + ) + for group in special_groups: + chair_emails = get_group_role_emails(group, ["chair", "delegate"]) + if chair_emails: + yield group.acronym + "-chairs", ["ietf"], list(chair_emails) + + +def get_group_email_aliases(acronym, group_type): + aliases = [] + group_queryset = Group.objects.all() + if acronym: + group_queryset = group_queryset.filter(acronym=acronym) + if group_type: + group_queryset = group_queryset.filter(type__slug=group_type) + for (alias, _, alist) in GroupAliasGenerator(group_queryset): + acro, _hyphen, alias_type = alias.partition("-") + expansion = ", ".join(sorted(alist)) + aliases.append({ + "acronym": acro, + "alias_type": f"-{alias_type}" if alias_type else "", + "expansion": expansion, + }) + return sorted(aliases, key=lambda a: a["acronym"]) + + +def role_holder_emails(): + """Get queryset of active Emails for group role holders""" + group_types_of_interest = [ + "ag", + "area", + "dir", + "iab", + "ietf", + "irtf", + "nomcom", + "rg", + "team", + "wg", + "rag", + ] + roles = Role.objects.filter( + group__state__slug="active", + group__type__in=group_types_of_interest, + ) + emails = Email.objects.filter(active=True).exclude( + address__startswith="unknown-email-" + ) + return emails.filter(person__role__in=roles).distinct() + + +def fill_in_charter_info(group, include_drafts=False): + group.areadirector = getattr(group.ad_role(),'email',None) + + personnel = {} + for r in Role.objects.filter(group=group).order_by('person__name').select_related("email", "person", "name"): + if r.name_id not in personnel: + personnel[r.name_id] = [] + personnel[r.name_id].append(r) + + if group.parent and group.parent.type_id == "area" and group.ad_role() and "ad" not in personnel: + ad_roles = list(Role.objects.filter(group=group.parent, name="ad", person=group.ad_role().person)) + if ad_roles: + personnel["ad"] = ad_roles + + group.personnel = [] + for role_name_slug, roles in personnel.items(): + label = roles[0].name.name + if len(roles) > 1: + if label.endswith("y"): + label = label[:-1] + "ies" + else: + label += "s" + + group.personnel.append((role_name_slug, label, roles)) + + group.personnel.sort(key=lambda t: t[2][0].name.order) + + milestone_state = "charter" if group.state_id == "proposed" else "active" + group.milestones = group.groupmilestone_set.filter(state=milestone_state) + if group.uses_milestone_dates: + group.milestones = group.milestones.order_by('resolved', 'due') + else: + group.milestones = group.milestones.order_by('resolved', 'order') + + if group.charter: + group.charter_text = get_charter_text(group) + else: + group.charter_text = "Not chartered yet." + group.charter_html = markdown.markdown(group.charter_text) + + +def fill_in_wg_roles(group): + def get_roles(slug, default): + for role_slug, label, roles in group.personnel: + if slug == role_slug: + return roles + return default + + group.chairs = get_roles("chair", []) + ads = get_roles("ad", []) + group.areadirector = ads[0] if ads else None + group.techadvisors = get_roles("techadv", []) + group.editors = get_roles("editor", []) + group.secretaries = get_roles("secr", []) + + +def fill_in_wg_drafts(group): + group.drafts = Document.objects.filter(type_id="draft", group=group).order_by("name") + group.rfcs = Document.objects.filter(type_id="rfc", group=group).order_by("rfc_number") + for rfc in group.rfcs: + # TODO: remote_field? + rfc.remote_field = RelatedDocument.objects.filter(source=rfc,relationship_id__in=['obs','updates']).distinct() + rfc.invrel = RelatedDocument.objects.filter(target=rfc,relationship_id__in=['obs','updates']).distinct() diff --git a/ietf/group/views.py b/ietf/group/views.py index 9f43028236..8561a5059f 100644 --- a/ietf/group/views.py +++ b/ietf/group/views.py @@ -1,5 +1,5 @@ # -*- coding: utf-8 -*- -# Copyright The IETF Trust 2009-2022, All Rights Reserved +# Copyright The IETF Trust 2009-2023, All Rights Reserved # # Portion Copyright (C) 2010 Nokia Corporation and/or its subsidiary(-ies). # All rights reserved. Contact: Pasi Eronen @@ -35,33 +35,42 @@ import copy +import csv import datetime import itertools -import io import math -import re import json +import types from collections import OrderedDict, defaultdict +from pathlib import Path from simple_history.utils import update_change_reason from django import forms from django.conf import settings from django.contrib.auth.decorators import login_required -from django.db.models import Q, Count -from django.http import HttpResponse, HttpResponseRedirect, Http404, JsonResponse +from django.db.models import Count, F, OuterRef, Prefetch, Q, Subquery, TextField, Value +from django.db.models.functions import Coalesce +from django.http import ( + HttpResponse, + HttpResponseRedirect, + Http404, + JsonResponse, + HttpResponseBadRequest, +) from django.shortcuts import render, redirect, get_object_or_404 from django.template.loader import render_to_string from django.urls import reverse as urlreverse from django.utils import timezone from django.utils.html import escape from django.views.decorators.cache import cache_page, cache_control +from django.urls import reverse import debug # pyflakes:ignore from ietf.community.models import CommunityList, EmailSubscription from ietf.community.utils import docs_tracked_by_community_list -from ietf.doc.models import DocTagName, State, DocAlias, RelatedDocument, Document +from ietf.doc.models import DocTagName, State, RelatedDocument, Document, DocEvent from ietf.doc.templatetags.ietf_filters import clean_whitespace from ietf.doc.utils import get_chartering_type, get_tags_for_stream_id from ietf.doc.utils_charter import charter_name_for_group, replace_charter_of_replaced_group @@ -72,18 +81,20 @@ AddUnavailablePeriodForm, EndUnavailablePeriodForm, ReviewSecretarySettingsForm, ) from ietf.group.mails import email_admin_re_charter, email_personnel_change, email_comment from ietf.group.models import ( Group, Role, GroupEvent, GroupStateTransitions, - ChangeStateGroupEvent, GroupFeatures ) -from ietf.group.utils import (get_charter_text, can_manage_all_groups_of_type, + ChangeStateGroupEvent, GroupFeatures, AppealArtifact ) +from ietf.group.utils import (can_manage_all_groups_of_type, milestone_reviewer_for_group_type, can_provide_status_update, can_manage_materials, group_attribute_change_desc, construct_group_menu_context, get_group_materials, save_group_in_history, can_manage_group, update_role_set, - get_group_or_404, setup_default_community_list_for_group, ) + get_group_or_404, setup_default_community_list_for_group, fill_in_charter_info, + get_group_email_aliases) # from ietf.ietfauth.utils import has_role, is_authorized_in_group from ietf.mailtrigger.utils import gather_relevant_expansions from ietf.meeting.helpers import get_meeting -from ietf.meeting.utils import group_sessions, add_event_info_to_session_qs +from ietf.meeting.models import ImportantDate, SchedTimeSessAssignment, SchedulingEvent +from ietf.meeting.utils import group_sessions from ietf.name.models import GroupTypeName, StreamName from ietf.person.models import Email, Person from ietf.review.models import (ReviewRequest, ReviewAssignment, ReviewerSettings, @@ -91,11 +102,9 @@ from ietf.review.policies import get_reviewer_queue_policy from ietf.review.utils import (can_manage_review_requests_for_team, can_access_review_stats_for_team, - extract_revision_ordered_review_requests_for_documents_and_replaced, assign_review_request_to_reviewer, close_review_request, - suggested_review_requests_for_team, unavailable_periods_to_list, current_unavailable_periods_for_reviewers, @@ -111,11 +120,12 @@ from ietf.name.models import ReviewAssignmentStateName from ietf.utils.mail import send_mail_text, parse_preformatted -from ietf.ietfauth.utils import user_is_person +from ietf.ietfauth.utils import user_is_person, role_required from ietf.dbtemplate.models import DBTemplate from ietf.mailtrigger.utils import gather_address_lists from ietf.mailtrigger.models import Recipient from ietf.settings import MAILING_LIST_INFO_URL +from ietf.utils.decorators import ignore_view_kwargs from ietf.utils.response import permission_denied from ietf.utils.text import strip_suffix from ietf.utils import markdown @@ -128,88 +138,17 @@ def roles(group, role_name): return Role.objects.filter(group=group, name=role_name).select_related("email", "person") -def fill_in_charter_info(group, include_drafts=False): - group.areadirector = getattr(group.ad_role(),'email',None) - - personnel = {} - for r in Role.objects.filter(group=group).order_by('person__name').select_related("email", "person", "name"): - if r.name_id not in personnel: - personnel[r.name_id] = [] - personnel[r.name_id].append(r) - - if group.parent and group.parent.type_id == "area" and group.ad_role() and "ad" not in personnel: - ad_roles = list(Role.objects.filter(group=group.parent, name="ad", person=group.ad_role().person)) - if ad_roles: - personnel["ad"] = ad_roles - - group.personnel = [] - for role_name_slug, roles in personnel.items(): - label = roles[0].name.name - if len(roles) > 1: - if label.endswith("y"): - label = label[:-1] + "ies" - else: - label += "s" - - group.personnel.append((role_name_slug, label, roles)) - - group.personnel.sort(key=lambda t: t[2][0].name.order) - - milestone_state = "charter" if group.state_id == "proposed" else "active" - group.milestones = group.groupmilestone_set.filter(state=milestone_state) - if group.uses_milestone_dates: - group.milestones = group.milestones.order_by('resolved', 'due') - else: - group.milestones = group.milestones.order_by('resolved', 'order') - - if group.charter: - group.charter_text = get_charter_text(group) - else: - group.charter_text = "Not chartered yet." - def extract_last_name(role): return role.person.name_parts()[3] -def fill_in_wg_roles(group): - def get_roles(slug, default): - for role_slug, label, roles in group.personnel: - if slug == role_slug: - return roles - return default - - group.chairs = get_roles("chair", []) - ads = get_roles("ad", []) - group.areadirector = ads[0] if ads else None - group.techadvisors = get_roles("techadv", []) - group.editors = get_roles("editor", []) - group.secretaries = get_roles("secr", []) - -def fill_in_wg_drafts(group): - aliases = DocAlias.objects.filter(docs__type="draft", docs__group=group).prefetch_related('docs').order_by("name") - group.drafts = [] - group.rfcs = [] - for a in aliases: - if a.name.startswith("draft"): - group.drafts.append(a) - else: - group.rfcs.append(a) - a.remote_field = RelatedDocument.objects.filter(source=a.document,relationship_id__in=['obs','updates']).distinct() - a.invrel = RelatedDocument.objects.filter(target=a,relationship_id__in=['obs','updates']).distinct() - - -def check_group_email_aliases(): - pattern = re.compile(r'expand-(.*?)(-\w+)@.*? +(.*)$') - tot_count = 0 - good_count = 0 - with io.open(settings.GROUP_VIRTUAL_PATH,"r") as virtual_file: - for line in virtual_file.readlines(): - m = pattern.match(line) - tot_count += 1 - if m: - good_count += 1 - if good_count > 50 and tot_count < 3*good_count: - return True - return False + +def response_from_file(fpath: Path) -> HttpResponse: + """Helper to shovel a file back in an HttpResponse""" + try: + content = fpath.read_bytes() + except IOError: + raise Http404 + return HttpResponse(content, content_type="text/plain; charset=utf-8") # --- View functions --------------------------------------------------- @@ -217,58 +156,26 @@ def check_group_email_aliases(): def wg_summary_area(request, group_type): if group_type != "wg": raise Http404 - areas = Group.objects.filter(type="area", state="active").order_by("name") - for area in areas: - area.groups = Group.objects.filter(parent=area, type="wg", state="active").order_by("acronym") - for group in area.groups: - group.chairs = sorted(roles(group, "chair"), key=extract_last_name) + return response_from_file(Path(settings.GROUP_SUMMARY_PATH) / "1wg-summary.txt") - areas = [a for a in areas if a.groups] - - return render(request, 'group/1wg-summary.txt', - { 'areas': areas }, - content_type='text/plain; charset=UTF-8') def wg_summary_acronym(request, group_type): if group_type != "wg": raise Http404 - areas = Group.objects.filter(type="area", state="active").order_by("name") - groups = Group.objects.filter(type="wg", state="active").order_by("acronym").select_related("parent") - for group in groups: - group.chairs = sorted(roles(group, "chair"), key=extract_last_name) - return render(request, 'group/1wg-summary-by-acronym.txt', - { 'areas': areas, - 'groups': groups }, - content_type='text/plain; charset=UTF-8') + return response_from_file(Path(settings.GROUP_SUMMARY_PATH) / "1wg-summary-by-acronym.txt") + -@cache_page ( 60 * 60, cache="slowpages" ) def wg_charters(request, group_type): if group_type != "wg": raise Http404 - areas = Group.objects.filter(type="area", state="active").order_by("name") - for area in areas: - area.groups = Group.objects.filter(parent=area, type="wg", state="active").order_by("name") - for group in area.groups: - fill_in_charter_info(group) - fill_in_wg_roles(group) - fill_in_wg_drafts(group) - return render(request, 'group/1wg-charters.txt', - { 'areas': areas }, - content_type='text/plain; charset=UTF-8') - -@cache_page ( 60 * 60, cache="slowpages" ) + return response_from_file(Path(settings.CHARTER_PATH) / "1wg-charters.txt") + + def wg_charters_by_acronym(request, group_type): if group_type != "wg": raise Http404 + return response_from_file(Path(settings.CHARTER_PATH) / "1wg-charters-by-acronym.txt") - groups = Group.objects.filter(type="wg", state="active").exclude(parent=None).order_by("acronym") - for group in groups: - fill_in_charter_info(group) - fill_in_wg_roles(group) - fill_in_wg_drafts(group) - return render(request, 'group/1wg-charters-by-acronym.txt', - { 'groups': groups }, - content_type='text/plain; charset=UTF-8') def active_groups(request, group_type=None): @@ -290,7 +197,7 @@ def active_groups(request, group_type=None): return active_dirs(request) elif group_type == "review": return active_review_dirs(request) - elif group_type in ("program", "iabasg"): + elif group_type in ("program", "iabasg","iabworkshop"): return active_iab(request) elif group_type == "adm": return active_adm(request) @@ -300,8 +207,28 @@ def active_groups(request, group_type=None): raise Http404 def active_group_types(request): - grouptypes = GroupTypeName.objects.filter(slug__in=['wg','rg','ag','rag','team','dir','review','area','program','iabasg','adm']).filter(group__state='active').annotate(group_count=Count('group')) - return render(request, 'group/active_groups.html', {'grouptypes':grouptypes}) + grouptypes = ( + GroupTypeName.objects.filter( + slug__in=[ + "wg", + "rg", + "ag", + "rag", + "team", + "dir", + "review", + "area", + "program", + "iabasg", + "iabworkshop" + "adm", + ] + ) + .filter(group__state="active") + .order_by('order', 'name') # default ordering ignored for "GROUP BY" queries, make it explicit + .annotate(group_count=Count("group")) + ) + return render(request, "group/active_groups.html", {"grouptypes": grouptypes}) def active_dirs(request): dirs = Group.objects.filter(type__in=['dir', 'review'], state="active").order_by("name") @@ -318,13 +245,22 @@ def active_review_dirs(request): return render(request, 'group/active_review_dirs.html', {'dirs' : dirs }) def active_teams(request): - teams = Group.objects.filter(type="team", state="active").order_by("name") + parent_type_order = {"area": 1, "adm": 3, None: 4} + + def team_sort_key(group): + type_id = group.parent.type_id if group.parent else None + return (parent_type_order.get(type_id, 2), group.parent.name if group.parent else "", group.name) + + teams = sorted( + Group.objects.filter(type="team", state="active").select_related("parent"), + key=team_sort_key, + ) for group in teams: group.chairs = sorted(roles(group, "chair"), key=extract_last_name) - return render(request, 'group/active_teams.html', {'teams' : teams }) + return render(request, 'group/active_teams.html', {'teams': teams}) def active_iab(request): - iabgroups = Group.objects.filter(type__in=("program","iabasg"), state="active").order_by("-type_id","name") + iabgroups = Group.objects.filter(type__in=("program","iabasg","iabworkshop"), state="active").order_by("-type_id","name") for group in iabgroups: group.leads = sorted(roles(group, "lead"), key=extract_last_name) return render(request, 'group/active_iabgroups.html', {'iabgroups' : iabgroups }) @@ -334,7 +270,7 @@ def active_adm(request): return render(request, 'group/active_adm.html', {'adm' : adm }) def active_rfced(request): - rfced = Group.objects.filter(type="rfcedtyp", state="active").order_by("parent", "name") + rfced = Group.objects.filter(type__in=["rfcedtyp", "edwg", "edappr"], state="active").order_by("parent", "name") return render(request, 'group/active_rfced.html', {'rfced' : rfced}) @@ -359,7 +295,7 @@ def active_wgs(request): if group.list_subscribe.startswith('http'): group.list_subscribe_url = group.list_subscribe elif group.list_email.endswith('@ietf.org'): - group.list_subscribe_url = MAILING_LIST_INFO_URL % {'list_addr':group.list_email.split('@')[0]} + group.list_subscribe_url = MAILING_LIST_INFO_URL % {'list_addr':group.list_email.split('@')[0].lower(),'domain':'ietf.org'} else: group.list_subscribe_url = "mailto:"+group.list_subscribe @@ -413,35 +349,86 @@ def chartering_groups(request): dict(charter_states=charter_states, group_types=group_types)) + def concluded_groups(request): sections = OrderedDict() - sections['WGs'] = Group.objects.filter(type='wg', state="conclude").select_related("state", "charter").order_by("parent__name","acronym") - sections['RGs'] = Group.objects.filter(type='rg', state="conclude").select_related("state", "charter").order_by("parent__name","acronym") - sections['BOFs'] = Group.objects.filter(type='wg', state="bof-conc").select_related("state", "charter").order_by("parent__name","acronym") - sections['AGs'] = Group.objects.filter(type='ag', state="conclude").select_related("state", "charter").order_by("parent__name","acronym") - sections['RAGs'] = Group.objects.filter(type='rag', state="conclude").select_related("state", "charter").order_by("parent__name","acronym") - sections['Directorates'] = Group.objects.filter(type='dir', state="conclude").select_related("state", "charter").order_by("parent__name","acronym") - sections['Review teams'] = Group.objects.filter(type='review', state="conclude").select_related("state", "charter").order_by("parent__name","acronym") - sections['Teams'] = Group.objects.filter(type='team', state="conclude").select_related("state", "charter").order_by("parent__name","acronym") - sections['Programs'] = Group.objects.filter(type='program', state="conclude").select_related("state", "charter").order_by("parent__name","acronym") + sections["WGs"] = ( + Group.objects.filter(type="wg", state="conclude") + .select_related("state", "charter") + .order_by("parent__name", "acronym") + ) + sections["RGs"] = ( + Group.objects.filter(type="rg", state="conclude") + .select_related("state", "charter") + .order_by("parent__name", "acronym") + ) + sections["BOFs"] = ( + Group.objects.filter(type="wg", state="bof-conc") + .select_related("state", "charter") + .order_by("parent__name", "acronym") + ) + sections["AGs"] = ( + Group.objects.filter(type="ag", state="conclude") + .select_related("state", "charter") + .order_by("parent__name", "acronym") + ) + sections["RAGs"] = ( + Group.objects.filter(type="rag", state="conclude") + .select_related("state", "charter") + .order_by("parent__name", "acronym") + ) + sections["Directorates"] = ( + Group.objects.filter(type="dir", state="conclude") + .select_related("state", "charter") + .order_by("parent__name", "acronym") + ) + sections["Review teams"] = ( + Group.objects.filter(type="review", state="conclude") + .select_related("state", "charter") + .order_by("parent__name", "acronym") + ) + sections["Teams"] = ( + Group.objects.filter(type="team", state="conclude") + .select_related("state", "charter") + .order_by("parent__name", "acronym") + ) + sections["Programs"] = ( + Group.objects.filter(type="program", state="conclude") + .select_related("state", "charter") + .order_by("parent__name", "acronym") + ) for name, groups in sections.items(): - # add start/conclusion date d = dict((g.pk, g) for g in groups) for g in groups: g.start_date = g.conclude_date = None - for e in ChangeStateGroupEvent.objects.filter(group__in=groups, state="active").order_by("-time"): + # Some older BOFs were created in the "active" state, so consider both "active" and "bof" + # ChangeStateGroupEvents when finding the start date. A group with _both_ "active" and "bof" + # events should not be in the "bof-conc" state so this shouldn't cause a problem (if it does, + # we'll need to clean up the data) + for e in ChangeStateGroupEvent.objects.filter( + group__in=groups, + state__in=["active", "bof"] if name == "BOFs" else ["active"], + ).order_by("-time"): d[e.group_id].start_date = e.time - for e in ChangeStateGroupEvent.objects.filter(group__in=groups, state="conclude").order_by("time"): + # Similarly, some older BOFs were concluded into the "conclude" state and the event was never + # fixed, so consider both "conclude" and "bof-conc" ChangeStateGroupEvents when finding the + # concluded date. A group with _both_ "conclude" and "bof-conc" events should not be in the + # "bof-conc" state so this shouldn't cause a problem (if it does, we'll need to clean up the + # data) + for e in ChangeStateGroupEvent.objects.filter( + group__in=groups, + state__in=["bof-conc", "conclude"] if name == "BOFs" else ["conclude"], + ).order_by("time"): d[e.group_id].conclude_date = e.time - return render(request, 'group/concluded_groups.html', - dict(sections=sections)) + return render(request, "group/concluded_groups.html", dict(sections=sections)) + def prepare_group_documents(request, group, clist): found_docs, meta = prepare_document_table(request, docs_tracked_by_community_list(clist), request.GET, max_results=500) @@ -454,8 +441,8 @@ def prepare_group_documents(request, group, clist): # non-WG drafts and call for WG adoption are considered related if (d.group != group or (d.stream_id and d.get_state_slug("draft-stream-%s" % d.stream_id) in ("c-adopt", "wg-cand"))): - if d.get_state_slug() != "expired": - d.search_heading = "Related Internet-Draft" + if (d.type_id == "draft" and d.get_state_slug() not in ["expired","rfc"]) or d.type_id == "rfc": + d.search_heading = "Related Internet-Drafts and RFCs" docs_related.append(d) else: if not (d.get_state_slug('draft-iesg') == "dead" or (d.stream_id and d.get_state_slug("draft-stream-%s" % d.stream_id) == "dead")): @@ -465,6 +452,47 @@ def prepare_group_documents(request, group, clist): return docs, meta, docs_related, meta_related +def get_leadership(group_type): + people = Person.objects.filter( + role__name__slug="chair", + role__group__type=group_type, + role__group__state__slug__in=("active", "bof", "proposed"), + ).distinct() + leaders = [] + for person in people: + parts = person.name_parts() + groups = [ + r.group.acronym + for r in person.role_set.filter( + name__slug="chair", + group__type=group_type, + group__state__slug__in=("active", "bof", "proposed"), + ) + ] + entry = {"name": "%s, %s" % (parts[3], parts[1]), "groups": ", ".join(groups)} + leaders.append(entry) + return sorted(leaders, key=lambda a: a["name"]) + + +def group_leadership(request, group_type=None): + context = {} + context["leaders"] = get_leadership(group_type) + context["group_type"] = group_type + return render(request, "group/group_leadership.html", context) + + +def group_leadership_csv(request, group_type=None): + leaders = get_leadership(group_type) + response = HttpResponse(content_type="text/csv") + response["Content-Disposition"] = ( + f'attachment; filename="group_leadership_{group_type}.csv"' + ) + writer = csv.writer(response, dialect=csv.excel, delimiter=str(",")) + writer.writerow(["Name", "Groups"]) + for leader in leaders: + writer.writerow([leader["name"], leader["groups"]]) + return response + def group_home(request, acronym, group_type=None): group = get_group_or_404(acronym, group_type) kwargs = dict(acronym=group.acronym) @@ -514,9 +542,8 @@ def group_documents_txt(request, acronym, group_type=None): rows = [] for d in itertools.chain(docs, docs_related): - rfc_number = d.rfc_number() - if rfc_number != None: - name = rfc_number + if d.type_id == "rfc": + name = str(d.rfc_number) else: name = "%s-%s" % (d.name, d.rev) @@ -544,6 +571,7 @@ def group_about(request, acronym, group_type=None): can_provide_update = can_provide_status_update(request.user, group) status_update = group.latest_event(type="status_update") + subgroups = Group.objects.filter(parent=group, state="active").exclude(type__slug__in=["sdo", "individ", "nomcom"]).order_by("type", "acronym") return render(request, 'group/group_about.html', construct_group_menu_context(request, group, "about", group_type, { @@ -556,6 +584,7 @@ def group_about(request, acronym, group_type=None): "charter_submit_url": charter_submit_url, "editable_roles": group.used_roles or group.features.default_used_roles, "closing_note": e, + "subgroups": subgroups, })) def all_status(request): @@ -582,17 +611,6 @@ def all_status(request): } ) -def group_about_rendertest(request, acronym, group_type=None): - group = get_group_or_404(acronym, group_type) - charter = None - if group.charter: - charter = get_charter_text(group) - try: - rendered = markdown.markdown(charter) - except Exception as e: - rendered = f'Markdown rendering failed: {e}' - return render(request, 'group/group_about_rendertest.html', {'group':group, 'charter':charter, 'rendered':rendered}) - def group_about_status(request, acronym, group_type=None): group = get_group_or_404(acronym, group_type) status_update = group.latest_event(type='status_update') @@ -653,21 +671,6 @@ def group_about_status_edit(request, acronym, group_type=None): } ) -def get_group_email_aliases(acronym, group_type): - if acronym: - pattern = re.compile(r'expand-(%s)(-\w+)@.*? +(.*)$'%acronym) - else: - pattern = re.compile(r'expand-(.*?)(-\w+)@.*? +(.*)$') - - aliases = [] - with io.open(settings.GROUP_VIRTUAL_PATH,"r") as virtual_file: - for line in virtual_file.readlines(): - m = pattern.match(line) - if m: - if acronym or not group_type or Group.objects.filter(acronym=m.group(1),type__slug=group_type): - aliases.append({'acronym':m.group(1),'alias_type':m.group(2),'expansion':m.group(3)}) - return aliases - def email(request, acronym, group_type=None): group = get_group_or_404(acronym, group_type) @@ -695,6 +698,61 @@ def history(request, acronym, group_type=None): "can_add_comment": can_add_comment, })) + +class RequestsHistoryParamsForm(forms.Form): + SINCE_CHOICES = ( + (None, "1 month"), + ("3m", "3 months"), + ("6m", "6 months"), + ("1y", "1 year"), + ("2y", "2 years"), + ("all", "All"), + ) + + reviewer_email = forms.EmailField(required=False) + since = forms.ChoiceField(choices=SINCE_CHOICES, required=False) + +def review_requests_history(request, acronym, group_type=None): + group = get_group_or_404(acronym, group_type) + if not group.features.has_reviews: + raise Http404 + + params = RequestsHistoryParamsForm(request.GET) + if not params.is_valid(): + return HttpResponseBadRequest("Invalid parameters") + + reviewer_email = params.cleaned_data["reviewer_email"] or None + if reviewer_email: + history = ReviewAssignment.history.model.objects.filter( + review_request__team__acronym=acronym, + reviewer=reviewer_email) + else: + history = ReviewAssignment.history.model.objects.filter( + review_request__team__acronym=acronym) + reviewer_email = '' + + since = params.cleaned_data["since"] or None + if since != "all": + date_limit = { + None: datetime.timedelta(days=31), + "3m": datetime.timedelta(days=31 * 3), + "6m": datetime.timedelta(days=180), + "1y": datetime.timedelta(days=365), + "2y": datetime.timedelta(days=2 * 365), + }[since] + + history = history.filter(review_request__time__gte=datetime_today(DEADLINE_TZINFO) - date_limit) + + return render(request, 'group/review_requests_history.html', + construct_group_menu_context(request, group, "reviews history", group_type, { + "group": group, + "acronym": acronym, + "history": history, + "since_choices": params.SINCE_CHOICES, + "since": since, + "reviewer_email": reviewer_email + })) + def materials(request, acronym, group_type=None): group = get_group_or_404(acronym, group_type) if not group.features.has_nonsession_materials: @@ -734,14 +792,31 @@ def dependencies(request, acronym, group_type=None): source__type="draft", relationship__slug__startswith="ref", ) - - both_rfcs = Q(source__states__slug="rfc", target__docs__states__slug="rfc") - inactive = Q(source__states__slug__in=["expired", "repl"]) + rfc_or_subseries = {"rfc", "bcp", "fyi", "std"} + both_rfcs = Q(source__type_id="rfc", target__type_id__in=rfc_or_subseries) + pre_rfc_draft_to_rfc = Q( + source__states__type="draft", + source__states__slug="rfc", + target__type_id__in=rfc_or_subseries, + ) + both_pre_rfcs = Q( + source__states__type="draft", + source__states__slug="rfc", + target__type_id="draft", + target__states__type="draft", + target__states__slug="rfc", + ) + inactive = Q( + source__states__type="draft", + source__states__slug__in=["expired", "repl"], + ) attractor = Q(target__name__in=["rfc5000", "rfc5741"]) - removed = Q(source__states__slug__in=["auth-rm", "ietf-rm"]) + removed = Q(source__states__type="draft", source__states__slug__in=["auth-rm", "ietf-rm"]) relations = ( RelatedDocument.objects.filter(references) .exclude(both_rfcs) + .exclude(pre_rfc_draft_to_rfc) + .exclude(both_pre_rfcs) .exclude(inactive) .exclude(attractor) .exclude(removed) @@ -749,29 +824,28 @@ def dependencies(request, acronym, group_type=None): links = set() for x in relations: - target_state = x.target.document.get_state_slug("draft") - if target_state != "rfc" or x.is_downref(): + always_include = x.target.type_id not in rfc_or_subseries and x.target.get_state_slug("draft") != "rfc" + if always_include or x.is_downref(): links.add(x) replacements = RelatedDocument.objects.filter( relationship__slug="replaces", - target__docs__in=[x.target.document for x in links], + target__in=[x.target for x in links], ) for x in replacements: links.add(x) - nodes = set([x.source for x in links]).union([x.target.document for x in links]) + nodes = set([x.source for x in links]).union([x.target for x in links]) graph = { "nodes": [ { - "id": x.canonical_name(), - "rfc": x.get_state("draft").slug == "rfc", - "post-wg": not x.get_state("draft-iesg").slug - in ["idexists", "watching", "dead"], - "expired": x.get_state("draft").slug == "expired", - "replaced": x.get_state("draft").slug == "repl", - "group": x.group.acronym if x.group.acronym != "none" else "", + "id": x.became_rfc().name if x.became_rfc() else x.name, + "rfc": x.type_id == "rfc" or x.became_rfc() is not None, + "post-wg": x.get_state_slug("draft-iesg") not in ["idexists", "dead"], + "expired": x.get_state_slug("draft") == "expired", + "replaced": x.get_state_slug("draft") == "repl", + "group": x.group.acronym if x.group and x.group.acronym != "none" else "", "url": x.get_absolute_url(), "level": x.intended_std_level.name if x.intended_std_level @@ -783,8 +857,8 @@ def dependencies(request, acronym, group_type=None): ], "links": [ { - "source": x.source.canonical_name(), - "target": x.target.document.canonical_name(), + "source": x.source.became_rfc().name if x.source.became_rfc() else x.source.name, + "target": x.target.became_rfc().name if x.target.became_rfc() else x.target.name, "rel": "downref" if x.is_downref() else x.relationship.slug, } for x in links @@ -807,35 +881,125 @@ def email_aliases(request, acronym=None, group_type=None): return render(request,'group/email_aliases.html',{'aliases':aliases,'ietf_domain':settings.IETF_DOMAIN,'group':group}) -def meetings(request, acronym=None, group_type=None): - group = get_group_or_404(acronym,group_type) if acronym else None +def meetings(request, acronym, group_type=None): + group = get_group_or_404(acronym, group_type) + + four_years_ago = timezone.now() - datetime.timedelta(days=4 * 365) + + stsas = SchedTimeSessAssignment.objects.filter( + session__type__in=["regular", "plenary", "other"], + session__group=group) + if group.acronym not in ["iab", "iesg"]: + stsas = stsas.filter(session__meeting__date__gt=four_years_ago) + stsas = stsas.annotate(sessionstatus=Coalesce( + Subquery( + SchedulingEvent.objects.filter( + session=OuterRef("session__pk") + ).order_by( + '-time', '-id' + ).values('status')[:1]), + Value(''), + output_field=TextField()) + ).filter( + sessionstatus__in=["sched", "schedw", "appr", "canceled"], + session__meeting__schedule=F("schedule") + ).distinct().select_related( + "session", "session__group", "session__group__parent", "session__meeting__type", "timeslot" + ).prefetch_related( + "session__materials", + "session__materials__states", + Prefetch("session__materials", + queryset=Document.objects.exclude(states__type=F("type"), states__slug='deleted').order_by('presentations__order').prefetch_related('states'), + to_attr="prefetched_active_materials" + ), + ) + + stsas = list(stsas) - four_years_ago = timezone.now()-datetime.timedelta(days=4*365) + for stsa in stsas: + stsa.session._otsa = stsa + stsa.session.official_timeslotassignment = types.MethodType(lambda self:self._otsa, stsa.session) + stsa.session.current_status = stsa.sessionstatus - sessions = add_event_info_to_session_qs( - group.session_set.filter( - meeting__date__gt=four_years_ago, - type__in=['regular','plenary','other'] + sessions = sorted( + set([stsa.session for stsa in stsas]), + key=lambda x: ( + x._otsa.timeslot.time, + x._otsa.timeslot.type_id, + x._otsa.session.group.parent.name if x._otsa.session.group.parent else None, + x._otsa.session.name ) - ).filter( - current_status__in=['sched','schedw','appr','canceled'], ) + meeting_seen = None + for s in sessions: + if s.meeting != meeting_seen: + meeting_seen = s.meeting + order = 1 + s._oim = order + s.order_in_meeting = types.MethodType(lambda self:self._oim, s) + order += 1 + + + revsub_dates_by_meeting = dict(ImportantDate.objects.filter(name_id="revsub", meeting__session__in=sessions).distinct().values_list("meeting_id","date")) + + for s in sessions: + s.order_number = s.order_in_meeting() + if s.meeting.pk in revsub_dates_by_meeting: + cutoff_date = revsub_dates_by_meeting[s.meeting.pk] + else: + cutoff_date = s.meeting.date + datetime.timedelta(days=s.meeting.submission_correction_day_offset) + s.cached_is_cutoff = date_today(datetime.UTC) > cutoff_date + future, in_progress, recent, past = group_sessions(sessions) - can_edit = group.has_role(request.user,group.features.groupman_roles) - can_always_edit = has_role(request.user,["Secretariat","Area Director"]) - - return render(request,'group/meetings.html', - construct_group_menu_context(request, group, "meetings", group_type, { - 'group':group, - 'future':future, - 'in_progress':in_progress, - 'recent':recent, - 'past':past, - 'can_edit':can_edit, - 'can_always_edit':can_always_edit, - })) + can_edit = group.has_role(request.user, group.features.groupman_roles) + can_always_edit = has_role(request.user, ["Secretariat", "Area Director"]) + + far_past = [] + if group.acronym in ["iab", "iesg"]: + recent_past = [] + for s in past: + if s.time >= four_years_ago: + recent_past.append(s) + else: + far_past.append(s) + past = recent_past + + # Add calendar actions + cal_actions = [] + + cal_actions.append(dict( + label='Download as .ics', + url=reverse('ietf.meeting.views.upcoming_ical')+"?show="+group.acronym) + ) + cal_actions.append(dict( + label='Subscribe with webcal', + url='webcal://'+request.get_host()+reverse('ietf.meeting.views.upcoming_ical')+"?show="+group.acronym) + ) + + return render( + request, + "group/meetings.html", + construct_group_menu_context( + request, + group, + "meetings", + group_type, + { + "group": group, + "future": future, + "in_progress": in_progress, + "recent": recent, + "past": past, + "far_past": far_past, + "can_edit": can_edit, + "can_always_edit": can_always_edit, + "cal_actions": cal_actions, + }, + ), + ) + def chair_photos(request, group_type=None): roles = sorted(Role.objects.filter(group__type=group_type, group__state='active', name_id='chair'),key=lambda x: x.person.last_name()+x.person.name+x.group.acronym) @@ -864,60 +1028,6 @@ def group_photos(request, group_type=None, acronym=None): 'group':group })) - -## XXX Remove after testing -# def get_or_create_initial_charter(group, group_type): -# charter_name = charter_name_for_group(group) -# -# try: -# charter = Document.objects.get(docalias__name=charter_name) -# except Document.DoesNotExist: -# charter = Document( -# name=charter_name, -# type_id="charter", -# title=group.name, -# group=group, -# abstract=group.name, -# rev="00-00", -# ) -# charter.save() -# charter.set_state(State.objects.get(used=True, type="charter", slug="notrev")) -# -# # Create an alias as well -# DocAlias.objects.create(name=charter.name).docs.add(charter) -# -# return charter -# -# @login_required -# def submit_initial_charter(request, group_type=None, acronym=None): -# -# # This needs refactoring. -# # The signature assumed you could have groups with the same name, but with different types, which we do not allow. -# # Consequently, this can be called with an existing group acronym and a type -# # that doesn't match the existing group type. The code below essentially ignores the group_type argument. -# # -# # If possible, the use of get_or_create_initial_charter should be moved -# # directly into charter_submit, and this function should go away. -# -# if acronym==None: -# raise Http404 -# -# group = get_object_or_404(Group, acronym=acronym) -# if not group.features.has_chartering_process: -# raise Http404 -# -# # This is where we start ignoring the passed in group_type -# group_type = group.type_id -# -# if not can_manage_group(request.user, group): -# permission_denied(request, "You don't have permission to access this view") -# -# if not group.charter: -# group.charter = get_or_create_initial_charter(group, group_type) -# group.save() -# -# return redirect('ietf.doc.views_charter.submit', name=group.charter.name, option="initcharter") - @login_required def edit(request, group_type=None, acronym=None, action="edit", field=None): """Edit or create a group, notifying parties as @@ -960,14 +1070,17 @@ def diff(attr, name): if not (can_manage_group(request.user, group) or group.has_role(request.user, group.features.groupman_roles)): permission_denied(request, "You don't have permission to access this view") + hide_parent = not has_role(request.user, ("Secretariat", "Area Director", "IRTF Chair")) else: # This allows ADs to create RG and the IRTF Chair to create WG, but we trust them not to if not has_role(request.user, ("Secretariat", "Area Director", "IRTF Chair")): permission_denied(request, "You don't have permission to access this view") - + hide_parent = False if request.method == 'POST': - form = GroupForm(request.POST, group=group, group_type=group_type, field=field) + form = GroupForm(request.POST, group=group, group_type=group_type, field=field, hide_parent=hide_parent) + if field and not form.fields: + permission_denied(request, "You don't have permission to edit this field") if form.is_valid(): clean = form.cleaned_data if new_group: @@ -1129,7 +1242,9 @@ def diff(attr, name): else: init = dict() - form = GroupForm(initial=init, group=group, group_type=group_type, field=field) + form = GroupForm(initial=init, group=group, group_type=group_type, field=field, hide_parent=hide_parent) + if field and not form.fields: + permission_denied(request, "You don't have permission to edit this field") return render(request, 'group/edit.html', dict(group=group, @@ -1283,6 +1398,8 @@ def streams(request): return render(request, 'group/index.html', {'streams':streams}) def stream_documents(request, acronym): + if acronym == "editorial": + return HttpResponseRedirect(urlreverse(group_documents, kwargs=dict(acronym="rswg"))) streams = [ s.slug for s in StreamName.objects.all().exclude(slug__in=['ietf', 'legacy']) ] if not acronym in streams: raise Http404("No such stream: %s" % acronym) @@ -1290,7 +1407,10 @@ def stream_documents(request, acronym): editable = has_role(request.user, "Secretariat") or group.has_role(request.user, "chair") stream = StreamName.objects.get(slug=acronym) - qs = Document.objects.filter(states__type="draft", states__slug__in=["active", "rfc"], stream=acronym) + qs = Document.objects.filter(stream=acronym).filter( + Q(type_id="draft", states__type="draft", states__slug="active") + | Q(type_id="rfc") + ).distinct() docs, meta = prepare_document_table(request, qs, max_results=1000) return render(request, 'group/stream_documents.html', {'stream':stream, 'docs':docs, 'meta':meta, 'editable':editable } ) @@ -1340,20 +1460,132 @@ def stream_edit(request, acronym): ) -@cache_control(public=True, max_age=30*60) +@cache_control(public=True, max_age=30 * 60) @cache_page(30 * 60) def group_menu_data(request): - groups = Group.objects.filter(state="active", parent__state="active").filter(Q(type__features__acts_like_wg=True)|Q(type_id__in=['program','iabasg'])|Q(parent__acronym='ietfadminllc')|Q(parent__acronym='rfceditor')).order_by("-type_id","acronym") + groups = ( + Group.objects.filter(state="active", parent__state="active") + .filter( + Q(type__features__acts_like_wg=True) + | Q(type_id__in=["program", "iabasg", "iabworkshop"]) + | Q(parent__acronym="ietfadminllc") + | Q(parent__acronym="rfceditor") + ) + .order_by("-type_id", "acronym") + .select_related("type") + ) groups_by_parent = defaultdict(list) for g in groups: - url = urlreverse("ietf.group.views.group_home", kwargs={ 'group_type': g.type_id, 'acronym': g.acronym }) -# groups_by_parent[g.parent_id].append({ 'acronym': g.acronym, 'name': escape(g.name), 'url': url }) - groups_by_parent[g.parent_id].append({ 'acronym': g.acronym, 'name': escape(g.name), 'type': escape(g.type.verbose_name or g.type.name), 'url': url }) + url = urlreverse( + "ietf.group.views.group_home", + kwargs={"group_type": g.type_id, "acronym": g.acronym}, + ) + # groups_by_parent[g.parent_id].append({ 'acronym': g.acronym, 'name': escape(g.name), 'url': url }) + groups_by_parent[g.parent_id].append( + { + "acronym": g.acronym, + "name": escape(g.name), + "type": escape(g.type.verbose_name or g.type.name), + "url": url, + } + ) + iab = Group.objects.get(acronym="iab") + groups_by_parent[iab.pk].insert( + 0, + { + "acronym": iab.acronym, + "name": iab.name, + "type": "Top Level Group", + "url": urlreverse( + "ietf.group.views.group_home", kwargs={"acronym": iab.acronym} + ), + }, + ) return JsonResponse(groups_by_parent) + +@cache_control(public=True, max_age=30 * 60) +@cache_page(30 * 60) +def group_stats_data(request, years="3", only_active=True): + when = timezone.now() - datetime.timedelta(days=int(years) * 365) + docs = ( + Document.objects.filter(type="draft", stream="ietf") + .filter( + Q(docevent__newrevisiondocevent__time__gte=when) + | Q(docevent__type="published_rfc", docevent__time__gte=when) + ) + .exclude(states__type="draft", states__slug="repl") + .distinct() + ) + + data = [] + for a in Group.objects.filter(type="area"): + if only_active and not a.is_active: + continue + + area_docs = docs.filter(group__parent=a).exclude(group__acronym="none") + if not area_docs: + continue + + area_page_cnt = 0 + area_doc_cnt = 0 + for wg in Group.objects.filter(type="wg", parent=a): + if only_active and not wg.is_active: + continue + + wg_docs = area_docs.filter(group=wg) + if not wg_docs: + continue + + wg_page_cnt = 0 + for doc in wg_docs: + # add doc data + data.append( + { + "id": doc.name, + "active": True, + "parent": wg.acronym, + "grandparent": a.acronym, + "pages": doc.pages, + "docs": 1, + } + ) + wg_page_cnt += doc.pages + + area_doc_cnt += len(wg_docs) + area_docs = area_docs.exclude(group=wg) + + # add WG data + data.append( + { + "id": wg.acronym, + "active": wg.is_active, + "parent": a.acronym, + "grandparent": "ietf", + "pages": wg_page_cnt, + "docs": len(wg_docs), + } + ) + area_page_cnt += wg_page_cnt + + # add area data + data.append( + { + "id": a.acronym, + "active": a.is_active, + "parent": "ietf", + "pages": area_page_cnt, + "docs": area_doc_cnt, + } + ) + + data.append({"id": "ietf", "active": True}) + return JsonResponse(data, safe=False) + + # --- Review views ----------------------------------------------------- def get_open_review_requests_for_team(team, assignment_status=None): @@ -2016,7 +2248,87 @@ def reset_next_reviewer(request, acronym, group_type=None): return render(request, 'group/reset_next_reviewer.html', { 'group':group, 'form': form,}) +def statements(request, acronym, group_type=None): + if not acronym in ["iab", "iesg"]: + raise Http404 + group = get_group_or_404(acronym, group_type) + statements = ( + group.document_set.filter(type_id="statement") + .annotate( + published=Subquery( + DocEvent.objects.filter(doc=OuterRef("pk"), type="published_statement") + .order_by("-time") + .values("time")[:1] + ) + ) + .annotate( + status=Subquery( + Document.states.through.objects.filter( + document_id=OuterRef("pk"), state__type="statement" + ).values_list("state__slug", flat=True)[:1] + ) + ) + .order_by("status", "-published") + ) + return render( + request, + "group/statements.html", + construct_group_menu_context( + request, + group, + "statements", + group_type, + { + "group": group, + "statements": statements, + }, + ), + ) +def appeals(request, acronym, group_type=None): + if not acronym in ["iab", "iesg"]: + raise Http404 + group = get_group_or_404(acronym, group_type) + appeals = group.appeal_set.all() + return render( + request, + "group/appeals.html", + construct_group_menu_context( + request, + group, + "appeals", + group_type, + { + "group": group, + "appeals": appeals, + }, + ), + ) - - +@ignore_view_kwargs("group_type") +def appeal_artifact(request, acronym, artifact_id): + artifact = get_object_or_404(AppealArtifact, pk=artifact_id) + if artifact.is_markdown(): + artifact_html = markdown.markdown(artifact.bits.tobytes().decode("utf-8")) + return render( + request, + "group/appeal_artifact.html", + dict(artifact=artifact, artifact_html=artifact_html) + ) + else: + return HttpResponse( + artifact.bits, + headers = { + "Content-Type": artifact.content_type, + "Content-Disposition": f'attachment; filename="{artifact.download_name()}"' + } + ) + +@role_required("Secretariat") +@ignore_view_kwargs("group_type") +def appeal_artifact_markdown(request, acronym, artifact_id): + artifact = get_object_or_404(AppealArtifact, pk=artifact_id) + if artifact.is_markdown(): + return HttpResponse(artifact.bits, content_type=artifact.content_type) + else: + raise Http404 diff --git a/ietf/secr/areas/__init__.py b/ietf/help/migrations/__init__.py similarity index 100% rename from ietf/secr/areas/__init__.py rename to ietf/help/migrations/__init__.py diff --git a/ietf/help/tests_views.py b/ietf/help/tests_views.py deleted file mode 100644 index ee80dad865..0000000000 --- a/ietf/help/tests_views.py +++ /dev/null @@ -1,21 +0,0 @@ -from pyquery import PyQuery - -from django.urls import reverse - -import debug # pyflakes:ignore - -from ietf.utils.test_utils import TestCase -from ietf.doc.models import StateType - -class HelpPageTests(TestCase): - - def test_state_index(self): - url = reverse('ietf.help.views.state_index') - r = self.client.get(url) - q = PyQuery(r.content) - content = [ e.text for e in q('#content table td a ') ] - names = StateType.objects.values_list('slug', flat=True) - # The following doesn't cover all doc types, only a selection - for name in names: - if not '-' in name: - self.assertIn(name, content) diff --git a/ietf/help/urls.py b/ietf/help/urls.py index f1cc625fa7..90ce7e12e9 100644 --- a/ietf/help/urls.py +++ b/ietf/help/urls.py @@ -2,10 +2,10 @@ from ietf.help import views from ietf.utils.urls import url +from django.views.generic import RedirectView urlpatterns = [ url(r'^state/(?P[-\w]+)/(?P[-\w]+)/?$', views.state), url(r'^state/(?P[-\w]+)/?$', views.state), - url(r'^state/?$', views.state_index), + url(r'^state/?$', RedirectView.as_view(url='/doc/help/state/', permanent=True)), ] - diff --git a/ietf/help/views.py b/ietf/help/views.py index 9c0060e991..493bf0dcf1 100644 --- a/ietf/help/views.py +++ b/ietf/help/views.py @@ -1,25 +1,11 @@ # Copyright The IETF Trust 2007, All Rights Reserved -import os - -from django.shortcuts import get_object_or_404, render - import debug # pyflakes:ignore -from ietf.doc.models import State, StateType from ietf.name.models import StreamName +from django.shortcuts import redirect -def state_index(request): - types = StateType.objects.all() - names = [ type.slug for type in types ] - for type in types: - if "-" in type.slug and type.slug.split('-',1)[0] in names: - type.stategroups = None - else: - groups = StateType.objects.filter(slug__startswith=type.slug) - type.stategroups = [ g.slug[len(type.slug)+1:] for g in groups if not g == type ] or "" - - return render(request, 'help/state_index.html', {"types": types}) +# This is just a redirect to the new URL under /doc; can probably go away eventually. def state(request, doc, type=None): if type: @@ -27,14 +13,5 @@ def state(request, doc, type=None): if type in streams: type = "stream-%s" % type slug = "%s-%s" % (doc,type) if type else doc - statetype = get_object_or_404(StateType, slug=slug) - states = State.objects.filter(used=True, type=statetype).order_by('order') - return render(request, 'help/states.html', {"doc": doc, "type": statetype, "states":states} ) - -def environment(request): - if request.is_secure(): - os.environ['SCHEME'] = "https" - else: - os.environ['SCHEME'] = "http" - os.environ["URL"] = request.build_absolute_uri(".") - return render(request, 'help/environment.html', {"env": os.environ} ) + return redirect('/doc/help/state/%s' % slug, permanent = True) + \ No newline at end of file diff --git a/ietf/idindex/.gitignore b/ietf/idindex/.gitignore deleted file mode 100644 index c7013ced9d..0000000000 --- a/ietf/idindex/.gitignore +++ /dev/null @@ -1,2 +0,0 @@ -/*.pyc -/settings_local.py diff --git a/ietf/idindex/index.py b/ietf/idindex/index.py index 441febdd96..19eb29d4da 100644 --- a/ietf/idindex/index.py +++ b/ietf/idindex/index.py @@ -14,7 +14,7 @@ import debug # pyflakes:ignore -from ietf.doc.models import Document, DocEvent, DocumentAuthor, RelatedDocument, DocAlias, State +from ietf.doc.models import Document, DocEvent, DocumentAuthor, RelatedDocument, State from ietf.doc.models import LastCallDocEvent, NewRevisionDocEvent from ietf.doc.models import IESG_SUBSTATE_TAGS from ietf.doc.templatetags.ietf_filters import clean_whitespace @@ -31,15 +31,18 @@ def formatted_rev_date(name): t = revision_time.get(name) return t.strftime("%Y-%m-%d") if t else "" - rfc_aliases = dict(DocAlias.objects.filter(name__startswith="rfc", - docs__states=State.objects.get(type="draft", slug="rfc")).values_list("docs__name", "name")) + rfcs = dict() + for rfc in Document.objects.filter(type_id="rfc"): + draft = rfc.came_from_draft() + if draft is not None: + rfcs[draft.name] = rfc.name - replacements = dict(RelatedDocument.objects.filter(target__docs__states=State.objects.get(type="draft", slug="repl"), + replacements = dict(RelatedDocument.objects.filter(target__states=State.objects.get(type="draft", slug="repl"), relationship="replaces").values_list("target__name", "source__name")) # we need a distinct to prevent the queries below from multiplying the result - all_ids = Document.objects.filter(type="draft").order_by('name').exclude(name__startswith="rfc").distinct() + all_ids = Document.objects.filter(type="draft").order_by('name').distinct() res = ["\nInternet-Drafts Status Summary\n"] @@ -48,7 +51,7 @@ def add_line(f1, f2, f3, f4): res.append(f1 + "\t" + f2 + "\t" + f3 + "\t" + f4) - inactive_states = ["idexists", "pub", "watching", "dead"] + inactive_states = ["idexists", "pub", "dead"] excludes = list(State.objects.filter(type="draft", slug__in=["rfc","repl"])) includes = list(State.objects.filter(type="draft-iesg").exclude(slug__in=inactive_states)) @@ -62,7 +65,7 @@ def add_line(f1, f2, f3, f4): state += "::" + "::".join(tags) add_line(d.name + "-" + d.rev, formatted_rev_date(d.name), - "In IESG processing - ID Tracker state <" + state + ">", + "In IESG processing - I-D Tracker state <" + state + ">", "", ) @@ -77,9 +80,9 @@ def add_line(f1, f2, f3, f4): last_field = "" if s.slug == "rfc": - a = rfc_aliases.get(name) - if a: - last_field = a[3:] + rfc = rfcs.get(name) + if rfc: + last_field = rfc[3:] # Rework this to take advantage of having the number at hand already. elif s.slug == "repl": state += " replaced by " + replacements.get(name, "0") @@ -108,14 +111,17 @@ def file_types_for_drafts(): def all_id2_txt(): # this returns a lot of data so try to be efficient - drafts = Document.objects.filter(type="draft").exclude(name__startswith="rfc").order_by('name') + drafts = Document.objects.filter(type="draft").order_by('name') drafts = drafts.select_related('group', 'group__parent', 'ad', 'intended_std_level', 'shepherd', ) drafts = drafts.prefetch_related("states") - rfc_aliases = dict(DocAlias.objects.filter(name__startswith="rfc", - docs__states=State.objects.get(type="draft", slug="rfc")).values_list("docs__name", "name")) + rfcs = dict() + for rfc in Document.objects.filter(type_id="rfc"): + draft = rfc.came_from_draft() + if draft is not None: + rfcs[draft.name] = rfc.name - replacements = dict(RelatedDocument.objects.filter(target__docs__states=State.objects.get(type="draft", slug="repl"), + replacements = dict(RelatedDocument.objects.filter(target__states=State.objects.get(type="draft", slug="repl"), relationship="replaces").values_list("target__name", "source__name")) revision_time = dict(DocEvent.objects.filter(type="new_revision", doc__name__startswith="draft-").order_by('time').values_list("doc__name", "time")) @@ -164,9 +170,9 @@ def all_id2_txt(): # 4 rfc_number = "" if state == "rfc": - a = rfc_aliases.get(d.name) - if a: - rfc_number = a[3:] + rfc = rfcs.get(d.name) + if rfc: + rfc_number = rfc[3:] fields.append(rfc_number) # 5 repl = "" @@ -270,7 +276,7 @@ def active_drafts_index_by_group(extra_values=()): groups = [g for g in groups_dict.values() if hasattr(g, "active_drafts")] groups.sort(key=lambda g: g.acronym) - fallback_time = datetime.datetime(1950, 1, 1, tzinfo=datetime.timezone.utc) + fallback_time = datetime.datetime(1950, 1, 1, tzinfo=datetime.UTC) for g in groups: g.active_drafts.sort(key=lambda d: d.get("initial_rev_time", fallback_time)) @@ -296,6 +302,6 @@ def id_index_txt(with_abstracts=False): return render_to_string("idindex/id_index.txt", { 'groups': groups, - 'time': timezone.now().astimezone(datetime.timezone.utc).strftime("%Y-%m-%d %H:%M:%S %Z"), + 'time': timezone.now().astimezone(datetime.UTC).strftime("%Y-%m-%d %H:%M:%S %Z"), 'with_abstracts': with_abstracts, }) diff --git a/ietf/idindex/tasks.py b/ietf/idindex/tasks.py new file mode 100644 index 0000000000..2f5f1871d7 --- /dev/null +++ b/ietf/idindex/tasks.py @@ -0,0 +1,99 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +# +# Celery task definitions +# +import os +import shutil + +import debug # pyflakes:ignore + +from celery import shared_task +from contextlib import AbstractContextManager +from pathlib import Path +from tempfile import NamedTemporaryFile +from typing import List + +from django.conf import settings + +from ietf.doc.storage_utils import store_file + +from .index import all_id_txt, all_id2_txt, id_index_txt + + +class TempFileManager(AbstractContextManager): + def __init__(self, tmpdir=None) -> None: + self.cleanup_list: set[Path] = set() + self.dir = tmpdir + + def make_temp_file(self, content): + with NamedTemporaryFile(mode="wt", delete=False, dir=self.dir) as tf: + tf_path = Path(tf.name) + self.cleanup_list.add(tf_path) + tf.write(content) + return tf_path + + def move_into_place(self, src_path: Path, dest_path: Path, hardlink_dirs: List[Path] = []): + shutil.move(src_path, dest_path) + dest_path.chmod(0o644) + self.cleanup_list.remove(src_path) + for path in hardlink_dirs: + target = path / dest_path.name + target.unlink(missing_ok=True) + os.link(dest_path, target) # until python>=3.10 + with dest_path.open("rb") as f: + store_file("indexes", dest_path.name, f, allow_overwrite=True) + + def cleanup(self): + for tf_path in self.cleanup_list: + tf_path.unlink(missing_ok=True) + + def __exit__(self, exc_type, exc_val, exc_tb): + self.cleanup() + return False # False: do not suppress the exception + + +@shared_task +def idindex_update_task(): + """Update I-D indexes""" + id_path = Path(settings.INTERNET_DRAFT_PATH) + derived_path = Path(settings.DERIVED_DIR) + download_path = Path(settings.ALL_ID_DOWNLOAD_DIR) + ftp_path = Path(settings.FTP_DIR) / "internet-drafts" + all_archive_path = Path(settings.INTERNET_ALL_DRAFTS_ARCHIVE_DIR) + + with TempFileManager() as tmp_mgr: + # Generate copies of new contents + all_id_content = all_id_txt() + all_id_tmpfile = tmp_mgr.make_temp_file(all_id_content) + derived_all_id_tmpfile = tmp_mgr.make_temp_file(all_id_content) + download_all_id_tmpfile = tmp_mgr.make_temp_file(all_id_content) + + id_index_content = id_index_txt() + id_index_tmpfile = tmp_mgr.make_temp_file(id_index_content) + derived_id_index_tmpfile = tmp_mgr.make_temp_file(id_index_content) + download_id_index_tmpfile = tmp_mgr.make_temp_file(id_index_content) + + id_abstracts_content = id_index_txt(with_abstracts=True) + id_abstracts_tmpfile = tmp_mgr.make_temp_file(id_abstracts_content) + derived_id_abstracts_tmpfile = tmp_mgr.make_temp_file(id_abstracts_content) + download_id_abstracts_tmpfile = tmp_mgr.make_temp_file(id_abstracts_content) + + all_id2_content = all_id2_txt() + all_id2_tmpfile = tmp_mgr.make_temp_file(all_id2_content) + derived_all_id2_tmpfile = tmp_mgr.make_temp_file(all_id2_content) + + # Move temp files as-atomically-as-possible into place + tmp_mgr.move_into_place(all_id_tmpfile, id_path / "all_id.txt", [ftp_path, all_archive_path]) + tmp_mgr.move_into_place(derived_all_id_tmpfile, derived_path / "all_id.txt") + tmp_mgr.move_into_place(download_all_id_tmpfile, download_path / "id-all.txt") + + tmp_mgr.move_into_place(id_index_tmpfile, id_path / "1id-index.txt", [ftp_path, all_archive_path]) + tmp_mgr.move_into_place(derived_id_index_tmpfile, derived_path / "1id-index.txt") + tmp_mgr.move_into_place(download_id_index_tmpfile, download_path / "id-index.txt") + + tmp_mgr.move_into_place(id_abstracts_tmpfile, id_path / "1id-abstracts.txt", [ftp_path, all_archive_path]) + tmp_mgr.move_into_place(derived_id_abstracts_tmpfile, derived_path / "1id-abstracts.txt") + tmp_mgr.move_into_place(download_id_abstracts_tmpfile, download_path / "id-abstract.txt") + + tmp_mgr.move_into_place(all_id2_tmpfile, id_path / "all_id2.txt", [ftp_path, all_archive_path]) + tmp_mgr.move_into_place(derived_all_id2_tmpfile, derived_path / "all_id2.txt") diff --git a/ietf/idindex/tests.py b/ietf/idindex/tests.py index f207fa5621..ba6100550d 100644 --- a/ietf/idindex/tests.py +++ b/ietf/idindex/tests.py @@ -3,19 +3,23 @@ import datetime +from unittest import mock from pathlib import Path +from tempfile import TemporaryDirectory from django.conf import settings from django.utils import timezone import debug # pyflakes:ignore -from ietf.doc.factories import WgDraftFactory -from ietf.doc.models import Document, DocAlias, RelatedDocument, State, LastCallDocEvent, NewRevisionDocEvent +from ietf.doc.factories import WgDraftFactory, RfcFactory +from ietf.doc.models import Document, RelatedDocument, State, LastCallDocEvent, NewRevisionDocEvent +from ietf.doc.storage_utils import retrieve_str from ietf.group.factories import GroupFactory from ietf.name.models import DocRelationshipName from ietf.idindex.index import all_id_txt, all_id2_txt, id_index_txt +from ietf.idindex.tasks import idindex_update_task, TempFileManager from ietf.person.factories import PersonFactory, EmailFactory from ietf.utils.test_utils import TestCase @@ -41,7 +45,8 @@ def test_all_id_txt(self): # published draft.set_state(State.objects.get(type="draft", slug="rfc")) - DocAlias.objects.create(name="rfc1234").docs.add(draft) + rfc = RfcFactory(rfc_number=1234) + draft.relateddocument_set.create(relationship_id="became_rfc", target=rfc) txt = all_id_txt() self.assertTrue(draft.name + "-" + draft.rev in txt) @@ -52,8 +57,13 @@ def test_all_id_txt(self): RelatedDocument.objects.create( relationship=DocRelationshipName.objects.get(slug="replaces"), - source=Document.objects.create(type_id="draft", rev="00", name="draft-test-replacement"), - target=draft.docalias.get(name__startswith="draft")) + source=Document.objects.create( + type_id="draft", + rev="00", + name="draft-test-replacement" + ), + target=draft + ) txt = all_id_txt() self.assertTrue(draft.name + "-" + draft.rev in txt) @@ -103,7 +113,8 @@ def get_fields(content): # test RFC draft.set_state(State.objects.get(type="draft", slug="rfc")) - DocAlias.objects.create(name="rfc1234").docs.add(draft) + rfc = RfcFactory(rfc_number=1234) + draft.relateddocument_set.create(relationship_id="became_rfc", target=rfc) t = get_fields(all_id2_txt()) self.assertEqual(t[4], "1234") @@ -111,8 +122,12 @@ def get_fields(content): draft.set_state(State.objects.get(type="draft", slug="repl")) RelatedDocument.objects.create( relationship=DocRelationshipName.objects.get(slug="replaces"), - source=Document.objects.create(type_id="draft", rev="00", name="draft-test-replacement"), - target=draft.docalias.get(name__startswith="draft")) + source=Document.objects.create( + type_id="draft", + rev="00", + name="draft-test-replacement" + ), + target=draft) t = get_fields(all_id2_txt()) self.assertEqual(t[5], "draft-test-replacement") @@ -140,3 +155,58 @@ def test_id_index_txt(self): txt = id_index_txt(with_abstracts=True) self.assertTrue(draft.abstract[:20] in txt) + + +class TaskTests(TestCase): + @mock.patch("ietf.idindex.tasks.all_id_txt") + @mock.patch("ietf.idindex.tasks.all_id2_txt") + @mock.patch("ietf.idindex.tasks.id_index_txt") + @mock.patch.object(TempFileManager, "__enter__") + def test_idindex_update_task( + self, + temp_file_mgr_enter_mock, + id_index_mock, + all_id2_mock, + all_id_mock, + ): + # Replace TempFileManager's __enter__() method with one that returns a mock. + # Pass a spec to the mock so we validate that only actual methods are called. + mgr_mock = mock.Mock(spec=TempFileManager) + temp_file_mgr_enter_mock.return_value = mgr_mock + + idindex_update_task() + + self.assertEqual(all_id_mock.call_count, 1) + self.assertEqual(all_id2_mock.call_count, 1) + self.assertEqual(id_index_mock.call_count, 2) + self.assertEqual(id_index_mock.call_args_list[0], (tuple(), dict())) + self.assertEqual( + id_index_mock.call_args_list[1], + (tuple(), {"with_abstracts": True}), + ) + self.assertEqual(mgr_mock.make_temp_file.call_count, 11) + self.assertEqual(mgr_mock.move_into_place.call_count, 11) + + def test_temp_file_manager(self): + with TemporaryDirectory() as temp_dir: + with TemporaryDirectory() as other_dir: + temp_path = Path(temp_dir) + other_path = Path(other_dir) + with TempFileManager(temp_path) as tfm: + path1 = tfm.make_temp_file("yay") + path2 = tfm.make_temp_file("boo") # do not keep this one + self.assertTrue(path1.exists()) + self.assertTrue(path2.exists()) + dest = temp_path / "yay.txt" + tfm.move_into_place(path1, dest, [other_path]) + # make sure things were cleaned up... + self.assertFalse(path1.exists()) # moved to dest + self.assertFalse(path2.exists()) # left behind + # check destination contents and permissions + self.assertEqual(dest.read_text(), "yay") + self.assertEqual( + retrieve_str("indexes", "yay.txt"), + "yay" + ) + self.assertEqual(dest.stat().st_mode & 0o777, 0o644) + self.assertTrue(dest.samefile(other_path / "yay.txt")) diff --git a/ietf/iesg/.gitignore b/ietf/iesg/.gitignore deleted file mode 100644 index c7013ced9d..0000000000 --- a/ietf/iesg/.gitignore +++ /dev/null @@ -1,2 +0,0 @@ -/*.pyc -/settings_local.py diff --git a/ietf/iesg/admin.py b/ietf/iesg/admin.py index 9ffb352c43..06bbaeb201 100644 --- a/ietf/iesg/admin.py +++ b/ietf/iesg/admin.py @@ -3,7 +3,7 @@ import debug # pyflakes:ignore from ietf.doc.models import TelechatDocEvent -from ietf.iesg.models import TelechatDate, TelechatAgendaItem +from ietf.iesg.models import TelechatDate, TelechatAgendaItem, TelechatAgendaContent class TelechatAgendaItemAdmin(admin.ModelAdmin): pass @@ -22,3 +22,6 @@ def save_model(self, request, obj, form, change): admin.site.register(TelechatDate, TelechatDateAdmin) +class TelechatAgendaContentAdmin(admin.ModelAdmin): + list_display = ('section',) +admin.site.register(TelechatAgendaContent, TelechatAgendaContentAdmin) diff --git a/ietf/iesg/agenda.py b/ietf/iesg/agenda.py index 113c3c5d1d..ace4c9ec40 100644 --- a/ietf/iesg/agenda.py +++ b/ietf/iesg/agenda.py @@ -4,7 +4,6 @@ # utilities for constructing agendas for IESG telechats -import io import datetime from collections import OrderedDict @@ -15,9 +14,9 @@ from ietf.doc.models import Document, LastCallDocEvent, ConsensusDocEvent from ietf.doc.utils_search import fill_in_telechat_date -from ietf.iesg.models import TelechatDate, TelechatAgendaItem +from ietf.iesg.models import TelechatDate, TelechatAgendaItem, TelechatAgendaContent from ietf.review.utils import review_assignments_to_list_for_docs -from ietf.utils.timezone import date_today +from ietf.utils.timezone import date_today, make_aware def get_agenda_date(date=None): if not date: @@ -26,8 +25,9 @@ def get_agenda_date(date=None): except IndexError: return date_today() else: + parsed_date = make_aware(datetime.datetime.strptime(date, "%Y-%m-%d"), settings.TIME_ZONE).date() try: - return TelechatDate.objects.active().get(date=datetime.datetime.strptime(date, "%Y-%m-%d").date()).date + return TelechatDate.objects.active().get(date=parsed_date).date except (ValueError, TelechatDate.DoesNotExist): raise Http404 @@ -66,7 +66,7 @@ def get_doc_section(doc): elif doc.type_id == 'statchg': protocol_action = False for relation in doc.relateddocument_set.filter(relationship__slug__in=('tops','tois','tohist','toinf','tobcp','toexp')): - if relation.relationship_id in ('tops','tois') or relation.target.document.std_level_id in ('std','ds','ps'): + if relation.relationship_id in ('tops','tois') or relation.target.std_level_id in ('std','ds','ps'): protocol_action = True if protocol_action: s = "2.3" @@ -133,26 +133,24 @@ def agenda_sections(): ('4.2', {'title':"WG rechartering"}), ('4.2.1', {'title':"Under evaluation for IETF review", 'docs':[]}), ('4.2.2', {'title':"Proposed for approval", 'docs':[]}), - ('5', {'title':"IAB news we can use"}), + ('5', {'title':"IESG Liaison News"}), ('6', {'title':"Management issues"}), ('7', {'title':"Any Other Business (WG News, New Proposals, etc.)"}), ]) def fill_in_agenda_administrivia(date, sections): - extra_info_files = ( - ("1.1", "roll_call", settings.IESG_ROLL_CALL_FILE), - ("1.3", "minutes", settings.IESG_MINUTES_FILE), - ("1.4", "action_items", settings.IESG_TASK_FILE), - ) + extra_info = ( + ("1.1", "roll_call"), + ("1.3", "minutes"), + ("1.4", "action_items"), + ) - for s, key, filename in extra_info_files: + for s, key in extra_info: try: - with io.open(filename, 'r', encoding='utf-8', errors='replace') as f: - t = f.read().strip() - except IOError: - t = "(Error reading %s)" % filename - - sections[s]["text"] = t + text = TelechatAgendaContent.objects.get(section__slug=key).text + except TelechatAgendaContent.DoesNotExist: + text = "" + sections[s]["text"] = text def fill_in_agenda_docs(date, sections, docs=None): if not docs: @@ -188,7 +186,7 @@ def fill_in_agenda_docs(date, sections, docs=None): doc.review_assignments = review_assignments_for_docs.get(doc.name, []) elif doc.type_id == "conflrev": - doc.conflictdoc = doc.relateddocument_set.get(relationship__slug='conflrev').target.document + doc.conflictdoc = doc.relateddocument_set.get(relationship__slug='conflrev').target elif doc.type_id == "charter": pass @@ -221,4 +219,4 @@ def agenda_data(date=None): fill_in_agenda_docs(date, sections) fill_in_agenda_management_issues(date, sections) - return { 'date': date.isoformat(), 'sections': sections } \ No newline at end of file + return { 'date': date.isoformat(), 'sections': sections } diff --git a/ietf/iesg/factories.py b/ietf/iesg/factories.py index a9c8bed56e..e49b08b583 100644 --- a/ietf/iesg/factories.py +++ b/ietf/iesg/factories.py @@ -4,7 +4,7 @@ import debug # pyflakes:ignore import factory -from ietf.iesg.models import TelechatAgendaItem +from ietf.iesg.models import TelechatAgendaItem, TelechatAgendaContent class IESGMgmtItemFactory(factory.django.DjangoModelFactory): @@ -14,3 +14,10 @@ class Meta: type = 3 text = factory.Faker('paragraph', nb_sentences=3) title = factory.Faker('sentence', nb_words=3) + + +class TelechatAgendaContentFactory(factory.django.DjangoModelFactory): + class Meta: + model = TelechatAgendaContent + + text = factory.Faker('paragraph', nb_sentences=5) diff --git a/ietf/iesg/migrations/0001_initial.py b/ietf/iesg/migrations/0001_initial.py new file mode 100644 index 0000000000..6ceb343d10 --- /dev/null +++ b/ietf/iesg/migrations/0001_initial.py @@ -0,0 +1,55 @@ +# Generated by Django 2.2.28 on 2023-03-20 19:22 + +from typing import List, Tuple +from django.db import migrations, models +import ietf.iesg.models + + +class Migration(migrations.Migration): + + initial = True + + dependencies: List[Tuple[str, str]] = [ + ] + + operations = [ + migrations.CreateModel( + name='Telechat', + fields=[ + ('telechat_id', models.IntegerField(primary_key=True, serialize=False)), + ('telechat_date', models.DateField(blank=True, null=True)), + ('minute_approved', models.IntegerField(blank=True, null=True)), + ('wg_news_txt', models.TextField(blank=True)), + ('iab_news_txt', models.TextField(blank=True)), + ('management_issue', models.TextField(blank=True)), + ('frozen', models.IntegerField(blank=True, null=True)), + ('mi_frozen', models.IntegerField(blank=True, null=True)), + ], + options={ + 'db_table': 'telechat', + }, + ), + migrations.CreateModel( + name='TelechatAgendaItem', + fields=[ + ('id', models.AutoField(db_column='template_id', primary_key=True, serialize=False)), + ('text', models.TextField(blank=True, db_column='template_text')), + ('type', models.IntegerField(choices=[(1, 'Any Other Business (WG News, New Proposals, etc.)'), (2, 'IAB News'), (3, 'Management Item')], db_column='template_type', default=3)), + ('title', models.CharField(db_column='template_title', max_length=255)), + ], + ), + migrations.CreateModel( + name='TelechatDate', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('date', models.DateField(default=ietf.iesg.models.next_telechat_date)), + ], + options={ + 'ordering': ['-date'], + }, + ), + migrations.AddIndex( + model_name='telechatdate', + index=models.Index(fields=['-date'], name='iesg_telech_date_a0e0ed_idx'), + ), + ] diff --git a/ietf/iesg/migrations/0002_telechatagendacontent.py b/ietf/iesg/migrations/0002_telechatagendacontent.py new file mode 100644 index 0000000000..b3f24017a3 --- /dev/null +++ b/ietf/iesg/migrations/0002_telechatagendacontent.py @@ -0,0 +1,23 @@ +# Generated by Django 2.2.28 on 2023-03-10 16:47 + +from django.db import migrations, models +import django.db.models.deletion + + +class Migration(migrations.Migration): + + dependencies = [ + ('name', '0002_telechatagendasectionname'), + ('iesg', '0001_initial'), + ] + + operations = [ + migrations.CreateModel( + name='TelechatAgendaContent', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('text', models.TextField(blank=True)), + ('section', models.ForeignKey(on_delete=django.db.models.deletion.PROTECT, to='name.TelechatAgendaSectionName')), + ], + ), + ] diff --git a/ietf/iesg/migrations/0003_delete_telechat.py b/ietf/iesg/migrations/0003_delete_telechat.py new file mode 100644 index 0000000000..6a09b88555 --- /dev/null +++ b/ietf/iesg/migrations/0003_delete_telechat.py @@ -0,0 +1,16 @@ +# Generated by Django 4.2.13 on 2024-06-21 20:40 + +from django.db import migrations + + +class Migration(migrations.Migration): + + dependencies = [ + ("iesg", "0002_telechatagendacontent"), + ] + + operations = [ + migrations.DeleteModel( + name="Telechat", + ), + ] diff --git a/ietf/secr/console/__init__.py b/ietf/iesg/migrations/__init__.py similarity index 100% rename from ietf/secr/console/__init__.py rename to ietf/iesg/migrations/__init__.py diff --git a/ietf/iesg/models.py b/ietf/iesg/models.py index 6a622e27e6..dcc8a9880b 100644 --- a/ietf/iesg/models.py +++ b/ietf/iesg/models.py @@ -39,6 +39,7 @@ from django.conf import settings from django.db import models +from ietf.name.models import TelechatAgendaSectionName from ietf.utils.timezone import date_today @@ -58,20 +59,6 @@ def __str__(self): type_name = self.TYPE_CHOICES_DICT.get(self.type, str(self.type)) return "%s: %s" % (type_name, self.title or "") -class Telechat(models.Model): - telechat_id = models.IntegerField(primary_key=True) - telechat_date = models.DateField(null=True, blank=True) - minute_approved = models.IntegerField(null=True, blank=True) - wg_news_txt = models.TextField(blank=True) - iab_news_txt = models.TextField(blank=True) - management_issue = models.TextField(blank=True) - frozen = models.IntegerField(null=True, blank=True) - mi_frozen = models.IntegerField(null=True, blank=True) - - class Meta: - db_table = 'telechat' - - def next_telechat_date(): dates = TelechatDate.objects.order_by("-date") if dates: @@ -95,3 +82,11 @@ class Meta: indexes = [ models.Index(fields=['-date',]), ] + + +class TelechatAgendaContent(models.Model): + section = models.ForeignKey(TelechatAgendaSectionName, on_delete=models.PROTECT) + text = models.TextField(blank=True) + + def __str__(self): + return f"{self.section.name} content" diff --git a/ietf/iesg/resources.py b/ietf/iesg/resources.py index a67dbc3a42..c28dcf51d3 100644 --- a/ietf/iesg/resources.py +++ b/ietf/iesg/resources.py @@ -3,13 +3,13 @@ # Autogenerated by the mkresources management command 2014-11-13 23:53 -from ietf.api import ModelResource -from tastypie.constants import ALL +from ietf.api import ModelResource, ToOneField +from tastypie.constants import ALL, ALL_WITH_RELATIONS from tastypie.cache import SimpleCache from ietf import api -from ietf.iesg.models import TelechatDate, Telechat, TelechatAgendaItem +from ietf.iesg.models import TelechatDate, TelechatAgendaItem, TelechatAgendaContent class TelechatDateResource(ModelResource): @@ -17,45 +17,57 @@ class Meta: cache = SimpleCache() queryset = TelechatDate.objects.all() serializer = api.Serializer() - #resource_name = 'telechatdate' - ordering = ['id', ] - filtering = { + # resource_name = 'telechatdate' + ordering = [ + "id", + ] + filtering = { "id": ALL, "date": ALL, } + + api.iesg.register(TelechatDateResource()) -class TelechatResource(ModelResource): - class Meta: - cache = SimpleCache() - queryset = Telechat.objects.all() - serializer = api.Serializer() - #resource_name = 'telechat' - ordering = ['tlechat_id', ] - filtering = { - "telechat_id": ALL, - "telechat_date": ALL, - "minute_approved": ALL, - "wg_news_txt": ALL, - "iab_news_txt": ALL, - "management_issue": ALL, - "frozen": ALL, - "mi_frozen": ALL, - } -api.iesg.register(TelechatResource()) class TelechatAgendaItemResource(ModelResource): class Meta: cache = SimpleCache() queryset = TelechatAgendaItem.objects.all() serializer = api.Serializer() - #resource_name = 'telechatagendaitem' - ordering = ['id', ] - filtering = { + # resource_name = 'telechatagendaitem' + ordering = [ + "id", + ] + filtering = { "id": ALL, "text": ALL, "type": ALL, "title": ALL, } + + api.iesg.register(TelechatAgendaItemResource()) +from ietf.name.resources import TelechatAgendaSectionNameResource + + +class TelechatAgendaContentResource(ModelResource): + section = ToOneField(TelechatAgendaSectionNameResource, "section") + + class Meta: + queryset = TelechatAgendaContent.objects.none() + serializer = api.Serializer() + cache = SimpleCache() + # resource_name = 'telechatagendacontent' + ordering = [ + "id", + ] + filtering = { + "id": ALL, + "text": ALL, + "section": ALL_WITH_RELATIONS, + } + + +api.iesg.register(TelechatAgendaContentResource()) diff --git a/ietf/iesg/tests.py b/ietf/iesg/tests.py index 5ecb4ed9de..e5fbe5da7b 100644 --- a/ietf/iesg/tests.py +++ b/ietf/iesg/tests.py @@ -2,6 +2,7 @@ # -*- coding: utf-8 -*- +from collections import Counter import datetime import io import tarfile @@ -17,20 +18,21 @@ import debug # pyflakes:ignore from ietf.doc.models import DocEvent, BallotPositionDocEvent, TelechatDocEvent -from ietf.doc.models import Document, DocAlias, State, RelatedDocument -from ietf.doc.factories import WgDraftFactory, IndividualDraftFactory, ConflictReviewFactory, BaseDocumentFactory, CharterFactory, WgRfcFactory, IndividualRfcFactory +from ietf.doc.models import Document, State, RelatedDocument +from ietf.doc.factories import BallotDocEventFactory, BallotPositionDocEventFactory, TelechatDocEventFactory, WgDraftFactory, IndividualDraftFactory, ConflictReviewFactory, BaseDocumentFactory, CharterFactory, WgRfcFactory, IndividualRfcFactory from ietf.doc.utils import create_ballot_if_not_open -from ietf.group.factories import RoleFactory, GroupFactory +from ietf.group.factories import RoleFactory, GroupFactory, DatedGroupMilestoneFactory, DatelessGroupMilestoneFactory from ietf.group.models import Group, GroupMilestone, Role -from ietf.iesg.agenda import get_agenda_date, agenda_data -from ietf.iesg.models import TelechatDate -from ietf.name.models import StreamName +from ietf.iesg.agenda import get_agenda_date, agenda_data, fill_in_agenda_administrivia, agenda_sections +from ietf.iesg.models import TelechatDate, TelechatAgendaContent +from ietf.iesg.utils import get_wg_dashboard_info +from ietf.name.models import StreamName, TelechatAgendaSectionName +from ietf.person.factories import PersonFactory from ietf.person.models import Person from ietf.utils.test_utils import TestCase, login_testing_unauthorized, unicontent -from ietf.iesg.factories import IESGMgmtItemFactory +from ietf.iesg.factories import IESGMgmtItemFactory, TelechatAgendaContentFactory from ietf.utils.timezone import date_today, DEADLINE_TZINFO - class IESGTests(TestCase): def test_feed(self): draft = WgDraftFactory(states=[('draft','active'),('draft-iesg','iesg-eva')],ad=Person.objects.get(user__username='ad')) @@ -52,6 +54,15 @@ def test_feed(self): self.assertContains(r, draft.name) self.assertContains(r, escape(pos.balloter.plain_name())) + # Mark draft as replaced + draft.set_state(State.objects.get(type="draft", slug="repl")) + + r = self.client.get(urlreverse("ietf.iesg.views.discusses")) + self.assertEqual(r.status_code, 200) + + self.assertNotContains(r, draft.name) + self.assertNotContains(r, escape(pos.balloter.plain_name())) + def test_milestones_needing_review(self): draft = WgDraftFactory() RoleFactory(name_id='ad',group=draft.group,person=Person.objects.get(user__username='ad')) @@ -71,7 +82,80 @@ def test_milestones_needing_review(self): r = self.client.get(url) self.assertEqual(r.status_code, 200) self.assertNotContains(r, m.desc) + + def test_milestones_needing_review_ordering(self): + dated_group = GroupFactory(uses_milestone_dates=True) + RoleFactory( + name_id='ad', + group=dated_group, + person=Person.objects.get(user__username='ad'), + ) + dated_milestones = [ + DatedGroupMilestoneFactory( + group=dated_group, + state_id="review", + desc="This is the description of one dated group milestone", + ), + DatedGroupMilestoneFactory( + group=dated_group, + state_id="review", + desc="This is the description of another dated group milestone", + ), + ] + dated_milestones[0].due -= datetime.timedelta(days=1) # make this one earlier + dated_milestones[0].save() + + dateless_group = GroupFactory(uses_milestone_dates=False) + RoleFactory( + name_id='ad', + group=dateless_group, + person=Person.objects.get(user__username='ad'), + ) + dateless_milestones = [ + DatelessGroupMilestoneFactory( + group=dateless_group, + state_id="review", + desc="This is the description of one dateless group milestone", + ), + DatelessGroupMilestoneFactory( + group=dateless_group, + state_id="review", + desc="This is the description of another dateless group milestone", + ), + ] + + url = urlreverse("ietf.iesg.views.milestones_needing_review") + self.client.login(username="ad", password="ad+password") + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + pq = PyQuery(r.content) + # check order-by-date + dated_tbody = pq(f'td:contains("{dated_milestones[0].desc}")').closest("tbody") + rows = list(dated_tbody.items("tr")) # keep as pyquery objects + self.assertTrue(rows[0].find('td:first:contains("Last")')) # Last milestone shown first + self.assertFalse(rows[0].find('td:first:contains("Next")')) + self.assertTrue(rows[0].find(f'td:contains("{dated_milestones[1].desc}")')) + self.assertFalse(rows[0].find(f'td:contains("{dated_milestones[0].desc}")')) + + self.assertFalse(rows[1].find('td:first:contains("Last")')) # Last milestone shown first + self.assertTrue(rows[1].find('td:first:contains("Next")')) + self.assertFalse(rows[1].find(f'td:contains("{dated_milestones[1].desc}")')) + self.assertTrue(rows[1].find(f'td:contains("{dated_milestones[0].desc}")')) + + # check order-by-order + dateless_tbody = pq(f'td:contains("{dateless_milestones[0].desc}")').closest("tbody") + rows = list(dateless_tbody.items("tr")) # keep as pyquery objects + self.assertTrue(rows[0].find('td:first:contains("Last")')) # Last milestone shown first + self.assertFalse(rows[0].find('td:first:contains("Next")')) + self.assertTrue(rows[0].find(f'td:contains("{dateless_milestones[1].desc}")')) + self.assertFalse(rows[0].find(f'td:contains("{dateless_milestones[0].desc}")')) + + self.assertFalse(rows[1].find('td:first:contains("Last")')) # Last milestone shown first + self.assertTrue(rows[1].find('td:first:contains("Next")')) + self.assertFalse(rows[1].find(f'td:contains("{dateless_milestones[1].desc}")')) + self.assertTrue(rows[1].find(f'td:contains("{dateless_milestones[0].desc}")')) + def test_review_decisions(self): draft = WgDraftFactory() @@ -96,13 +180,1600 @@ def test_photos(self): ads = Role.objects.filter(group__type='area', group__state='active', name_id='ad') self.assertEqual(len(q('.photo')), ads.count()) + def test_ietf_activity(self): + url = urlreverse("ietf.iesg.views.ietf_activity") + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + + def test_working_groups(self): + # Clean away the wasted built-for-every-test noise + Group.objects.filter(type__in=["wg", "area"]).delete() + + ( + area_summary, + area_totals, + ad_summary, + noad_summary, + ad_totals, + noad_totals, + totals, + wg_summary, + ) = get_wg_dashboard_info() + self.assertEqual(area_summary, []) + self.assertEqual( + area_totals, {"group_count": 0, "doc_count": 0, "page_count": 0} + ) + self.assertEqual(ad_summary, []) + self.assertEqual(noad_summary, []) + self.assertEqual( + ad_totals, + { + "ad_group_count": 0, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + }, + ) + self.assertEqual( + noad_totals, + { + "ad_group_count": 0, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + }, + ) + self.assertEqual( + totals, + { + "group_count": 0, + "doc_count": 0, + "page_count": 0, + "groups_with_docs_count": 0, + }, + ) + self.assertEqual(wg_summary, []) + + # Construct Areas with WGs similar in shape to a real moment of the IETF + + # Note that this test construciton uses the first letter of the wg acronyms + # for convenience to switch on whether groups have documents with assigned ADs. + # (Search for ` if wg_acronym[0] > "g"`) + # There's no other significance to the names of the area directors or the + # acronyms of the areas and groups other than being distinct. Taking the + # values from sets of similar things hopefully helps with debugging the tests. + + areas = {} + for area_acronym in ["red", "orange", "yellow", "green", "blue", "violet"]: + areas[area_acronym] = GroupFactory(type_id="area", acronym=area_acronym) + for ad, area, wgs in [ + ("Alpha", "red", ["bassoon"]), + ("Bravo", "orange", ["celesta"]), + ("Charlie", "orange", ["clarinet", "cymbals"]), + ("Delta", "yellow", ["flute"]), + ("Echo", "yellow", ["glockenspiel"]), + ("Foxtrot", "green", ["gong", "guitar"]), + ("Golf", "green", ["harp"]), + ("Hotel", "blue", ["harpsichord"]), + ("Indigo", "blue", ["oboe", "organ"]), + ("Juliet", "violet", ["piano"]), + ("Kilo", "violet", ["piccolo"]), + ("Lima", "violet", ["saxophone", "tambourine"]), + ]: + p = Person.objects.filter(name=ad).first() or PersonFactory(name=ad) + RoleFactory(group=areas[area], person=p, name_id="ad") + for wg in wgs: + g = GroupFactory(acronym=wg, type_id="wg", parent=areas[area]) + RoleFactory(group=g, person=p, name_id="ad") + + # Some ADs have out of area groups + g = GroupFactory(acronym="timpani", parent=areas["orange"]) + RoleFactory(group=g, person=Person.objects.get(name="Juliet"), name_id="ad") + + ( + area_summary, + area_totals, + ad_summary, + noad_summary, + ad_totals, + noad_totals, + totals, + wg_summary, + ) = get_wg_dashboard_info() + + # checks for the expected result with area sorted by name + self.assertEqual( + area_summary, + [ + { + "area": "blue", + "groups_in_area": 3, + "groups_with_docs": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "area": "green", + "groups_in_area": 3, + "groups_with_docs": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "area": "orange", + "groups_in_area": 4, + "groups_with_docs": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "area": "red", + "groups_in_area": 1, + "groups_with_docs": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "area": "violet", + "groups_in_area": 4, + "groups_with_docs": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "area": "yellow", + "groups_in_area": 2, + "groups_with_docs": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + ], + ) + self.assertEqual( + area_totals, {"group_count": 0, "doc_count": 0, "page_count": 0} + ) + self.assertEqual( + ad_summary, + [ + { + "ad": "Alpha", + "area": "red", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Bravo", + "area": "orange", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Charlie", + "area": "orange", + "ad_group_count": 2, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Delta", + "area": "yellow", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Echo", + "area": "yellow", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Foxtrot", + "area": "green", + "ad_group_count": 2, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Golf", + "area": "green", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Hotel", + "area": "blue", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Indigo", + "area": "blue", + "ad_group_count": 2, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Juliet", + "area": "orange", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Juliet", + "area": "violet", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Kilo", + "area": "violet", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Lima", + "area": "violet", + "ad_group_count": 2, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + ], + ) + self.assertEqual( + noad_summary, + [ + { + "ad": "Alpha", + "area": "red", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Bravo", + "area": "orange", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Charlie", + "area": "orange", + "ad_group_count": 2, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Delta", + "area": "yellow", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Echo", + "area": "yellow", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Foxtrot", + "area": "green", + "ad_group_count": 2, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Golf", + "area": "green", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Hotel", + "area": "blue", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Indigo", + "area": "blue", + "ad_group_count": 2, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Juliet", + "area": "orange", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Juliet", + "area": "violet", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Kilo", + "area": "violet", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + { + "ad": "Lima", + "area": "violet", + "ad_group_count": 2, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0, + "doc_percent": 0, + "page_percent": 0, + }, + ], + ) + self.assertEqual( + ad_totals, + { + "ad_group_count": 17, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + }, + ) + self.assertEqual( + noad_totals, + { + "ad_group_count": 17, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + }, + ) + self.assertEqual( + totals, + { + "group_count": 17, + "doc_count": 0, + "page_count": 0, + "groups_with_docs_count": 0, + }, + ) + self.assertEqual( + wg_summary, + [ + { + "wg": "bassoon", + "area": "red", + "ad": "Alpha", + "doc_count": 0, + "page_count": 0, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "celesta", + "area": "orange", + "ad": "Bravo", + "doc_count": 0, + "page_count": 0, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "clarinet", + "area": "orange", + "ad": "Charlie", + "doc_count": 0, + "page_count": 0, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "cymbals", + "area": "orange", + "ad": "Charlie", + "doc_count": 0, + "page_count": 0, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "flute", + "area": "yellow", + "ad": "Delta", + "doc_count": 0, + "page_count": 0, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "glockenspiel", + "area": "yellow", + "ad": "Echo", + "doc_count": 0, + "page_count": 0, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "gong", + "area": "green", + "ad": "Foxtrot", + "doc_count": 0, + "page_count": 0, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "guitar", + "area": "green", + "ad": "Foxtrot", + "doc_count": 0, + "page_count": 0, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "harp", + "area": "green", + "ad": "Golf", + "doc_count": 0, + "page_count": 0, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "harpsichord", + "area": "blue", + "ad": "Hotel", + "doc_count": 0, + "page_count": 0, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "oboe", + "area": "blue", + "ad": "Indigo", + "doc_count": 0, + "page_count": 0, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "organ", + "area": "blue", + "ad": "Indigo", + "doc_count": 0, + "page_count": 0, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "piano", + "area": "violet", + "ad": "Juliet", + "doc_count": 0, + "page_count": 0, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "piccolo", + "area": "violet", + "ad": "Kilo", + "doc_count": 0, + "page_count": 0, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "saxophone", + "area": "violet", + "ad": "Lima", + "doc_count": 0, + "page_count": 0, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "tambourine", + "area": "violet", + "ad": "Lima", + "doc_count": 0, + "page_count": 0, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "timpani", + "area": "orange", + "ad": "Juliet", + "doc_count": 0, + "page_count": 0, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + ], + ) + + # As seen above, all doc and page counts are currently 0 + + # We'll give a group a document but not assign it to its AD + WgDraftFactory( + group=Group.objects.get(acronym="saxophone"), pages=len("saxophone") + ) + ( + area_summary, + area_totals, + ad_summary, + noad_summary, + ad_totals, + noad_totals, + totals, + wg_summary, + ) = get_wg_dashboard_info() + count_violet_dicts = 0 + for d in area_summary: + if d["area"] == "violet": + count_violet_dicts += 1 + self.assertEqual(d["groups_with_docs"], 1) + self.assertEqual(d["doc_count"], 1) + self.assertEqual(d["page_count"], 9) + self.assertEqual(d["group_percent"], 100.0) + self.assertEqual(d["doc_percent"], 100.0) + self.assertEqual(d["page_percent"], 100.0) + else: + self.assertEqual(d["groups_with_docs"], 0) + self.assertEqual(d["doc_count"], 0) + self.assertEqual(d["page_count"], 0) + self.assertEqual(d["group_percent"], 0) + self.assertEqual(d["doc_percent"], 0) + self.assertEqual(d["page_percent"], 0) + self.assertEqual(count_violet_dicts, 1) + + self.assertEqual( + area_totals, {"group_count": 1, "doc_count": 1, "page_count": 9} + ) + + # No AD has this document, even though it's in Lima's group + count_lima_dicts = 0 + for d in ad_summary: + if d["ad"] == "Lima": + count_lima_dicts += 1 + self.assertEqual(d["doc_group_count"], 0) + self.assertEqual(d["doc_count"], 0) + self.assertEqual(d["page_count"], 0) + self.assertEqual(d["group_percent"], 0) + self.assertEqual(d["doc_percent"], 0) + self.assertEqual(d["page_percent"], 0) + self.assertEqual(count_lima_dicts, 1) + + # It's in Lima's group, so normally it will eventually land on Lima + count_lima_dicts = 0 + for d in noad_summary: + if d["ad"] == "Lima": + count_lima_dicts += 1 + self.assertEqual(d["doc_group_count"], 1) + self.assertEqual(d["doc_count"], 1) + self.assertEqual(d["page_count"], 9) + self.assertEqual(d["group_percent"], 100.0) + self.assertEqual(d["doc_percent"], 100.0) + self.assertEqual(d["page_percent"], 100.0) + else: + self.assertEqual(d["doc_group_count"], 0) + self.assertEqual(d["doc_count"], 0) + self.assertEqual(d["page_count"], 0) + self.assertEqual(d["group_percent"], 0) + self.assertEqual(d["doc_percent"], 0) + self.assertEqual(d["page_percent"], 0) + self.assertEqual(count_lima_dicts, 1) + + self.assertEqual( + ad_totals, + { + "ad_group_count": 17, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + }, + ) + self.assertEqual( + noad_totals, + { + "ad_group_count": 17, + "doc_group_count": 1, + "doc_count": 1, + "page_count": 9, + }, + ) + self.assertEqual( + totals, + { + "group_count": 17, + "doc_count": 1, + "page_count": 9, + "groups_with_docs_count": 1, + }, + ) + + count_sax_dicts = 0 + for d in wg_summary: + if d["wg"] == "saxophone": + count_sax_dicts += 1 + self.assertEqual(d["doc_count"], 1) + self.assertEqual(d["page_count"], 9) + else: + self.assertEqual(d["doc_count"], 0) + self.assertEqual(d["page_count"], 0) + self.assertEqual(count_sax_dicts, 1) + + # Assign that doc to Lima + self.assertEqual(Document.objects.count(), 1) + Document.objects.all().update(ad=Person.objects.get(name="Lima")) + ( + area_summary, + area_totals, + ad_summary, + noad_summary, + ad_totals, + noad_totals, + totals, + wg_summary, + ) = get_wg_dashboard_info() + count_violet_dicts = 0 + for d in area_summary: + if d["area"] == "violet": + count_violet_dicts += 1 + self.assertEqual(d["groups_with_docs"], 1) + self.assertEqual(d["doc_count"], 1) + self.assertEqual(d["page_count"], 9) + self.assertEqual(d["group_percent"], 100.0) + self.assertEqual(d["doc_percent"], 100.0) + self.assertEqual(d["page_percent"], 100.0) + else: + self.assertEqual(d["groups_with_docs"], 0) + self.assertEqual(d["doc_count"], 0) + self.assertEqual(d["page_count"], 0) + self.assertEqual(d["group_percent"], 0) + self.assertEqual(d["doc_percent"], 0) + self.assertEqual(d["page_percent"], 0) + self.assertEqual(count_violet_dicts, 1) + + self.assertEqual( + area_totals, {"group_count": 1, "doc_count": 1, "page_count": 9} + ) + + # This time it will show up as a doc assigned to Lima + count_lima_dicts = 0 + for d in ad_summary: + if d["ad"] == "Lima": + count_lima_dicts += 1 + self.assertEqual(d["doc_group_count"], 1) + self.assertEqual(d["doc_count"], 1) + self.assertEqual(d["page_count"], 9) + self.assertEqual(d["group_percent"], 100.0) + self.assertEqual(d["doc_percent"], 100.0) + self.assertEqual(d["page_percent"], 100.0) + else: + self.assertEqual(d["doc_group_count"], 0) + self.assertEqual(d["doc_count"], 0) + self.assertEqual(d["page_count"], 0) + self.assertEqual(d["group_percent"], 0) + self.assertEqual(d["doc_percent"], 0) + self.assertEqual(d["page_percent"], 0) + self.assertEqual(count_lima_dicts, 1) + + # and there will be no noad documents + count_lima_dicts = 0 + for d in noad_summary: + if d["ad"] == "Lima": + count_lima_dicts += 1 + self.assertEqual(d["doc_group_count"], 0) + self.assertEqual(d["doc_count"], 0) + self.assertEqual(d["page_count"], 0) + self.assertEqual(d["group_percent"], 0) + self.assertEqual(d["doc_percent"], 0) + self.assertEqual(d["page_percent"], 0) + self.assertEqual(count_lima_dicts, 1) + + self.assertEqual( + ad_totals, + { + "ad_group_count": 17, + "doc_group_count": 1, + "doc_count": 1, + "page_count": 9, + }, + ) + self.assertEqual( + noad_totals, + { + "ad_group_count": 17, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + }, + ) + self.assertEqual( + totals, + { + "group_count": 17, + "doc_count": 1, + "page_count": 9, + "groups_with_docs_count": 1, + }, + ) + + count_sax_dicts = 0 + for d in wg_summary: + if d["wg"] == "saxophone": + count_sax_dicts += 1 + self.assertEqual(d["doc_count"], 1) + self.assertEqual(d["page_count"], 9) + else: + self.assertEqual(d["doc_count"], 0) + self.assertEqual(d["page_count"], 0) + self.assertEqual(count_sax_dicts, 1) + + # Now give Lima a document in a group that's not in their area: + WgDraftFactory( + group=Group.objects.get(acronym="gong"), + pages=len("gong"), + ad=Person.objects.get(name="Lima"), + ) + ( + area_summary, + area_totals, + ad_summary, + noad_summary, + ad_totals, + noad_totals, + totals, + wg_summary, + ) = get_wg_dashboard_info() + seen_dicts = Counter([d["area"] for d in area_summary]) + for d in areas: + self.assertEqual(seen_dicts[area], 1 if area in ["violet", "green"] else 0) + for d in area_summary: + if d["area"] in ["violet", "green"]: + self.assertEqual(d["doc_count"], 1) + self.assertEqual(d["page_count"], 9 if d["area"] == "violet" else 4) + self.assertEqual(d["group_percent"], 50) + self.assertEqual(d["doc_percent"], 50) + self.assertEqual( + d["page_percent"], + 100 * 9 / 13 if d["area"] == "violet" else 100 * 4 / 13, + ) + else: + self.assertEqual(d["doc_count"], 0) + self.assertEqual(d["page_count"], 0) + self.assertEqual(d["group_percent"], 0) + self.assertEqual(d["doc_percent"], 0) + self.assertEqual(d["page_percent"], 0) + + self.assertEqual( + area_totals, {"group_count": 2, "doc_count": 2, "page_count": 13} + ) + + for d in ad_summary: + if d["ad"] == "Lima": + self.assertEqual(d["doc_group_count"], 1) + self.assertEqual(d["doc_count"], 1) + self.assertEqual(d["page_count"], 9 if d["area"] == "violet" else 4) + self.assertEqual(d["group_percent"], 50) + self.assertEqual(d["doc_percent"], 50) + self.assertEqual( + d["page_percent"], + 100 * 9 / 13 if d["area"] == "violet" else 100 * 4 / 13, + ) + else: + self.assertEqual(d["doc_group_count"], 0) + self.assertEqual( + d["doc_count"], 0 + ) # Note in particular this is 0 for Foxtrot + self.assertEqual(d["page_count"], 0) + self.assertEqual(d["group_percent"], 0) + self.assertEqual(d["doc_percent"], 0) + self.assertEqual(d["page_percent"], 0) + + for d in wg_summary: + if d["wg"] == "gong": + # Lima's doc in gong above counts at the dict for gong even though the ad reported there is Foxtrot. + self.assertEqual( + d, + { + "wg": "gong", + "area": "green", + "ad": "Foxtrot", + "doc_count": 1, + "page_count": 4, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + ) + elif d["ad"] == "Lima": + self.assertEqual( + d["area"], "violet" + ) # The out of area assignment is not reflected in the wg_summary at all. + + # Now pile on a lot of documents + for wg_acronym in [ + "bassoon", + "celesta", + "clarinet", + "cymbals", + "flute", + "glockenspiel", + "gong", + "guitar", + "harp", + "harpsichord", + "oboe", + "organ", + "piano", + "piccolo", + "saxophone", + "tambourine", + "timpani", + ]: + if wg_acronym in ["bassoon", "celesta"]: + continue # Those WGs have no docs + # The rest have a doc that's not assigned to any ad + WgDraftFactory( + group=Group.objects.get(acronym=wg_acronym), pages=len(wg_acronym) + ) + if wg_acronym[0] > "g": + # Some have a doc assigned to the responsible ad + WgDraftFactory( + group=Group.objects.get(acronym=wg_acronym), + pages=len(wg_acronym), + ad=Role.objects.get(name_id="ad", group__acronym=wg_acronym).person, + ) + # The other AD for an area might be covering a doc + WgDraftFactory( + group=Group.objects.get(acronym="saxophone"), + pages=len("saxophone"), + ad=Person.objects.get(name="Juliet"), + ) + # An Ad not associated with the group or the area is responsible for a doc + WgDraftFactory( + group=Group.objects.get(acronym="bassoon"), + pages=len("bassoon"), + ad=Person.objects.get(name="Juliet"), + ) + + ( + area_summary, + area_totals, + ad_summary, + noad_summary, + ad_totals, + noad_totals, + totals, + wg_summary, + ) = get_wg_dashboard_info() + + self.assertEqual( + area_summary, + [ + { + "area": "blue", + "groups_in_area": 3, + "groups_with_docs": 3, + "doc_count": 6, + "page_count": 40, + "group_percent": 18.75, + "doc_percent": 21.428571428571427, + "page_percent": 20.51282051282051, + }, + { + "area": "green", + "groups_in_area": 3, + "groups_with_docs": 3, + "doc_count": 5, + "page_count": 22, + "group_percent": 18.75, + "doc_percent": 17.857142857142858, + "page_percent": 11.282051282051283, + }, + { + "area": "orange", + "groups_in_area": 4, + "groups_with_docs": 3, + "doc_count": 4, + "page_count": 29, + "group_percent": 18.75, + "doc_percent": 14.285714285714285, + "page_percent": 14.871794871794872, + }, + { + "area": "red", + "groups_in_area": 1, + "groups_with_docs": 1, + "doc_count": 1, + "page_count": 7, + "group_percent": 6.25, + "doc_percent": 3.571428571428571, + "page_percent": 3.5897435897435894, + }, + { + "area": "violet", + "groups_in_area": 4, + "groups_with_docs": 4, + "doc_count": 10, + "page_count": 80, + "group_percent": 25.0, + "doc_percent": 35.714285714285715, + "page_percent": 41.02564102564102, + }, + { + "area": "yellow", + "groups_in_area": 2, + "groups_with_docs": 2, + "doc_count": 2, + "page_count": 17, + "group_percent": 12.5, + "doc_percent": 7.142857142857142, + "page_percent": 8.717948717948717, + }, + ], + ) + self.assertEqual( + area_totals, {"group_count": 16, "doc_count": 28, "page_count": 195} + ) + self.assertEqual( + ad_summary, + [ + { + "ad": "Alpha", + "area": "red", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0.0, + "doc_percent": 0.0, + "page_percent": 0.0, + }, + { + "ad": "Bravo", + "area": "orange", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0.0, + "doc_percent": 0.0, + "page_percent": 0.0, + }, + { + "ad": "Charlie", + "area": "orange", + "ad_group_count": 2, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0.0, + "doc_percent": 0.0, + "page_percent": 0.0, + }, + { + "ad": "Delta", + "area": "yellow", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0.0, + "doc_percent": 0.0, + "page_percent": 0.0, + }, + { + "ad": "Echo", + "area": "yellow", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0.0, + "doc_percent": 0.0, + "page_percent": 0.0, + }, + { + "ad": "Foxtrot", + "area": "green", + "ad_group_count": 2, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0.0, + "doc_percent": 0.0, + "page_percent": 0.0, + }, + { + "ad": "Golf", + "area": "green", + "ad_group_count": 1, + "doc_group_count": 1, + "doc_count": 1, + "page_count": 4, + "group_percent": 8.333333333333332, + "doc_percent": 7.6923076923076925, + "page_percent": 4.395604395604396, + }, + { + "ad": "Hotel", + "area": "blue", + "ad_group_count": 1, + "doc_group_count": 1, + "doc_count": 1, + "page_count": 11, + "group_percent": 8.333333333333332, + "doc_percent": 7.6923076923076925, + "page_percent": 12.087912087912088, + }, + { + "ad": "Indigo", + "area": "blue", + "ad_group_count": 2, + "doc_group_count": 2, + "doc_count": 2, + "page_count": 9, + "group_percent": 16.666666666666664, + "doc_percent": 15.384615384615385, + "page_percent": 9.89010989010989, + }, + { + "ad": "Juliet", + "area": "orange", + "ad_group_count": 1, + "doc_group_count": 1, + "doc_count": 1, + "page_count": 7, + "group_percent": 8.333333333333332, + "doc_percent": 7.6923076923076925, + "page_percent": 7.6923076923076925, + }, + { + "ad": "Juliet", + "area": "red", + "ad_group_count": 0, + "doc_group_count": 1, + "doc_count": 1, + "page_count": 7, + "group_percent": 8.333333333333332, + "doc_percent": 7.6923076923076925, + "page_percent": 7.6923076923076925, + }, + { + "ad": "Juliet", + "area": "violet", + "ad_group_count": 1, + "doc_group_count": 2, + "doc_count": 2, + "page_count": 14, + "group_percent": 16.666666666666664, + "doc_percent": 15.384615384615385, + "page_percent": 15.384615384615385, + }, + { + "ad": "Kilo", + "area": "violet", + "ad_group_count": 1, + "doc_group_count": 1, + "doc_count": 1, + "page_count": 7, + "group_percent": 8.333333333333332, + "doc_percent": 7.6923076923076925, + "page_percent": 7.6923076923076925, + }, + { + "ad": "Lima", + "area": "green", + "ad_group_count": 0, + "doc_group_count": 1, + "doc_count": 1, + "page_count": 4, + "group_percent": 8.333333333333332, + "doc_percent": 7.6923076923076925, + "page_percent": 4.395604395604396, + }, + { + "ad": "Lima", + "area": "violet", + "ad_group_count": 2, + "doc_group_count": 2, + "doc_count": 3, + "page_count": 28, + "group_percent": 16.666666666666664, + "doc_percent": 23.076923076923077, + "page_percent": 30.76923076923077, + }, + ], + ) + self.assertEqual( + noad_summary, + [ + { + "ad": "Alpha", + "area": "red", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0.0, + "doc_percent": 0.0, + "page_percent": 0.0, + }, + { + "ad": "Bravo", + "area": "orange", + "ad_group_count": 1, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + "group_percent": 0.0, + "doc_percent": 0.0, + "page_percent": 0.0, + }, + { + "ad": "Charlie", + "area": "orange", + "ad_group_count": 2, + "doc_group_count": 2, + "doc_count": 2, + "page_count": 15, + "group_percent": 13.333333333333334, + "doc_percent": 13.333333333333334, + "page_percent": 14.423076923076922, + }, + { + "ad": "Delta", + "area": "yellow", + "ad_group_count": 1, + "doc_group_count": 1, + "doc_count": 1, + "page_count": 5, + "group_percent": 6.666666666666667, + "doc_percent": 6.666666666666667, + "page_percent": 4.807692307692308, + }, + { + "ad": "Echo", + "area": "yellow", + "ad_group_count": 1, + "doc_group_count": 1, + "doc_count": 1, + "page_count": 12, + "group_percent": 6.666666666666667, + "doc_percent": 6.666666666666667, + "page_percent": 11.538461538461538, + }, + { + "ad": "Foxtrot", + "area": "green", + "ad_group_count": 2, + "doc_group_count": 2, + "doc_count": 2, + "page_count": 10, + "group_percent": 13.333333333333334, + "doc_percent": 13.333333333333334, + "page_percent": 9.615384615384617, + }, + { + "ad": "Golf", + "area": "green", + "ad_group_count": 1, + "doc_group_count": 1, + "doc_count": 1, + "page_count": 4, + "group_percent": 6.666666666666667, + "doc_percent": 6.666666666666667, + "page_percent": 3.8461538461538463, + }, + { + "ad": "Hotel", + "area": "blue", + "ad_group_count": 1, + "doc_group_count": 1, + "doc_count": 1, + "page_count": 11, + "group_percent": 6.666666666666667, + "doc_percent": 6.666666666666667, + "page_percent": 10.576923076923077, + }, + { + "ad": "Indigo", + "area": "blue", + "ad_group_count": 2, + "doc_group_count": 2, + "doc_count": 2, + "page_count": 9, + "group_percent": 13.333333333333334, + "doc_percent": 13.333333333333334, + "page_percent": 8.653846153846153, + }, + { + "ad": "Juliet", + "area": "orange", + "ad_group_count": 1, + "doc_group_count": 1, + "doc_count": 1, + "page_count": 7, + "group_percent": 6.666666666666667, + "doc_percent": 6.666666666666667, + "page_percent": 6.730769230769231, + }, + { + "ad": "Juliet", + "area": "violet", + "ad_group_count": 1, + "doc_group_count": 1, + "doc_count": 1, + "page_count": 5, + "group_percent": 6.666666666666667, + "doc_percent": 6.666666666666667, + "page_percent": 4.807692307692308, + }, + { + "ad": "Kilo", + "area": "violet", + "ad_group_count": 1, + "doc_group_count": 1, + "doc_count": 1, + "page_count": 7, + "group_percent": 6.666666666666667, + "doc_percent": 6.666666666666667, + "page_percent": 6.730769230769231, + }, + { + "ad": "Lima", + "area": "violet", + "ad_group_count": 2, + "doc_group_count": 2, + "doc_count": 2, + "page_count": 19, + "group_percent": 13.333333333333334, + "doc_percent": 13.333333333333334, + "page_percent": 18.269230769230766, + }, + ], + ) + self.assertEqual( + ad_totals, + { + "ad_group_count": 17, + "doc_group_count": 12, + "doc_count": 13, + "page_count": 91, + }, + ) + self.assertEqual( + noad_totals, + { + "ad_group_count": 17, + "doc_group_count": 15, + "doc_count": 15, + "page_count": 104, + }, + ) + self.assertEqual( + totals, + { + "group_count": 17, + "doc_count": 28, + "page_count": 195, + "groups_with_docs_count": 16, + }, + ) + self.assertEqual( + wg_summary, + [ + { + "wg": "bassoon", + "area": "red", + "ad": "Alpha", + "doc_count": 1, + "page_count": 7, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "celesta", + "area": "orange", + "ad": "Bravo", + "doc_count": 0, + "page_count": 0, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "clarinet", + "area": "orange", + "ad": "Charlie", + "doc_count": 1, + "page_count": 8, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "cymbals", + "area": "orange", + "ad": "Charlie", + "doc_count": 1, + "page_count": 7, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "flute", + "area": "yellow", + "ad": "Delta", + "doc_count": 1, + "page_count": 5, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "glockenspiel", + "area": "yellow", + "ad": "Echo", + "doc_count": 1, + "page_count": 12, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "gong", + "area": "green", + "ad": "Foxtrot", + "doc_count": 2, + "page_count": 8, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "guitar", + "area": "green", + "ad": "Foxtrot", + "doc_count": 1, + "page_count": 6, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "harp", + "area": "green", + "ad": "Golf", + "doc_count": 2, + "page_count": 8, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "harpsichord", + "area": "blue", + "ad": "Hotel", + "doc_count": 2, + "page_count": 22, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "oboe", + "area": "blue", + "ad": "Indigo", + "doc_count": 2, + "page_count": 8, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "organ", + "area": "blue", + "ad": "Indigo", + "doc_count": 2, + "page_count": 10, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "piano", + "area": "violet", + "ad": "Juliet", + "doc_count": 2, + "page_count": 10, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "piccolo", + "area": "violet", + "ad": "Kilo", + "doc_count": 2, + "page_count": 14, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "saxophone", + "area": "violet", + "ad": "Lima", + "doc_count": 4, + "page_count": 36, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "tambourine", + "area": "violet", + "ad": "Lima", + "doc_count": 2, + "page_count": 20, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + { + "wg": "timpani", + "area": "orange", + "ad": "Juliet", + "doc_count": 2, + "page_count": 14, + "rfc_count": 0, + "recent_rfc_count": 0, + }, + ], + ) + + # Make sure the view doesn't _crash_ - the template is a dead-simple rendering of the dicts, but this test doesn't prove that + url = urlreverse("ietf.iesg.views.working_groups") + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + + class IESGAgendaTests(TestCase): def setUp(self): super().setUp() mars = GroupFactory(acronym='mars',parent=Group.objects.get(acronym='farfut')) wgdraft = WgDraftFactory(name='draft-ietf-mars-test', group=mars, intended_std_level_id='ps') - rfc = IndividualRfcFactory.create(stream_id='irtf', other_aliases=['rfc6666',], states=[('draft','rfc'),('draft-iesg','pub')], std_level_id='inf', ) - wgdraft.relateddocument_set.create(target=rfc.docalias.get(name='rfc6666'), relationship_id='refnorm') + rfc = IndividualRfcFactory.create(stream_id='irtf', rfc_number=6666, std_level_id='inf', ) + wgdraft.relateddocument_set.create(target=rfc, relationship_id='refnorm') ise_draft = IndividualDraftFactory(name='draft-imaginary-independent-submission') ise_draft.stream = StreamName.objects.get(slug="ise") ise_draft.save_with_history([DocEvent(doc=ise_draft, rev=ise_draft.rev, type="changed_stream", by=Person.objects.get(user__username="secretary"), desc="Test")]) @@ -135,6 +1806,16 @@ def setUp(self): for i in range(0, 10): self.mgmt_items.append(IESGMgmtItemFactory()) + def test_fill_in_agenda_administrivia(self): + roll_call = TelechatAgendaContentFactory(section_id='roll_call') + minutes = TelechatAgendaContentFactory(section_id='minutes') + action_items = TelechatAgendaContentFactory(section_id='action_items') + sections = agenda_sections() + fill_in_agenda_administrivia(None, sections) # n.b., date parameter is unused at present + self.assertIn(roll_call.text, sections["1.1"]["text"]) + self.assertIn(minutes.text, sections["1.3"]["text"]) + self.assertIn(action_items.text, sections["1.4"]["text"]) + def test_fill_in_agenda_docs(self): draft = self.telechat_docs["ietf_draft"] statchg = self.telechat_docs["statchg"] @@ -222,7 +1903,7 @@ def test_fill_in_agenda_docs(self): relation = RelatedDocument.objects.create( source=statchg, - target=DocAlias.objects.filter(name__startswith='rfc', docs__std_level="ps")[0], + target=Document.objects.filter(type_id="rfc", std_level="ps").first(), relationship_id="tohist") statchg.group = Group.objects.get(acronym="mars") @@ -240,7 +1921,7 @@ def test_fill_in_agenda_docs(self): self.assertTrue(statchg in agenda_data(date_str)["sections"]["2.3.3"]["docs"]) # 3.3 document status changes - relation.target = DocAlias.objects.filter(name__startswith='rfc', docs__std_level="inf")[0] + relation.target = Document.objects.filter(type_id="rfc", std_level="inf").first() relation.save() statchg.group = Group.objects.get(acronym="mars") @@ -331,9 +2012,14 @@ def test_agenda_json(self): self.assertTrue(r.json()) def test_agenda(self): + action_items = TelechatAgendaContentFactory(section_id='action_items') r = self.client.get(urlreverse("ietf.iesg.views.agenda")) self.assertEqual(r.status_code, 200) + self.assertContains(r, action_items.text) + + q = PyQuery(r.content) + for k, d in self.telechat_docs.items(): if d.type_id == "charter": self.assertContains(r, d.group.name, msg_prefix="%s '%s' not in response" % (k, d.group.name)) @@ -342,6 +2028,18 @@ def test_agenda(self): self.assertContains(r, d.name, msg_prefix="%s '%s' not in response" % (k, d.name)) self.assertContains(r, d.title, msg_prefix="%s '%s' title not in response" % (k, d.title)) + if d.type_id in ["charter", "draft"]: + if d.group.parent is None: + continue + wg_url = urlreverse("ietf.group.views.active_groups", kwargs=dict(group_type="wg")) + href = f"{wg_url}#{d.group.parent.acronym.upper()}" + texts = [elem.text.strip() for elem in q(f'a[href="{href}"]')] + self.assertGreater(len(texts), 0) + if d.type_id == "charter": + self.assertTrue(any(t == d.group.parent.acronym.upper() for t in texts)) + elif d.type_id == "draft": + self.assertTrue(any(t == f"({d.group.parent.acronym.upper()})" for t in texts)) + for i, mi in enumerate(self.mgmt_items, start=1): s = "6." + str(i) self.assertContains(r, s, msg_prefix="Section '%s' not in response" % s) @@ -350,6 +2048,29 @@ def test_agenda(self): # Make sure the sort places 6.9 before 6.10 self.assertLess(r.content.find(b"6.9"), r.content.find(b"6.10")) + def test_agenda_restricted_sections(self): + r = self.client.get(urlreverse("ietf.iesg.views.agenda")) + # not logged in + for section_id in ("roll_call", "minutes"): + self.assertNotContains( + r, urlreverse("ietf.iesg.views.telechat_agenda_content_view", kwargs={"section": section_id}) + ) + + self.client.login(username="plain", password="plain+password") + for section_id in ("roll_call", "minutes"): + self.assertNotContains( + r, urlreverse("ietf.iesg.views.telechat_agenda_content_view", kwargs={"section": section_id}) + ) + + for username in ("ad", "secretary", "iab chair"): + self.client.login(username=username, password=f"{username}+password") + r = self.client.get(urlreverse("ietf.iesg.views.agenda")) + self.assertEqual(r.status_code, 200) + for section_id in ("roll_call", "minutes"): + self.assertContains( + r, urlreverse("ietf.iesg.views.telechat_agenda_content_view", kwargs={"section": section_id}) + ) + def test_agenda_txt(self): r = self.client.get(urlreverse("ietf.iesg.views.agenda_txt")) self.assertEqual(r.status_code, 200) @@ -415,12 +2136,13 @@ def test_agenda_documents_txt(self): def test_agenda_documents(self): url = urlreverse("ietf.iesg.views.agenda_documents") r = self.client.get(url) + self.assertEqual(r.status_code, 200) for k, d in self.telechat_docs.items(): self.assertContains(r, d.name, msg_prefix="%s '%s' not in response" % (k, d.name, )) - self.assertContains(r, d.title, msg_prefix="%s '%s' title not in response" % (k, d.title, )) - + self.assertContains(r, d.title, msg_prefix="%s '%s' not in response" % (k, d.title, )) + def test_past_documents(self): url = urlreverse("ietf.iesg.views.past_documents") # We haven't put any documents on past telechats, so this should be empty @@ -495,6 +2217,66 @@ def test_admin_change(self): draft = Document.objects.get(name="draft-ietf-mars-test") self.assertEqual(draft.telechat_date(),today) +class IESGAgendaTelechatPagesTests(TestCase): + def setUp(self): + super().setUp() + # make_immutable_test_data made a set of future telechats - only need one + # We'll take the "next" one + self.telechat_date = get_agenda_date() + # make_immutable_test_data made and area with only one ad - give it another + ad = Person.objects.get(user__username="ad") + adrole = Role.objects.get(person=ad, name="ad") + ad2 = RoleFactory(group=adrole.group, name_id="ad").person + self.ads=[ad,ad2] + + # Make some drafts + docs = [ + WgDraftFactory(pages=2, states=[('draft-iesg','iesg-eva'),]), + IndividualDraftFactory(pages=20, states=[('draft-iesg','iesg-eva'),]), + WgDraftFactory(pages=200, states=[('draft-iesg','iesg-eva'),]), + ] + # Put them on the telechat + for doc in docs: + TelechatDocEventFactory(doc=doc, telechat_date=self.telechat_date) + # Give them ballots + ballots = [BallotDocEventFactory(doc=doc) for doc in docs] + + # Give the "ad" Area-Director a discuss on one + BallotPositionDocEventFactory(balloter=ad, doc=docs[0], pos_id="discuss", ballot=ballots[0]) + # and a "norecord" position on another + BallotPositionDocEventFactory(balloter=ad, doc=docs[1], pos_id="norecord", ballot=ballots[1]) + # Now "ad" should have 220 pages left to ballot on. + # Every other ad should have 222 pages left to ballot on. + + def test_ad_pages_left_to_ballot_on(self): + url = urlreverse("ietf.iesg.views.agenda_documents") + + # A non-AD user won't get "pages left" + response = self.client.get(url) + telechat = response.context["telechats"][0] + self.assertEqual(telechat["date"], self.telechat_date) + self.assertEqual(telechat["ad_pages_left_to_ballot_on"],0) + self.assertNotContains(response,"pages left to ballot on") + + username=self.ads[0].user.username + self.assertTrue(self.client.login(username=username, password=f"{username}+password")) + + response = self.client.get(url) + telechat = response.context["telechats"][0] + self.assertEqual(telechat["ad_pages_left_to_ballot_on"],220) + self.assertContains(response,"220 pages left to ballot on") + + self.client.logout() + username=self.ads[1].user.username + self.assertTrue(self.client.login(username=username, password=f"{username}+password")) + + response = self.client.get(url) + telechat = response.context["telechats"][0] + self.assertEqual(telechat["ad_pages_left_to_ballot_on"],222) + + + + class RescheduleOnAgendaTests(TestCase): def test_reschedule(self): draft = WgDraftFactory() @@ -542,4 +2324,105 @@ def test_reschedule(self): self.assertTrue(draft.latest_event(TelechatDocEvent, "scheduled_for_telechat")) self.assertEqual(draft.latest_event(TelechatDocEvent, "scheduled_for_telechat").telechat_date, d) self.assertTrue(not draft.latest_event(TelechatDocEvent, "scheduled_for_telechat").returning_item) - self.assertEqual(draft.docevent_set.count(), events_before + 1) \ No newline at end of file + self.assertEqual(draft.docevent_set.count(), events_before + 1) + + +class TelechatAgendaContentTests(TestCase): + def test_telechat_agenda_content_view(self): + self.client.login(username="ad", password="ad+password") + r = self.client.get(urlreverse("ietf.iesg.views.telechat_agenda_content_view", kwargs={"section": "fake"})) + self.assertEqual(r.status_code, 404, "Nonexistent section should 404") + for section in TelechatAgendaSectionName.objects.filter(used=True).values_list("slug", flat=True): + r = self.client.get( + urlreverse("ietf.iesg.views.telechat_agenda_content_view", kwargs={"section": section}) + ) + self.assertEqual(r.status_code, 404, "Section with no content should 404") + for section in TelechatAgendaSectionName.objects.filter(used=True).values_list("slug", flat=True): + content = TelechatAgendaContentFactory(section_id=section).text + r = self.client.get( + urlreverse("ietf.iesg.views.telechat_agenda_content_view", kwargs={"section": section}) + ) + self.assertContains(r, content, status_code=200) + self.assertEqual(r.get("Content-Type", None), "text/plain; charset=utf-8") + + def test_telechat_agenda_content_view_permissions(self): + for section in TelechatAgendaSectionName.objects.filter(used=True).values_list("slug", flat=True): + TelechatAgendaContentFactory(section_id=section) + url = urlreverse("ietf.iesg.views.telechat_agenda_content_view", kwargs={"section": section}) + self.client.logout() + login_testing_unauthorized(self, "plain", url) + login_testing_unauthorized(self, "ad", url) + self.assertEqual(self.client.get(url).status_code, 200) + self.client.login(username="iab chair", password="iab chair+password") + self.assertEqual(self.client.get(url).status_code, 200) + self.client.login(username="secretary", password="secretary+password") + self.assertEqual(self.client.get(url).status_code, 200) + + def test_telechat_agenda_content_edit(self): + for section in TelechatAgendaSectionName.objects.filter(used=True): + self.assertFalse(TelechatAgendaContent.objects.filter(section=section).exists()) + url = urlreverse("ietf.iesg.views.telechat_agenda_content_edit", kwargs={"section": section.slug}) + self.client.logout() + login_testing_unauthorized(self, "plain", url, method="get") + login_testing_unauthorized(self, "ad", url, method="get") + login_testing_unauthorized(self, "iab chair", url, method="get") + login_testing_unauthorized(self, "secretary", url, method="get") + r = self.client.get(url) + self.assertContains(r, str(section), status_code=200) + + self.client.logout() + login_testing_unauthorized(self, "plain", url, method="post") + login_testing_unauthorized(self, "ad", url, method="post") + login_testing_unauthorized(self, "iab chair", url, method="post") + login_testing_unauthorized(self, "secretary", url, method="post") + r = self.client.post(url, {"text": "This is some content"}) + self.assertRedirects(r, urlreverse("ietf.iesg.views.telechat_agenda_content_manage")) + contents = TelechatAgendaContent.objects.filter(section=section) + self.assertEqual(contents.count(), 1) + self.assertEqual(contents.first().text, "This is some content") + + self.client.logout() + login_testing_unauthorized(self, "plain", url, method="post") + login_testing_unauthorized(self, "ad", url, method="post") + login_testing_unauthorized(self, "iab chair", url, method="post") + login_testing_unauthorized(self, "secretary", url, method="post") + r = self.client.post(url, {"text": "This is some different content"}) + self.assertRedirects(r, urlreverse("ietf.iesg.views.telechat_agenda_content_manage")) + contents = TelechatAgendaContent.objects.filter(section=section) + self.assertEqual(contents.count(), 1) + self.assertEqual(contents.first().text, "This is some different content") + + def test_telechat_agenda_content_manage(self): + url = urlreverse("ietf.iesg.views.telechat_agenda_content_manage") + login_testing_unauthorized(self, "plain", url) + login_testing_unauthorized(self, "ad", url) + login_testing_unauthorized(self, "iab chair", url) + login_testing_unauthorized(self, "secretary", url) + self.assertEqual(TelechatAgendaContent.objects.count(), 0) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + pq = PyQuery(r.content) + for section in TelechatAgendaSectionName.objects.filter(used=True): + # check that there's a tab even when section is empty + nav_button = pq(f"button.nav-link#{section.slug}-tab") + self.assertEqual(nav_button.text(), str(section)) + edit_url = urlreverse("ietf.iesg.views.telechat_agenda_content_edit", kwargs={"section": section.pk}) + edit_button = pq(f'div#{section.slug} a[href="{edit_url}"]') + self.assertEqual(len(edit_button), 1) + self.assertIn(f"No {section}", pq(f"div#{section.slug}").text()) + # and create a section for the next test + TelechatAgendaContentFactory(section_id=section.slug) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + pq = PyQuery(r.content) + for section in TelechatAgendaSectionName.objects.filter(used=True): + # check that there's a tab with the content + nav_button = pq(f"button.nav-link#{section.slug}-tab") + self.assertEqual(nav_button.text(), str(section)) + edit_url = urlreverse("ietf.iesg.views.telechat_agenda_content_edit", kwargs={"section": section.pk}) + edit_button = pq(f'div#{section.slug} a[href="{edit_url}"]') + self.assertEqual(len(edit_button), 1) + self.assertIn( + TelechatAgendaContent.objects.get(section=section).text, pq(f"div#{section.slug}").text() + ) + diff --git a/ietf/iesg/urls.py b/ietf/iesg/urls.py index fa6f3dbf73..5fd9dea0cc 100644 --- a/ietf/iesg/urls.py +++ b/ietf/iesg/urls.py @@ -52,9 +52,14 @@ url(r'^agenda/documents.txt$', views.agenda_documents_txt), url(r'^agenda/documents/$', views.agenda_documents), + url(r'^agenda/sections$', views.telechat_agenda_content_manage), + url(r'^agenda/section/(?P
[a-z_]+)$', views.telechat_agenda_content_view), + url(r'^agenda/section/(?P
[a-z_]+)/edit$', views.telechat_agenda_content_edit), url(r'^past/documents/$', views.past_documents), url(r'^agenda/telechat-(?:%(date)s-)?docs.tgz' % settings.URL_REGEXPS, views.telechat_docs_tarfile), url(r'^discusses/$', views.discusses), + url(r'^ietf-activity/$', views.ietf_activity), + url(r'^working-groups/$', views.working_groups), url(r'^milestones/$', views.milestones_needing_review), url(r'^photos/$', views.photos), -] \ No newline at end of file +] diff --git a/ietf/iesg/utils.py b/ietf/iesg/utils.py index 4ddc9cb404..1d24ecac8e 100644 --- a/ietf/iesg/utils.py +++ b/ietf/iesg/utils.py @@ -1,56 +1,321 @@ -from collections import namedtuple +from collections import Counter, defaultdict, namedtuple -import debug # pyflakes:ignore +import datetime + +import debug # pyflakes:ignore + +from django.db import models +from django.utils import timezone from ietf.doc.models import Document, STATUSCHANGE_RELATIONS from ietf.doc.utils_search import fill_in_telechat_date +from ietf.group.models import Group from ietf.iesg.agenda import get_doc_section +from ietf.person.utils import get_active_ads +from ietf.utils.unicodenormalize import normalize_for_sorting +TelechatPageCount = namedtuple( + "TelechatPageCount", + ["for_approval", "for_action", "related", "ad_pages_left_to_ballot_on"], +) -TelechatPageCount = namedtuple('TelechatPageCount',['for_approval','for_action','related']) -def telechat_page_count(date=None, docs=None): +def telechat_page_count(date=None, docs=None, ad=None): if not date and not docs: - return TelechatPageCount(0, 0, 0) + return TelechatPageCount(0, 0, 0, 0) if not docs: - candidates = Document.objects.filter(docevent__telechatdocevent__telechat_date=date).distinct() + candidates = Document.objects.filter( + docevent__telechatdocevent__telechat_date=date + ).distinct() fill_in_telechat_date(candidates) - docs = [ doc for doc in candidates if doc.telechat_date()==date ] + docs = [doc for doc in candidates if doc.telechat_date() == date] + + for_action = [d for d in docs if get_doc_section(d).endswith(".3")] - for_action =[d for d in docs if get_doc_section(d).endswith('.3')] + for_approval = set(docs) - set(for_action) - for_approval = set(docs)-set(for_action) + drafts = [d for d in for_approval if d.type_id == "draft"] - drafts = [d for d in for_approval if d.type_id == 'draft'] + ad_pages_left_to_ballot_on = 0 + pages_for_approval = 0 - pages_for_approval = sum([d.pages or 0 for d in drafts]) + for draft in drafts: + pages_for_approval += draft.pages or 0 + if ad: + ballot = draft.active_ballot() + if ballot: + positions = ballot.active_balloter_positions() + ad_position = positions.get(ad, None) + if ad_position is None or ad_position.pos_id == "norecord": + ad_pages_left_to_ballot_on += draft.pages or 0 pages_for_action = 0 for d in for_action: - if d.type_id == 'draft': + if d.type_id == "draft": pages_for_action += d.pages or 0 - elif d.type_id == 'statchg': + elif d.type_id == "statchg": for rel in d.related_that_doc(STATUSCHANGE_RELATIONS): - pages_for_action += rel.document.pages or 0 - elif d.type_id == 'conflrev': - for rel in d.related_that_doc('conflrev'): - pages_for_action += rel.document.pages or 0 + pages_for_action += rel.pages or 0 + elif d.type_id == "conflrev": + for rel in d.related_that_doc("conflrev"): + pages_for_action += rel.pages or 0 else: pass related_pages = 0 - for d in for_approval-set(drafts): - if d.type_id == 'statchg': + for d in for_approval - set(drafts): + if d.type_id == "statchg": for rel in d.related_that_doc(STATUSCHANGE_RELATIONS): - related_pages += rel.document.pages or 0 - elif d.type_id == 'conflrev': - for rel in d.related_that_doc('conflrev'): - related_pages += rel.document.pages or 0 + related_pages += rel.pages or 0 + elif d.type_id == "conflrev": + for rel in d.related_that_doc("conflrev"): + related_pages += rel.pages or 0 else: # There's really nothing to rely on to give a reading load estimate for charters pass - - return TelechatPageCount(for_approval=pages_for_approval, - for_action=pages_for_action, - related=related_pages) + + return TelechatPageCount( + for_approval=pages_for_approval, + for_action=pages_for_action, + related=related_pages, + ad_pages_left_to_ballot_on=ad_pages_left_to_ballot_on, + ) + + +def get_wg_dashboard_info(): + docs = ( + Document.objects.filter( + group__type="wg", + group__state="active", + states__type="draft", + states__slug="active", + ) + .filter(models.Q(ad__isnull=True) | models.Q(ad__in=get_active_ads())) + .distinct() + .prefetch_related("group", "group__parent") + .exclude( + states__type="draft-stream-ietf", + states__slug__in=["c-adopt", "wg-cand", "dead", "parked", "info"], + ) + ) + groups = Group.objects.filter(state="active", type="wg") + areas = Group.objects.filter(state="active", type="area") + + total_group_count = groups.count() + total_doc_count = docs.count() + total_page_count = docs.aggregate(models.Sum("pages"))["pages__sum"] or 0 + totals = { + "group_count": total_group_count, + "doc_count": total_doc_count, + "page_count": total_page_count, + } + + # Since this view is primarily about counting subsets of the above docs query and the + # expected number of returned documents is just under 1000 typically - do the totaling + # work in python rather than asking the db to do it. + + groups_for_area = defaultdict(set) + pages_for_area = defaultdict(lambda: 0) + docs_for_area = defaultdict(lambda: 0) + groups_for_ad = defaultdict(lambda: defaultdict(set)) + pages_for_ad = defaultdict(lambda: defaultdict(lambda: 0)) + docs_for_ad = defaultdict(lambda: defaultdict(lambda: 0)) + groups_for_noad = defaultdict(lambda: defaultdict(set)) + pages_for_noad = defaultdict(lambda: defaultdict(lambda: 0)) + docs_for_noad = defaultdict(lambda: defaultdict(lambda: 0)) + docs_for_wg = defaultdict(lambda: 0) + pages_for_wg = defaultdict(lambda: 0) + groups_total = set() + pages_total = 0 + docs_total = 0 + + responsible_for_group = defaultdict(lambda: defaultdict(lambda: "None")) + responsible_count = defaultdict(lambda: defaultdict(lambda: 0)) + for group in groups: + responsible = f"{', '.join([r.person.plain_name() for r in group.role_set.filter(name_id='ad')])}" + docs_for_noad[responsible][group.parent.acronym] = ( + 0 # Ensure these keys are present later + ) + docs_for_ad[responsible][group.parent.acronym] = 0 + responsible_for_group[group.acronym][group.parent.acronym] = responsible + responsible_count[responsible][group.parent.acronym] += 1 + + for doc in docs: + docs_for_wg[doc.group] += 1 + pages_for_wg[doc.group] += doc.pages + groups_for_area[doc.group.area.acronym].add(doc.group.acronym) + pages_for_area[doc.group.area.acronym] += doc.pages + docs_for_area[doc.group.area.acronym] += 1 + + if doc.ad is None: + responsible = responsible_for_group[doc.group.acronym][ + doc.group.parent.acronym + ] + groups_for_noad[responsible][doc.group.parent.acronym].add( + doc.group.acronym + ) + pages_for_noad[responsible][doc.group.parent.acronym] += doc.pages + docs_for_noad[responsible][doc.group.parent.acronym] += 1 + else: + responsible = f"{doc.ad.plain_name()}" + groups_for_ad[responsible][doc.group.parent.acronym].add(doc.group.acronym) + pages_for_ad[responsible][doc.group.parent.acronym] += doc.pages + docs_for_ad[responsible][doc.group.parent.acronym] += 1 + + docs_total += 1 + groups_total.add(doc.group.acronym) + pages_total += doc.pages + + groups_total = len(groups_total) + totals["groups_with_docs_count"] = groups_total + + area_summary = [] + + for area in areas: + group_count = len(groups_for_area[area.acronym]) + doc_count = docs_for_area[area.acronym] + page_count = pages_for_area[area.acronym] + area_summary.append( + { + "area": area.acronym, + "groups_in_area": groups.filter(parent=area).count(), + "groups_with_docs": group_count, + "doc_count": doc_count, + "page_count": page_count, + "group_percent": group_count / groups_total * 100 + if groups_total != 0 + else 0, + "doc_percent": doc_count / docs_total * 100 if docs_total != 0 else 0, + "page_percent": page_count / pages_total * 100 + if pages_total != 0 + else 0, + } + ) + area_summary.sort(key=lambda r: r["area"]) + area_totals = { + "group_count": groups_total, + "doc_count": docs_total, + "page_count": pages_total, + } + + noad_summary = [] + noad_totals = { + "ad_group_count": 0, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + } + for ad in docs_for_noad: + for area in docs_for_noad[ad]: + noad_totals["ad_group_count"] += responsible_count[ad][area] + noad_totals["doc_group_count"] += len(groups_for_noad[ad][area]) + noad_totals["doc_count"] += docs_for_noad[ad][area] + noad_totals["page_count"] += pages_for_noad[ad][area] + for ad in docs_for_noad: + for area in docs_for_noad[ad]: + noad_summary.append( + { + "ad": ad, + "area": area, + "ad_group_count": responsible_count[ad][area], + "doc_group_count": len(groups_for_noad[ad][area]), + "doc_count": docs_for_noad[ad][area], + "page_count": pages_for_noad[ad][area], + "group_percent": len(groups_for_noad[ad][area]) + / noad_totals["doc_group_count"] + * 100 + if noad_totals["doc_group_count"] != 0 + else 0, + "doc_percent": docs_for_noad[ad][area] + / noad_totals["doc_count"] + * 100 + if noad_totals["doc_count"] != 0 + else 0, + "page_percent": pages_for_noad[ad][area] + / noad_totals["page_count"] + * 100 + if noad_totals["page_count"] != 0 + else 0, + } + ) + noad_summary.sort(key=lambda r: (normalize_for_sorting(r["ad"]), r["area"])) + + ad_summary = [] + ad_totals = { + "ad_group_count": 0, + "doc_group_count": 0, + "doc_count": 0, + "page_count": 0, + } + for ad in docs_for_ad: + for area in docs_for_ad[ad]: + ad_totals["ad_group_count"] += responsible_count[ad][area] + ad_totals["doc_group_count"] += len(groups_for_ad[ad][area]) + ad_totals["doc_count"] += docs_for_ad[ad][area] + ad_totals["page_count"] += pages_for_ad[ad][area] + for ad in docs_for_ad: + for area in docs_for_ad[ad]: + ad_summary.append( + { + "ad": ad, + "area": area, + "ad_group_count": responsible_count[ad][area], + "doc_group_count": len(groups_for_ad[ad][area]), + "doc_count": docs_for_ad[ad][area], + "page_count": pages_for_ad[ad][area], + "group_percent": len(groups_for_ad[ad][area]) + / ad_totals["doc_group_count"] + * 100 + if ad_totals["doc_group_count"] != 0 + else 0, + "doc_percent": docs_for_ad[ad][area] / ad_totals["doc_count"] * 100 + if ad_totals["doc_count"] != 0 + else 0, + "page_percent": pages_for_ad[ad][area] + / ad_totals["page_count"] + * 100 + if ad_totals["page_count"] != 0 + else 0, + } + ) + ad_summary.sort(key=lambda r: (normalize_for_sorting(r["ad"]), r["area"])) + + rfc_counter = Counter( + Document.objects.filter(type="rfc").values_list("group__acronym", flat=True) + ) + recent_rfc_counter = Counter( + Document.objects.filter( + type="rfc", + docevent__type="published_rfc", + docevent__time__gte=timezone.now() - datetime.timedelta(weeks=104), + ).values_list("group__acronym", flat=True) + ) + for wg in set(groups) - set(docs_for_wg.keys()): + docs_for_wg[wg] += 0 + pages_for_wg[wg] += 0 + wg_summary = [] + for wg in docs_for_wg: + wg_summary.append( + { + "wg": wg.acronym, + "area": wg.parent.acronym, + "ad": responsible_for_group[wg.acronym][wg.parent.acronym], + "doc_count": docs_for_wg[wg], + "page_count": pages_for_wg[wg], + "rfc_count": rfc_counter[wg.acronym], + "recent_rfc_count": recent_rfc_counter[wg.acronym], + } + ) + wg_summary.sort(key=lambda r: (r["wg"], r["area"])) + + return ( + area_summary, + area_totals, + ad_summary, + noad_summary, + ad_totals, + noad_totals, + totals, + wg_summary, + ) diff --git a/ietf/iesg/views.py b/ietf/iesg/views.py index 5334e8a85f..f03afb9fc1 100644 --- a/ietf/iesg/views.py +++ b/ietf/iesg/views.py @@ -41,13 +41,15 @@ import os import tarfile import time +from dateutil import relativedelta from django import forms from django.conf import settings from django.db import models from django.http import HttpResponse -from django.shortcuts import render, redirect +from django.shortcuts import render, redirect, get_object_or_404 from django.contrib.sites.models import Site +from django.urls import reverse as urlreverse from django.utils.encoding import force_bytes #from django.views.decorators.cache import cache_page #from django.views.decorators.vary import vary_on_cookie @@ -58,12 +60,15 @@ from ietf.doc.utils import update_telechat, augment_events_with_revision from ietf.group.models import GroupMilestone, Role from ietf.iesg.agenda import agenda_data, agenda_sections, fill_in_agenda_docs, get_agenda_date -from ietf.iesg.models import TelechatDate -from ietf.iesg.utils import telechat_page_count +from ietf.iesg.models import TelechatDate, TelechatAgendaContent +from ietf.iesg.utils import get_wg_dashboard_info, telechat_page_count from ietf.ietfauth.utils import has_role, role_required, user_is_person +from ietf.name.models import TelechatAgendaSectionName from ietf.person.models import Person +from ietf.meeting.utils import get_activity_stats from ietf.doc.utils_search import fill_in_document_table_attributes, fill_in_telechat_date from ietf.utils.timezone import date_today, datetime_from_date +from ietf.utils.unicodenormalize import normalize_for_sorting def review_decisions(request, year=None): events = DocEvent.objects.filter(type__in=("iesg_disapproved", "iesg_approved")) @@ -97,7 +102,7 @@ def agenda_json(request, date=None): res = { "telechat-date": str(data["date"]), - "as-of": str(datetime.datetime.utcnow()), + "as-of": str(datetime.datetime.now(datetime.UTC)), "page-counts": telechat_page_count(date=get_agenda_date(date))._asdict(), "sections": {}, } @@ -118,7 +123,7 @@ def agenda_json(request, date=None): for doc in docs: wginfo = { - 'docname': doc.canonical_name(), + 'docname': doc.name, 'rev': doc.rev, 'wgname': doc.group.name, 'acronym': doc.group.acronym, @@ -133,13 +138,11 @@ def agenda_json(request, date=None): for doc in docs: docinfo = { - 'docname':doc.canonical_name(), + 'docname':doc.name, 'title':doc.title, 'ad':doc.ad.name if doc.ad else None, } - if doc.note: - docinfo['note'] = doc.note defer = doc.active_defer_event() if defer: docinfo['defer-by'] = defer.by.name @@ -147,8 +150,8 @@ def agenda_json(request, date=None): if doc.type_id == "draft": docinfo['rev'] = doc.rev docinfo['intended-std-level'] = str(doc.intended_std_level) - if doc.rfc_number(): - docinfo['rfc-number'] = doc.rfc_number() + if doc.type_id == "rfc": + docinfo['rfc-number'] = doc.rfc_number iana_state = doc.get_state("draft-iana-review") if iana_state and iana_state.slug in ("not-ok", "changed", "need-rev"): @@ -168,8 +171,8 @@ def agenda_json(request, date=None): elif doc.type_id == 'conflrev': docinfo['rev'] = doc.rev - td = doc.relateddocument_set.get(relationship__slug='conflrev').target.document - docinfo['target-docname'] = td.canonical_name() + td = doc.relateddocument_set.get(relationship__slug='conflrev').target + docinfo['target-docname'] = td.name docinfo['target-title'] = td.title docinfo['target-rev'] = td.rev docinfo['intended-std-level'] = str(td.intended_std_level) @@ -195,10 +198,18 @@ def agenda(request, date=None): data = agenda_data(date) if has_role(request.user, ["Area Director", "IAB Chair", "Secretariat"]): - data["sections"]["1.1"]["title"] = data["sections"]["1.1"]["title"].replace("Roll call", 'Roll Call' % settings.IESG_ROLL_CALL_URL ) - data["sections"]["1.3"]["title"] = data["sections"]["1.3"]["title"].replace("minutes", 'Minutes' % settings.IESG_MINUTES_URL) + data["sections"]["1.1"]["title"] = data["sections"]["1.1"]["title"].replace( + "Roll call", + 'Roll Call'.format( + urlreverse("ietf.iesg.views.telechat_agenda_content_view", kwargs={"section": "roll_call"}) + ) + ) + data["sections"]["1.3"]["title"] = data["sections"]["1.3"]["title"].replace( + "minutes", + 'Minutes'.format( + urlreverse("ietf.iesg.views.telechat_agenda_content_view", kwargs={"section": "minutes"}) + )) - request.session['ballot_edit_return_point'] = request.path_info return render(request, "iesg/agenda.html", { "date": data["date"], "sections": sorted(data["sections"].items(), key=lambda x:[int(p) for p in x[0].split('.')]), @@ -211,7 +222,7 @@ def agenda_txt(request, date=None): "date": data["date"], "sections": sorted(data["sections"].items(), key=lambda x:[int(p) for p in x[0].split('.')]), "domain": Site.objects.get_current().domain, - }, content_type="text/plain; charset=%s"%settings.DEFAULT_CHARSET) + }, content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}") @role_required('Area Director', 'Secretariat') def agenda_moderator_package(request, date=None): @@ -267,16 +278,23 @@ def leaf_section(num, section): @role_required('Area Director', 'Secretariat') def agenda_package(request, date=None): data = agenda_data(date) - return render(request, "iesg/agenda_package.txt", { + return render( + request, + "iesg/agenda_package.txt", + { "date": data["date"], "sections": sorted(data["sections"].items()), "roll_call": data["sections"]["1.1"]["text"], - "roll_call_url": settings.IESG_ROLL_CALL_URL, "minutes": data["sections"]["1.3"]["text"], - "minutes_url": settings.IESG_MINUTES_URL, - "management_items": [(num, section) for num, section in data["sections"].items() if "6" < num < "7"], + "management_items": [ + (num, section) + for num, section in data["sections"].items() + if "6" < num < "7" + ], "domain": Site.objects.get_current().domain, - }, content_type='text/plain') + }, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) def agenda_documents_txt(request): @@ -307,7 +325,10 @@ def agenda_documents_txt(request): d.rev, ) rows.append("\t".join(row)) - return HttpResponse("\n".join(rows), content_type='text/plain') + return HttpResponse( + "\n".join(rows), + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) class RescheduleForm(forms.Form): telechat_date = forms.TypedChoiceField(coerce=lambda x: datetime.datetime.strptime(x, '%Y-%m-%d').date(), empty_value=None, required=False) @@ -352,6 +373,8 @@ def handle_reschedule_form(request, doc, dates, status): return form def agenda_documents(request): + ad = request.user.person if has_role(request.user, "Area Director") else None + dates = list(TelechatDate.objects.active().order_by('date').values_list("date", flat=True)[:4]) docs_by_date = dict((d, []) for d in dates) @@ -381,15 +404,17 @@ def agenda_documents(request): # the search_result_row view to display them (which expects them) fill_in_document_table_attributes(docs_by_date[date], have_telechat_date=True) fill_in_agenda_docs(date, sections, docs_by_date[date]) - pages = telechat_page_count(docs=docs_by_date[date]).for_approval - + page_count = telechat_page_count(docs=docs_by_date[date], ad=ad) + pages = page_count.for_approval + telechats.append({ "date": date, "pages": pages, + "ad_pages_left_to_ballot_on": page_count.ad_pages_left_to_ballot_on, "sections": sorted((num, section) for num, section in sections.items() if "2" <= num < "5") }) - request.session['ballot_edit_return_point'] = request.path_info + return render(request, 'iesg/agenda_documents.html', { 'telechats': telechats }) def past_documents(request): @@ -474,6 +499,7 @@ def discusses(request): models.Q(states__type__in=("statchg", "conflrev"), states__slug__in=("iesgeval", "defer")), docevent__ballotpositiondocevent__pos__blocking=True) + possible_docs = possible_docs.exclude(states__in=State.objects.filter(type="draft", slug="repl")) possible_docs = possible_docs.select_related("stream", "group", "ad").distinct() docs = [] @@ -517,10 +543,12 @@ def milestones_needing_review(request): ad_list.append(ad) ad.groups_needing_review = sorted(groups, key=lambda g: g.acronym) for g, milestones in groups.items(): - g.milestones_needing_review = sorted(milestones, key=lambda m: m.due) + g.milestones_needing_review = sorted( + milestones, key=lambda m: m.due if m.group.uses_milestone_dates else m.order + ) return render(request, 'iesg/milestones_needing_review.html', - dict(ads=sorted(ad_list, key=lambda ad: ad.plain_name()),)) + dict(ads=sorted(ad_list, key=lambda ad: normalize_for_sorting(ad.plain_name())),)) def photos(request): roles = sorted(Role.objects.filter(group__type='area', group__state='active', name_id='ad'),key=lambda x: "" if x.group.acronym=="gen" else x.group.acronym) @@ -528,4 +556,84 @@ def photos(request): role.last_initial = role.person.last_name()[0] return render(request, 'iesg/photos.html', {'group_type': 'IESG', 'role': '', 'roles': roles }) - \ No newline at end of file +def month_choices(): + choices = [(str(n).zfill(2), str(n).zfill(2)) for n in range(1, 13)] + return choices + +def year_choices(): + this_year = date_today().year + choices = [(str(n), str(n)) for n in range(this_year, 2009, -1)] + return choices + +class ActivityForm(forms.Form): + month = forms.ChoiceField(choices=month_choices, help_text='Month', required=True) + year = forms.ChoiceField(choices=year_choices, help_text='Year', required=True) + +def ietf_activity(request): + # default date range for last month + today = date_today() + edate = today.replace(day=1) + sdate = (edate - datetime.timedelta(days=1)).replace(day=1) + if request.method == 'GET': + form = ActivityForm(request.GET) + if form.is_valid(): + month = form.cleaned_data['month'] + year = form.cleaned_data['year'] + sdate = datetime.date(int(year), int(month), 1) + edate = sdate + relativedelta.relativedelta(months=1) + + # always pass back an unbound form to avoid annoying is-valid styling + form = ActivityForm(initial={'month': str(sdate.month).zfill(2), 'year': sdate.year}) + context = get_activity_stats(sdate, edate) + context['form'] = form + return render(request, "iesg/ietf_activity_report.html", context) + + +class TelechatAgendaContentForm(forms.Form): + text = forms.CharField(max_length=100_000, widget=forms.Textarea, required=False) + + +@role_required("Secretariat") +def telechat_agenda_content_edit(request, section): + section = get_object_or_404(TelechatAgendaSectionName, slug=section, used=True) + content = TelechatAgendaContent.objects.filter(section=section).first() + initial = {"text": content.text} if content else {} + if request.method == "POST": + form = TelechatAgendaContentForm(data=request.POST, initial=initial) + if form.is_valid(): + TelechatAgendaContent.objects.update_or_create( + section=section, defaults={"text": form.cleaned_data["text"]} + ) + return redirect("ietf.iesg.views.telechat_agenda_content_manage") + else: + form = TelechatAgendaContentForm(initial=initial) + return render(request, "iesg/telechat_agenda_content_edit.html", {"section": section, "form": form}) + + +@role_required("Secretariat") +def telechat_agenda_content_manage(request): + # Fill in any missing instances with empty stand-ins. The edit view will create persistent instances if needed. + contents = [ + TelechatAgendaContent.objects.filter(section=section).first() or TelechatAgendaContent(section=section) + for section in TelechatAgendaSectionName.objects.filter(used=True) + ] + return render(request, "iesg/telechat_agenda_content_manage.html", {"contents": contents}) + + +@role_required("Secretariat", "IAB Chair", "Area Director") +def telechat_agenda_content_view(request, section): + content = get_object_or_404(TelechatAgendaContent, section__slug=section, section__used=True) + return HttpResponse( + content=content.text, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) + +def working_groups(request): + + area_summary, area_totals, ad_summary, noad_summary, ad_totals, noad_totals, totals, wg_summary = get_wg_dashboard_info() + + return render( + request, + "iesg/working_groups.html", + dict(area_summary=area_summary, area_totals=area_totals, ad_summary=ad_summary, noad_summary=noad_summary, ad_totals=ad_totals, noad_totals=noad_totals, totals=totals, wg_summary=wg_summary), + ) diff --git a/ietf/ietfauth/.gitignore b/ietf/ietfauth/.gitignore deleted file mode 100644 index 8ca2ec2c8b..0000000000 --- a/ietf/ietfauth/.gitignore +++ /dev/null @@ -1,2 +0,0 @@ -/*.swp -/*.pyc diff --git a/ietf/ietfauth/admin.py b/ietf/ietfauth/admin.py new file mode 100644 index 0000000000..c2914f9efa --- /dev/null +++ b/ietf/ietfauth/admin.py @@ -0,0 +1,136 @@ +# Copyright The IETF Trust 2025, All Rights Reserved +import datetime + +from django.conf import settings +from django.contrib import admin, messages +from django.contrib.admin import action +from django.contrib.admin.actions import delete_selected as default_delete_selected +from django.contrib.auth.admin import UserAdmin +from django.contrib.auth.models import User +from django.utils import timezone + + +# Replace default UserAdmin with our custom one +admin.site.unregister(User) + + +class AgeListFilter(admin.SimpleListFilter): + title = "account age" + parameter_name = "age" + + def lookups(self, request, model_admin): + return [ + ("1day", "> 1 day"), + ("3days", "> 3 days"), + ("1week", "> 1 week"), + ("1month", "> 1 month"), + ("1year", "> 1 year"), + ] + + def queryset(self, request, queryset): + deltas = { + "1day": datetime.timedelta(days=1), + "3days": datetime.timedelta(days=3), + "1week": datetime.timedelta(weeks=1), + "1month": datetime.timedelta(days=30), + "1year": datetime.timedelta(days=365), + } + if self.value(): + return queryset.filter(date_joined__lt=timezone.now()-deltas[self.value()]) + return queryset + + +@admin.register(User) +class CustomUserAdmin(UserAdmin): + list_display = ( + "username", + "person", + "date_joined", + "last_login", + "is_staff", + ) + list_filter = list(UserAdmin.list_filter) + [ + AgeListFilter, + ("person", admin.EmptyFieldListFilter), + ] + actions = ["delete_selected"] + + @action( + permissions=["delete"], description="Delete personless %(verbose_name_plural)s" + ) + def delete_selected(self, request, queryset): + """Delete selected action restricted to Users with a null Person field + + This displaces the default delete_selected action with a safer one that will + only delete personless Users. It is done this way instead of by introducing + a new action so that we can simply hand off to the default action (imported + as default_delete_selected()) without having to adjust its template (and maybe + other things) to make it work with a different action name. + """ + already_confirmed = bool(request.POST.get("post")) + personless_queryset = queryset.filter(person__isnull=True) + original_count = queryset.count() + personless_count = personless_queryset.count() + if personless_count > original_count: + # Refuse to act if the count increased! + self.message_user( + request, + ( + "Limiting the selection to Users without a Person INCREASED the " + "count from {} to {}. This should not happen and probably means a " + "concurrent change to the database affected this request. Please " + "try again.".format(original_count, personless_count) + ), + level=messages.ERROR, + ) + return None # return to changelist + + # Display warning/info if this is showing the confirmation page + if not already_confirmed: + if personless_count < original_count: + self.message_user( + request, + ( + "Limiting the selection to Users without a Person reduced the " + "count from {} to {}. Only {} will be deleted.".format( + original_count, personless_count, personless_count + ) + ), + level=messages.WARNING, + ) + else: + self.message_user( + request, + "Confirmed that all selected Users had no Persons.", + ) + + # Django limits the number of fields in a request. The delete form itself + # includes a few metadata fields, so give it a little padding. The default + # limit is 1000 and everything will break if it's a small number, so not + # bothering to check that it's > 10. + max_count = settings.DATA_UPLOAD_MAX_NUMBER_FIELDS - 10 + if personless_count > max_count: + self.message_user( + request, + ( + f"Only {max_count} Users can be deleted at once. Will only delete " + f"the first {max_count} selected Personless Users." + ), + level=messages.WARNING, + ) + # delete() doesn't like a queryset limited via [:max_count], so do an + # equivalent filter. + last_to_delete = personless_queryset.order_by("pk")[max_count] + personless_queryset = personless_queryset.filter(pk__lt=last_to_delete.pk) + + if already_confirmed and personless_count != original_count: + # After confirmation, none of the above filtering should change anything. + # Refuse to delete if the DB moved underneath us. + self.message_user( + request, + "Queryset count changed, nothing deleted. Please try again.", + level=messages.ERROR, + ) + return None + + return default_delete_selected(self, request, personless_queryset) diff --git a/ietf/ietfauth/backends.py b/ietf/ietfauth/backends.py new file mode 100644 index 0000000000..ad34ca9a74 --- /dev/null +++ b/ietf/ietfauth/backends.py @@ -0,0 +1,21 @@ + +# From https://simpleisbetterthancomplex.com/tutorial/2017/02/06/how-to-implement-case-insensitive-username.html +from django.contrib.auth import get_user_model +from django.contrib.auth.backends import ModelBackend + + +class CaseInsensitiveModelBackend(ModelBackend): + def authenticate(self, request, username=None, password=None, **kwargs): + UserModel = get_user_model() + if username is None: + username = kwargs.get(UserModel.USERNAME_FIELD) + try: + case_insensitive_username_field = '{}__iexact'.format(UserModel.USERNAME_FIELD) + user = UserModel._default_manager.get(**{case_insensitive_username_field: username}) + except UserModel.DoesNotExist: + # Run the default password hasher once to reduce the timing + # difference between an existing and a non-existing user (#20760). + UserModel().set_password(password) + else: + if user.check_password(password) and self.user_can_authenticate(user): + return user diff --git a/ietf/ietfauth/factories.py b/ietf/ietfauth/factories.py index b68b0cecef..5e827791a1 100644 --- a/ietf/ietfauth/factories.py +++ b/ietf/ietfauth/factories.py @@ -14,6 +14,7 @@ class OidClientRecordFactory(factory.django.DjangoModelFactory): class Meta: model = OidClientRecord + skip_postgeneration_save = True name = factory.Faker('company') owner = factory.SubFactory(UserFactory) diff --git a/ietf/ietfauth/forms.py b/ietf/ietfauth/forms.py index dab5ce374b..41828f2bf6 100644 --- a/ietf/ietfauth/forms.py +++ b/ietf/ietfauth/forms.py @@ -3,23 +3,23 @@ import re + from unidecode import unidecode from django import forms -from django.conf import settings +from django.contrib.auth.models import User +from django.contrib.auth import password_validation from django.core.exceptions import ValidationError from django.db import models -from django.contrib.auth.models import User -from django.utils.html import mark_safe # type:ignore -from django.urls import reverse as urlreverse - -from django_password_strength.widgets import PasswordStrengthInput, PasswordConfirmationInput - -import debug # pyflakes:ignore from ietf.person.models import Person, Email from ietf.mailinglists.models import Allowlisted from ietf.utils.text import isascii +from .password_validation import StrongPasswordValidator + +from .validators import prevent_at_symbol, prevent_system_name, prevent_anonymous_name, is_allowed_address +from .widgets import PasswordStrengthInput, PasswordConfirmationInput + class RegistrationForm(forms.Form): email = forms.EmailField(label="Your email (lowercase)") @@ -30,45 +30,65 @@ def clean_email(self): return email if email.lower() != email: raise forms.ValidationError('The supplied address contained uppercase letters. Please use a lowercase email address.') - if User.objects.filter(username=email).exists(): - raise forms.ValidationError('An account with the email address you provided already exists.') return email +class PasswordStrengthField(forms.CharField): + widget = PasswordStrengthInput( + attrs={ + "class": "password_strength", + "data-disable-strength-enforcement": "", # usually removed in init + } + ) + + def __init__(self, *args, **kwargs): + super().__init__(*args, **kwargs) + for pwval in password_validation.get_default_password_validators(): + if isinstance(pwval, password_validation.MinimumLengthValidator): + self.widget.attrs["minlength"] = pwval.min_length + elif isinstance(pwval, StrongPasswordValidator): + self.widget.attrs.pop( + "data-disable-strength-enforcement", None + ) + + + class PasswordForm(forms.Form): - password = forms.CharField(widget=PasswordStrengthInput(attrs={'class':'password_strength'})) + password = PasswordStrengthField() password_confirmation = forms.CharField(widget=PasswordConfirmationInput( confirm_with='password', attrs={'class':'password_confirmation'}), help_text="Enter the same password as above, for verification.",) - + + def __init__(self, *args, user=None, **kwargs): + # user is a kw-only argument to avoid interfering with the signature + # when this class is mixed with ModelForm in PersonPasswordForm + self.user = user + super().__init__(*args, **kwargs) def clean_password_confirmation(self): - password = self.cleaned_data.get("password", "") - password_confirmation = self.cleaned_data["password_confirmation"] + # clean fields here rather than a clean() method so validation is + # still enforced in PersonPasswordForm without having to override its + # clean() method + password = self.cleaned_data.get("password") + password_confirmation = self.cleaned_data.get("password_confirmation") if password != password_confirmation: - raise forms.ValidationError("The two password fields didn't match.") + raise ValidationError( + "The password confirmation is different than the new password" + ) + try: + password_validation.validate_password(password_confirmation, self.user) + except ValidationError as err: + self.add_error("password", err) return password_confirmation + def ascii_cleaner(supposedly_ascii): outside_printable_ascii_pattern = r'[^\x20-\x7F]' if re.search(outside_printable_ascii_pattern, supposedly_ascii): raise forms.ValidationError("Only unaccented Latin characters are allowed.") return supposedly_ascii -def prevent_at_symbol(name): - if "@" in name: - raise forms.ValidationError("Please fill in name - this looks like an email address (@ is not allowed in names).") - -def prevent_system_name(name): - name_without_spaces = name.replace(" ", "").replace("\t", "") - if "(system)" in name_without_spaces.lower(): - raise forms.ValidationError("Please pick another name - this name is reserved.") - -def prevent_anonymous_name(name): - name_without_spaces = name.replace(" ", "").replace("\t", "") - if "anonymous" in name_without_spaces.lower(): - raise forms.ValidationError("Please pick another name - this name is reserved.") class PersonPasswordForm(forms.ModelForm, PasswordForm): @@ -159,20 +179,7 @@ def clean(self): class NewEmailForm(forms.Form): - new_email = forms.EmailField(label="New email address", required=False) - - def clean_new_email(self): - email = self.cleaned_data.get("new_email", "") - if email: - existing = Email.objects.filter(address=email).first() - if existing: - raise forms.ValidationError("Email address '%s' is already assigned to account '%s' (%s)" % (existing, existing.person and existing.person.user, existing.person)) - - for pat in settings.EXCLUDED_PERSONAL_EMAIL_REGEX_PATTERNS: - if re.search(pat, email): - raise ValidationError("This email address is not valid in a datatracker account") - - return email + new_email = forms.EmailField(label="New email address", required=False, validators=[is_allowed_address]) class RoleEmailForm(forms.Form): @@ -192,13 +199,6 @@ def __init__(self, role, *args, **kwargs): class ResetPasswordForm(forms.Form): username = forms.EmailField(label="Your email (lowercase)") - def clean_username(self): - import ietf.ietfauth.views - username = self.cleaned_data["username"] - if not User.objects.filter(username=username).exists(): - raise forms.ValidationError(mark_safe("Didn't find a matching account. If you don't have an account yet, you can create one.".format(urlreverse(ietf.ietfauth.views.create_account)))) - return username - class TestEmailForm(forms.Form): email = forms.EmailField(required=False) @@ -208,33 +208,21 @@ class Meta: model = Allowlisted exclude = ['by', 'time' ] - -from django import forms - -class ChangePasswordForm(forms.Form): +class ChangePasswordForm(PasswordForm): current_password = forms.CharField(widget=forms.PasswordInput) + field_order = ["current_password", "password", "password_confirmation"] - new_password = forms.CharField(widget=PasswordStrengthInput(attrs={'class':'password_strength'})) - new_password_confirmation = forms.CharField(widget=PasswordConfirmationInput( - confirm_with='new_password', - attrs={'class':'password_confirmation'})) - - def __init__(self, user, data=None): - self.user = user - super(ChangePasswordForm, self).__init__(data) + def __init__(self, user, *args, **kwargs): + # user arg is optional in superclass, but required for this form + super().__init__(*args, user=user, **kwargs) def clean_current_password(self): - password = self.cleaned_data.get('current_password', None) + # n.b., password = None is handled by check_password and results in a failed check + password = self.cleaned_data.get("current_password", None) if not self.user.check_password(password): - raise ValidationError('Invalid password') + raise ValidationError("Invalid password") return password - - def clean(self): - new_password = self.cleaned_data.get('new_password', None) - conf_password = self.cleaned_data.get('new_password_confirmation', None) - if not new_password == conf_password: - raise ValidationError("The password confirmation is different than the new password") class ChangeUsernameForm(forms.Form): @@ -257,6 +245,6 @@ def clean_password(self): def clean_username(self): username = self.cleaned_data['username'] - if User.objects.filter(username=username).exists(): + if User.objects.filter(username__iexact=username).exists(): raise ValidationError("A login with that username already exists. Please contact the secretariat to get this resolved.") return username diff --git a/ietf/ietfauth/htpasswd.py b/ietf/ietfauth/htpasswd.py deleted file mode 100644 index 3716d98600..0000000000 --- a/ietf/ietfauth/htpasswd.py +++ /dev/null @@ -1,30 +0,0 @@ -# Copyright The IETF Trust 2016-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -import io -import subprocess, hashlib -from django.utils.encoding import force_bytes - -from django.conf import settings - -def update_htpasswd_file(username, password): - if getattr(settings, 'USE_PYTHON_HTDIGEST', None): - pass_file = settings.HTPASSWD_FILE - realm = settings.HTDIGEST_REALM - prefix = force_bytes('%s:%s:' % (username, realm)) - key = force_bytes(hashlib.md5(prefix + force_bytes(password)).hexdigest()) - f = io.open(pass_file, 'r+b') - pos = f.tell() - line = f.readline() - while line: - if line.startswith(prefix): - break - pos=f.tell() - line = f.readline() - f.seek(pos) - f.write(b'%s%s\n' % (prefix, key)) - f.close() - else: - p = subprocess.Popen([settings.HTPASSWD_COMMAND, "-b", settings.HTPASSWD_FILE, username, password], stdout=subprocess.PIPE, stderr=subprocess.PIPE) - stdout, stderr = p.communicate() diff --git a/ietf/ietfauth/management/commands/send_apikey_usage_emails.py b/ietf/ietfauth/management/commands/send_apikey_usage_emails.py deleted file mode 100644 index d3fce1bcc2..0000000000 --- a/ietf/ietfauth/management/commands/send_apikey_usage_emails.py +++ /dev/null @@ -1,52 +0,0 @@ -# Copyright The IETF Trust 2017-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -import datetime - -from textwrap import dedent - -from django.conf import settings -from django.core.management.base import BaseCommand -from django.utils import timezone - -import debug # pyflakes:ignore - -from ietf.person.models import PersonalApiKey, PersonApiKeyEvent -from ietf.utils.mail import send_mail - - -class Command(BaseCommand): - """ - Send out emails to all persons who have personal API keys about usage. - - Usage is show over the given period, where the default period is 7 days. - """ - - help = dedent(__doc__).strip() - - def add_arguments(self, parser): - parser.add_argument('-d', '--days', dest='days', type=int, default=7, - help='The period over which to show usage.') - - def handle(self, *filenames, **options): - """ - """ - - self.verbosity = int(options.get('verbosity')) - days = options.get('days') - - keys = PersonalApiKey.objects.filter(valid=True) - for key in keys: - earliest = timezone.now() - datetime.timedelta(days=days) - events = PersonApiKeyEvent.objects.filter(key=key, time__gt=earliest) - count = events.count() - events = events[:32] - if count: - key_name = key.hash()[:8] - subject = "API key usage for key '%s' for the last %s days" %(key_name, days) - to = key.person.email_address() - frm = settings.DEFAULT_FROM_EMAIL - send_mail(None, to, frm, subject, 'utils/apikey_usage_report.txt', {'person':key.person, - 'days':days, 'key':key, 'key_name':key_name, 'count':count, 'events':events, } ) - diff --git a/ietf/ietfauth/password_validation.py b/ietf/ietfauth/password_validation.py new file mode 100644 index 0000000000..bfed4a784e --- /dev/null +++ b/ietf/ietfauth/password_validation.py @@ -0,0 +1,23 @@ +# Copyright The IETF Trust 2025, All Rights Reserved +from django.core.exceptions import ValidationError +from zxcvbn import zxcvbn + + +class StrongPasswordValidator: + message = "This password does not meet complexity requirements and is easily guessable." + code = "weak" + min_zxcvbn_score = 3 + + def __init__(self, message=None, code=None, min_zxcvbn_score=None): + if message is not None: + self.message = message + if code is not None: + self.code = code + if min_zxcvbn_score is not None: + self.min_zxcvbn_score = min_zxcvbn_score + + def validate(self, password, user=None): + """Validate that a password is strong enough""" + strength_report = zxcvbn(password[:72], max_length=72) + if strength_report["score"] < self.min_zxcvbn_score: + raise ValidationError(message=self.message, code=self.code) diff --git a/ietf/ietfauth/tests.py b/ietf/ietfauth/tests.py index 9fd75d06de..a77e5bd5d5 100644 --- a/ietf/ietfauth/tests.py +++ b/ietf/ietfauth/tests.py @@ -1,15 +1,12 @@ -# Copyright The IETF Trust 2009-2022, All Rights Reserved +# Copyright The IETF Trust 2009-2023, All Rights Reserved # -*- coding: utf-8 -*- import datetime -import io import logging # pyflakes:ignore -import os import re import requests import requests_mock -import shutil import time import urllib @@ -21,106 +18,102 @@ from oic.utils.authn.client import CLIENT_AUTHN_METHOD from oidc_provider.models import RSAKey from pyquery import PyQuery -from unittest import skipIf from urllib.parse import urlsplit +import django.core.signing from django.urls import reverse as urlreverse from django.contrib.auth.models import User from django.conf import settings -from django.template.loader import render_to_string +from django.template.loader import render_to_string from django.utils import timezone import debug # pyflakes:ignore from ietf.group.factories import GroupFactory, RoleFactory from ietf.group.models import Group, Role, RoleName -from ietf.ietfauth.htpasswd import update_htpasswd_file from ietf.ietfauth.utils import has_role -from ietf.mailinglists.models import Subscribed -from ietf.meeting.factories import MeetingFactory +from ietf.meeting.factories import MeetingFactory, RegistrationFactory, RegistrationTicketFactory from ietf.nomcom.factories import NomComFactory -from ietf.person.factories import PersonFactory, EmailFactory, UserFactory -from ietf.person.models import Person, Email, PersonalApiKey +from ietf.person.factories import PersonFactory, EmailFactory, UserFactory, PersonalApiKeyFactory +from ietf.person.models import Person, Email +from ietf.person.tasks import send_apikey_usage_emails_task from ietf.review.factories import ReviewRequestFactory, ReviewAssignmentFactory from ietf.review.models import ReviewWish, UnavailablePeriod -from ietf.stats.models import MeetingRegistration -from ietf.utils.decorators import skip_coverage from ietf.utils.mail import outbox, empty_outbox, get_payload_text from ietf.utils.test_utils import TestCase, login_testing_unauthorized from ietf.utils.timezone import date_today -import ietf.ietfauth.views - -if os.path.exists(settings.HTPASSWD_COMMAND): - skip_htpasswd_command = False - skip_message = "" -else: - skip_htpasswd_command = True - skip_message = ("Skipping htpasswd test: The binary for htpasswd wasn't found in the\n " - "location indicated in settings.py.") - print(" "+skip_message) - class IetfAuthTests(TestCase): - def setUp(self): - super().setUp() - self.saved_use_python_htdigest = getattr(settings, "USE_PYTHON_HTDIGEST", None) - settings.USE_PYTHON_HTDIGEST = True - - self.saved_htpasswd_file = settings.HTPASSWD_FILE - self.htpasswd_dir = self.tempdir('htpasswd') - settings.HTPASSWD_FILE = os.path.join(self.htpasswd_dir, "htpasswd") - io.open(settings.HTPASSWD_FILE, 'a').close() # create empty file - - self.saved_htdigest_realm = getattr(settings, "HTDIGEST_REALM", None) - settings.HTDIGEST_REALM = "test-realm" - - def tearDown(self): - shutil.rmtree(self.htpasswd_dir) - settings.USE_PYTHON_HTDIGEST = self.saved_use_python_htdigest - settings.HTPASSWD_FILE = self.saved_htpasswd_file - settings.HTDIGEST_REALM = self.saved_htdigest_realm - super().tearDown() def test_index(self): - self.assertEqual(self.client.get(urlreverse(ietf.ietfauth.views.index)).status_code, 200) + self.assertEqual(self.client.get(urlreverse("ietf.ietfauth.views.index")).status_code, 200) def test_login_and_logout(self): PersonFactory(user__username='plain') # try logging in without a next - r = self.client.get(urlreverse(ietf.ietfauth.views.login)) + r = self.client.get(urlreverse("ietf.ietfauth.views.login")) self.assertEqual(r.status_code, 200) - r = self.client.post(urlreverse(ietf.ietfauth.views.login), {"username":"plain", "password":"plain+password"}) + r = self.client.post(urlreverse("ietf.ietfauth.views.login"), {"username":"plain", "password":"plain+password"}) self.assertEqual(r.status_code, 302) - self.assertEqual(urlsplit(r["Location"])[2], urlreverse(ietf.ietfauth.views.profile)) + self.assertEqual(urlsplit(r["Location"])[2], urlreverse("ietf.ietfauth.views.profile")) # try logging out - r = self.client.get(urlreverse('django.contrib.auth.views.logout')) + r = self.client.post(urlreverse('django.contrib.auth.views.logout'), {}) self.assertEqual(r.status_code, 200) self.assertNotContains(r, "accounts/logout") - r = self.client.get(urlreverse(ietf.ietfauth.views.profile)) + r = self.client.get(urlreverse("ietf.ietfauth.views.profile")) self.assertEqual(r.status_code, 302) - self.assertEqual(urlsplit(r["Location"])[2], urlreverse(ietf.ietfauth.views.login)) + self.assertEqual(urlsplit(r["Location"])[2], urlreverse("ietf.ietfauth.views.login")) # try logging in with a next - r = self.client.post(urlreverse(ietf.ietfauth.views.login) + "?next=/foobar", {"username":"plain", "password":"plain+password"}) + r = self.client.post(urlreverse("ietf.ietfauth.views.login") + "?next=/foobar", {"username":"plain", "password":"plain+password"}) self.assertEqual(r.status_code, 302) self.assertEqual(urlsplit(r["Location"])[2], "/foobar") + def test_login_button(self): + PersonFactory(user__username='plain') + + def _test_login(url): + # try mashing the sign-in button repeatedly + r = self.client.get(url) + if r.status_code == 302: + r = self.client.get(r["Location"]) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + login_url = q("a:Contains('Sign in')").attr("href") + self.assertEqual(login_url, "/accounts/login/?next=" + url) + r = self.client.get(login_url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + login_url = q("a:Contains('Sign in')").attr("href") + self.assertEqual(login_url, "/accounts/login/?next=" + url) + + # try logging in with the provided next + r = self.client.post(login_url, {"username":"plain", "password":"plain+password"}) + self.assertEqual(r.status_code, 302) + self.assertEqual(urlsplit(r["Location"])[2], url) + self.client.logout() + + # try with a trivial next + _test_login("/") + # try with a next that requires login + _test_login(urlreverse("ietf.ietfauth.views.profile")) + def test_login_with_different_email(self): person = PersonFactory(user__username='plain') email = EmailFactory(person=person) # try logging in without a next - r = self.client.get(urlreverse(ietf.ietfauth.views.login)) + r = self.client.get(urlreverse("ietf.ietfauth.views.login")) self.assertEqual(r.status_code, 200) - r = self.client.post(urlreverse(ietf.ietfauth.views.login), {"username":email, "password":"plain+password"}) + r = self.client.post(urlreverse("ietf.ietfauth.views.login"), {"username":email, "password":"plain+password"}) self.assertEqual(r.status_code, 302) - self.assertEqual(urlsplit(r["Location"])[2], urlreverse(ietf.ietfauth.views.profile)) + self.assertEqual(urlsplit(r["Location"])[2], urlreverse("ietf.ietfauth.views.profile")) def extract_confirm_url(self, confirm_email): # dig out confirm_email link @@ -134,20 +127,11 @@ def extract_confirm_url(self, confirm_email): return confirm_url - def username_in_htpasswd_file(self, username): - with io.open(settings.HTPASSWD_FILE) as f: - for l in f: - if l.startswith(username + ":"): - return True - with io.open(settings.HTPASSWD_FILE) as f: - print(f.read()) - - return False # For the lowered barrier to account creation period, we are disabling this kind of failure # def test_create_account_failure(self): - # url = urlreverse(ietf.ietfauth.views.create_account) + # url = urlreverse("ietf.ietfauth.views.create_account") # # get # r = self.client.get(url) @@ -165,8 +149,8 @@ def test_create_account_failure_template(self): r = render_to_string('registration/manual.html', { 'account_request_email': settings.ACCOUNT_REQUEST_EMAIL }) self.assertTrue("Additional Assistance Required" in r) - def register_and_verify(self, email): - url = urlreverse(ietf.ietfauth.views.create_account) + def register(self, email): + url = urlreverse("ietf.ietfauth.views.create_account") # register email empty_outbox() @@ -175,59 +159,63 @@ def register_and_verify(self, email): self.assertContains(r, "Account request received") self.assertEqual(len(outbox), 1) + def register_and_verify(self, email): + self.register(email) + # go to confirm page confirm_url = self.extract_confirm_url(outbox[-1]) r = self.client.get(confirm_url) self.assertEqual(r.status_code, 200) # password mismatch - r = self.client.post(confirm_url, { 'password': 'secret', 'password_confirmation': 'nosecret' }) + r = self.client.post( + confirm_url, { + "password": "secret-and-secure", + "password_confirmation": "not-secret-or-secure", + } + ) + self.assertEqual(r.status_code, 200) + self.assertEqual(User.objects.filter(username=email).count(), 0) + + # weak password + r = self.client.post( + confirm_url, { + "password": "password1234", + "password_confirmation": "password1234", + } + ) self.assertEqual(r.status_code, 200) self.assertEqual(User.objects.filter(username=email).count(), 0) # confirm - r = self.client.post(confirm_url, { 'name': 'User Name', 'ascii': 'User Name', 'password': 'secret', 'password_confirmation': 'secret' }) + r = self.client.post( + confirm_url, + { + "name": "User Name", + "ascii": "User Name", + "password": "secret-and-secure", + "password_confirmation": "secret-and-secure", + }, + ) self.assertEqual(r.status_code, 200) self.assertEqual(User.objects.filter(username=email).count(), 1) self.assertEqual(Person.objects.filter(user__username=email).count(), 1) self.assertEqual(Email.objects.filter(person__user__username=email).count(), 1) - self.assertTrue(self.username_in_htpasswd_file(email)) - - def test_create_allowlisted_account(self): + # This also tests new account creation. + def test_create_existing_account(self): + # create account once email = "new-account@example.com" - - # add allowlist entry - r = self.client.post(urlreverse(ietf.ietfauth.views.login), {"username":"secretary", "password":"secretary+password"}) - self.assertEqual(r.status_code, 302) - self.assertEqual(urlsplit(r["Location"])[2], urlreverse(ietf.ietfauth.views.profile)) - - r = self.client.get(urlreverse(ietf.ietfauth.views.add_account_allowlist)) - self.assertEqual(r.status_code, 200) - self.assertContains(r, "Add an allowlist entry") - - r = self.client.post(urlreverse(ietf.ietfauth.views.add_account_allowlist), {"email": email}) - self.assertEqual(r.status_code, 200) - self.assertContains(r, "Allowlist entry creation successful") - - # log out - r = self.client.get(urlreverse('django.contrib.auth.views.logout')) - self.assertEqual(r.status_code, 200) - - # register and verify allowlisted email self.register_and_verify(email) + # create account again + self.register(email) - def test_create_subscribed_account(self): - # verify creation with email in subscribed list - saved_delay = settings.LIST_ACCOUNT_DELAY - settings.LIST_ACCOUNT_DELAY = 1 - email = "subscribed@example.com" - s = Subscribed(email=email) - s.save() - time.sleep(1.1) - self.register_and_verify(email) - settings.LIST_ACCOUNT_DELAY = saved_delay + # check notification + note = get_payload_text(outbox[-1]) + self.assertIn(email, note) + self.assertIn("A datatracker account for that email already exists", note) + self.assertIn(urlreverse("ietf.ietfauth.views.password_reset"), note) def test_ietfauth_profile(self): EmailFactory(person__user__username='plain') @@ -236,7 +224,7 @@ def test_ietfauth_profile(self): username = "plain" email_address = Email.objects.filter(person__user__username=username).first().address - url = urlreverse(ietf.ietfauth.views.profile) + url = urlreverse("ietf.ietfauth.views.profile") login_testing_unauthorized(self, username, url) @@ -317,11 +305,14 @@ def test_ietfauth_profile(self): self.assertEqual(r.status_code, 200) self.assertEqual(Email.objects.filter(address=new_email_address, person__user__username=username, active=1).count(), 1) - # check that we can't re-add it - that would give a duplicate - r = self.client.get(confirm_url) + # try and add it again + empty_outbox() + r = self.client.post(url, with_new_email_address) self.assertEqual(r.status_code, 200) - q = PyQuery(r.content) - self.assertEqual(len(q('[name="action"][value="confirm"]')), 0) + self.assertEqual(len(outbox), 1) + note = get_payload_text(outbox[-1]) + self.assertIn(new_email_address, note) + self.assertIn("already associated with your account", note) pronoundish = base_data.copy() pronoundish["pronouns_freetext"] = "baz/boom" @@ -381,6 +372,21 @@ def test_ietfauth_profile(self): self.assertEqual(len(updated_roles), 1) self.assertEqual(updated_roles[0].email_id, new_email_address) + def test_email_case_insensitive_protection(self): + EmailFactory(address="TestAddress@example.net") + person = PersonFactory() + url = urlreverse("ietf.ietfauth.views.profile") + login_testing_unauthorized(self, person.user.username, url) + + data = { + "name": person.name, + "plain": person.plain, + "ascii": person.ascii, + "active_emails": [e.address for e in person.email_set.filter(active=True)], + "new_email": "testaddress@example.net", + } + r = self.client.post(url, data) + self.assertContains(r, "A confirmation email has been sent to", status_code=200) def test_nomcom_dressing_on_profile(self): url = urlreverse('ietf.ietfauth.views.profile') @@ -408,38 +414,39 @@ def test_nomcom_dressing_on_profile(self): self.assertFalse(q('#volunteer-button')) self.assertTrue(q('#volunteered')) - def test_reset_password(self): - url = urlreverse(ietf.ietfauth.views.password_reset) - email = 'someone@example.com' - password = 'foobar' - - user = User.objects.create(username=email, email=email) - user.set_password(password) + WEAK_PASSWORD="password1234" + VALID_PASSWORD = "complex-and-long-valid-password" + ANOTHER_VALID_PASSWORD = "very-complicated-and-lengthy-password" + url = urlreverse("ietf.ietfauth.views.password_reset") + email = "someone@example.com" + + user = PersonFactory(user__email=email).user + user.set_password(VALID_PASSWORD) user.save() - p = Person.objects.create(name="Some One", ascii="Some One", user=user) - Email.objects.create(address=user.username, person=p, origin=user.username) - + # get r = self.client.get(url) self.assertEqual(r.status_code, 200) - # ask for reset, wrong username - r = self.client.post(url, { 'username': "nobody@example.com" }) + # ask for reset, wrong username (form should not fail) + r = self.client.post(url, {"username": "nobody@example.com"}) self.assertEqual(r.status_code, 200) q = PyQuery(r.content) - self.assertTrue(len(q("form .is-invalid")) > 0) + self.assertTrue(len(q("form .is-invalid")) == 0) # ask for reset empty_outbox() - r = self.client.post(url, { 'username': user.username }) + r = self.client.post(url, {"username": user.username}) self.assertEqual(r.status_code, 200) self.assertEqual(len(outbox), 1) # goto change password page, logged in as someone else confirm_url = self.extract_confirm_url(outbox[-1]) other_user = UserFactory() - self.client.login(username=other_user.username, password=other_user.username + '+password') + self.client.login( + username=other_user.username, password=other_user.username + "+password" + ) r = self.client.get(confirm_url) self.assertEqual(r.status_code, 403) @@ -448,21 +455,47 @@ def test_reset_password(self): r = self.client.get(confirm_url) self.assertEqual(r.status_code, 200) q = PyQuery(r.content) - self.assertNotIn(user.username, q('.nav').text(), - 'user should not appear signed in while resetting password') + self.assertNotIn( + user.username, + q(".nav").text(), + "user should not appear signed in while resetting password", + ) # password mismatch - r = self.client.post(confirm_url, { 'password': 'secret', 'password_confirmation': 'nosecret' }) + r = self.client.post( + confirm_url, + { + "password": ANOTHER_VALID_PASSWORD, + "password_confirmation": ANOTHER_VALID_PASSWORD[::-1], + }, + ) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertTrue(len(q("form .is-invalid")) > 0) + + # weak password + r = self.client.post( + confirm_url, + { + "password": WEAK_PASSWORD, + "password_confirmation": WEAK_PASSWORD, + }, + ) self.assertEqual(r.status_code, 200) q = PyQuery(r.content) self.assertTrue(len(q("form .is-invalid")) > 0) # confirm - r = self.client.post(confirm_url, { 'password': 'secret', 'password_confirmation': 'secret' }) + r = self.client.post( + confirm_url, + { + "password": ANOTHER_VALID_PASSWORD, + "password_confirmation": ANOTHER_VALID_PASSWORD, + }, + ) self.assertEqual(r.status_code, 200) q = PyQuery(r.content) self.assertEqual(len(q("form .is-invalid")), 0) - self.assertTrue(self.username_in_htpasswd_file(user.username)) # reuse reset url r = self.client.get(confirm_url) @@ -470,15 +503,18 @@ def test_reset_password(self): # login after reset request empty_outbox() - user.set_password(password) + user.set_password(VALID_PASSWORD) user.save() - r = self.client.post(url, { 'username': user.username }) + r = self.client.post(url, {"username": user.username}) self.assertEqual(r.status_code, 200) self.assertEqual(len(outbox), 1) confirm_url = self.extract_confirm_url(outbox[-1]) - r = self.client.post(urlreverse(ietf.ietfauth.views.login), {'username': email, 'password': password}) + r = self.client.post( + urlreverse("ietf.ietfauth.views.login"), + {"username": email, "password": VALID_PASSWORD}, + ) r = self.client.get(confirm_url) self.assertEqual(r.status_code, 404) @@ -486,17 +522,83 @@ def test_reset_password(self): # change password after reset request empty_outbox() - r = self.client.post(url, { 'username': user.username }) + r = self.client.post(url, {"username": user.username}) self.assertEqual(r.status_code, 200) self.assertEqual(len(outbox), 1) confirm_url = self.extract_confirm_url(outbox[-1]) - user.set_password('newpassword') + user.set_password(ANOTHER_VALID_PASSWORD) user.save() r = self.client.get(confirm_url) self.assertEqual(r.status_code, 404) + def test_reset_password_without_person(self): + """No password reset for account without a person""" + url = urlreverse('ietf.ietfauth.views.password_reset') + user = UserFactory() + user.set_password('some password') + user.save() + empty_outbox() + r = self.client.post(url, { 'username': user.username}) + self.assertContains(r, 'We have sent you an email with instructions', status_code=200) + q = PyQuery(r.content) + self.assertTrue(len(q("form .is-invalid")) == 0) + self.assertEqual(len(outbox), 0) + + def test_reset_password_address_handling(self): + """Reset password links are only sent to known, active addresses""" + url = urlreverse('ietf.ietfauth.views.password_reset') + person = PersonFactory() + person.email_set.update(active=False) + empty_outbox() + r = self.client.post(url, { 'username': person.user.username}) + self.assertContains(r, 'We have sent you an email with instructions', status_code=200) + q = PyQuery(r.content) + self.assertTrue(len(q("form .is-invalid")) == 0) + self.assertEqual(len(outbox), 0) + + active_address = EmailFactory(person=person).address + r = self.client.post(url, {'username': person.user.username}) + self.assertContains(r, 'We have sent you an email with instructions', status_code=200) + self.assertEqual(len(outbox), 1) + to = outbox[0].get('To') + self.assertIn(active_address, to) + self.assertNotIn(person.user.username, to) + + def test_reset_password_without_username(self): + """Reset password using non-username email address""" + url = urlreverse('ietf.ietfauth.views.password_reset') + person = PersonFactory() + secondary_address = EmailFactory(person=person).address + inactive_secondary_address = EmailFactory(person=person, active=False).address + empty_outbox() + r = self.client.post(url, { 'username': secondary_address}) + self.assertContains(r, 'We have sent you an email with instructions', status_code=200) + self.assertEqual(len(outbox), 1) + to = outbox[0].get('To') + self.assertIn(person.user.username, to) + self.assertIn(secondary_address, to) + self.assertNotIn(inactive_secondary_address, to) + + def test_reset_password_without_user(self): + """Reset password using email address for person without a user account""" + url = urlreverse('ietf.ietfauth.views.password_reset') + email = EmailFactory() + person = email.person + # Remove the user object from the person to get a Email/Person without User: + person.user = None + person.save() + # Remove the remaining User record, since reset_password looks for that by username: + User.objects.filter(username__iexact=email.address).delete() + empty_outbox() + r = self.client.post(url, { 'username': email.address }) + self.assertEqual(len(outbox), 1) + lastReceivedEmail = outbox[-1] + self.assertIn(email.address, lastReceivedEmail.get('To')) + self.assertTrue(lastReceivedEmail.get('Subject').startswith("Confirm password reset")) + self.assertContains(r, "Your password reset request has been successfully received", status_code=200) + def test_review_overview(self): review_req = ReviewRequestFactory() assignment = ReviewAssignmentFactory(review_request=review_req,reviewer=EmailFactory(person__user__username='reviewer')) @@ -512,7 +614,7 @@ def test_review_overview(self): availability="unavailable", ) - url = urlreverse(ietf.ietfauth.views.review_overview) + url = urlreverse("ietf.ietfauth.views.review_overview") login_testing_unauthorized(self, reviewer.user.username, url) @@ -538,117 +640,176 @@ def test_review_overview(self): self.assertEqual(r.status_code, 302) self.assertEqual(ReviewWish.objects.filter(doc=doc, team=review_req.team).count(), 0) - def test_htpasswd_file_with_python(self): - # make sure we test both Python and call-out to binary - settings.USE_PYTHON_HTDIGEST = True - - update_htpasswd_file("foo", "passwd") - self.assertTrue(self.username_in_htpasswd_file("foo")) - - @skipIf(skip_htpasswd_command, skip_message) - @skip_coverage - def test_htpasswd_file_with_htpasswd_binary(self): - # make sure we test both Python and call-out to binary - settings.USE_PYTHON_HTDIGEST = False - - update_htpasswd_file("foo", "passwd") - self.assertTrue(self.username_in_htpasswd_file("foo")) - - def test_change_password(self): - - chpw_url = urlreverse(ietf.ietfauth.views.change_password) - prof_url = urlreverse(ietf.ietfauth.views.profile) - login_url = urlreverse(ietf.ietfauth.views.login) - redir_url = '%s?next=%s' % (login_url, chpw_url) + VALID_PASSWORD = "complex-and-long-valid-password" + ANOTHER_VALID_PASSWORD = "very-complicated-and-lengthy-password" + chpw_url = urlreverse("ietf.ietfauth.views.change_password") + prof_url = urlreverse("ietf.ietfauth.views.profile") + login_url = urlreverse("ietf.ietfauth.views.login") + redir_url = "%s?next=%s" % (login_url, chpw_url) # get without logging in r = self.client.get(chpw_url) self.assertRedirects(r, redir_url) - user = User.objects.create(username="someone@example.com", email="someone@example.com") - user.set_password("password") + user = User.objects.create( + username="someone@example.com", email="someone@example.com" + ) + user.set_password(VALID_PASSWORD) user.save() p = Person.objects.create(name="Some One", ascii="Some One", user=user) Email.objects.create(address=user.username, person=p, origin=user.username) # log in - r = self.client.post(redir_url, {"username":user.username, "password":"password"}) + r = self.client.post( + redir_url, {"username": user.username, "password": VALID_PASSWORD} + ) self.assertRedirects(r, chpw_url) # wrong current password - r = self.client.post(chpw_url, {"current_password": "fiddlesticks", - "new_password": "foobar", - "new_password_confirmation": "foobar", - }) + r = self.client.post( + chpw_url, + { + "current_password": "fiddlesticks", + "password": ANOTHER_VALID_PASSWORD, + "password_confirmation": ANOTHER_VALID_PASSWORD, + }, + ) self.assertEqual(r.status_code, 200) - self.assertFormError(r, 'form', 'current_password', 'Invalid password') + self.assertFormError(r.context["form"], "current_password", "Invalid password") # mismatching new passwords - r = self.client.post(chpw_url, {"current_password": "password", - "new_password": "foobar", - "new_password_confirmation": "barfoo", - }) + r = self.client.post( + chpw_url, + { + "current_password": VALID_PASSWORD, + "password": ANOTHER_VALID_PASSWORD, + "password_confirmation": ANOTHER_VALID_PASSWORD[::-1], + }, + ) self.assertEqual(r.status_code, 200) - self.assertFormError(r, 'form', None, "The password confirmation is different than the new password") + self.assertFormError( + r.context["form"], + "password_confirmation", + "The password confirmation is different than the new password", + ) + + # password too short + r = self.client.post( + chpw_url, + { + "current_password": VALID_PASSWORD, + "password": "sh0rtpw0rd", + "password_confirmation": "sh0rtpw0rd", + } + ) + self.assertEqual(r.status_code, 200) + self.assertFormError( + r.context["form"], + "password", + "This password is too short. It must contain at least " + f"{settings.PASSWORD_POLICY_MIN_LENGTH} characters." + ) + + # password too simple + r = self.client.post( + chpw_url, + { + "current_password": VALID_PASSWORD, + "password": "passwordpassword", + "password_confirmation": "passwordpassword", + } + ) + self.assertEqual(r.status_code, 200) + self.assertFormError( + r.context["form"], + "password", + "This password does not meet complexity requirements " + "and is easily guessable." + ) # correct password change - r = self.client.post(chpw_url, {"current_password": "password", - "new_password": "foobar", - "new_password_confirmation": "foobar", - }) + r = self.client.post( + chpw_url, + { + "current_password": VALID_PASSWORD, + "password": ANOTHER_VALID_PASSWORD, + "password_confirmation": ANOTHER_VALID_PASSWORD, + }, + ) self.assertRedirects(r, prof_url) # refresh user object user = User.objects.get(username="someone@example.com") - self.assertTrue(user.check_password('foobar')) + self.assertTrue(user.check_password(ANOTHER_VALID_PASSWORD)) def test_change_username(self): - - chun_url = urlreverse(ietf.ietfauth.views.change_username) - prof_url = urlreverse(ietf.ietfauth.views.profile) - login_url = urlreverse(ietf.ietfauth.views.login) - redir_url = '%s?next=%s' % (login_url, chun_url) + VALID_PASSWORD = "complex-and-long-valid-password" + chun_url = urlreverse("ietf.ietfauth.views.change_username") + prof_url = urlreverse("ietf.ietfauth.views.profile") + login_url = urlreverse("ietf.ietfauth.views.login") + redir_url = "%s?next=%s" % (login_url, chun_url) # get without logging in r = self.client.get(chun_url) self.assertRedirects(r, redir_url) - user = User.objects.create(username="someone@example.com", email="someone@example.com") - user.set_password("password") + user = User.objects.create( + username="someone@example.com", email="someone@example.com" + ) + user.set_password(VALID_PASSWORD) user.save() p = Person.objects.create(name="Some One", ascii="Some One", user=user) Email.objects.create(address=user.username, person=p, origin=user.username) - Email.objects.create(address="othername@example.org", person=p, origin=user.username) + Email.objects.create( + address="othername@example.org", person=p, origin=user.username + ) # log in - r = self.client.post(redir_url, {"username":user.username, "password":"password"}) + r = self.client.post( + redir_url, {"username": user.username, "password": VALID_PASSWORD} + ) self.assertRedirects(r, chun_url) # wrong username - r = self.client.post(chun_url, {"username": "fiddlesticks", - "password": "password", - }) + r = self.client.post( + chun_url, + { + "username": "fiddlesticks", + "password": VALID_PASSWORD, + }, + ) self.assertEqual(r.status_code, 200) - self.assertFormError(r, 'form', 'username', - "Select a valid choice. fiddlesticks is not one of the available choices.") + self.assertFormError( + r.context["form"], + "username", + "Select a valid choice. fiddlesticks is not one of the available choices.", + ) # wrong password - r = self.client.post(chun_url, {"username": "othername@example.org", - "password": "foobar", - }) + r = self.client.post( + chun_url, + { + "username": "othername@example.org", + "password": "foobar", + }, + ) self.assertEqual(r.status_code, 200) - self.assertFormError(r, 'form', 'password', 'Invalid password') + self.assertFormError(r.context["form"], "password", "Invalid password") # correct username change - r = self.client.post(chun_url, {"username": "othername@example.org", - "password": "password", - }) + r = self.client.post( + chun_url, + { + "username": "othername@example.org", + "password": VALID_PASSWORD, + }, + ) self.assertRedirects(r, prof_url) # refresh user object prev = user user = User.objects.get(username="othername@example.org") self.assertEqual(prev, user) - self.assertTrue(user.check_password('password')) + self.assertTrue(user.check_password(VALID_PASSWORD)) def test_apikey_management(self): # Create a person with a role that will give at least one valid apikey @@ -692,8 +853,20 @@ def test_apikey_management(self): url = urlreverse('ietf.ietfauth.views.apikey_disable') r = self.client.get(url) + self.assertEqual(r.status_code, 200) self.assertContains(r, 'Disable a personal API key') self.assertContains(r, 'Key') + + # Try to delete something that doesn't exist + r = self.client.post(url, {'hash': key.hash()+'bad'}) + self.assertEqual(r.status_code, 200) + self.assertContains(r,"Key validation failed; key not disabled") + + # Try to delete someone else's key + otherkey = PersonalApiKeyFactory() + r = self.client.post(url, {'hash': otherkey.hash()}) + self.assertEqual(r.status_code, 200) + self.assertContains(r,"Key validation failed; key not disabled") # Delete a key r = self.client.post(url, {'hash': key.hash()}) @@ -746,9 +919,8 @@ def test_apikey_errors(self): self.assertContains(r, 'Invalid apikey', status_code=403) # invalid apikey (invalidated api key) - unauthorized_url = urlreverse('ietf.api.views.app_auth') - invalidated_apikey = PersonalApiKey.objects.create( - endpoint=unauthorized_url, person=person, valid=False) + unauthorized_url = urlreverse('ietf.api.views.app_auth', kwargs={'app': 'authortools'}) + invalidated_apikey = PersonalApiKeyFactory(endpoint=unauthorized_url, person=person, valid=False) r = self.client.post(unauthorized_url, {'apikey': invalidated_apikey.hash()}) self.assertContains(r, 'Invalid apikey', status_code=403) @@ -761,15 +933,16 @@ def test_apikey_errors(self): person.user.save() # endpoint mismatch - key2 = PersonalApiKey.objects.create(person=person, endpoint='/') + key2 = PersonalApiKeyFactory( + person=person, + endpoint='/', + validate_model=False, # allow invalid endpoint + ) r = self.client.post(key.endpoint, {'apikey':key2.hash(), 'dummy':'dummy',}) self.assertContains(r, 'Apikey endpoint mismatch', status_code=400) key2.delete() def test_send_apikey_report(self): - from ietf.ietfauth.management.commands.send_apikey_usage_emails import Command - from ietf.utils.mail import outbox, empty_outbox - person = RoleFactory(name_id='secr', group__acronym='secretariat').person url = urlreverse('ietf.ietfauth.views.apikey_create') @@ -794,9 +967,8 @@ def test_send_apikey_report(self): date = str(date_today()) empty_outbox() - cmd = Command() - cmd.handle(verbosity=0, days=7) - + send_apikey_usage_emails_task(days=7) + self.assertEqual(len(outbox), len(endpoints)) for mail in outbox: body = get_payload_text(mail) @@ -844,6 +1016,72 @@ def test_edit_person_extresources(self): self.assertEqual(person.personextresource_set.get(name__slug='github_repo').display_name, 'Some display text') self.assertIn(person.personextresource_set.first().name.slug, str(person.personextresource_set.first())) + def test_confirm_new_email(self): + person = PersonFactory() + valid_auth = django.core.signing.dumps( + [person.user.username, "new_email@example.com"], salt="add_email" + ) + invalid_auth = django.core.signing.dumps( + [person.user.username, "not_this_one@example.com"], salt="pepper" + ) + + # Test that we check the salt + r = self.client.get( + urlreverse("ietf.ietfauth.views.confirm_new_email", kwargs={"auth": invalid_auth}) + ) + self.assertEqual(r.status_code, 404) + r = self.client.post( + urlreverse("ietf.ietfauth.views.confirm_new_email", kwargs={"auth": invalid_auth}) + ) + self.assertEqual(r.status_code, 404) + + # Now check that the valid auth works + self.assertFalse( + person.email_set.filter(address__icontains="new_email@example.com").exists() + ) + confirm_url = urlreverse( + "ietf.ietfauth.views.confirm_new_email", kwargs={"auth": valid_auth} + ) + r = self.client.get(confirm_url) + self.assertContains(r, urllib.parse.quote(confirm_url), status_code=200) + r = self.client.post(confirm_url, data={"action": "confirm"}) + self.assertContains(r, "has been updated", status_code=200) + self.assertTrue( + person.email_set.filter(address__icontains="new_email@example.com").exists() + ) + + # Authorizing a second time should be handled gracefully + r = self.client.post(confirm_url, data={"action": "confirm"}) + self.assertContains(r, "already includes", status_code=200) + + # Another person should not be able to add the same address and should be told so, + # whether they use the same or different letter case + other_person = PersonFactory() + other_auth = django.core.signing.dumps( + [other_person.user.username, "new_email@example.com"], salt="add_email" + ) + r = self.client.post( + urlreverse("ietf.ietfauth.views.confirm_new_email", kwargs={"auth": other_auth}), + data={"action": "confirm"}, + ) + self.assertContains(r, "in use by another user", status_code=200) + + other_auth = django.core.signing.dumps( + [other_person.user.username, "NeW_eMaIl@eXaMpLe.CoM"], salt="add_email" + ) + r = self.client.post( + urlreverse("ietf.ietfauth.views.confirm_new_email", kwargs={"auth": other_auth}), + data={"action": "confirm"}, + ) + + self.assertContains(r, "in use by another user", status_code=200) + self.assertFalse( + other_person.email_set.filter(address__icontains="new_email@example.com").exists() + ) + self.assertTrue( + person.email_set.filter(address__icontains="new_email@example.com").exists() + ) + class OpenIDConnectTests(TestCase): def request_matcher(self, request): @@ -909,11 +1147,15 @@ def test_oidc_code_auth(self): EmailFactory(person=person) email_list = person.email_set.all().values_list('address', flat=True) meeting = MeetingFactory(type_id='ietf', date=date_today()) - MeetingRegistration.objects.create( - meeting=meeting, person=None, first_name=person.first_name(), last_name=person.last_name(), - email=email_list[0], ticket_type='full_week', reg_type='remote', affiliation='Some Company', - ) - + reg_person = RegistrationFactory( + meeting=meeting, + person=person, + first_name=person.first_name(), + last_name=person.last_name(), + email=email_list[0], + affiliation='Some Company', + with_ticket={'attendance_type_id': 'remote', 'ticket_type_id': 'week_pass'}, + ) # Get access authorisation session = {} session["state"] = rndstr() @@ -966,35 +1208,48 @@ def test_oidc_code_auth(self): for key in ['iss', 'sub', 'aud', 'exp', 'iat', 'auth_time', 'nonce', 'at_hash']: self.assertIn(key, access_token_info['id_token']) - # Get userinfo, check keys present + # Get userinfo, check keys present, most common scenario userinfo = client.do_user_info_request(state=params["state"], scope=args['scope']) for key in [ 'email', 'family_name', 'given_name', 'meeting', 'name', 'pronouns', 'roles', 'ticket_type', 'reg_type', 'affiliation', 'picture', 'dots', ]: self.assertIn(key, userinfo) self.assertTrue(userinfo[key]) self.assertIn('remote', set(userinfo['reg_type'].split())) - self.assertNotIn('hackathon', set(userinfo['reg_type'].split())) + self.assertNotIn('hackathon_onsite', set(userinfo['reg_type'].split())) self.assertIn(active_group.acronym, [i[1] for i in userinfo['roles']]) self.assertNotIn(closed_group.acronym, [i[1] for i in userinfo['roles']]) - # Create another registration, with a different email - MeetingRegistration.objects.create( - meeting=meeting, person=None, first_name=person.first_name(), last_name=person.last_name(), - email=email_list[1], ticket_type='one_day', reg_type='hackathon', affiliation='Some Company, Inc', - ) + # Create a registration, with only email, no person (rare if at all) + reg_person.delete() + reg_email = RegistrationFactory( + meeting=meeting, + person=None, + first_name=person.first_name(), + last_name=person.last_name(), + email=email_list[1], + affiliation='Some Company, Inc', + with_ticket={'attendance_type_id': 'hackathon_onsite', 'ticket_type_id': 'one_day'}, + ) userinfo = client.do_user_info_request(state=params["state"], scope=args['scope']) - self.assertIn('hackathon', set(userinfo['reg_type'].split())) - self.assertIn('remote', set(userinfo['reg_type'].split())) - self.assertIn('full_week', set(userinfo['ticket_type'].split())) - self.assertIn('Some Company', userinfo['affiliation']) - - # Create a third registration, with a composite reg type - MeetingRegistration.objects.create( - meeting=meeting, person=None, first_name=person.first_name(), last_name=person.last_name(), - email=email_list[1], ticket_type='one_day', reg_type='hackathon remote', affiliation='Some Company, Inc', - ) + self.assertIn('hackathon_onsite', set(userinfo['reg_type'].split())) + self.assertNotIn('remote', set(userinfo['reg_type'].split())) + self.assertIn('one_day', set(userinfo['ticket_type'].split())) + self.assertIn('Some Company, Inc', userinfo['affiliation']) + + # Test with multiple tickets + reg_email.delete() + creg = RegistrationFactory( + meeting=meeting, + person=None, + first_name=person.first_name(), + last_name=person.last_name(), + email=email_list[1], + affiliation='Some Company, Inc', + with_ticket={'attendance_type_id': 'hackathon_remote', 'ticket_type_id': 'week_pass'}, + ) + RegistrationTicketFactory(registration=creg, attendance_type_id='remote', ticket_type_id='week_pass') userinfo = client.do_user_info_request(state=params["state"], scope=args['scope']) - self.assertEqual(set(userinfo['reg_type'].split()), set(['remote', 'hackathon'])) + self.assertEqual(set(userinfo['reg_type'].split()), set(['remote', 'hackathon_remote'])) # Check that ending a session works r = client.do_end_session_request(state=params["state"], scope=args['scope']) diff --git a/ietf/ietfauth/urls.py b/ietf/ietfauth/urls.py index 56daae0535..7493fe5c97 100644 --- a/ietf/ietfauth/urls.py +++ b/ietf/ietfauth/urls.py @@ -14,7 +14,7 @@ url(r'^confirmnewemail/(?P[^/]+)/$', views.confirm_new_email), url(r'^create/$', views.create_account), url(r'^create/confirm/(?P[^/]+)/$', views.confirm_account), - url(r'^login/$', views.login), + url(r'^login/$', views.AnyEmailLoginView.as_view(), name="ietf.ietfauth.views.login"), url(r'^logout/$', LogoutView.as_view(), name="django.contrib.auth.views.logout"), url(r'^password/$', views.change_password), url(r'^profile/$', views.profile), @@ -24,5 +24,4 @@ url(r'^review/$', views.review_overview), url(r'^testemail/$', views.test_email), url(r'^username/$', views.change_username), - url(r'^allowlist/add/?$', views.add_account_allowlist), ] diff --git a/ietf/ietfauth/utils.py b/ietf/ietfauth/utils.py index 43353ca533..0df667fbd2 100644 --- a/ietf/ietfauth/utils.py +++ b/ietf/ietfauth/utils.py @@ -7,23 +7,25 @@ import oidc_provider.lib.claims -from functools import wraps +from functools import wraps, WRAPPER_ASSIGNMENTS +from urllib.parse import quote as urlquote from django.conf import settings from django.contrib.auth import REDIRECT_FIELD_NAME +from django.contrib.sites.models import Site +from django.core import signing from django.core.exceptions import PermissionDenied from django.db.models import Q from django.http import HttpResponseRedirect from django.shortcuts import get_object_or_404 -from django.utils.decorators import available_attrs -from django.utils.http import urlquote import debug # pyflakes:ignore from ietf.group.models import Role, GroupFeatures -from ietf.person.models import Person +from ietf.person.models import Email, Person from ietf.person.utils import get_dots from ietf.doc.utils_bofreq import bofreq_editors +from ietf.utils.mail import send_mail def user_is_person(user, person): """Test whether user is associated with person.""" @@ -39,9 +41,10 @@ def has_role(user, role_names, *args, **kwargs): """Determines whether user has any of the given standard roles given. Role names must be a list or, in case of a single value, a string.""" - if not isinstance(role_names, (list, tuple)): - role_names = [ role_names ] - + extra_role_qs = kwargs.get("extra_role_qs", None) + if not isinstance(role_names, (list, tuple, set)): + role_names = [role_names] + if not user or not user.is_authenticated: return False @@ -49,7 +52,13 @@ def has_role(user, role_names, *args, **kwargs): if not hasattr(user, "roles_check_cache"): user.roles_check_cache = {} - key = frozenset(role_names) + keynames = set(role_names) + if extra_role_qs: + keynames.update(set(extra_role_qs.keys())) + year = kwargs.get("year", None) + if year is not None: + keynames.add(f"nomcomyear{year}") + key = frozenset(keynames) if key not in user.roles_check_cache: try: person = user.person @@ -57,51 +66,123 @@ def has_role(user, role_names, *args, **kwargs): return False role_qs = { - "Area Director": Q(person=person, name__in=("pre-ad", "ad"), group__type="area", group__state="active"), - "Secretariat": Q(person=person, name="secr", group__acronym="secretariat"), - "IAB" : Q(person=person, name="member", group__acronym="iab"), - "IANA": Q(person=person, name="auth", group__acronym="iana"), - "RFC Editor": Q(person=person, name="auth", group__acronym="rpc"), - "ISE" : Q(person=person, name="chair", group__acronym="ise"), - "IAD": Q(person=person, name="admdir", group__acronym="ietf"), - "IETF Chair": Q(person=person, name="chair", group__acronym="ietf"), - "IETF Trust Chair": Q(person=person, name="chair", group__acronym="ietf-trust"), - "IRTF Chair": Q(person=person, name="chair", group__acronym="irtf"), - "IAB Chair": Q(person=person, name="chair", group__acronym="iab"), - "IAB Executive Director": Q(person=person, name="execdir", group__acronym="iab"), - "IAB Group Chair": Q(person=person, name="chair", group__type="iab", group__state="active"), - "IAOC Chair": Q(person=person, name="chair", group__acronym="iaoc"), - "WG Chair": Q(person=person,name="chair", group__type="wg", group__state__in=["active","bof", "proposed"]), - "WG Secretary": Q(person=person,name="secr", group__type="wg", group__state__in=["active","bof", "proposed"]), - "RG Chair": Q(person=person,name="chair", group__type="rg", group__state__in=["active","proposed"]), - "RG Secretary": Q(person=person,name="secr", group__type="rg", group__state__in=["active","proposed"]), - "AG Secretary": Q(person=person,name="secr", group__type="ag", group__state__in=["active"]), - "RAG Secretary": Q(person=person,name="secr", group__type="rag", group__state__in=["active"]), - "Team Chair": Q(person=person,name="chair", group__type="team", group__state="active"), - "Program Lead": Q(person=person,name="lead", group__type="program", group__state="active"), - "Program Secretary": Q(person=person,name="secr", group__type="program", group__state="active"), - "Program Chair": Q(person=person,name="chair", group__type="program", group__state="active"), - "Nomcom Chair": Q(person=person, name="chair", group__type="nomcom", group__acronym__icontains=kwargs.get('year', '0000')), - "Nomcom Advisor": Q(person=person, name="advisor", group__type="nomcom", group__acronym__icontains=kwargs.get('year', '0000')), - "Nomcom": Q(person=person, group__type="nomcom", group__acronym__icontains=kwargs.get('year', '0000')), - "Liaison Manager": Q(person=person,name="liaiman",group__type="sdo",group__state="active", ), - "Authorized Individual": Q(person=person,name="auth",group__type="sdo",group__state="active", ), - "Recording Manager": Q(person=person,name="recman",group__type="ietf",group__state="active", ), - "Reviewer": Q(person=person, name="reviewer", group__state="active"), - "Review Team Secretary": Q(person=person, name="secr", group__reviewteamsettings__isnull=False,group__state="active", ), - "IRSG Member": (Q(person=person, name="member", group__acronym="irsg") | Q(person=person, name="chair", group__acronym="irtf") | Q(person=person, name="atlarge", group__acronym="irsg")), - "Robot": Q(person=person, name="robot", group__acronym="secretariat"), - } - - filter_expr = Q(pk__in=[]) # ensure empty set is returned if no other terms are added + "Area Director": Q( + name__in=("pre-ad", "ad"), group__type="area", group__state="active" + ), + "Secretariat": Q(name="secr", group__acronym="secretariat"), + "IAB": Q(name="member", group__acronym="iab"), + "IANA": Q(name="auth", group__acronym="iana"), + "RFC Editor": Q(name="auth", group__acronym="rpc"), + "ISE": Q(name="chair", group__acronym="ise"), + "IAD": Q(name="admdir", group__acronym="ietf"), + "IETF Chair": Q(name="chair", group__acronym="ietf"), + "IETF Trust Chair": Q(name="chair", group__acronym="ietf-trust"), + "IRTF Chair": Q(name="chair", group__acronym="irtf"), + "RSAB Chair": Q(name="chair", group__acronym="rsab"), + "IAB Chair": Q(name="chair", group__acronym="iab"), + "IAB Executive Director": Q(name="execdir", group__acronym="iab"), + "IAB Group Chair": Q( + name="chair", group__type="iab", group__state="active" + ), + "IAOC Chair": Q(name="chair", group__acronym="iaoc"), + "WG Chair": Q( + name="chair", + group__type="wg", + group__state__in=["active", "bof", "proposed"], + ), + "WG Secretary": Q( + name="secr", + group__type="wg", + group__state__in=["active", "bof", "proposed"], + ), + "RG Chair": Q( + name="chair", group__type="rg", group__state__in=["active", "proposed"] + ), + "RG Secretary": Q( + name="secr", group__type="rg", group__state__in=["active", "proposed"] + ), + "AG Secretary": Q( + name="secr", group__type="ag", group__state__in=["active"] + ), + "RAG Secretary": Q( + name="secr", group__type="rag", group__state__in=["active"] + ), + "Team Chair": Q(name="chair", group__type="team", group__state="active"), + "Program Lead": Q( + name="lead", group__type="program", group__state="active" + ), + "Program Secretary": Q( + name="secr", group__type="program", group__state="active" + ), + "Program Chair": Q( + name="chair", group__type="program", group__state="active" + ), + "EDWG Chair": Q(name="chair", group__type="edwg", group__state="active"), + "Nomcom Chair": Q( + name="chair", + group__type="nomcom", + group__acronym__icontains=kwargs.get("year", "0000"), + ), + "Nomcom Advisor": Q( + name="advisor", + group__type="nomcom", + group__acronym__icontains=kwargs.get("year", "0000"), + ), + "Nomcom": Q( + group__type="nomcom", + group__acronym__icontains=kwargs.get("year", "0000"), + ), + "Liaison Manager": Q( + name="liaiman", + group__type="sdo", + group__state="active", + ), + "Liaison Coordinator": Q( + name="liaison_coordinator", + group__acronym="iab", + ), + "Authorized Individual": Q( + name="auth", + group__type="sdo", + group__state="active", + ), + "Recording Manager": Q( + name="recman", + group__type="ietf", + group__state="active", + ), + "Reviewer": Q(name="reviewer", group__state="active"), + "Review Team Secretary": Q( + name="secr", + group__reviewteamsettings__isnull=False, + group__state="active", + ), + "IRSG Member": ( + Q(name="member", group__acronym="irsg") + | Q(name="chair", group__acronym="irtf") + | Q(name="atlarge", group__acronym="irsg") + ), + "RSAB Member": Q(name="member", group__acronym="rsab"), + "Robot": Q(name="robot", group__acronym="secretariat"), + } + + filter_expr = Q( + pk__in=[] + ) # ensure empty set is returned if no other terms are added for r in role_names: filter_expr |= role_qs[r] + if extra_role_qs: + for r in extra_role_qs: + filter_expr |= extra_role_qs[r] - user.roles_check_cache[key] = bool(Role.objects.filter(filter_expr).exists()) + user.roles_check_cache[key] = bool( + Role.objects.filter(person=person).filter(filter_expr).exists() + ) return user.roles_check_cache[key] + # convenient decorator def passes_test_decorator(test_func, message): @@ -110,7 +191,7 @@ def passes_test_decorator(test_func, message): error. The test function should be on the form fn(user) -> true/false.""" def decorate(view_func): - @wraps(view_func, assigned=available_attrs(view_func)) + @wraps(view_func, assigned=WRAPPER_ASSIGNMENTS) def inner(request, *args, **kwargs): if not request.user.is_authenticated: return HttpResponseRedirect('%s?%s=%s' % (settings.LOGIN_URL, REDIRECT_FIELD_NAME, urlquote(request.get_full_path()))) @@ -130,9 +211,9 @@ def role_required(*role_names): # specific permissions + def is_authorized_in_doc_stream(user, doc): - """Return whether user is authorized to perform stream duties on - document.""" + """Is user authorized to perform stream duties on doc?""" if has_role(user, ["Secretariat"]): return True @@ -163,6 +244,10 @@ def is_authorized_in_doc_stream(user, doc): if doc.group.type.slug == 'individ': docman_roles = GroupFeatures.objects.get(type_id="ietf").docman_roles group_req = Q(group__acronym=doc.stream.slug) + elif doc.stream.slug == "editorial": + group_req = Q(group=doc.group) | Q(group__acronym='rsab') + if doc.group.type.slug in ("individ", "edappr"): + docman_roles = GroupFeatures.objects.get(type_id="edappr").docman_roles else: group_req = Q() # no group constraint for other cases @@ -202,7 +287,7 @@ def is_individual_draft_author(user, doc): if not hasattr(user, 'person'): return False - if user.person in doc.authors(): + if user.person in doc.author_persons(): return True return False @@ -217,7 +302,7 @@ def is_bofreq_editor(user, doc): def openid_userinfo(claims, user): # Populate claims dict. person = get_object_or_404(Person, user=user) - email = person.email() + email = person.email_allowing_inactive() if person.photo: photo_url = person.cdn_photo_url() else: @@ -265,13 +350,14 @@ def scope_pronouns(self): ) def scope_registration(self): + # import here to avoid circular imports from ietf.meeting.helpers import get_current_ietf_meeting - from ietf.stats.models import MeetingRegistration + from ietf.meeting.models import Registration meeting = get_current_ietf_meeting() person = self.user.person email_list = person.email_set.values_list('address') q = Q(person=person, meeting=meeting) | Q(email__in=email_list, meeting=meeting) - regs = MeetingRegistration.objects.filter(q).distinct() + regs = Registration.objects.filter(q).distinct() for reg in regs: if not reg.person_id: reg.person = person @@ -282,16 +368,82 @@ def scope_registration(self): ticket_types = set([]) reg_types = set([]) for reg in regs: - ticket_types.add(reg.ticket_type) - reg_types.add(reg.reg_type) + for ticket in reg.tickets.all(): + ticket_types.add(ticket.ticket_type.slug) + reg_types.add(ticket.attendance_type.slug) info = { - 'meeting': meeting.number, + 'meeting': meeting.number, # full_week, one_day, student: - 'ticket_type': ' '.join(ticket_types), + 'ticket_type': ' '.join(ticket_types), # onsite, remote, hackathon_onsite, hackathon_remote: - 'reg_type': ' '.join(reg_types), - 'affiliation': ([ reg.affiliation for reg in regs if reg.affiliation ] or [''])[0], + 'reg_type': ' '.join(reg_types), + 'affiliation': ([reg.affiliation for reg in regs if reg.affiliation] or [''])[0], } return info - + +def can_request_rfc_publication(user, doc): + """Answers whether this user has an appropriate role to send this document to the RFC Editor for publication as an RFC. + + This not take anything but the stream of the document into account. + + NOTE: This intentionally always returns False for IETF stream documents. + The publication request process for the IETF stream is handled by the + secretariat at ietf.doc.views_ballot.approve_ballot""" + + if doc.stream_id == "irtf": + return has_role(user, ("Secretariat", "IRTF Chair")) + elif doc.stream_id == "editorial": + return has_role(user, ("Secretariat", "RSAB Chair")) + elif doc.stream_id == "ise": + return has_role(user, ("Secretariat", "ISE")) + elif doc.stream_id == "iab": + return has_role(user, ("Secretariat", "IAB Chair")) + elif doc.stream_id == "ietf": + return False # See the docstring + else: + return False + + +def send_new_email_confirmation_request(person: Person, address: str): + """Request confirmation of a new email address + + If the email address is already in use, sends an alert to it. If not, sends a confirmation request. + By design, does not indicate which was sent. This is intended to make it a bit harder to scrape addresses + with a mindless bot. + """ + auth = signing.dumps([person.user.username, address], salt="add_email") + domain = Site.objects.get_current().domain + from_email = settings.DEFAULT_FROM_EMAIL + + existing = Email.objects.filter(address=address).first() + if existing: + subject = f"Attempt to add your email address by {person.name}" + send_mail( + None, + address, + from_email, + subject, + "registration/add_email_exists_email.txt", + { + "domain": domain, + "email": address, + "person": person, + }, + ) + else: + subject = f"Confirm email address for {person.name}" + send_mail( + None, + address, + from_email, + subject, + "registration/add_email_email.txt", + { + "domain": domain, + "auth": auth, + "email": address, + "person": person, + "expire": settings.DAYS_TO_EXPIRE_REGISTRATION_LINK, + }, + ) diff --git a/ietf/ietfauth/validators.py b/ietf/ietfauth/validators.py new file mode 100644 index 0000000000..84684f34d5 --- /dev/null +++ b/ietf/ietfauth/validators.py @@ -0,0 +1,34 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +import re + +from django import forms +from django.conf import settings +from django.core.exceptions import ValidationError + + +def prevent_at_symbol(name): + if "@" in name: + raise forms.ValidationError( + "Please fill in name - this looks like an email address (@ is not allowed in names)." + ) + + +def prevent_system_name(name): + name_without_spaces = name.replace(" ", "").replace("\t", "") + if "(system)" in name_without_spaces.lower(): + raise forms.ValidationError("Please pick another name - this name is reserved.") + + +def prevent_anonymous_name(name): + name_without_spaces = name.replace(" ", "").replace("\t", "") + if "anonymous" in name_without_spaces.lower(): + raise forms.ValidationError("Please pick another name - this name is reserved.") + + +def is_allowed_address(value): + """Validate that an address complies with datatracker requirements""" + for pat in settings.EXCLUDED_PERSONAL_EMAIL_REGEX_PATTERNS: + if re.search(pat, value): + raise ValidationError( + "This email address is not valid in a datatracker account" + ) diff --git a/ietf/ietfauth/views.py b/ietf/ietfauth/views.py index f89bc01494..b5256b14f8 100644 --- a/ietf/ietfauth/views.py +++ b/ietf/ietfauth/views.py @@ -38,14 +38,14 @@ import importlib # needed if we revert to higher barrier for account creation -#from datetime import datetime as DateTime, timedelta as TimeDelta, date as Date +# from datetime import datetime as DateTime, timedelta as TimeDelta, date as Date from collections import defaultdict import django.core.signing from django import forms from django.contrib import messages from django.conf import settings -from django.contrib.auth import update_session_auth_hash, logout, authenticate +from django.contrib.auth import logout, update_session_auth_hash, password_validation from django.contrib.auth.decorators import login_required from django.contrib.auth.forms import AuthenticationForm from django.contrib.auth.hashers import identify_hasher @@ -53,6 +53,7 @@ from django.contrib.auth.views import LoginView from django.contrib.sites.models import Site from django.core.exceptions import ObjectDoesNotExist, ValidationError +from django.db import IntegrityError from django.urls import reverse as urlreverse from django.http import Http404, HttpResponseRedirect, HttpResponseForbidden from django.shortcuts import render, redirect, get_object_or_404 @@ -62,11 +63,9 @@ from ietf.group.models import Role, Group from ietf.ietfauth.forms import ( RegistrationForm, PasswordForm, ResetPasswordForm, TestEmailForm, - AllowlistForm, ChangePasswordForm, get_person_form, RoleEmailForm, + ChangePasswordForm, get_person_form, RoleEmailForm, NewEmailForm, ChangeUsernameForm, PersonPasswordForm) -from ietf.ietfauth.htpasswd import update_htpasswd_file -from ietf.ietfauth.utils import role_required, has_role -from ietf.mailinglists.models import Allowlisted +from ietf.ietfauth.utils import has_role, send_new_email_confirmation_request from ietf.name.models import ExtResourceName from ietf.nomcom.models import NomCom from ietf.person.models import Person, Email, Alias, PersonalApiKey, PERSON_API_KEY_VALUES @@ -81,7 +80,6 @@ # These are needed if we revert to the higher bar for account creation - def index(request): return render(request, 'registration/index.html') @@ -98,7 +96,7 @@ def index(request): # def ietf_login(request): # if not request.user.is_authenticated: # return HttpResponse("Not authenticated?", status=500) -# +# # redirect_to = request.REQUEST.get(REDIRECT_FIELD_NAME, '') # request.session.set_test_cookie() # return HttpResponseRedirect('/accounts/loggedin/?%s=%s' % (REDIRECT_FIELD_NAME, urlquote(redirect_to))) @@ -112,33 +110,67 @@ def index(request): # redirect_to = settings.LOGIN_REDIRECT_URL # return HttpResponseRedirect(redirect_to) + def create_account(request): - to_email = None + new_account_email = None - if request.method == 'POST': + if request.method == "POST": form = RegistrationForm(request.POST) if form.is_valid(): - to_email = form.cleaned_data['email'] # This will be lowercase if form.is_valid() - - # For the IETF 113 Registration period (at least) we are lowering the barriers for account creation - # to the simple email round-trip check - send_account_creation_email(request, to_email) - - # The following is what to revert to should that lowered barrier prove problematic - # existing = Subscribed.objects.filter(email=to_email).first() - # ok_to_create = ( Allowlisted.objects.filter(email=to_email).exists() - # or existing and (existing.time + TimeDelta(seconds=settings.LIST_ACCOUNT_DELAY)) < DateTime.now() ) - # if ok_to_create: - # send_account_creation_email(request, to_email) - # else: - # return render(request, 'registration/manual.html', { 'account_request_email': settings.ACCOUNT_REQUEST_EMAIL }) + new_account_email = form.cleaned_data[ + "email" + ] # This will be lowercase if form.is_valid() + email_is_known = False # do we already know of the new_account_email address? + + # Find an existing Person to contact, if one exists + person_to_contact = None + user = User.objects.filter(username__iexact=new_account_email).first() + if user is not None: + email_is_known = True + try: + person_to_contact = user.person + except User.person.RelatedObjectDoesNotExist: + # User.person is a OneToOneField so it raises an exception if the field is null + pass # leave person_to_contact as None + if person_to_contact is None: + email = Email.objects.filter(address__iexact=new_account_email).first() + if email is not None: + email_is_known = True + # Email.person is a ForeignKey, so its value is None if the field is null + person_to_contact = email.person + # Get a "good" email to contact the existing Person + to_email = person_to_contact.email_address() if person_to_contact else None + + if to_email: + # We have a "good" email - send instructions to it + send_account_creation_exists_email(request, new_account_email, to_email) + elif email_is_known: + # Either a User or an Email matching new_account_email is in the system but we do not have a + # "good" email to use to contact its owner. Fail so the user can contact the secretariat to sort + # things out. + form.add_error( + "email", + ValidationError( + f"Unable to create account for {new_account_email}. Please contact " + f"the Secretariat at {settings.SECRETARIAT_SUPPORT_EMAIL} for assistance." + ), + ) + new_account_email = None # Indicate to the template that we failed to create the requested account + else: + send_account_creation_email(request, new_account_email) + else: form = RegistrationForm() - return render(request, 'registration/create.html', { - 'form': form, - 'to_email': to_email, - }) + return render( + request, + "registration/create.html", + { + "form": form, + "to_email": new_account_email, + }, + ) + def send_account_creation_email(request, to_email): auth = django.core.signing.dumps(to_email, salt="create_account") @@ -153,13 +185,30 @@ def send_account_creation_email(request, to_email): }) +def send_account_creation_exists_email(request, new_account_email, to_email): + domain = Site.objects.get_current().domain + subject = "Attempted account creation at %s" % domain + from_email = settings.DEFAULT_FROM_EMAIL + send_mail( + request, + to_email, + from_email, + subject, + "registration/creation_exists_email.txt", + { + "domain": domain, + "username": new_account_email, + }, + ) + + def confirm_account(request, auth): try: email = django.core.signing.loads(auth, salt="create_account", max_age=settings.DAYS_TO_EXPIRE_REGISTRATION_LINK * 24 * 60 * 60) except django.core.signing.BadSignature: raise Http404("Invalid or expired auth") - if User.objects.filter(username=email).exists(): + if User.objects.filter(username__iexact=email).exists(): return redirect(profile) success = False @@ -171,8 +220,6 @@ def confirm_account(request, auth): user = User.objects.create(username=email, email=email) user.set_password(password) user.save() - # password is also stored in htpasswd file - update_htpasswd_file(email, password) # make sure the rest of the person infrastructure is # well-connected @@ -249,23 +296,8 @@ def profile(request): to_email = f.cleaned_data["new_email"] if not to_email: continue - email_confirmations.append(to_email) - - auth = django.core.signing.dumps([person.user.username, to_email], salt="add_email") - - domain = Site.objects.get_current().domain - subject = 'Confirm email address for %s' % person.name - from_email = settings.DEFAULT_FROM_EMAIL - - send_mail(request, to_email, from_email, subject, 'registration/add_email_email.txt', { - 'domain': domain, - 'auth': auth, - 'email': to_email, - 'person': person, - 'expire': settings.DAYS_TO_EXPIRE_REGISTRATION_LINK, - }) - + send_new_email_confirmation_request(person, to_email) for r in roles: e = r.email_form.cleaned_data["email"] @@ -390,15 +422,26 @@ def confirm_new_email(request, auth): except django.core.signing.BadSignature: raise Http404("Invalid or expired auth") - person = get_object_or_404(Person, user__username=username) + person = get_object_or_404(Person, user__username__iexact=username) # do another round of validation since the situation may have # changed since submitting the request form = NewEmailForm({ "new_email": email }) can_confirm = form.is_valid() and email new_email_obj = None + created = False if request.method == 'POST' and can_confirm and request.POST.get("action") == "confirm": - new_email_obj = Email.objects.create(address=email, person=person, origin=username) + try: + new_email_obj, created = Email.objects.get_or_create( + address=email, + person=person, + defaults={'origin': username}, + ) + except IntegrityError: + can_confirm = False + form.add_error( + None, "Email address is in use by another user. Please contact the secretariat for assistance." + ) return render(request, 'registration/confirm_new_email.html', { 'username': username, @@ -406,6 +449,7 @@ def confirm_new_email(request, auth): 'can_confirm': can_confirm, 'form': form, 'new_email_obj': new_email_obj, + 'already_confirmed': new_email_obj and not created, }) def password_reset(request): @@ -413,31 +457,50 @@ def password_reset(request): if request.method == 'POST': form = ResetPasswordForm(request.POST) if form.is_valid(): - username = form.cleaned_data['username'] - - data = { 'username': username } - if User.objects.filter(username=username).exists(): - user = User.objects.get(username=username) - data['password'] = user.password and user.password[-4:] - if user.last_login: - data['last_login'] = user.last_login.timestamp() - else: - data['last_login'] = None - - auth = django.core.signing.dumps(data, salt="password_reset") - - domain = Site.objects.get_current().domain - subject = 'Confirm password reset at %s' % domain - from_email = settings.DEFAULT_FROM_EMAIL - to_email = username # form validation makes sure that this is an email address - - send_mail(request, to_email, from_email, subject, 'registration/password_reset_email.txt', { - 'domain': domain, - 'auth': auth, - 'username': username, - 'expire': settings.MINUTES_TO_EXPIRE_RESET_PASSWORD_LINK, - }) + submitted_username = form.cleaned_data['username'] + # The form validation checks that a matching User exists. Add the person__isnull check + # because the OneToOne field does not gracefully handle checks for user.person is Null. + # If we don't get a User here, we know it's because there's no related Person. + # We still report that the action succeeded, so we're not leaking the existence of user + # email addresses. + user = User.objects.filter(username__iexact=submitted_username, person__isnull=False).first() + if not user: + # try to find user ID from the email address + email = Email.objects.filter(address=submitted_username).first() + if email and email.person: + if email.person.user: + user = email.person.user + else: + # Create a User record with this (conditioned by way of Email) username + # Don't bother setting the name or email fields on User - rely on the + # Person pointer. + user = User.objects.create( + username=email.address.lower(), + is_active=True, + ) + email.person.user = user + email.person.save() + if user and user.person.email_set.filter(active=True).exists(): + data = { + 'username': user.username, + 'password': user.password and user.password[-4:], + 'last_login': user.last_login.timestamp() if user.last_login else None, + } + auth = django.core.signing.dumps(data, salt="password_reset") + domain = Site.objects.get_current().domain + subject = 'Confirm password reset at %s' % domain + from_email = settings.DEFAULT_FROM_EMAIL + # Send email to addresses from the database, NOT to the address from the form. + # This prevents unicode spoofing tricks (https://nvd.nist.gov/vuln/detail/CVE-2019-19844). + to_emails = list(set(email.address for email in user.person.email_set.filter(active=True))) + to_emails.sort() + send_mail(request, to_emails, from_email, subject, 'registration/password_reset_email.txt', { + 'domain': domain, + 'auth': auth, + 'username': submitted_username, + 'expire': settings.MINUTES_TO_EXPIRE_RESET_PASSWORD_LINK, + }) success = True else: form = ResetPasswordForm() @@ -454,11 +517,11 @@ def confirm_password_reset(request, auth): password = data['password'] last_login = None if data['last_login']: - last_login = datetime.datetime.fromtimestamp(data['last_login'], datetime.timezone.utc) + last_login = datetime.datetime.fromtimestamp(data['last_login'], datetime.UTC) except django.core.signing.BadSignature: raise Http404("Invalid or expired auth") - user = get_object_or_404(User, username=username, password__endswith=password, last_login=last_login) + user = get_object_or_404(User, username__iexact=username, password__endswith=password, last_login=last_login) if request.user.is_authenticated and request.user != user: return HttpResponseForbidden( f'This password reset link is not for the signed-in user. ' @@ -466,18 +529,16 @@ def confirm_password_reset(request, auth): ) success = False if request.method == 'POST': - form = PasswordForm(request.POST) + form = PasswordForm(user=user, data=request.POST) if form.is_valid(): password = form.cleaned_data["password"] user.set_password(password) user.save() - # password is also stored in htpasswd file - update_htpasswd_file(user.username, password) success = True else: - form = PasswordForm() + form = PasswordForm(user=user) hlibname, hashername = settings.PASSWORD_HASHERS[0].rsplit('.',1) hlib = importlib.import_module(hlibname) @@ -519,23 +580,6 @@ def test_email(request): return r -@role_required('Secretariat') -def add_account_allowlist(request): - success = False - if request.method == 'POST': - form = AllowlistForm(request.POST) - if form.is_valid(): - email = form.cleaned_data['email'] - entry = Allowlisted(email=email, by=request.user.person) - entry.save() - success = True - else: - form = AllowlistForm() - - return render(request, 'ietfauth/allowlist_form.html', { - 'form': form, - 'success': success, - }) class AddReviewWishForm(forms.Form): doc = SearchableDocumentField(label="Document", doc_type="draft") @@ -625,12 +669,10 @@ def change_password(request): if request.method == 'POST': form = ChangePasswordForm(user, request.POST) if form.is_valid(): - new_password = form.cleaned_data["new_password"] + new_password = form.cleaned_data["password"] user.set_password(new_password) user.save() - # password is also stored in htpasswd file - update_htpasswd_file(user.username, new_password) # keep the session update_session_auth_hash(request, user) @@ -652,7 +694,7 @@ def change_password(request): 'hasher': hasher, }) - + @login_required @person_required def change_username(request): @@ -667,13 +709,10 @@ def change_username(request): form = ChangeUsernameForm(user, request.POST) if form.is_valid(): new_username = form.cleaned_data["username"] - password = form.cleaned_data["password"] assert new_username in emails user.username = new_username.lower() user.save() - # password is also stored in htpasswd file - update_htpasswd_file(user.username, password) # keep the session update_session_auth_hash(request, user) @@ -688,53 +727,79 @@ def change_username(request): return render(request, 'registration/change_username.html', {'form': form}) - -def login(request, extra_context=None): - """ - This login function is a wrapper around django's login() for the purpose - of providing a notification if the user's password has been cleared. The - warning will be triggered if the password field has been set to something - which is not recognized as a valid password hash. +class AnyEmailAuthenticationForm(AuthenticationForm): + """AuthenticationForm that allows any email address as the username + + Also performs a check for a cleared password field and provides a helpful error message + if that applies to the user attempting to log in. """ - - if request.method == "POST": - form = AuthenticationForm(request, data=request.POST) - username = form.data.get('username') - user = User.objects.filter(username=username).first() - if not user: - # try to find user ID from the email address + _unauthenticated_user = None + + def clean_username(self): + username = self.cleaned_data.get("username", None) + if username is None: + raise self.get_invalid_login_error() + user = User.objects.filter(username__iexact=username).first() + if user is None: email = Email.objects.filter(address=username).first() - if email and email.person and email.person.user: - u2 = email.person.user - # be conservative, only accept this if login is valid - if u2: - pw = form.data.get('password') - au = authenticate(request, username=u2.username, password=pw) - if au: - # kludge to change the querydict - q2 = request.POST.copy() - q2['username'] = u2.username - request.POST = q2 - user = u2 - # - if user: - try: - identify_hasher(user.password) + if email and email.person: + user = email.person.user # might be None + if user is None: + raise self.get_invalid_login_error() + self._unauthenticated_user = user # remember this for the clean() method + return user.username + + def clean(self): + if self._unauthenticated_user is not None: + try: + identify_hasher(self._unauthenticated_user.password) except ValueError: - extra_context = {"alert": - "Note: Your password has been cleared because " - "of possible password leakage. " - "Please use the password reset link below " - "to set a new password for your account.", - } - response = LoginView.as_view(extra_context=extra_context)(request) - if isinstance(response, HttpResponseRedirect) and user and user.is_authenticated: - try: - user.person - except Person.DoesNotExist: - logout(request) - response = render(request, 'registration/missing_person.html') - return response + self.add_error( + "password", + 'Your password has been cleared because of possible password leakage. ' + 'Please use the "Forgot your password?" button below to set a new password ' + 'for your account.', + ) + return super().clean() + + def confirm_login_allowed(self, user): + """Check whether a successfully authenticated user is permitted to log in""" + super().confirm_login_allowed(user) + # Optionally enforce password validation + if getattr(settings, "PASSWORD_POLICY_ENFORCE_AT_LOGIN", False): + try: + password_validation.validate_password( + self.cleaned_data["password"], user + ) + except ValidationError: + raise ValidationError( + # dict mapping field to error / error list + { + "__all__": ValidationError( + 'You entered your password correctly, but it does not ' + 'meet our current length and complexity requirements. ' + 'Please use the "Forgot your password?" button below to ' + 'set a new password for your account.' + ), + } + ) + + +class AnyEmailLoginView(LoginView): + """LoginView that allows any email address as the username + + Redirects to the missing_person page instead of logging in if the user does not have a Person + """ + form_class = AnyEmailAuthenticationForm + + def form_valid(self, form): + """Security check complete. Log the user in if they have a Person.""" + user = form.get_user() # user has authenticated at this point + if not hasattr(user, "person"): + logout(self.request) # should not be logged in yet, but just in case... + return render(self.request, "registration/missing_person.html") + return super().form_valid(form) + @login_required @person_required @@ -770,11 +835,11 @@ class Meta: @person_required def apikey_disable(request): person = request.user.person - choices = [ (k.hash(), str(k)) for k in person.apikeys.all() ] + choices = [ (k.hash(), str(k)) for k in person.apikeys.exclude(valid=False) ] # class KeyDeleteForm(forms.Form): hash = forms.ChoiceField(label='Key', choices=choices) - def clean_key(self): + def clean_hash(self): hash = force_bytes(self.cleaned_data['hash']) key = PersonalApiKey.validate_key(hash) if key and key.person == request.user.person: @@ -785,7 +850,7 @@ def clean_key(self): if request.method == 'POST': form = KeyDeleteForm(request.POST) if form.is_valid(): - hash = force_bytes(form.data['hash']) + hash = force_bytes(form.cleaned_data['hash']) key = PersonalApiKey.validate_key(hash) key.valid = False key.save() diff --git a/ietf/ietfauth/widgets.py b/ietf/ietfauth/widgets.py new file mode 100644 index 0000000000..fd7fa16726 --- /dev/null +++ b/ietf/ietfauth/widgets.py @@ -0,0 +1,115 @@ +from django.forms import PasswordInput +from django.utils.safestring import mark_safe +from django.utils.translation import gettext as _ + +# The PasswordStrengthInput and PasswordConfirmationInput widgets come from the +# django-password-strength project, https://pypi.org/project/django-password-strength/ +# +# Original license: +# +# Copyright © 2015 A.J. May and individual contributors. All rights reserved. +# +# Redistribution and use in source and binary forms, with or without modification, are permitted provided that the +# following conditions are met: +# +# 1. Redistributions of source code must retain the above copyright notice, this list of conditions and the following +# disclaimer. +# +# 2. Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the +# following disclaimer in the documentation and/or other materials provided with the distribution. +# +# 3. Neither the name of the copyright holder nor the names of its contributors may be used to endorse or promote +# products derived from this software without specific prior written permission. +# +# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, +# INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE +# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, +# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR +# SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, +# WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE +# USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. +# + +class PasswordStrengthInput(PasswordInput): + """ + Form widget to show the user how strong his/her password is. + """ + + def render(self, name, value, attrs=None, renderer=None): + strength_markup = """ +
+
+
+
+ +
+ """.format( + _("Warning"), + _( + 'This password would take to crack.' + ), + ) + + try: + self.attrs["class"] = "%s password_strength".strip() % self.attrs["class"] + except KeyError: + self.attrs["class"] = "password_strength" + + return mark_safe( + super(PasswordInput, self).render(name, value, attrs, renderer) + + strength_markup + ) + + class Media: + js = ( + "ietf/js/zxcvbn.js", + "ietf/js/password_strength.js", + ) + + +class PasswordConfirmationInput(PasswordInput): + """ + Form widget to confirm the users password by letting him/her type it again. + """ + + def __init__(self, confirm_with=None, attrs=None, render_value=False): + super(PasswordConfirmationInput, self).__init__(attrs, render_value) + self.confirm_with = confirm_with + + def render(self, name, value, attrs=None, renderer=None): + if self.confirm_with: + self.attrs["data-confirm-with"] = "id_%s" % self.confirm_with + + confirmation_markup = """ + + """ % ( + _("Warning"), + _("Your passwords don't match."), + ) + + try: + self.attrs["class"] = ( + "%s password_confirmation".strip() % self.attrs["class"] + ) + except KeyError: + self.attrs["class"] = "password_confirmation" + + return mark_safe( + super(PasswordInput, self).render(name, value, attrs, renderer) + + confirmation_markup + ) diff --git a/ietf/ipr/.gitignore b/ietf/ipr/.gitignore deleted file mode 100644 index c7013ced9d..0000000000 --- a/ietf/ipr/.gitignore +++ /dev/null @@ -1,2 +0,0 @@ -/*.pyc -/settings_local.py diff --git a/ietf/ipr/admin.py b/ietf/ipr/admin.py index a0185f58c6..d6a320203b 100644 --- a/ietf/ipr/admin.py +++ b/ietf/ipr/admin.py @@ -1,13 +1,23 @@ -# Copyright The IETF Trust 2010-2020, All Rights Reserved +# Copyright The IETF Trust 2010-2025, All Rights Reserved # -*- coding: utf-8 -*- from django import forms from django.contrib import admin from ietf.name.models import DocRelationshipName -from ietf.ipr.models import (IprDisclosureBase, IprDocRel, IprEvent, - RelatedIpr, HolderIprDisclosure, ThirdPartyIprDisclosure, GenericIprDisclosure, - NonDocSpecificIprDisclosure, LegacyMigrationIprEvent) +from ietf.ipr.models import ( + IprDisclosureBase, + IprDocRel, + IprEvent, + RelatedIpr, + HolderIprDisclosure, + RemovedIprDisclosure, + ThirdPartyIprDisclosure, + GenericIprDisclosure, + NonDocSpecificIprDisclosure, + LegacyMigrationIprEvent, +) +from ietf.utils.admin import SaferTabularInline # ------------------------------------------------------ # ModelAdmins @@ -20,13 +30,13 @@ class Meta: 'sections':forms.TextInput, } -class IprDocRelInline(admin.TabularInline): +class IprDocRelInline(SaferTabularInline): model = IprDocRel form = IprDocRelAdminForm raw_id_fields = ['document'] extra = 1 -class RelatedIprInline(admin.TabularInline): +class RelatedIprInline(SaferTabularInline): model = RelatedIpr raw_id_fields = ['target'] fk_name = 'source' @@ -94,7 +104,7 @@ class IprDocRelAdmin(admin.ModelAdmin): class RelatedIprAdmin(admin.ModelAdmin): list_display = ['source', 'target', 'relationship', ] - search_fields = ['source__name', 'target__name', 'target__docs__name', ] + search_fields = ['source__name', 'target__name', ] raw_id_fields = ['source', 'target', ] admin.site.register(RelatedIpr, RelatedIprAdmin) @@ -110,3 +120,9 @@ class LegacyMigrationIprEventAdmin(admin.ModelAdmin): list_filter = ['time', 'type', 'response_due'] raw_id_fields = ['by', 'disclosure', 'message', 'in_reply_to'] admin.site.register(LegacyMigrationIprEvent, LegacyMigrationIprEventAdmin) + +class RemovedIprDisclosureAdmin(admin.ModelAdmin): + pass + + +admin.site.register(RemovedIprDisclosure, RemovedIprDisclosureAdmin) diff --git a/ietf/ipr/factories.py b/ietf/ipr/factories.py index e32090a36b..8a8a740158 100644 --- a/ietf/ipr/factories.py +++ b/ietf/ipr/factories.py @@ -1,9 +1,10 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved +# Copyright The IETF Trust 2018-2023, All Rights Reserved # -*- coding: utf-8 -*- import datetime import factory +from faker import Faker from django.utils import timezone @@ -13,17 +14,19 @@ ) def _fake_patent_info(): + fake = Faker() return "Date: %s\nNotes: %s\nTitle: %s\nNumber: %s\nInventor: %s\n" % ( (timezone.now()-datetime.timedelta(days=365)).strftime("%Y-%m-%d"), - factory.Faker('paragraph'), - factory.Faker('sentence', nb_words=8), + fake.paragraph(), + fake.sentence(nb_words=8), 'US9999999', - factory.Faker('name'), + fake.name(), ) class IprDisclosureBaseFactory(factory.django.DjangoModelFactory): class Meta: model = IprDisclosureBase + skip_postgeneration_save = True by = factory.SubFactory('ietf.person.factories.PersonFactory') compliant = True @@ -39,7 +42,7 @@ def docs(self, create, extracted, **kwargs): return if extracted: for doc in extracted: - IprDocRel.objects.create(disclosure=self,document=doc.docalias.first()) + IprDocRel.objects.create(disclosure=self,document=doc) @factory.post_generation def updates(self, create, extracted, **kwargs): @@ -93,3 +96,11 @@ class Meta: disclosure = factory.SubFactory(IprDisclosureBaseFactory) desc = factory.Faker('sentence') +class IprDocRelFactory(factory.django.DjangoModelFactory): + class Meta: + model = IprDocRel + + disclosure = factory.SubFactory(HolderIprDisclosureFactory) + document = factory.SubFactory("ietf.doc.factories.IndividualDraftFactory") + revisions = "00" + sections = "" diff --git a/ietf/ipr/feeds.py b/ietf/ipr/feeds.py index b5f4c4e6a4..e8a1c739f2 100644 --- a/ietf/ipr/feeds.py +++ b/ietf/ipr/feeds.py @@ -1,12 +1,13 @@ -# Copyright The IETF Trust 2007-2020, All Rights Reserved +# Copyright The IETF Trust 2007-2023, All Rights Reserved # -*- coding: utf-8 -*- +from django.conf import settings from django.contrib.syndication.views import Feed from django.utils.feedgenerator import Atom1Feed from django.urls import reverse_lazy from django.utils.safestring import mark_safe -from django.utils.encoding import force_text +from django.utils.encoding import force_str from ietf.ipr.models import IprDisclosureBase @@ -19,13 +20,13 @@ class LatestIprDisclosuresFeed(Feed): feed_url = "/feed/ipr/" def items(self): - return IprDisclosureBase.objects.filter(state__in=('posted','removed')).order_by('-time')[:30] + return IprDisclosureBase.objects.filter(state__in=settings.PUBLISH_IPR_STATES).order_by('-time')[:30] def item_title(self, item): return mark_safe(item.title) def item_description(self, item): - return force_text(item.title) + return force_str(item.title) def item_pubdate(self, item): return item.time diff --git a/ietf/ipr/forms.py b/ietf/ipr/forms.py index 06248e2a73..dac34bddf6 100644 --- a/ietf/ipr/forms.py +++ b/ietf/ipr/forms.py @@ -1,4 +1,4 @@ -# Copyright The IETF Trust 2014-2020, All Rights Reserved +# Copyright The IETF Trust 2014-2023, All Rights Reserved # -*- coding: utf-8 -*- @@ -14,7 +14,7 @@ import debug # pyflakes:ignore from ietf.group.models import Group -from ietf.doc.fields import SearchableDocAliasField +from ietf.doc.fields import SearchableDocumentField from ietf.ipr.mail import utc_from_string from ietf.ipr.fields import SearchableIprDisclosuresField from ietf.ipr.models import (IprDocRel, IprDisclosureBase, HolderIprDisclosure, @@ -95,7 +95,7 @@ def clean(self): return self.cleaned_data class DraftForm(forms.ModelForm): - document = SearchableDocAliasField(label="I-D name/RFC number", required=True, doc_type="draft") + document = SearchableDocumentField(label="I-D name/RFC number", required=True, doc_type="all") class Meta: model = IprDocRel @@ -105,6 +105,17 @@ class Meta: } help_texts = { 'sections': 'Sections' } + def clean(self): + cleaned_data = super().clean() + revisions = cleaned_data.get("revisions") + document = cleaned_data.get("document") + if not document: + self.add_error("document", "Identifying the Internet-Draft or RFC for this disclosure is required.") + elif not document.name.startswith("rfc"): + if revisions is None or revisions.strip() == "": + self.add_error("revisions", "Revisions of this Internet-Draft for which this disclosure is relevant must be specified.") + return cleaned_data + patent_number_help_text = "Enter one or more comma-separated patent publication or application numbers as two-letter country code and serial number, e.g.: US62/123456 or WO2017123456. Do not include thousands-separator commas in serial numbers. It is preferable to use individual disclosures for each patent, even if this field permits multiple patents to be listed, in order to get inventor, title, and date information below correct." validate_patent_number = RegexValidator( regex=(r"^(" @@ -327,7 +338,19 @@ def clean(self): return cleaned_data + class HolderIprDisclosureForm(IprDisclosureFormBase): + is_blanket_disclosure = forms.BooleanField( + label=mark_safe( + 'This is a blanket IPR disclosure ' + '(see Section 5.4.3 of RFC 8179)' + ), + help_text="In satisfaction of its disclosure obligations, Patent Holder commits to license all of " + "IPR (as defined in RFC 8179) that would have required disclosure under RFC 8179 on a " + "royalty-free (and otherwise reasonable and non-discriminatory) basis. Patent Holder " + "confirms that all other terms and conditions are described in this IPR disclosure.", + required=False, + ) licensing = CustomModelChoiceField(IprLicenseTypeName.objects.all(), widget=forms.RadioSelect,empty_label=None) @@ -345,6 +368,15 @@ def __init__(self, *args, **kwargs): else: # entering new disclosure self.fields['licensing'].queryset = IprLicenseTypeName.objects.exclude(slug='none-selected') + + if self.data.get("is_blanket_disclosure", False): + # for a blanket disclosure, patent details are not required + self.fields["patent_number"].required = False + self.fields["patent_inventor"].required = False + self.fields["patent_title"].required = False + self.fields["patent_date"].required = False + # n.b., self.fields["patent_notes"] is never required + def clean(self): cleaned_data = super(HolderIprDisclosureForm, self).clean() @@ -413,7 +445,7 @@ def save(self, *args, **kwargs): class SearchForm(forms.Form): state = forms.MultipleChoiceField(choices=[], widget=forms.CheckboxSelectMultiple,required=False) - draft = forms.CharField(label="Draft name", max_length=128, required=False) + draft = forms.CharField(label="Internet-Draft name", max_length=128, required=False) rfc = forms.IntegerField(label="RFC number", required=False) holder = forms.CharField(label="Name of patent owner/applicant", max_length=128,required=False) patent = forms.CharField(label="Text in patent information", max_length=128,required=False) @@ -428,4 +460,4 @@ def __init__(self, *args, **kwargs): class StateForm(forms.Form): state = forms.ModelChoiceField(queryset=IprDisclosureStateName.objects,label="New State",empty_label=None) comment = forms.CharField(required=False, widget=forms.Textarea, help_text="You may add a comment to be included in the disclosure history.", strip=False) - private = forms.BooleanField(label="Private comment", required=False, help_text="If this box is checked the comment will not appear in the disclosure's public history view.") \ No newline at end of file + private = forms.BooleanField(label="Private comment", required=False, help_text="If this box is checked the comment will not appear in the disclosure's public history view.") diff --git a/ietf/ipr/mail.py b/ietf/ipr/mail.py index f1d8039db9..9bef751b95 100644 --- a/ietf/ipr/mail.py +++ b/ietf/ipr/mail.py @@ -12,7 +12,7 @@ from email.utils import parsedate_tz from django.template.loader import render_to_string -from django.utils.encoding import force_text, force_bytes +from django.utils.encoding import force_str, force_bytes import debug # pyflakes:ignore @@ -66,9 +66,9 @@ def utc_from_string(s): if date is None: return None elif is_aware(date): - return date.astimezone(datetime.timezone.utc) + return date.astimezone(datetime.UTC) else: - return date.replace(tzinfo=datetime.timezone.utc) + return date.replace(tzinfo=datetime.UTC) # ---------------------------------------------------------------- # Email Functions @@ -102,7 +102,7 @@ def get_reply_to(): address with "plus addressing" using a random string. Guaranteed to be unique""" local,domain = get_base_ipr_request_address().split('@') while True: - rand = force_text(base64.urlsafe_b64encode(os.urandom(12))) + rand = force_str(base64.urlsafe_b64encode(os.urandom(12))) address = "{}+{}@{}".format(local,rand,domain) q = Message.objects.filter(reply_to=address) if not q: @@ -171,31 +171,44 @@ def message_from_message(message,by=None): ) return msg + +class UndeliverableIprResponseError(Exception): + """Response email could not be delivered and should be treated as an error""" + + def process_response_email(msg): - """Saves an incoming message. msg=string. Message "To" field is expected to - be in the format ietf-ipr+[identifier]@ietf.org. Expect to find a message with - a matching value in the reply_to field, associated to an IPR disclosure through - IprEvent. Create a Message object for the incoming message and associate it to - the original message via new IprEvent""" + """Save an incoming IPR response email message + + Message "To" field is expected to be in the format ietf-ipr+[identifier]@ietf.org. If + the address or identifier is missing, the message will be silently dropped. + + Expect to find a message with a matching value in the reply_to field, associated to an + IPR disclosure through IprEvent. If it cannot be matched, raises UndeliverableIprResponseError + + Creates a Message object for the incoming message and associates it to + the original message via new IprEvent + """ message = message_from_bytes(force_bytes(msg)) to = message.get('To', '') # exit if this isn't a response we're interested in (with plus addressing) - local,domain = get_base_ipr_request_address().split('@') + local, domain = get_base_ipr_request_address().split('@') if not re.match(r'^{}\+[a-zA-Z0-9_\-]{}@{}'.format(local,'{16}',domain),to): - return None - + _from = message.get("From", "") + log(f"Ignoring IPR email without a message identifier from {_from} to {to}") + return + try: to_message = Message.objects.get(reply_to=to) except Message.DoesNotExist: log('Error finding matching message ({})'.format(to)) - return None + raise UndeliverableIprResponseError(f"Unable to find message matching {to}") try: disclosure = to_message.msgevents.first().disclosure except: log('Error processing message ({})'.format(to)) - return None + raise UndeliverableIprResponseError("Error processing message for {to}") ietf_message = message_from_message(message) IprEvent.objects.create( @@ -207,4 +220,4 @@ def process_response_email(msg): ) log("Received IPR email from %s" % ietf_message.frm) - return ietf_message + diff --git a/ietf/ipr/management/commands/generate_draft_recursive_txt.py b/ietf/ipr/management/commands/generate_draft_recursive_txt.py deleted file mode 100644 index 240fc5348e..0000000000 --- a/ietf/ipr/management/commands/generate_draft_recursive_txt.py +++ /dev/null @@ -1,17 +0,0 @@ -# Copyright The IETF Trust 2014-2021, All Rights Reserved -# -*- coding: utf-8 -*- - - -from django.core.management.base import BaseCommand, CommandError - -from ietf.ipr.utils import generate_draft_recursive_txt - - -class Command(BaseCommand): - help = ("Generate machine-readable list of IPR disclosures by draft name (recursive)") - - def handle(self, *args, **options): - try: - generate_draft_recursive_txt() - except (ValueError, IOError) as e: - raise CommandError(e) diff --git a/ietf/ipr/management/commands/process_email.py b/ietf/ipr/management/commands/process_email.py index d04b54b216..616cade5c4 100644 --- a/ietf/ipr/management/commands/process_email.py +++ b/ietf/ipr/management/commands/process_email.py @@ -9,7 +9,7 @@ from django.core.management import CommandError from ietf.utils.management.base import EmailOnFailureCommand -from ietf.ipr.mail import process_response_email +from ietf.ipr.mail import process_response_email, UndeliverableIprResponseError import debug # pyflakes:ignore @@ -23,11 +23,15 @@ def add_arguments(self, parser): def handle(self, *args, **options): email = options.get('email', None) - binary_input = io.open(email, 'rb') if email else sys.stdin.buffer - self.msg_bytes = binary_input.read() + if email: + binary_input = io.open(email, 'rb') + self.msg_bytes = binary_input.read() + binary_input.close() + else: + self.msg_bytes = sys.stdin.buffer.read() try: process_response_email(self.msg_bytes) - except ValueError as e: + except (ValueError, UndeliverableIprResponseError) as e: raise CommandError(e) failure_subject = 'Error during ipr email processing' @@ -44,4 +48,4 @@ def make_failure_message(self, error, **extra): 'application', 'octet-stream', # mime type filename='original-message', ) - return msg \ No newline at end of file + return msg diff --git a/ietf/ipr/management/tests.py b/ietf/ipr/management/tests.py index d84b0cfef4..d7acd65042 100644 --- a/ietf/ipr/management/tests.py +++ b/ietf/ipr/management/tests.py @@ -1,7 +1,7 @@ # Copyright The IETF Trust 2021, All Rights Reserved # -*- coding: utf-8 -*- """Tests of ipr management commands""" -import mock +from unittest import mock import sys from django.core.management import call_command diff --git a/ietf/ipr/migrations/0001_initial.py b/ietf/ipr/migrations/0001_initial.py index e370d3d5ad..b204fd4012 100644 --- a/ietf/ipr/migrations/0001_initial.py +++ b/ietf/ipr/migrations/0001_initial.py @@ -1,7 +1,4 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-20 10:52 - +# Generated by Django 2.2.28 on 2023-03-20 19:22 from django.db import migrations, models import django.db.models.deletion @@ -32,15 +29,11 @@ class Migration(migrations.Migration): ('submitter_email', models.EmailField(blank=True, max_length=254)), ('time', models.DateTimeField(auto_now_add=True)), ('title', models.CharField(blank=True, max_length=255)), + ('by', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), ], - ), - migrations.CreateModel( - name='IprDocRel', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('sections', models.TextField(blank=True)), - ('revisions', models.CharField(blank=True, max_length=16)), - ], + options={ + 'ordering': ['-time', '-id'], + }, ), migrations.CreateModel( name='IprEvent', @@ -49,18 +42,16 @@ class Migration(migrations.Migration): ('time', models.DateTimeField(auto_now_add=True)), ('desc', models.TextField()), ('response_due', models.DateTimeField(blank=True, null=True)), + ('by', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), + ('disclosure', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='ipr.IprDisclosureBase')), + ('in_reply_to', ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='irtoevents', to='message.Message')), + ('message', ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='msgevents', to='message.Message')), + ('type', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.IprEventTypeName')), ], options={ 'ordering': ['-time', '-id'], }, ), - migrations.CreateModel( - name='RelatedIpr', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('relationship', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.DocRelationshipName')), - ], - ), migrations.CreateModel( name='GenericIprDisclosure', fields=[ @@ -86,7 +77,6 @@ class Migration(migrations.Migration): ('holder_contact_info', models.TextField(blank=True, help_text='Address, phone, etc.')), ('licensing_comments', models.TextField(blank=True)), ('submitter_claims_all_terms_disclosed', models.BooleanField(default=False)), - ('licensing', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.IprLicenseTypeName')), ], bases=('ipr.iprdisclosurebase',), ), @@ -122,55 +112,24 @@ class Migration(migrations.Migration): ], bases=('ipr.iprdisclosurebase',), ), - migrations.AddField( - model_name='relatedipr', - name='source', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='relatedipr_source_set', to='ipr.IprDisclosureBase'), - ), - migrations.AddField( - model_name='relatedipr', - name='target', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='relatedipr_target_set', to='ipr.IprDisclosureBase'), - ), - migrations.AddField( - model_name='iprevent', - name='by', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person'), - ), - migrations.AddField( - model_name='iprevent', - name='disclosure', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='ipr.IprDisclosureBase'), - ), - migrations.AddField( - model_name='iprevent', - name='in_reply_to', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='irtoevents', to='message.Message'), - ), - migrations.AddField( - model_name='iprevent', - name='message', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='msgevents', to='message.Message'), - ), - migrations.AddField( - model_name='iprevent', - name='type', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.IprEventTypeName'), - ), - migrations.AddField( - model_name='iprdocrel', - name='disclosure', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='ipr.IprDisclosureBase'), - ), - migrations.AddField( - model_name='iprdocrel', - name='document', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.DocAlias'), + migrations.CreateModel( + name='RelatedIpr', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('relationship', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.DocRelationshipName')), + ('source', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='relatedipr_source_set', to='ipr.IprDisclosureBase')), + ('target', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='relatedipr_target_set', to='ipr.IprDisclosureBase')), + ], ), - migrations.AddField( - model_name='iprdisclosurebase', - name='by', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person'), + migrations.CreateModel( + name='IprDocRel', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('sections', models.TextField(blank=True)), + ('revisions', models.CharField(blank=True, max_length=16)), + ('disclosure', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='ipr.IprDisclosureBase')), + ('document', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.DocAlias')), + ], ), migrations.AddField( model_name='iprdisclosurebase', @@ -187,4 +146,17 @@ class Migration(migrations.Migration): name='state', field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.IprDisclosureStateName'), ), + migrations.AddIndex( + model_name='iprevent', + index=models.Index(fields=['-time', '-id'], name='ipr_ipreven_time_9630c4_idx'), + ), + migrations.AddIndex( + model_name='iprdisclosurebase', + index=models.Index(fields=['-time', '-id'], name='ipr_iprdisc_time_846a78_idx'), + ), + migrations.AddField( + model_name='holderiprdisclosure', + name='licensing', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.IprLicenseTypeName'), + ), ] diff --git a/ietf/ipr/migrations/0002_auto_20180225_1207.py b/ietf/ipr/migrations/0002_auto_20180225_1207.py deleted file mode 100644 index 3f34a491d8..0000000000 --- a/ietf/ipr/migrations/0002_auto_20180225_1207.py +++ /dev/null @@ -1,20 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-25 12:07 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('ipr', '0001_initial'), - ] - - operations = [ - migrations.AlterModelOptions( - name='iprdisclosurebase', - options={'ordering': ['-time', '-id']}, - ), - ] diff --git a/ietf/ipr/migrations/0002_iprdocrel_no_aliases.py b/ietf/ipr/migrations/0002_iprdocrel_no_aliases.py new file mode 100644 index 0000000000..bcfc73a320 --- /dev/null +++ b/ietf/ipr/migrations/0002_iprdocrel_no_aliases.py @@ -0,0 +1,104 @@ +# Generated by Django 4.2.2 on 2023-06-16 13:40 + +from django.db import migrations +import django.db.models.deletion +from django.db.models import F, Subquery, OuterRef, ManyToManyField, CharField +import ietf.utils.models + +def forward(apps, schema_editor): + IprDocRel = apps.get_model("ipr", "IprDocRel") + DocAlias = apps.get_model("doc", "DocAlias") + document_subquery = Subquery( + DocAlias.objects.filter( + pk=OuterRef("deprecated_document") + ).values("docs")[:1] + ) + name_subquery = Subquery( + DocAlias.objects.filter( + pk=OuterRef("deprecated_document") + ).values("name")[:1] + ) + IprDocRel.objects.annotate( + firstdoc=document_subquery, + aliasname=name_subquery, + ).update( + document=F("firstdoc"), + originaldocumentaliasname=F("aliasname"), + ) + # This might not be right - we may need here (and in the relateddocument migrations) to pay attention to + # whether the name being pointed to is and rfc name or a draft name and point to the right object instead... + +def reverse(apps, schema_editor): + pass + +class Migration(migrations.Migration): + dependencies = [ + ("ipr", "0001_initial"), + ("doc", "0016_relate_hist_no_aliases") + ] + + operations = [ + migrations.AlterField( + model_name='iprdocrel', + name='document', + field=ietf.utils.models.ForeignKey( + db_index=False, + on_delete=django.db.models.deletion.CASCADE, + to='doc.docalias', + ), + ), + migrations.RenameField( + model_name="iprdocrel", + old_name="document", + new_name="deprecated_document" + ), + migrations.AlterField( + model_name='iprdocrel', + name='deprecated_document', + field=ietf.utils.models.ForeignKey( + db_index=True, + on_delete=django.db.models.deletion.CASCADE, + to='doc.docalias', + ), + ), + migrations.AddField( + model_name="iprdocrel", + name="document", + field=ietf.utils.models.ForeignKey( + default=1, # A lie, but a convenient one - no iprdocrel objects point here. + on_delete=django.db.models.deletion.CASCADE, + to="doc.document", + db_index=False, + ), + preserve_default=False, + ), + migrations.AddField( + model_name="iprdocrel", + name="originaldocumentaliasname", + field=CharField(max_length=255,null=True,blank=True), + preserve_default=True, + ), + migrations.RunPython(forward, reverse), + migrations.AlterField( + model_name="iprdocrel", + name="document", + field=ietf.utils.models.ForeignKey( + on_delete=django.db.models.deletion.CASCADE, + to="doc.document", + db_index=True, + ), + ), + migrations.AlterField( + model_name='iprdisclosurebase', + name='docs', + field=ManyToManyField(through='ipr.IprDocRel', to='doc.Document'), + ), + migrations.RemoveField( + model_name="iprdocrel", + name="deprecated_document", + field=ietf.utils.models.ForeignKey( + on_delete=django.db.models.deletion.CASCADE, + to='doc.DocAlias', + ), + ), + ] diff --git a/ietf/ipr/migrations/0003_add_ipdocrel_document2_fk.py b/ietf/ipr/migrations/0003_add_ipdocrel_document2_fk.py deleted file mode 100644 index e5ac7f4413..0000000000 --- a/ietf/ipr/migrations/0003_add_ipdocrel_document2_fk.py +++ /dev/null @@ -1,29 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-08 11:58 - - -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0015_1_add_fk_to_document_id'), - ('ipr', '0002_auto_20180225_1207'), - ] - - operations = [ - migrations.AddField( - model_name='iprdocrel', - name='document2', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.DocAlias', to_field=b'id'), - ), - migrations.AlterField( - model_name='iprdocrel', - name='document', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='old_ipr', to='doc.DocAlias', to_field=b'name'), - ), - ] diff --git a/ietf/ipr/migrations/0003_alter_iprdisclosurebase_docs.py b/ietf/ipr/migrations/0003_alter_iprdisclosurebase_docs.py new file mode 100644 index 0000000000..23b349f567 --- /dev/null +++ b/ietf/ipr/migrations/0003_alter_iprdisclosurebase_docs.py @@ -0,0 +1,18 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations, models + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0017_delete_docalias"), + ("ipr", "0002_iprdocrel_no_aliases"), + ] + + operations = [ + migrations.AlterField( + model_name="iprdisclosurebase", + name="docs", + field=models.ManyToManyField(through="ipr.IprDocRel", to="doc.document"), + ), + ] diff --git a/ietf/ipr/migrations/0004_holderiprdisclosure_is_blanket_disclosure.py b/ietf/ipr/migrations/0004_holderiprdisclosure_is_blanket_disclosure.py new file mode 100644 index 0000000000..66282b3cd5 --- /dev/null +++ b/ietf/ipr/migrations/0004_holderiprdisclosure_is_blanket_disclosure.py @@ -0,0 +1,16 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +from django.db import migrations, models + + +class Migration(migrations.Migration): + dependencies = [ + ("ipr", "0003_alter_iprdisclosurebase_docs"), + ] + + operations = [ + migrations.AddField( + model_name="holderiprdisclosure", + name="is_blanket_disclosure", + field=models.BooleanField(default=False), + ), + ] diff --git a/ietf/ipr/migrations/0004_remove_iprdocrel_document.py b/ietf/ipr/migrations/0004_remove_iprdocrel_document.py deleted file mode 100644 index d5af9d9473..0000000000 --- a/ietf/ipr/migrations/0004_remove_iprdocrel_document.py +++ /dev/null @@ -1,20 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-20 09:53 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('ipr', '0003_add_ipdocrel_document2_fk'), - ] - - operations = [ - migrations.RemoveField( - model_name='iprdocrel', - name='document', - ), - ] diff --git a/ietf/ipr/migrations/0005_removediprdisclosure.py b/ietf/ipr/migrations/0005_removediprdisclosure.py new file mode 100644 index 0000000000..400a264579 --- /dev/null +++ b/ietf/ipr/migrations/0005_removediprdisclosure.py @@ -0,0 +1,28 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +from django.db import migrations, models + + +class Migration(migrations.Migration): + dependencies = [ + ("ipr", "0004_holderiprdisclosure_is_blanket_disclosure"), + ] + + operations = [ + migrations.CreateModel( + name="RemovedIprDisclosure", + fields=[ + ( + "id", + models.AutoField( + auto_created=True, + primary_key=True, + serialize=False, + verbose_name="ID", + ), + ), + ("removed_id", models.PositiveBigIntegerField(unique=True)), + ("reason", models.TextField()), + ], + ), + ] diff --git a/ietf/ipr/migrations/0005_rename_field_document2.py b/ietf/ipr/migrations/0005_rename_field_document2.py deleted file mode 100644 index 4fef4c2274..0000000000 --- a/ietf/ipr/migrations/0005_rename_field_document2.py +++ /dev/null @@ -1,22 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-21 05:31 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0019_rename_field_document2'), - ('ipr', '0004_remove_iprdocrel_document'), - ] - - operations = [ - migrations.RenameField( - model_name='iprdocrel', - old_name='document2', - new_name='document', - ), - ] diff --git a/ietf/ipr/migrations/0006_already_removed_ipr.py b/ietf/ipr/migrations/0006_already_removed_ipr.py new file mode 100644 index 0000000000..0e2dbc63eb --- /dev/null +++ b/ietf/ipr/migrations/0006_already_removed_ipr.py @@ -0,0 +1,24 @@ +# Copyright The IETF Trust 2025, All Rights Reserved +from django.db import migrations + + +def forward(apps, schema_editor): + RemovedIprDisclosure = apps.get_model("ipr", "RemovedIprDisclosure") + for id in (6544, 6068): + RemovedIprDisclosure.objects.create( + removed_id=id, + reason="This IPR disclosure was removed as objectively false.", + ) + + +def reverse(apps, schema_editor): + RemovedIprDisclosure = apps.get_model("ipr", "RemovedIprDisclosure") + RemovedIprDisclosure.objects.all().delete() + + +class Migration(migrations.Migration): + dependencies = [ + ("ipr", "0005_removediprdisclosure"), + ] + + operations = [migrations.RunPython(forward, reverse)] diff --git a/ietf/ipr/migrations/0006_document_primary_key_cleanup.py b/ietf/ipr/migrations/0006_document_primary_key_cleanup.py deleted file mode 100644 index df8f66c549..0000000000 --- a/ietf/ipr/migrations/0006_document_primary_key_cleanup.py +++ /dev/null @@ -1,23 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-06-10 03:42 - - -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('ipr', '0005_rename_field_document2'), - ] - - operations = [ - migrations.AlterField( - model_name='iprdocrel', - name='document', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.DocAlias'), - ), - ] diff --git a/ietf/ipr/migrations/0007_create_ipr_doc_events.py b/ietf/ipr/migrations/0007_create_ipr_doc_events.py deleted file mode 100644 index 395bc11ce9..0000000000 --- a/ietf/ipr/migrations/0007_create_ipr_doc_events.py +++ /dev/null @@ -1,68 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.27 on 2020-01-17 12:32 - - -from django.db import migrations - - -def create_or_delete_ipr_doc_events(apps, delete=False): - """Create or delete DocEvents for IprEvents - - Mostly duplicates IprEvent.create_doc_events(). This is necessary - because model methods, including custom save() methods, are not - available to migrations. - """ - IprEvent = apps.get_model('ipr', 'IprEvent') - DocEvent = apps.get_model('doc', 'DocEvent') - - # Map from self.type_id to DocEvent.EVENT_TYPES for types that - # should be logged as DocEvents - event_type_map = { - 'posted': 'posted_related_ipr', - 'removed': 'removed_related_ipr', - } - - for ipr_event in IprEvent.objects.filter(type_id__in=event_type_map): - related_docs = set() # related docs, no duplicates - for alias in ipr_event.disclosure.docs.all(): - related_docs.update(alias.docs.all()) - for doc in related_docs: - kwargs = dict( - type=event_type_map[ipr_event.type_id], - time=ipr_event.time, - by=ipr_event.by, - doc=doc, - rev='', - desc='%s related IPR disclosure: %s' % (ipr_event.type.name, - ipr_event.disclosure.title), - ) - events = DocEvent.objects.filter(**kwargs) # get existing events - if delete: - events.delete() - elif len(events) == 0: - DocEvent.objects.create(**kwargs) # create if did not exist - -def forward(apps, schema_editor): - """Create a DocEvent for each 'posted' or 'removed' IprEvent""" - create_or_delete_ipr_doc_events(apps, delete=False) - -def reverse(apps, schema_editor): - """Delete DocEvents that would be created by the forward migration - - This removes data, but only data that can be regenerated by running - the forward migration. - """ - create_or_delete_ipr_doc_events(apps, delete=True) - -class Migration(migrations.Migration): - - dependencies = [ - ('ipr', '0006_document_primary_key_cleanup'), - # Ensure the DocEvent types we need exist - ('doc', '0029_add_ipr_event_types'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/ipr/migrations/0008_auto_20201109_0439.py b/ietf/ipr/migrations/0008_auto_20201109_0439.py deleted file mode 100644 index d8140031c4..0000000000 --- a/ietf/ipr/migrations/0008_auto_20201109_0439.py +++ /dev/null @@ -1,21 +0,0 @@ -# Generated by Django 2.2.17 on 2020-11-09 04:39 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('ipr', '0007_create_ipr_doc_events'), - ] - - operations = [ - migrations.AddIndex( - model_name='iprdisclosurebase', - index=models.Index(fields=['-time', '-id'], name='ipr_iprdisc_time_846a78_idx'), - ), - migrations.AddIndex( - model_name='iprevent', - index=models.Index(fields=['-time', '-id'], name='ipr_ipreven_time_9630c4_idx'), - ), - ] diff --git a/ietf/ipr/models.py b/ietf/ipr/models.py index 2eb588b8bc..ea148c2704 100644 --- a/ietf/ipr/models.py +++ b/ietf/ipr/models.py @@ -1,13 +1,14 @@ -# Copyright The IETF Trust 2007-2020, All Rights Reserved +# Copyright The IETF Trust 2007-2025, All Rights Reserved # -*- coding: utf-8 -*- from django.conf import settings +from django.core.exceptions import ValidationError from django.db import models from django.urls import reverse from django.utils import timezone -from ietf.doc.models import DocAlias, DocEvent +from ietf.doc.models import Document, DocEvent from ietf.name.models import DocRelationshipName,IprDisclosureStateName,IprLicenseTypeName,IprEventTypeName from ietf.person.models import Person from ietf.message.models import Message @@ -16,11 +17,11 @@ class IprDisclosureBase(models.Model): by = ForeignKey(Person) # who was logged in, or System if nobody was logged in compliant = models.BooleanField("Complies to RFC3979", default=True) - docs = models.ManyToManyField(DocAlias, through='IprDocRel') + docs = models.ManyToManyField(Document, through='ipr.IprDocRel') holder_legal_name = models.CharField(max_length=255) notes = models.TextField("Additional notes", blank=True) other_designations = models.CharField("Designations for other contributions", blank=True, max_length=255) - rel = models.ManyToManyField('self', through='RelatedIpr', symmetrical=False) + rel = models.ManyToManyField('self', through='ipr.RelatedIpr', symmetrical=False) state = ForeignKey(IprDisclosureStateName) submitter_name = models.CharField(max_length=255,blank=True) submitter_email = models.EmailField(blank=True) @@ -124,17 +125,30 @@ def is_thirdparty(self): class HolderIprDisclosure(IprDisclosureBase): - ietfer_name = models.CharField(max_length=255, blank=True) # "Whose Personal Belief Triggered..." - ietfer_contact_email = models.EmailField(blank=True) - ietfer_contact_info = models.TextField(blank=True) - patent_info = models.TextField() - has_patent_pending = models.BooleanField(default=False) - holder_contact_email = models.EmailField() - holder_contact_name = models.CharField(max_length=255) - holder_contact_info = models.TextField(blank=True, help_text="Address, phone, etc.") - licensing = ForeignKey(IprLicenseTypeName) - licensing_comments = models.TextField(blank=True) + ietfer_name = models.CharField( + max_length=255, blank=True + ) # "Whose Personal Belief Triggered..." + ietfer_contact_email = models.EmailField(blank=True) + ietfer_contact_info = models.TextField(blank=True) + patent_info = models.TextField() + has_patent_pending = models.BooleanField(default=False) + holder_contact_email = models.EmailField() + holder_contact_name = models.CharField(max_length=255) + holder_contact_info = models.TextField(blank=True, help_text="Address, phone, etc.") + licensing = ForeignKey(IprLicenseTypeName) + licensing_comments = models.TextField(blank=True) submitter_claims_all_terms_disclosed = models.BooleanField(default=False) + is_blanket_disclosure = models.BooleanField(default=False) + + def clean(self): + if self.is_blanket_disclosure: + # If the IprLicenseTypeName does not exist, we have a serious problem and a 500 response is ok, + # so not handling failure of the `get()` + royalty_free_licensing = IprLicenseTypeName.objects.get(slug="royalty-free") + if self.licensing_id != royalty_free_licensing.pk: + raise ValidationError( + f'Must select "{royalty_free_licensing.desc}" for a blanket IPR disclosure.') + class ThirdPartyIprDisclosure(IprDisclosureBase): ietfer_name = models.CharField(max_length=255) # "Whose Personal Belief Triggered..." @@ -160,9 +174,10 @@ class GenericIprDisclosure(IprDisclosureBase): class IprDocRel(models.Model): disclosure = ForeignKey(IprDisclosureBase) - document = ForeignKey(DocAlias) + document = ForeignKey(Document) sections = models.TextField(blank=True) revisions = models.CharField(max_length=16,blank=True) # allows strings like 01-07 + originaldocumentaliasname = models.CharField(max_length=255, null=True, blank=True) def doc_type(self): name = self.document.name @@ -175,7 +190,7 @@ def doc_type(self): def formatted_name(self): name = self.document.name - if name.startswith("rfc"): + if len(name) >= 3 and name[:3] in ("rfc", "bcp", "fyi", "std"): return name.upper() #elif self.revisions: # return "%s-%s" % (name, self.revisions) @@ -231,12 +246,10 @@ def create_doc_events(self): event_type_map = { 'posted': 'posted_related_ipr', 'removed': 'removed_related_ipr', + 'removed_objfalse': 'removed_objfalse_related_ipr', } if self.type_id in event_type_map: - related_docs = set() # related docs, no duplicates - for alias in self.disclosure.docs.all(): - related_docs.update(alias.docs.all()) - for doc in related_docs: + for doc in self.disclosure.docs.distinct(): DocEvent.objects.create( type=event_type_map[self.type_id], time=self.time, @@ -257,3 +270,7 @@ class LegacyMigrationIprEvent(IprEvent): """A subclass of IprEvent specifically for capturing contents of legacy_url_0, the text of a disclosure submitted by email""" pass + +class RemovedIprDisclosure(models.Model): + removed_id = models.PositiveBigIntegerField(unique=True) + reason = models.TextField() diff --git a/ietf/ipr/resources.py b/ietf/ipr/resources.py index 665b0ab02f..c4d2c436e6 100644 --- a/ietf/ipr/resources.py +++ b/ietf/ipr/resources.py @@ -1,4 +1,4 @@ -# Copyright The IETF Trust 2015-2019, All Rights Reserved +# Copyright The IETF Trust 2015-2025, All Rights Reserved # -*- coding: utf-8 -*- # Autogenerated by the mkresources management command 2015-03-21 14:05 PDT @@ -11,16 +11,16 @@ from ietf import api -from ietf.ipr.models import ( IprDisclosureBase, IprDocRel, HolderIprDisclosure, ThirdPartyIprDisclosure, +from ietf.ipr.models import ( IprDisclosureBase, IprDocRel, HolderIprDisclosure, RemovedIprDisclosure, ThirdPartyIprDisclosure, RelatedIpr, NonDocSpecificIprDisclosure, GenericIprDisclosure, IprEvent, LegacyMigrationIprEvent ) from ietf.person.resources import PersonResource from ietf.name.resources import IprDisclosureStateNameResource -from ietf.doc.resources import DocAliasResource +from ietf.doc.resources import DocumentResource class IprDisclosureBaseResource(ModelResource): by = ToOneField(PersonResource, 'by') state = ToOneField(IprDisclosureStateNameResource, 'state') - docs = ToManyField(DocAliasResource, 'docs', null=True) + docs = ToManyField(DocumentResource, 'docs', null=True) rel = ToManyField('ietf.ipr.resources.IprDisclosureBaseResource', 'rel', null=True) class Meta: queryset = IprDisclosureBase.objects.all() @@ -45,10 +45,9 @@ class Meta: } api.ipr.register(IprDisclosureBaseResource()) -from ietf.doc.resources import DocAliasResource class IprDocRelResource(ModelResource): disclosure = ToOneField(IprDisclosureBaseResource, 'disclosure') - document = ToOneField(DocAliasResource, 'document') + document = ToOneField(DocumentResource, 'document') class Meta: cache = SimpleCache() queryset = IprDocRel.objects.all() @@ -66,13 +65,12 @@ class Meta: from ietf.person.resources import PersonResource from ietf.name.resources import IprDisclosureStateNameResource, IprLicenseTypeNameResource -from ietf.doc.resources import DocAliasResource class HolderIprDisclosureResource(ModelResource): by = ToOneField(PersonResource, 'by') state = ToOneField(IprDisclosureStateNameResource, 'state') iprdisclosurebase_ptr = ToOneField(IprDisclosureBaseResource, 'iprdisclosurebase_ptr') licensing = ToOneField(IprLicenseTypeNameResource, 'licensing') - docs = ToManyField(DocAliasResource, 'docs', null=True) + docs = ToManyField(DocumentResource, 'docs', null=True) rel = ToManyField(IprDisclosureBaseResource, 'rel', null=True) class Meta: cache = SimpleCache() @@ -111,12 +109,11 @@ class Meta: from ietf.person.resources import PersonResource from ietf.name.resources import IprDisclosureStateNameResource -from ietf.doc.resources import DocAliasResource class ThirdPartyIprDisclosureResource(ModelResource): by = ToOneField(PersonResource, 'by') state = ToOneField(IprDisclosureStateNameResource, 'state') iprdisclosurebase_ptr = ToOneField(IprDisclosureBaseResource, 'iprdisclosurebase_ptr') - docs = ToManyField(DocAliasResource, 'docs', null=True) + docs = ToManyField(DocumentResource, 'docs', null=True) rel = ToManyField(IprDisclosureBaseResource, 'rel', null=True) class Meta: cache = SimpleCache() @@ -168,12 +165,11 @@ class Meta: from ietf.person.resources import PersonResource from ietf.name.resources import IprDisclosureStateNameResource -from ietf.doc.resources import DocAliasResource class NonDocSpecificIprDisclosureResource(ModelResource): by = ToOneField(PersonResource, 'by') state = ToOneField(IprDisclosureStateNameResource, 'state') iprdisclosurebase_ptr = ToOneField(IprDisclosureBaseResource, 'iprdisclosurebase_ptr') - docs = ToManyField(DocAliasResource, 'docs', null=True) + docs = ToManyField(DocumentResource, 'docs', null=True) rel = ToManyField(IprDisclosureBaseResource, 'rel', null=True) class Meta: cache = SimpleCache() @@ -207,12 +203,11 @@ class Meta: from ietf.person.resources import PersonResource from ietf.name.resources import IprDisclosureStateNameResource -from ietf.doc.resources import DocAliasResource class GenericIprDisclosureResource(ModelResource): by = ToOneField(PersonResource, 'by') state = ToOneField(IprDisclosureStateNameResource, 'state') iprdisclosurebase_ptr = ToOneField(IprDisclosureBaseResource, 'iprdisclosurebase_ptr') - docs = ToManyField(DocAliasResource, 'docs', null=True) + docs = ToManyField(DocumentResource, 'docs', null=True) rel = ToManyField(IprDisclosureBaseResource, 'rel', null=True) class Meta: cache = SimpleCache() @@ -300,3 +295,18 @@ class Meta: } api.ipr.register(LegacyMigrationIprEventResource()) + + +class RemovedIprDisclosureResource(ModelResource): + class Meta: + queryset = RemovedIprDisclosure.objects.all() + serializer = api.Serializer() + cache = SimpleCache() + #resource_name = 'removediprdisclosure' + ordering = ['id', ] + filtering = { + "id": ALL, + "removed_id": ALL, + "reason": ALL, + } +api.ipr.register(RemovedIprDisclosureResource()) diff --git a/ietf/ipr/sitemaps.py b/ietf/ipr/sitemaps.py index 0b6ce42829..543272e8f1 100644 --- a/ietf/ipr/sitemaps.py +++ b/ietf/ipr/sitemaps.py @@ -1,11 +1,12 @@ -# Copyright The IETF Trust 2007-2019, All Rights Reserved +# Copyright The IETF Trust 2007-2023, All Rights Reserved # +from django.conf import settings from django.contrib.sitemaps import GenericSitemap from ietf.ipr.models import IprDisclosureBase # changefreq is "never except when it gets updated or withdrawn" # so skip giving one -queryset = IprDisclosureBase.objects.filter(state__in=('posted','removed')) +queryset = IprDisclosureBase.objects.filter(state__in=settings.PUBLISH_IPR_STATES) archive = {'queryset':queryset, 'date_field': 'time', 'allow_empty':True } IPRMap = GenericSitemap(archive) # type: ignore diff --git a/ietf/ipr/templatetags/ipr_filters.py b/ietf/ipr/templatetags/ipr_filters.py index 21d5579bf9..8b3b420c41 100644 --- a/ietf/ipr/templatetags/ipr_filters.py +++ b/ietf/ipr/templatetags/ipr_filters.py @@ -1,10 +1,14 @@ -# Copyright The IETF Trust 2014-2020, All Rights Reserved +# Copyright The IETF Trust 2014-2023, All Rights Reserved # -*- coding: utf-8 -*- +import debug # pyflakes: ignore + from django import template from django.utils.html import format_html +from ietf.doc.models import NewRevisionDocEvent + register = template.Library() @@ -26,3 +30,37 @@ def render_message_for_history(msg): @register.filter def to_class_name(value): return value.__class__.__name__ + +def draft_rev_at_time(iprdocrel): + draft = iprdocrel.document + event = iprdocrel.disclosure.get_latest_event_posted() + if event is None: + return ("","The Internet-Draft's revision at the time this disclosure was posted could not be determined.") + time = event.time + if not NewRevisionDocEvent.objects.filter(doc=draft).exists(): + return ("","The Internet-Draft's revision at the time this disclosure was posted could not be determined.") + rev_event_before = NewRevisionDocEvent.objects.filter(doc=draft, time__lte=time).order_by('-time').first() + if rev_event_before is None: + return ("","The Internet-Draft's initial submission was after this disclosure was posted.") + else: + return(rev_event_before.rev, "") + +@register.filter +def no_revisions_message(iprdocrel): + draft = iprdocrel.document + if draft.type_id != "draft" or iprdocrel.revisions.strip() != "": + return "" + rev_at_time, exception = draft_rev_at_time(iprdocrel) + current_rev = draft.rev + + first_line = "No revisions for this Internet-Draft were specified in this disclosure." + contact_line = "Contact the discloser or patent holder if there are questions about which revisions this disclosure pertains to." + + if current_rev == "00": + return f"{first_line} However, there is only one revision of this Internet-Draft." + + if rev_at_time: + return f"{first_line} The Internet-Draft's revision was {rev_at_time} at the time this disclosure was posted. {contact_line}" + else: + return f"{first_line} {exception} {contact_line}" + diff --git a/ietf/ipr/tests.py b/ietf/ipr/tests.py index fadbb4290e..53a599e2de 100644 --- a/ietf/ipr/tests.py +++ b/ietf/ipr/tests.py @@ -1,31 +1,54 @@ -# Copyright The IETF Trust 2009-2020, All Rights Reserved +# Copyright The IETF Trust 2009-2023, All Rights Reserved # -*- coding: utf-8 -*- import datetime - +import json +from unittest import mock +import re from pyquery import PyQuery from urllib.parse import quote, urlparse from zoneinfo import ZoneInfo from django.conf import settings +from django.test.utils import override_settings from django.urls import reverse as urlreverse from django.utils import timezone +from django.db.models import Max + import debug # pyflakes:ignore -from ietf.doc.models import DocAlias -from ietf.doc.factories import DocumentFactory, WgDraftFactory, WgRfcFactory +from ietf.api.views import EmailIngestionError +from ietf.doc.factories import ( + DocumentFactory, + WgDraftFactory, + WgRfcFactory, + RfcFactory, + NewRevisionDocEventFactory +) +from ietf.doc.utils import prettify_std_name from ietf.group.factories import RoleFactory -from ietf.ipr.factories import HolderIprDisclosureFactory, GenericIprDisclosureFactory, IprEventFactory +from ietf.ipr.factories import ( + HolderIprDisclosureFactory, + GenericIprDisclosureFactory, + IprDisclosureBaseFactory, + IprDocRelFactory, + IprEventFactory, + ThirdPartyIprDisclosureFactory +) +from ietf.ipr.forms import DraftForm, HolderIprDisclosureForm from ietf.ipr.mail import (process_response_email, get_reply_to, get_update_submitter_emails, - get_pseudo_submitter, get_holders, get_update_cc_addrs) -from ietf.ipr.models import (IprDisclosureBase,GenericIprDisclosure,HolderIprDisclosure, - ThirdPartyIprDisclosure) -from ietf.ipr.utils import get_genitive, get_ipr_summary + get_pseudo_submitter, get_holders, get_update_cc_addrs, UndeliverableIprResponseError) +from ietf.ipr.models import (IprDisclosureBase, GenericIprDisclosure, HolderIprDisclosure, RemovedIprDisclosure, + ThirdPartyIprDisclosure, IprEvent) +from ietf.ipr.templatetags.ipr_filters import no_revisions_message +from ietf.ipr.utils import get_genitive, get_ipr_summary, ingest_response_email from ietf.mailtrigger.utils import gather_address_lists +from ietf.message.factories import MessageFactory from ietf.message.models import Message +from ietf.person.factories import PersonFactory from ietf.utils.mail import outbox, empty_outbox, get_payload_text from ietf.utils.test_utils import TestCase, login_testing_unauthorized from ietf.utils.text import text_to_dict @@ -86,9 +109,46 @@ def test_get_update_submitter_emails(self): self.assertTrue(messages[0].startswith('To: %s' % ipr.submitter_email)) def test_showlist(self): + for disc_factory_type in (HolderIprDisclosureFactory, GenericIprDisclosureFactory, ThirdPartyIprDisclosureFactory): + ipr = disc_factory_type(state_id="removed") + r = self.client.get(urlreverse("ietf.ipr.views.showlist")) + self.assertContains(r, ipr.title) + self.assertContains(r, "removed at the request of the submitter") + self.assertNotContains(r, "removed as objectively false") + ipr.state_id="posted" + ipr.save() + r = self.client.get(urlreverse("ietf.ipr.views.showlist")) + self.assertContains(r, ipr.title) + self.assertNotContains(r, "removed at the request of the submitter") + self.assertNotContains(r, "removed as objectively false") + ipr.state_id="removed_objfalse" + ipr.save() + r = self.client.get(urlreverse("ietf.ipr.views.showlist")) + self.assertContains(r, ipr.title) + self.assertNotContains(r, "removed at the request of the submitter") + self.assertContains(r, "removed as objectively false") + ipr.delete() + + def test_show_delete(self): ipr = HolderIprDisclosureFactory() - r = self.client.get(urlreverse("ietf.ipr.views.showlist")) - self.assertContains(r, ipr.title) + removed = RemovedIprDisclosure.objects.create( + removed_id=ipr.pk, reason="Removed for reasons" + ) + url = urlreverse("ietf.ipr.views.show", kwargs=dict(id=removed.removed_id)) + r = self.client.get(url) + self.assertContains(r, "Removed for reasons") + q = PyQuery(r.content) + self.assertEqual(len(q("#deletion_warning")), 0) + self.client.login(username="secretary", password="secretary+password") + r = self.client.get(url) + self.assertContains(r, "Removed for reasons") + q = PyQuery(r.content) + self.assertEqual(len(q("#deletion_warning")), 1) + ipr.delete() + r = self.client.get(url) + self.assertContains(r, "Removed for reasons") + q = PyQuery(r.content) + self.assertEqual(len(q("#deletion_warning")), 0) def test_show_posted(self): ipr = HolderIprDisclosureFactory() @@ -115,17 +175,16 @@ def test_show_removed(self): r = self.client.get(urlreverse("ietf.ipr.views.show", kwargs=dict(id=ipr.pk))) self.assertContains(r, 'This IPR disclosure was removed') + def test_show_removed_objfalse(self): + ipr = HolderIprDisclosureFactory(state_id='removed_objfalse') + r = self.client.get(urlreverse("ietf.ipr.views.show", kwargs=dict(id=ipr.pk))) + self.assertContains(r, 'This IPR disclosure was removed as objectively false') + def test_ipr_history(self): ipr = HolderIprDisclosureFactory() r = self.client.get(urlreverse("ietf.ipr.views.history", kwargs=dict(id=ipr.pk))) self.assertContains(r, ipr.title) - def test_iprs_for_drafts(self): - draft=WgDraftFactory() - ipr = HolderIprDisclosureFactory(docs=[draft,]) - r = self.client.get(urlreverse("ietf.ipr.views.by_draft_txt")) - self.assertContains(r, draft.name) - self.assertContains(r, str(ipr.pk)) def test_about(self): r = self.client.get(urlreverse("ietf.ipr.views.about")) @@ -147,21 +206,56 @@ def test_search(self): r = self.client.get(url + "?submit=draft&id=%s" % draft.name) self.assertContains(r, ipr.title) + # find by id, mixed case letters + r = self.client.get(url + "?submit=draft&id=%s" % draft.name.swapcase()) + self.assertContains(r, ipr.title) + # find draft r = self.client.get(url + "?submit=draft&draft=%s" % draft.name) self.assertContains(r, ipr.title) + # find draft, mixed case letters + r = self.client.get(url + "?submit=draft&draft=%s" % draft.name.swapcase()) + self.assertContains(r, ipr.title) + # search + select document r = self.client.get(url + "?submit=draft&draft=draft") self.assertContains(r, draft.name) self.assertNotContains(r, ipr.title) - DocAlias.objects.create(name="rfc321").docs.add(draft) + rfc = RfcFactory(rfc_number=321) + draft.relateddocument_set.create(relationship_id="became_rfc",target=rfc) # find RFC r = self.client.get(url + "?submit=rfc&rfc=321") self.assertContains(r, ipr.title) + rfc_new = RfcFactory(rfc_number=322) + rfc_new.relateddocument_set.create(relationship_id="obs", target=rfc) + + # find RFC 322 which obsoletes RFC 321 whose draft has IPR + r = self.client.get(url + "?submit=rfc&rfc=322") + self.assertContains(r, ipr.title) + self.assertContains(r, "Total number of IPR disclosures found: 1") + self.assertContains(r, "Total number of documents searched: 3.") + self.assertContains( + r, + f'Results for {prettify_std_name(rfc_new.name)} ("{rfc_new.title}")', + html=True, + ) + self.assertContains( + r, + f'Results for {prettify_std_name(rfc.name)} ("{rfc.title}"), ' + f'which was obsoleted by {prettify_std_name(rfc_new.name)} ("{rfc_new.title}")', + html=True, + ) + self.assertContains( + r, + f'Results for {prettify_std_name(draft.name)} ("{draft.title}"), ' + f'which became rfc {prettify_std_name(rfc.name)} ("{rfc.title}")', + html=True, + ) + # find by patent owner r = self.client.get(url + "?submit=holder&holder=%s" % ipr.holder_legal_name) self.assertContains(r, ipr.title) @@ -185,6 +279,24 @@ def test_search(self): r = self.client.get(url + "?submit=iprtitle&iprtitle=%s" % quote(ipr.title)) self.assertContains(r, ipr.title) + def test_search_null_characters(self): + """IPR search gracefully rejects null characters in parameters""" + # Not a combinatorially exhaustive set, but tries to exercise all the parameters + bad_params = [ + "option=document_search&document_search=draft-\x00stuff" + "submit=dra\x00ft", + "submit=draft&id=some\x00id", + "submit=draft&id_document_tag=some\x00id", + "submit=draft&id=someid&state=re\x00moved", + "submit=draft&id=someid&state=posted&state=re\x00moved", + "submit=draft&id=someid&state=removed&draft=draft-no\x00tvalid", + "submit=rfc&rfc=rfc\x00123", + ] + url = urlreverse("ietf.ipr.views.search") + for query_params in bad_params: + r = self.client.get(f"{url}?{query_params}") + self.assertEqual(r.status_code, 400, f"querystring '{query_params}' should be rejected") + def test_feed(self): ipr = HolderIprDisclosureFactory() r = self.client.get("/feed/ipr/") @@ -197,16 +309,16 @@ def test_sitemap(self): def test_new_generic(self): """Ensure new-generic redirects to new-general""" - url = urlreverse("ietf.ipr.views.new", kwargs={ "type": "generic" }) + url = urlreverse("ietf.ipr.views.new", kwargs={ "_type": "generic" }) r = self.client.get(url) self.assertEqual(r.status_code,302) - self.assertEqual(urlparse(r["Location"]).path, urlreverse("ietf.ipr.views.new", kwargs={ "type": "general"})) + self.assertEqual(urlparse(r["Location"]).path, urlreverse("ietf.ipr.views.new", kwargs={ "_type": "general"})) def test_new_general(self): """Add a new general disclosure. Note: submitter does not need to be logged in. """ - url = urlreverse("ietf.ipr.views.new", kwargs={ "type": "general" }) + url = urlreverse("ietf.ipr.views.new", kwargs={ "_type": "general" }) # invalid post r = self.client.post(url, { @@ -243,8 +355,8 @@ def test_new_specific(self): """Add a new specific disclosure. Note: submitter does not need to be logged in. """ draft = WgDraftFactory() - WgRfcFactory() - url = urlreverse("ietf.ipr.views.new", kwargs={ "type": "specific" }) + rfc = WgRfcFactory() + url = urlreverse("ietf.ipr.views.new", kwargs={ "_type": "specific" }) # successful post empty_outbox() @@ -257,9 +369,9 @@ def test_new_specific(self): "ietfer_contact_info": "555-555-0101", "iprdocrel_set-TOTAL_FORMS": 2, "iprdocrel_set-INITIAL_FORMS": 0, - "iprdocrel_set-0-document": draft.docalias.first().pk, + "iprdocrel_set-0-document": draft.pk, "iprdocrel_set-0-revisions": '00', - "iprdocrel_set-1-document": DocAlias.objects.filter(name__startswith="rfc").first().pk, + "iprdocrel_set-1-document": rfc.pk, "patent_number": "SE12345678901", "patent_inventor": "A. Nonymous", "patent_title": "A method of transferring bits", @@ -297,12 +409,44 @@ def test_new_specific(self): r = self.client.post(url, data) self.assertContains(r, "Your IPR disclosure has been submitted", msg_prefix="Checked patent number: %s" % patent_number) + def test_new_specific_no_revision(self): + draft = WgDraftFactory() + rfc = WgRfcFactory() + url = urlreverse("ietf.ipr.views.new", kwargs={ "_type": "specific" }) + + # successful post + empty_outbox() + data = { + "holder_legal_name": "Test Legal", + "holder_contact_name": "Test Holder", + "holder_contact_email": "test@holder.com", + "holder_contact_info": "555-555-0100", + "ietfer_name": "Test Participant", + "ietfer_contact_info": "555-555-0101", + "iprdocrel_set-TOTAL_FORMS": 2, + "iprdocrel_set-INITIAL_FORMS": 0, + "iprdocrel_set-0-document": draft.pk, + "iprdocrel_set-1-document": rfc.pk, + "patent_number": "SE12345678901", + "patent_inventor": "A. Nonymous", + "patent_title": "A method of transferring bits", + "patent_date": "2000-01-01", + "has_patent_pending": False, + "licensing": "royalty-free", + "submitter_name": "Test Holder", + "submitter_email": "test@holder.com", + } + r = self.client.post(url, data) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertTrue(q("#id_iprdocrel_set-0-revisions").hasClass("is-invalid")) + def test_new_thirdparty(self): """Add a new third-party disclosure. Note: submitter does not need to be logged in. """ draft = WgDraftFactory() - WgRfcFactory() - url = urlreverse("ietf.ipr.views.new", kwargs={ "type": "third-party" }) + rfc = WgRfcFactory() + url = urlreverse("ietf.ipr.views.new", kwargs={ "_type": "third-party" }) # successful post empty_outbox() @@ -313,9 +457,9 @@ def test_new_thirdparty(self): "ietfer_contact_info": "555-555-0101", "iprdocrel_set-TOTAL_FORMS": 2, "iprdocrel_set-INITIAL_FORMS": 0, - "iprdocrel_set-0-document": draft.docalias.first().pk, + "iprdocrel_set-0-document": draft.pk, "iprdocrel_set-0-revisions": '00', - "iprdocrel_set-1-document": DocAlias.objects.filter(name__startswith="rfc").first().pk, + "iprdocrel_set-1-document": rfc.pk, "patent_number": "SE12345678901", "patent_inventor": "A. Nonymous", "patent_title": "A method of transferring bits", @@ -349,7 +493,7 @@ def test_edit(self): r = self.client.get(url) self.assertContains(r, original_ipr.holder_legal_name) - #url = urlreverse("ietf.ipr.views.new", kwargs={ "type": "specific" }) + #url = urlreverse("ietf.ipr.views.new", kwargs={ "_type": "specific" }) # successful post empty_outbox() post_data = { @@ -360,7 +504,7 @@ def test_edit(self): "holder_legal_name": "Test Legal", "ietfer_contact_info": "555-555-0101", "ietfer_name": "Test Participant", - "iprdocrel_set-0-document": draft.docalias.first().pk, + "iprdocrel_set-0-document": draft.pk, "iprdocrel_set-0-revisions": '00', "iprdocrel_set-INITIAL_FORMS": 0, "iprdocrel_set-TOTAL_FORMS": 1, @@ -388,7 +532,7 @@ def test_edit(self): def test_update(self): draft = WgDraftFactory() - WgRfcFactory() + rfc = WgRfcFactory() original_ipr = HolderIprDisclosureFactory(docs=[draft,]) # get @@ -396,7 +540,7 @@ def test_update(self): r = self.client.get(url) self.assertContains(r, original_ipr.title) - #url = urlreverse("ietf.ipr.views.new", kwargs={ "type": "specific" }) + #url = urlreverse("ietf.ipr.views.new", kwargs={ "_type": "specific" }) # successful post empty_outbox() r = self.client.post(url, { @@ -409,9 +553,9 @@ def test_update(self): "ietfer_contact_info": "555-555-0101", "iprdocrel_set-TOTAL_FORMS": 2, "iprdocrel_set-INITIAL_FORMS": 0, - "iprdocrel_set-0-document": draft.docalias.first().pk, + "iprdocrel_set-0-document": draft.pk, "iprdocrel_set-0-revisions": '00', - "iprdocrel_set-1-document": DocAlias.objects.filter(name__startswith="rfc").first().pk, + "iprdocrel_set-1-document": rfc.pk, "patent_number": "SE12345678901", "patent_inventor": "A. Nonymous", "patent_title": "A method of transferring bits", @@ -436,7 +580,7 @@ def test_update(self): def test_update_bad_post(self): draft = WgDraftFactory() - url = urlreverse("ietf.ipr.views.new", kwargs={ "type": "specific" }) + url = urlreverse("ietf.ipr.views.new", kwargs={ "_type": "specific" }) empty_outbox() r = self.client.post(url, { @@ -446,7 +590,7 @@ def test_update_bad_post(self): "holder_contact_email": "test@holder.com", "iprdocrel_set-TOTAL_FORMS": 1, "iprdocrel_set-INITIAL_FORMS": 0, - "iprdocrel_set-0-document": draft.docalias.first().pk, + "iprdocrel_set-0-document": draft.pk, "iprdocrel_set-0-revisions": '00', "patent_number": "SE12345678901", "patent_inventor": "A. Nonymous", @@ -525,7 +669,7 @@ def test_admin_removed(self): self.client.login(username="secretary", password="secretary+password") # test for presence of pending ipr - num = IprDisclosureBase.objects.filter(state__in=('removed','rejected')).count() + num = IprDisclosureBase.objects.filter(state__in=('removed','removed_objfalse','rejected')).count() r = self.client.get(url) self.assertEqual(r.status_code,200) @@ -583,7 +727,7 @@ def test_notify(self): get_payload_text(outbox[len_before + 1]).replace('\n', ' ') ) self.assertIn(f'{settings.IDTRACKER_BASE_URL}{urlreverse("ietf.ipr.views.showlist")}', get_payload_text(outbox[len_before]).replace('\n',' ')) - self.assertIn(f'{settings.IDTRACKER_BASE_URL}{urlreverse("ietf.ipr.views.history",kwargs=dict(id=ipr.pk))}', get_payload_text(outbox[len_before+1]).replace('\n',' ')) + self.assertIn(f'{settings.IDTRACKER_BASE_URL}{urlreverse("ietf.ipr.views.show",kwargs=dict(id=ipr.pk))}', get_payload_text(outbox[len_before+1]).replace('\n',' ')) def test_notify_generic(self): RoleFactory(name_id='ad',group__acronym='gen') @@ -605,8 +749,8 @@ def test_notify_generic(self): ) self.assertIn(f'{settings.IDTRACKER_BASE_URL}{urlreverse("ietf.ipr.views.showlist")}', get_payload_text(outbox[1]).replace('\n',' ')) - def send_ipr_email_helper(self): - ipr = HolderIprDisclosureFactory() + def send_ipr_email_helper(self) -> tuple[str, IprEvent, HolderIprDisclosure]: + ipr = HolderIprDisclosureFactory.create() # call create() explicitly so mypy sees correct type url = urlreverse('ietf.ipr.views.email',kwargs={ "id": ipr.id }) self.client.login(username="secretary", password="secretary+password") yesterday = date_today() - datetime.timedelta(1) @@ -623,10 +767,11 @@ def send_ipr_email_helper(self): q = Message.objects.filter(reply_to=data['reply_to']) self.assertEqual(q.count(),1) event = q[0].msgevents.first() + assert event is not None self.assertTrue(event.response_past_due()) self.assertEqual(len(outbox), 1) self.assertTrue('joe@test.com' in outbox[0]['To']) - return data['reply_to'], event + return data['reply_to'], event, ipr uninteresting_ipr_message_strings = [ ("To: {to}\nCc: {cc}\nFrom: joe@test.com\nDate: {date}\nSubject: test\n"), @@ -640,34 +785,46 @@ def send_ipr_email_helper(self): def test_process_response_email(self): # first send a mail - reply_to, event = self.send_ipr_email_helper() + reply_to, event, _ = self.send_ipr_email_helper() # test process response uninteresting messages addrs = gather_address_lists('ipr_disclosure_submitted').as_strings() for message_string in self.uninteresting_ipr_message_strings: - result = process_response_email( + process_response_email( message_string.format( to=addrs.to, cc=addrs.cc, date=timezone.now().ctime() ) ) - self.assertIsNone(result) - + # test process response message_string = """To: {} From: joe@test.com Date: {} Subject: test """.format(reply_to, timezone.now().ctime()) - result = process_response_email(message_string) - - self.assertIsInstance(result, Message) + process_response_email(message_string) self.assertFalse(event.response_past_due()) + # test with an unmatchable message identifier + bad_reply_to = re.sub( + r"\+.{16}@", + '+0123456789abcdef@', + reply_to, + ) + self.assertNotEqual(reply_to, bad_reply_to) + message_string = f"""To: {bad_reply_to} + From: joe@test.com + Date: {timezone.now().ctime()} + Subject: test + """ + with self.assertRaises(UndeliverableIprResponseError): + process_response_email(message_string) + def test_process_response_email_with_invalid_encoding(self): """Interesting emails with invalid encoding should be handled""" - reply_to, _ = self.send_ipr_email_helper() + reply_to, _, disclosure = self.send_ipr_email_helper() # test process response message_string = """To: {} From: joe@test.com @@ -675,8 +832,8 @@ def test_process_response_email_with_invalid_encoding(self): Subject: test """.format(reply_to, timezone.now().ctime()) message_bytes = message_string.encode('utf8') + b'\nInvalid stuff: \xfe\xff\n' - result = process_response_email(message_bytes) - self.assertIsInstance(result, Message) + process_response_email(message_bytes) + result = IprEvent.objects.filter(disclosure=disclosure).first().message # newest # \ufffd is a rhombus character with an inverse ?, used to replace invalid characters self.assertEqual(result.body, 'Invalid stuff: \ufffd\ufffd\n\n', # not sure where the extra \n is from 'Invalid characters should be replaced with \ufffd characters') @@ -691,8 +848,45 @@ def test_process_response_email_uninteresting_with_invalid_encoding(self): cc=addrs.cc, date=timezone.now().ctime(), ).encode('utf8') + b'\nInvalid stuff: \xfe\xff\n' - result = process_response_email(message_bytes) - self.assertIsNone(result) + process_response_email(message_bytes) + + @override_settings(ADMINS=(("Some Admin", "admin@example.com"),)) + @mock.patch("ietf.ipr.utils.process_response_email") + def test_ingest_response_email(self, mock_process_response_email): + message = b"What a nice message" + mock_process_response_email.side_effect = ValueError("ouch!") + with self.assertRaises(EmailIngestionError) as context: + ingest_response_email(message) + self.assertIsNone(context.exception.email_recipients) # default recipients + self.assertIsNotNone(context.exception.email_body) # body set + self.assertIsNotNone(context.exception.email_original_message) # original message attached + self.assertEqual(context.exception.email_attach_traceback, True) + self.assertTrue(mock_process_response_email.called) + self.assertEqual(mock_process_response_email.call_args, mock.call(message)) + mock_process_response_email.reset_mock() + + mock_process_response_email.side_effect = UndeliverableIprResponseError + mock_process_response_email.return_value = None + with self.assertRaises(EmailIngestionError) as context: + ingest_response_email(message) + self.assertIsNone(context.exception.as_emailmessage()) # should not send an email on a clean rejection + self.assertTrue(mock_process_response_email.called) + self.assertEqual(mock_process_response_email.call_args, mock.call(message)) + mock_process_response_email.reset_mock() + + mock_process_response_email.side_effect = None + mock_process_response_email.return_value = None # ignored message + ingest_response_email(message) # should not raise an exception + self.assertIsNone(context.exception.as_emailmessage()) # should not send an email on ignored message + self.assertTrue(mock_process_response_email.called) + self.assertEqual(mock_process_response_email.call_args, mock.call(message)) + mock_process_response_email.reset_mock() + + # successful operation + mock_process_response_email.return_value = MessageFactory() + ingest_response_email(message) + self.assertTrue(mock_process_response_email.called) + self.assertEqual(mock_process_response_email.call_args, mock.call(message)) def test_ajax_search(self): url = urlreverse('ietf.ipr.views.ajax_search') @@ -710,9 +904,9 @@ def test_edit_using_factory(self): post_data = { 'iprdocrel_set-TOTAL_FORMS' : 1, 'iprdocrel_set-INITIAL_FORMS' : 0, - 'iprdocrel_set-0-id': disclosure.pk, + 'iprdocrel_set-0-id': '', "iprdocrel_set-0-document": disclosure.docs.first().pk, - "iprdocrel_set-0-revisions": disclosure.docs.first().document.rev, + "iprdocrel_set-0-revisions": disclosure.docs.first().rev, 'holder_legal_name': disclosure.holder_legal_name, 'patent_number': patent_dict['Number'], 'patent_title': patent_dict['Title'], @@ -734,18 +928,28 @@ def test_docevent_creation(self): 'New Document already has a "posted_related_ipr" DocEvent') self.assertEqual(0, doc.docevent_set.filter(type='removed_related_ipr').count(), 'New Document already has a "removed_related_ipr" DocEvent') + self.assertEqual(0, doc.docevent_set.filter(type='removed_objfalse_related_ipr').count(), + 'New Document already has a "removed_objfalse_related_ipr" DocEvent') # A 'posted' IprEvent must create a corresponding DocEvent IprEventFactory(type_id='posted', disclosure=ipr) self.assertEqual(1, doc.docevent_set.filter(type='posted_related_ipr').count(), 'Creating "posted" IprEvent did not create a "posted_related_ipr" DocEvent') self.assertEqual(0, doc.docevent_set.filter(type='removed_related_ipr').count(), 'Creating "posted" IprEvent created a "removed_related_ipr" DocEvent') + self.assertEqual(0, doc.docevent_set.filter(type='removed_objfalse_related_ipr').count(), + 'Creating "posted" IprEvent created a "removed_objfalse_related_ipr" DocEvent') # A 'removed' IprEvent must create a corresponding DocEvent IprEventFactory(type_id='removed', disclosure=ipr) self.assertEqual(1, doc.docevent_set.filter(type='posted_related_ipr').count(), 'Creating "removed" IprEvent created a "posted_related_ipr" DocEvent') self.assertEqual(1, doc.docevent_set.filter(type='removed_related_ipr').count(), 'Creating "removed" IprEvent did not create a "removed_related_ipr" DocEvent') + # A 'removed_objfalse' IprEvent must create a corresponding DocEvent + IprEventFactory(type_id='removed_objfalse', disclosure=ipr) + self.assertEqual(1, doc.docevent_set.filter(type='posted_related_ipr').count(), + 'Creating "removed_objfalse" IprEvent created a "posted_related_ipr" DocEvent') + self.assertEqual(1, doc.docevent_set.filter(type='removed_objfalse_related_ipr').count(), + 'Creating "removed_objfalse" IprEvent did not create a "removed_objfalse_related_ipr" DocEvent') # The DocEvent descriptions must refer to the IprEvents posted_docevent = doc.docevent_set.filter(type='posted_related_ipr').first() self.assertIn(ipr.title, posted_docevent.desc, @@ -753,4 +957,236 @@ def test_docevent_creation(self): removed_docevent = doc.docevent_set.filter(type='removed_related_ipr').first() self.assertIn(ipr.title, removed_docevent.desc, 'IprDisclosure title does not appear in DocEvent desc when removed') + removed_objfalse_docevent = doc.docevent_set.filter(type='removed_objfalse_related_ipr').first() + self.assertIn(ipr.title, removed_objfalse_docevent.desc, + 'IprDisclosure title does not appear in DocEvent desc when removed as objectively false') + + def test_no_revisions_message(self): + draft = WgDraftFactory(rev="02") + now = timezone.now() + for rev in range(0,3): + NewRevisionDocEventFactory(doc=draft, rev=f"{rev:02d}", time=now-datetime.timedelta(days=30*(2-rev))) + + # Disclosure has non-empty revisions field on its related draft + iprdocrel = IprDocRelFactory(document=draft) + IprEventFactory(type_id="posted",time=now,disclosure=iprdocrel.disclosure) + self.assertEqual( + no_revisions_message(iprdocrel), + "" + ) + + # Disclosure has more than one revision, none called out, disclosure after submissions + iprdocrel = IprDocRelFactory(document=draft, revisions="") + IprEventFactory(type_id="posted",time=now,disclosure=iprdocrel.disclosure) + self.assertEqual( + no_revisions_message(iprdocrel), + "No revisions for this Internet-Draft were specified in this disclosure. The Internet-Draft's revision was 02 at the time this disclosure was posted. Contact the discloser or patent holder if there are questions about which revisions this disclosure pertains to." + ) + + # Disclosure has more than one revision, none called out, disclosure after 01 + iprdocrel = IprDocRelFactory(document=draft, revisions="") + e = IprEventFactory(type_id="posted",disclosure=iprdocrel.disclosure) + e.time = now-datetime.timedelta(days=15) + e.save() + self.assertEqual( + no_revisions_message(iprdocrel), + "No revisions for this Internet-Draft were specified in this disclosure. The Internet-Draft's revision was 01 at the time this disclosure was posted. Contact the discloser or patent holder if there are questions about which revisions this disclosure pertains to." + ) + + # Disclosure has more than one revision, none called out, disclosure was before the 00 + iprdocrel = IprDocRelFactory(document=draft, revisions="") + e = IprEventFactory(type_id="posted",disclosure=iprdocrel.disclosure) + e.time = now-datetime.timedelta(days=180) + e.save() + self.assertEqual( + no_revisions_message(iprdocrel), + "No revisions for this Internet-Draft were specified in this disclosure. The Internet-Draft's initial submission was after this disclosure was posted. Contact the discloser or patent holder if there are questions about which revisions this disclosure pertains to." + ) + + # disclosed draft has no NewRevisionDocEvents + draft = WgDraftFactory(rev="20") + draft.docevent_set.all().delete() + iprdocrel = IprDocRelFactory(document=draft, revisions="") + IprEventFactory(type_id="posted",disclosure=iprdocrel.disclosure) + self.assertEqual( + no_revisions_message(iprdocrel), + "No revisions for this Internet-Draft were specified in this disclosure. The Internet-Draft's revision at the time this disclosure was posted could not be determined. Contact the discloser or patent holder if there are questions about which revisions this disclosure pertains to." + ) + + # disclosed draft has only one revision + draft = WgDraftFactory(rev="00") + iprdocrel = IprDocRelFactory(document=draft, revisions="") + IprEventFactory(type_id="posted",disclosure=iprdocrel.disclosure) + self.assertEqual( + no_revisions_message(iprdocrel), + "No revisions for this Internet-Draft were specified in this disclosure. However, there is only one revision of this Internet-Draft." + ) + + +class DraftFormTests(TestCase): + def setUp(self): + super().setUp() + self.disclosure = IprDisclosureBaseFactory() + self.draft = WgDraftFactory.create_batch(10)[-1] + self.rfc = RfcFactory() + + def test_revisions_valid(self): + post_data = { + # n.b., "document" is a SearchableDocumentField, which is a multiple choice field limited + # to a single choice. Its value must be an array of pks with one element. + "document": [str(self.draft.pk)], + "disclosure": str(self.disclosure.pk), + } + # The revisions field is just a char field that allows descriptions of the applicable + # document revisions. It's usually just a rev or "00-02", but the form allows anything + # not empty. The secretariat will review the value before the disclosure is posted so + # minimal validation is ok here. + self.assertTrue(DraftForm(post_data | {"revisions": "00"}).is_valid()) + self.assertTrue(DraftForm(post_data | {"revisions": "00-02"}).is_valid()) + self.assertTrue(DraftForm(post_data | {"revisions": "01,03, 05"}).is_valid()) + self.assertTrue(DraftForm(post_data | {"revisions": "all but 01"}).is_valid()) + # RFC instead of draft - allow empty / missing revisions + post_data["document"] = [str(self.rfc.pk)] + self.assertTrue(DraftForm(post_data).is_valid()) + self.assertTrue(DraftForm(post_data | {"revisions": ""}).is_valid()) + + def test_revisions_invalid(self): + missing_rev_error_msg = ( + "Revisions of this Internet-Draft for which this disclosure is relevant must be specified." + ) + null_char_error_msg = "Null characters are not allowed." + + post_data = { + # n.b., "document" is a SearchableDocumentField, which is a multiple choice field limited + # to a single choice. Its value must be an array of pks with one element. + "document": [str(self.draft.pk)], + "disclosure": str(self.disclosure.pk), + } + self.assertFormError( + DraftForm(post_data), "revisions", missing_rev_error_msg + ) + self.assertFormError( + DraftForm(post_data | {"revisions": ""}), "revisions", missing_rev_error_msg + ) + self.assertFormError( + DraftForm(post_data | {"revisions": "1\x00"}), + "revisions", + [null_char_error_msg, missing_rev_error_msg], + ) + # RFC instead of draft still validates the revisions field + self.assertFormError( + DraftForm(post_data | {"document": [str(self.rfc.pk)], "revisions": "1\x00"}), + "revisions", + null_char_error_msg, + ) + + +class HolderIprDisclosureFormTests(TestCase): + def setUp(self): + super().setUp() + # Checkboxes that are False are left out of the Form data, not sent back at all. These are + # commented out - if they were checked, their value would be "on". + self.data = { + "holder_legal_name": "Test Legal", + "holder_contact_name": "Test Holder", + "holder_contact_email": "test@holder.com", + "holder_contact_info": "555-555-0100", + "ietfer_name": "Test Participant", + "ietfer_contact_info": "555-555-0101", + "iprdocrel_set-TOTAL_FORMS": 2, + "iprdocrel_set-INITIAL_FORMS": 0, + "iprdocrel_set-0-document": "1234", # fake id - validates but won't save() + "iprdocrel_set-0-revisions": '00', + "iprdocrel_set-1-document": "4567", # fake id - validates but won't save() + # "is_blanket_disclosure": "on", + "patent_number": "SE12345678901", + "patent_inventor": "A. Nonymous", + "patent_title": "A method of transferring bits", + "patent_date": "2000-01-01", + # "has_patent_pending": "on", + "licensing": "reasonable", + "submitter_name": "Test Holder", + "submitter_email": "test@holder.com", + } + + def test_blanket_disclosure_licensing_restrictions(self): + """when is_blanket_disclosure is True only royalty-free licensing is valid + Most of the form functionality is tested via the views in IprTests above. More thorough testing + of validation ought to move here so we don't have to exercise the whole Django plumbing repeatedly. + """ + self.assertTrue(HolderIprDisclosureForm(data=self.data).is_valid()) + self.data["is_blanket_disclosure"] = "on" + self.assertFalse(HolderIprDisclosureForm(data=self.data).is_valid()) + self.data["licensing"] = "royalty-free" + self.assertTrue(HolderIprDisclosureForm(data=self.data).is_valid()) + + def test_patent_details_required_unless_blanket(self): + self.assertTrue(HolderIprDisclosureForm(data=self.data).is_valid()) + patent_fields = ["patent_number", "patent_inventor", "patent_title", "patent_date"] + # any of the fields being missing should invalidate the form + for pf in patent_fields: + val = self.data.pop(pf) + self.assertFalse(HolderIprDisclosureForm(data=self.data).is_valid()) + self.data[pf] = val + + # should be optional if is_blanket_disclosure is True + self.data["is_blanket_disclosure"] = "on" + self.data["licensing"] = "royalty-free" # also needed for a blanket disclosure + for pf in patent_fields: + val = self.data.pop(pf) + self.assertTrue(HolderIprDisclosureForm(data=self.data).is_valid()) + self.data[pf] = val + +class JsonSnapshotTests(TestCase): + def test_json_snapshot(self): + h = HolderIprDisclosureFactory() + url = urlreverse("ietf.ipr.views.json_snapshot", kwargs=dict(id=h.id)) + login_testing_unauthorized(self, "secretary", url) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + dump = json.loads(r.content) + self.assertCountEqual( + [o["model"] for o in dump], + ["ipr.holderiprdisclosure", "ipr.iprdisclosurebase", "person.person"], + ) + h.docs.add(WgRfcFactory()) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + dump = json.loads(r.content) + self.assertCountEqual( + [o["model"] for o in dump], + [ + "ipr.holderiprdisclosure", + "ipr.iprdisclosurebase", + "ipr.iprdocrel", + "person.person", + ], + ) + IprEventFactory( + disclosure=h, + message=MessageFactory(by=PersonFactory()), + in_reply_to=MessageFactory(), + ) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + dump = json.loads(r.content) + self.assertCountEqual( + [o["model"] for o in dump], + [ + "ipr.holderiprdisclosure", + "ipr.iprdisclosurebase", + "ipr.iprdocrel", + "ipr.iprevent", + "message.message", + "message.message", + "person.person", + "person.person", + "person.person", + "person.person", + ], + ) + no_such_ipr_id = IprDisclosureBase.objects.aggregate(Max("id"))["id__max"] + 1 + url = urlreverse("ietf.ipr.views.json_snapshot", kwargs=dict(id=no_such_ipr_id)) + r = self.client.get(url) + self.assertEqual(r.status_code, 404) diff --git a/ietf/ipr/urls.py b/ietf/ipr/urls.py index 6f7b2d4080..2c8a26c624 100644 --- a/ietf/ipr/urls.py +++ b/ietf/ipr/urls.py @@ -12,8 +12,6 @@ url(r'^admin/$', RedirectView.as_view(url=reverse_lazy('ietf.ipr.views.admin',kwargs={'state':'pending'}), permanent=True)), url(r'^admin/(?Ppending|removed|parked)/$', views.admin), url(r'^ajax/search/$', views.ajax_search), - url(r'^by-draft/$', views.by_draft_txt), - url(r'^by-draft-recursive/$', views.by_draft_recursive_txt), url(r'^(?P\d+)/$', views.show), url(r'^(?P\d+)/addcomment/$', views.add_comment), url(r'^(?P\d+)/addemail/$', views.add_email), @@ -23,8 +21,9 @@ url(r'^(?P\d+)/notify/(?Pupdate|posted)/$', views.notify), url(r'^(?P\d+)/post/$', views.post), url(r'^(?P\d+)/state/$', views.state), + url(r'^(?P\d+)/json-snapshot/$', views.json_snapshot), url(r'^update/$', RedirectView.as_view(url=reverse_lazy('ietf.ipr.views.showlist'), permanent=True)), url(r'^update/(?P\d+)/$', views.update), - url(r'^new-(?P(specific|generic|general|third-party))/$', views.new), + url(r'^new-(?P<_type>(specific|generic|general|third-party))/$', views.new), url(r'^search/$', views.search), ] diff --git a/ietf/ipr/utils.py b/ietf/ipr/utils.py index 43f5494b1c..bcbb052260 100644 --- a/ietf/ipr/utils.py +++ b/ietf/ipr/utils.py @@ -1,9 +1,16 @@ -# Copyright The IETF Trust 2014-2020, All Rights Reserved +# Copyright The IETF Trust 2014-2025, All Rights Reserved # -*- coding: utf-8 -*- -from ietf.ipr.models import IprDocRel +import json +import debug # pyflakes:ignore + +from textwrap import dedent + +from django.core import serializers -import debug # pyflakes:ignore +from ietf.ipr.mail import process_response_email, UndeliverableIprResponseError + +from ietf.ipr.models import IprDocRel def get_genitive(name): """Return the genitive form of name""" @@ -32,60 +39,69 @@ def get_ipr_summary(disclosure): return summary if len(summary) <= 128 else summary[:125]+'...' -def iprs_from_docs(aliases,**kwargs): - """Returns a list of IPRs related to doc aliases""" +def iprs_from_docs(docs,**kwargs): + """Returns a list of IPRs related to docs""" iprdocrels = [] - for alias in aliases: - for document in alias.docs.all(): - if document.ipr(**kwargs): - iprdocrels += document.ipr(**kwargs) + for document in docs: + if document.ipr(**kwargs): + iprdocrels += document.ipr(**kwargs) return list(set([i.disclosure for i in iprdocrels])) -def related_docs(alias, relationship=('replaces', 'obs')): +def related_docs(doc, relationship=('replaces', 'obs'), reverse_relationship=("became_rfc",)): """Returns list of related documents""" - results = [] - for doc in alias.docs.all(): - results += list(doc.docalias.all()) - - rels = [] - for doc in alias.docs.all(): - rels += list(doc.all_relations_that_doc(relationship)) - - for rel in rels: - rel_aliases = list(rel.target.document.docalias.all()) - - for x in rel_aliases: - x.related = rel - x.relation = rel.relationship.revname - results += rel_aliases - - return list(set(results)) + results = [doc] + rels = doc.all_relations_that_doc(relationship) -def generate_draft_recursive_txt(): - docipr = {} + for rel in rels: + rel.target.related = rel + rel.target.relation = rel.relationship.revname + results += [x.target for x in rels] - for o in IprDocRel.objects.filter(disclosure__state='posted').select_related('document'): - alias = o.document - name = alias.name - for document in alias.docs.all(): - related = set(document.docalias.all()) | set(document.all_related_that_doc(('obs', 'replaces'))) - for alias in related: - name = alias.name - if name.startswith("rfc"): - name = name.upper() - if not name in docipr: - docipr[name] = [] - docipr[name].append(o.disclosure_id) + rev_rels = doc.all_relations_that(reverse_relationship) + for rel in rev_rels: + rel.source.related = rel + rel.source.relation = rel.relationship.name + results += [x.source for x in rev_rels] - lines = [ "# Machine-readable list of IPR disclosures by draft name" ] - for name, iprs in docipr.items(): - lines.append(name + "\t" + "\t".join(str(ipr_id) for ipr_id in sorted(iprs))) + return list(set(results)) - data = '\n'.join(lines) - filename = '/a/ietfdata/derived/ipr_draft_recursive.txt' - with open(filename, 'w') as f: - f.write(data) +def ingest_response_email(message: bytes): + from ietf.api.views import EmailIngestionError # avoid circular import + try: + process_response_email(message) + except UndeliverableIprResponseError: + # Message was rejected due to some problem the sender can fix, so bounce but don't send + # an email to the admins + raise EmailIngestionError("IPR response rejected", email_body=None) + except Exception as err: + # Message was rejected due to an unhandled exception. This is likely something + # the admins need to address, so send them a copy of the email. + raise EmailIngestionError( + "Datatracker IPR email ingestion error", + email_body=dedent("""\ + An error occurred while ingesting IPR email into the Datatracker. The original message is attached. + + {error_summary} + """), + email_original_message=message, + email_attach_traceback=True, + ) from err + +def json_dump_disclosure(disclosure): + objs = set() + objs.add(disclosure) + objs.add(disclosure.iprdisclosurebase_ptr) + objs.add(disclosure.by) + objs.update(IprDocRel.objects.filter(disclosure=disclosure)) + objs.update(disclosure.iprevent_set.all()) + objs.update([i.by for i in disclosure.iprevent_set.all()]) + objs.update([i.message for i in disclosure.iprevent_set.all() if i.message ]) + objs.update([i.message.by for i in disclosure.iprevent_set.all() if i.message ]) + objs.update([i.in_reply_to for i in disclosure.iprevent_set.all() if i.in_reply_to ]) + objs.update([i.in_reply_to.by for i in disclosure.iprevent_set.all() if i.in_reply_to ]) + objs = sorted(list(objs),key=lambda o:o.__class__.__name__) + return json.dumps(json.loads(serializers.serialize("json",objs)),indent=4) diff --git a/ietf/ipr/views.py b/ietf/ipr/views.py index 2450e53137..0a43ff2c27 100644 --- a/ietf/ipr/views.py +++ b/ietf/ipr/views.py @@ -1,4 +1,4 @@ -# Copyright The IETF Trust 2007-2022, All Rights Reserved +# Copyright The IETF Trust 2007-2023, All Rights Reserved # -*- coding: utf-8 -*- @@ -10,7 +10,7 @@ from django.db.models import Q from django.forms.models import inlineformset_factory, model_to_dict from django.forms.formsets import formset_factory -from django.http import HttpResponse, Http404, HttpResponseRedirect +from django.http import HttpResponse, Http404, HttpResponseRedirect, HttpResponseBadRequest from django.shortcuts import render, get_object_or_404, redirect from django.template.loader import render_to_string from django.urls import reverse as urlreverse @@ -18,7 +18,7 @@ import debug # pyflakes:ignore -from ietf.doc.models import DocAlias +from ietf.doc.models import Document from ietf.group.models import Role, Group from ietf.ietfauth.utils import role_required, has_role from ietf.ipr.mail import (message_from_message, get_reply_to, get_update_submitter_emails) @@ -28,17 +28,17 @@ AddCommentForm, AddEmailForm, NotifyForm, StateForm, NonDocSpecificIprDisclosureForm, GenericIprDisclosureForm) from ietf.ipr.models import (IprDisclosureStateName, IprDisclosureBase, - HolderIprDisclosure, GenericIprDisclosure, ThirdPartyIprDisclosure, + HolderIprDisclosure, GenericIprDisclosure, RemovedIprDisclosure, ThirdPartyIprDisclosure, NonDocSpecificIprDisclosure, IprDocRel, RelatedIpr,IprEvent) from ietf.ipr.utils import (get_genitive, get_ipr_summary, - iprs_from_docs, related_docs) + iprs_from_docs, json_dump_disclosure, related_docs) from ietf.mailtrigger.utils import gather_address_lists from ietf.message.models import Message from ietf.message.utils import infer_message from ietf.name.models import IprLicenseTypeName from ietf.person.models import Person -from ietf.secr.utils.document import get_rfc_num, is_draft +from ietf.utils import log from ietf.utils.draft_search import normalize_draftname from ietf.utils.mail import send_mail, send_mail_message from ietf.utils.response import permission_denied @@ -69,16 +69,20 @@ def get_document_emails(ipr): has been posted""" messages = [] for rel in ipr.iprdocrel_set.all(): - doc = rel.document.document + doc = rel.document - if is_draft(doc): + if doc.type_id=="draft": doc_info = 'Internet-Draft entitled "{}" ({})'.format(doc.title,doc.name) + elif doc.type_id=="rfc": + doc_info = 'RFC entitled "{}" (RFC{})'.format(doc.title, doc.rfc_number) else: - doc_info = 'RFC entitled "{}" (RFC{})'.format(doc.title,get_rfc_num(doc)) + log.unreachable("2023-08-15") + return "" addrs = gather_address_lists('ipr_posted_on_doc',doc=doc).as_strings(compact=False) - author_names = ', '.join(a.person.name for a in doc.documentauthor_set.select_related("person")) + # Get a list of author names for the salutation in the body of the email + author_names = ', '.join(doc.author_names()) context = dict( settings=settings, @@ -149,13 +153,13 @@ def ipr_rfc_number(disclosureDate, thirdPartyDisclosureFlag): # RFC publication date comes from the RFC Editor announcement ipr_rfc_pub_datetime = { - 1310 : datetime.datetime(1992, 3, 13, 0, 0, tzinfo=datetime.timezone.utc), - 1802 : datetime.datetime(1994, 3, 23, 0, 0, tzinfo=datetime.timezone.utc), - 2026 : datetime.datetime(1996, 10, 29, 0, 0, tzinfo=datetime.timezone.utc), - 3668 : datetime.datetime(2004, 2, 18, 0, 0, tzinfo=datetime.timezone.utc), - 3979 : datetime.datetime(2005, 3, 2, 2, 23, tzinfo=datetime.timezone.utc), - 4879 : datetime.datetime(2007, 4, 10, 18, 21, tzinfo=datetime.timezone.utc), - 8179 : datetime.datetime(2017, 5, 31, 23, 1, tzinfo=datetime.timezone.utc), + 1310 : datetime.datetime(1992, 3, 13, 0, 0, tzinfo=datetime.UTC), + 1802 : datetime.datetime(1994, 3, 23, 0, 0, tzinfo=datetime.UTC), + 2026 : datetime.datetime(1996, 10, 29, 0, 0, tzinfo=datetime.UTC), + 3668 : datetime.datetime(2004, 2, 18, 0, 0, tzinfo=datetime.UTC), + 3979 : datetime.datetime(2005, 3, 2, 2, 23, tzinfo=datetime.UTC), + 4879 : datetime.datetime(2007, 4, 10, 18, 21, tzinfo=datetime.UTC), + 8179 : datetime.datetime(2017, 5, 31, 23, 1, tzinfo=datetime.UTC), } if disclosureDate < ipr_rfc_pub_datetime[1310]: @@ -269,7 +273,7 @@ def add_email(request, id): @role_required('Secretariat',) def admin(request, state): """Administrative disclosure listing. For non-posted disclosures""" - states = IprDisclosureStateName.objects.filter(slug__in=[state, "rejected"] if state == "removed" else [state]) + states = IprDisclosureStateName.objects.filter(slug__in=[state, "rejected", "removed_objfalse"] if state == "removed" else [state]) if not states: raise Http404 @@ -442,58 +446,35 @@ def history(request, id): 'selected_tab_entry':'history' }) -def by_draft_txt(request): - docipr = {} - for o in IprDocRel.objects.filter(disclosure__state='posted').select_related('document'): - name = o.document.name - if name.startswith("rfc"): - name = name.upper() - - if not name in docipr: - docipr[name] = [] - - docipr[name].append(o.disclosure_id) - - lines = [ "# Machine-readable list of IPR disclosures by draft name" ] - for name, iprs in docipr.items(): - lines.append(name + "\t" + "\t".join(str(ipr_id) for ipr_id in sorted(iprs))) - - return HttpResponse("\n".join(lines), content_type="text/plain; charset=%s"%settings.DEFAULT_CHARSET) - -def by_draft_recursive_txt(request): - """Returns machine-readable list of IPR disclosures by draft name, recursive. - NOTE: this view is expensive and should be removed _after_ tools.ietf.org is retired, - including util function and management commands that generate the content for - this view.""" - - with open('/a/ietfdata/derived/ipr_draft_recursive.txt') as f: - content = f.read() - return HttpResponse(content, content_type="text/plain; charset=%s"%settings.DEFAULT_CHARSET) - - -def new(request, type, updates=None): +def new(request, _type, updates=None): """Submit a new IPR Disclosure. If the updates field != None, this disclosure updates one or more other disclosures.""" # Note that URL patterns won't ever send updates - updates is only non-null when called from code # This odd construct flipping generic and general allows the URLs to say 'general' while having a minimal impact on the code. # A cleanup to change the code to switch on type 'general' should follow. - if type == 'generic' and updates: # Only happens when called directly from the updates view + if ( + _type == "generic" and updates + ): # Only happens when called directly from the updates view pass - elif type == 'generic': - return HttpResponseRedirect(urlreverse('ietf.ipr.views.new',kwargs=dict(type='general'))) - elif type == 'general': - type = 'generic' + elif _type == "generic": + return HttpResponseRedirect( + urlreverse("ietf.ipr.views.new", kwargs=dict(_type="general")) + ) + elif _type == "general": + _type = "generic" else: pass # 1 to show initially + the template - DraftFormset = inlineformset_factory(IprDisclosureBase, IprDocRel, form=DraftForm, can_delete=False, extra=1 + 1) + DraftFormset = inlineformset_factory( + IprDisclosureBase, IprDocRel, form=DraftForm, can_delete=False, extra=1 + 1 + ) - if request.method == 'POST': - form = ipr_form_mapping[type](request.POST) - if type != 'generic': + if request.method == "POST": + form = ipr_form_mapping[_type](request.POST) + if _type != "generic": draft_formset = DraftFormset(request.POST, instance=IprDisclosureBase()) else: draft_formset = None @@ -502,72 +483,92 @@ def new(request, type, updates=None): person = Person.objects.get(name="(System)") else: person = request.user.person - + # check formset validity - if type != 'generic': + if _type != "generic": valid_formsets = draft_formset.is_valid() else: valid_formsets = True - + if form.is_valid() and valid_formsets: - if 'updates' in form.cleaned_data: - updates = form.cleaned_data['updates'] - del form.cleaned_data['updates'] + if "updates" in form.cleaned_data: + updates = form.cleaned_data["updates"] + del form.cleaned_data["updates"] disclosure = form.save(commit=False) disclosure.by = person - disclosure.state = IprDisclosureStateName.objects.get(slug='pending') + disclosure.state = IprDisclosureStateName.objects.get(slug="pending") disclosure.save() - - if type != 'generic': + + if _type != "generic": draft_formset = DraftFormset(request.POST, instance=disclosure) draft_formset.save() set_disclosure_title(disclosure) disclosure.save() - + if updates: for ipr in updates: - RelatedIpr.objects.create(source=disclosure,target=ipr,relationship_id='updates') - + RelatedIpr.objects.create( + source=disclosure, target=ipr, relationship_id="updates" + ) + # create IprEvent IprEvent.objects.create( - type_id='submitted', + type_id="submitted", by=person, disclosure=disclosure, - desc="Disclosure Submitted") + desc="Disclosure Submitted", + ) # send email notification - (to, cc) = gather_address_lists('ipr_disclosure_submitted') - send_mail(request, to, ('IPR Submitter App', 'ietf-ipr@ietf.org'), - 'New IPR Submission Notification', + (to, cc) = gather_address_lists("ipr_disclosure_submitted") + send_mail( + request, + to, + ("IPR Submitter App", "ietf-ipr@ietf.org"), + "New IPR Submission Notification", "ipr/new_update_email.txt", - {"ipr": disclosure,}, - cc=cc) - + { + "ipr": disclosure, + }, + cc=cc, + ) + return render(request, "ipr/submitted.html") else: if updates: original = IprDisclosureBase(id=updates).get_child() initial = model_to_dict(original) - initial.update({'updates':str(updates), }) - patent_info = text_to_dict(initial.get('patent_info', '')) + initial.update( + { + "updates": str(updates), + } + ) + patent_info = text_to_dict(initial.get("patent_info", "")) if list(patent_info.keys()): - patent_dict = dict([ ('patent_'+k.lower(), v) for k,v in list(patent_info.items()) ]) + patent_dict = dict( + [("patent_" + k.lower(), v) for k, v in list(patent_info.items())] + ) else: - patent_dict = {'patent_notes': initial.get('patent_info', '')} + patent_dict = {"patent_notes": initial.get("patent_info", "")} initial.update(patent_dict) - form = ipr_form_mapping[type](initial=initial) + form = ipr_form_mapping[_type](initial=initial) else: - form = ipr_form_mapping[type]() - disclosure = IprDisclosureBase() # dummy disclosure for inlineformset + form = ipr_form_mapping[_type]() + disclosure = IprDisclosureBase() # dummy disclosure for inlineformset draft_formset = DraftFormset(instance=disclosure) - return render(request, "ipr/details_edit.html", { - 'form': form, - 'draft_formset':draft_formset, - 'type':type, - }) + return render( + request, + "ipr/details_edit.html", + { + "form": form, + "draft_formset": draft_formset, + "type": _type, + }, + ) + @role_required('Secretariat',) def notify(request, id, type): @@ -629,11 +630,16 @@ def post(request, id): def search(request): search_type = request.GET.get("submit") + if search_type and "\x00" in search_type: + return HttpResponseBadRequest("Null characters are not allowed") + # query field q = '' # legacy support if not search_type and request.GET.get("option", None) == "document_search": docname = request.GET.get("document_search", "") + if docname and "\x00" in docname: + return HttpResponseBadRequest("Null characters are not allowed") if docname.startswith("draft-"): search_type = "draft" q = docname @@ -643,18 +649,24 @@ def search(request): if search_type: form = SearchForm(request.GET) docid = request.GET.get("id") or request.GET.get("id_document_tag") or "" + if docid and "\x00" in docid: + return HttpResponseBadRequest("Null characters are not allowed") docs = doc = None iprs = [] related_iprs = [] # set states - states = request.GET.getlist('state',('posted','removed')) + states = request.GET.getlist('state',settings.PUBLISH_IPR_STATES) + if any("\x00" in state for state in states if state): + return HttpResponseBadRequest("Null characters are not allowed") if states == ['all']: states = IprDisclosureStateName.objects.values_list('slug',flat=True) # get query field if request.GET.get(search_type): q = request.GET.get(search_type) + if q and "\x00" in q: + return HttpResponseBadRequest("Null characters are not allowed") if q or docid: # Search by RFC number or draft-identifier @@ -663,23 +675,53 @@ def search(request): doc = q if docid: - start = DocAlias.objects.filter(name=docid) + start = Document.objects.filter(name__iexact=docid) else: if search_type == "draft": q = normalize_draftname(q) - start = DocAlias.objects.filter(name__contains=q, name__startswith="draft") + start = Document.objects.filter(name__icontains=q, name__startswith="draft") elif search_type == "rfc": - start = DocAlias.objects.filter(name="rfc%s" % q.lstrip("0")) + start = Document.objects.filter(name="rfc%s" % q.lstrip("0")) # one match if len(start) == 1: first = start[0] - doc = first.document - docs = related_docs(first) - iprs = iprs_from_docs(docs,states=states) + doc = first + docs = set([first]) + docs.update( + related_docs( + first, relationship=("replaces", "obs"), reverse_relationship=() + ) + ) + docs.update( + set( + [ + draft + for drafts in [ + related_docs( + d, relationship=(), reverse_relationship=("became_rfc",) + ) + for d in docs + ] + for draft in drafts + ] + ) + ) + docs.discard(None) + docs = sorted( + docs, + key=lambda d: ( + d.rfc_number if d.rfc_number is not None else 0, + d.became_rfc().rfc_number if d.became_rfc() else 0, + ), + reverse=True, + ) + iprs = iprs_from_docs(docs, states=states) template = "ipr/search_doc_result.html" - updated_docs = related_docs(first, ('updates',)) - related_iprs = list(set(iprs_from_docs(updated_docs, states=states)) - set(iprs)) + updated_docs = related_docs(first, ("updates",)) + related_iprs = list( + set(iprs_from_docs(updated_docs, states=states)) - set(iprs) + ) # multiple matches, select just one elif start: docs = start @@ -706,27 +748,27 @@ def search(request): # Search by wg acronym # Document list with IPRs elif search_type == "group": - docs = list(DocAlias.objects.filter(docs__group=q)) + docs = list(Document.objects.filter(group=q)) related = [] for doc in docs: doc.product_of_this_wg = True related += related_docs(doc) iprs = iprs_from_docs(list(set(docs+related)),states=states) - docs = [ doc for doc in docs if doc.document.ipr() ] - docs = sorted(docs, key=lambda x: max([ipr.disclosure.time for ipr in x.document.ipr()]), reverse=True) + docs = [ doc for doc in docs if doc.ipr() ] + docs = sorted(docs, key=lambda x: max([ipr.disclosure.time for ipr in x.ipr()]), reverse=True) template = "ipr/search_wg_result.html" q = Group.objects.get(id=q).acronym # make acronym for use in template # Search by rfc and id title # Document list with IPRs elif search_type == "doctitle": - docs = list(DocAlias.objects.filter(docs__title__icontains=q)) + docs = list(Document.objects.filter(title__icontains=q)) related = [] for doc in docs: related += related_docs(doc) iprs = iprs_from_docs(list(set(docs+related)),states=states) - docs = [ doc for doc in docs if doc.document.ipr() ] - docs = sorted(docs, key=lambda x: max([ipr.disclosure.time for ipr in x.document.ipr()]), reverse=True) + docs = [ doc for doc in docs if doc.ipr() ] + docs = sorted(docs, key=lambda x: max([ipr.disclosure.time for ipr in x.ipr()]), reverse=True) template = "ipr/search_doctitle_result.html" # Search by title of IPR disclosure @@ -776,9 +818,16 @@ def get_details_tabs(ipr, selected): def show(request, id): """View of individual declaration""" - ipr = get_object_or_404(IprDisclosureBase, id=id).get_child() + ipr = IprDisclosureBase.objects.filter(id=id) + removed = RemovedIprDisclosure.objects.filter(removed_id=id) + if removed.exists(): + return render(request, "ipr/deleted.html", {"removed": removed.get(), "ipr": ipr}) + if not ipr.exists(): + raise Http404 + else: + ipr = ipr.get().get_child() if not has_role(request.user, 'Secretariat'): - if ipr.state.slug == 'removed': + if ipr.state.slug in ['removed', 'removed_objfalse']: return render(request, "ipr/removed.html", { 'ipr': ipr }) @@ -801,10 +850,10 @@ def show(request, id): def showlist(request): """List all disclosures by type, posted only""" - generic = GenericIprDisclosure.objects.filter(state__in=('posted','removed')).prefetch_related('relatedipr_source_set__target','relatedipr_target_set__source').order_by('-time') - specific = HolderIprDisclosure.objects.filter(state__in=('posted','removed')).prefetch_related('relatedipr_source_set__target','relatedipr_target_set__source').order_by('-time') - thirdpty = ThirdPartyIprDisclosure.objects.filter(state__in=('posted','removed')).prefetch_related('relatedipr_source_set__target','relatedipr_target_set__source').order_by('-time') - nondocspecific = NonDocSpecificIprDisclosure.objects.filter(state__in=('posted','removed')).prefetch_related('relatedipr_source_set__target','relatedipr_target_set__source').order_by('-time') + generic = GenericIprDisclosure.objects.filter(state__in=settings.PUBLISH_IPR_STATES).prefetch_related('relatedipr_source_set__target','relatedipr_target_set__source').order_by('-time') + specific = HolderIprDisclosure.objects.filter(state__in=settings.PUBLISH_IPR_STATES).prefetch_related('relatedipr_source_set__target','relatedipr_target_set__source').order_by('-time') + thirdpty = ThirdPartyIprDisclosure.objects.filter(state__in=settings.PUBLISH_IPR_STATES).prefetch_related('relatedipr_source_set__target','relatedipr_target_set__source').order_by('-time') + nondocspecific = NonDocSpecificIprDisclosure.objects.filter(state__in=settings.PUBLISH_IPR_STATES).prefetch_related('relatedipr_source_set__target','relatedipr_target_set__source').order_by('-time') # combine nondocspecific with generic and re-sort generic = itertools.chain(generic,nondocspecific) @@ -860,3 +909,8 @@ def update(request, id): child = ipr.get_child() type = class_to_type[child.__class__.__name__] return new(request, type, updates=id) + +@role_required("Secretariat") +def json_snapshot(request, id): + obj = get_object_or_404(IprDisclosureBase,id=id).get_child() + return HttpResponse(json_dump_disclosure(obj),content_type="application/json") diff --git a/ietf/liaisons/.gitignore b/ietf/liaisons/.gitignore deleted file mode 100644 index c7013ced9d..0000000000 --- a/ietf/liaisons/.gitignore +++ /dev/null @@ -1,2 +0,0 @@ -/*.pyc -/settings_local.py diff --git a/ietf/liaisons/admin.py b/ietf/liaisons/admin.py index c7cb7a4dae..d873cce536 100644 --- a/ietf/liaisons/admin.py +++ b/ietf/liaisons/admin.py @@ -7,15 +7,16 @@ from ietf.liaisons.models import ( LiaisonStatement, LiaisonStatementEvent, RelatedLiaisonStatement, LiaisonStatementAttachment ) +from ietf.utils.admin import SaferTabularInline -class RelatedLiaisonStatementInline(admin.TabularInline): +class RelatedLiaisonStatementInline(SaferTabularInline): model = RelatedLiaisonStatement fk_name = 'source' raw_id_fields = ['target'] extra = 1 -class LiaisonStatementAttachmentInline(admin.TabularInline): +class LiaisonStatementAttachmentInline(SaferTabularInline): model = LiaisonStatementAttachment raw_id_fields = ['document'] extra = 1 @@ -24,7 +25,7 @@ class LiaisonStatementAdmin(admin.ModelAdmin): list_display = ['id', 'title', 'submitted', 'from_groups_short_display', 'purpose', 'related_to'] list_display_links = ['id', 'title'] ordering = ('title', ) - raw_id_fields = ('from_contact', 'attachments', 'from_groups', 'to_groups') + raw_id_fields = ('attachments', 'from_groups', 'to_groups') #filter_horizontal = ('from_groups', 'to_groups') inlines = [ RelatedLiaisonStatementInline, LiaisonStatementAttachmentInline ] @@ -50,4 +51,4 @@ class LiaisonStatementEventAdmin(admin.ModelAdmin): raw_id_fields = ["statement", "by"] admin.site.register(LiaisonStatement, LiaisonStatementAdmin) -admin.site.register(LiaisonStatementEvent, LiaisonStatementEventAdmin) \ No newline at end of file +admin.site.register(LiaisonStatementEvent, LiaisonStatementEventAdmin) diff --git a/ietf/liaisons/factories.py b/ietf/liaisons/factories.py index 61788817eb..ca588236e3 100644 --- a/ietf/liaisons/factories.py +++ b/ietf/liaisons/factories.py @@ -1,14 +1,15 @@ import factory from ietf.group.factories import GroupFactory -from ietf.liaisons.models import LiaisonStatement, LiaisonStatementEvent, LiaisonStatementAttachment +from ietf.liaisons.models import LiaisonStatement, LiaisonStatementEvent, LiaisonStatementAttachment, RelatedLiaisonStatement class LiaisonStatementFactory(factory.django.DjangoModelFactory): class Meta: model = LiaisonStatement + skip_postgeneration_save = True title = factory.Faker('sentence') - from_contact = factory.SubFactory('ietf.person.factories.EmailFactory') + from_contact = factory.Faker('email') purpose_id = 'comment' body = factory.Faker('paragraph') state_id = 'posted' @@ -49,3 +50,12 @@ class Meta: type_id='liai-att', # TODO: Make name more convenient (the default now is to try to generate a draftname) ) + + +class RelatedLiaisonStatementFactory(factory.django.DjangoModelFactory): + class Meta: + model = RelatedLiaisonStatement + + source = factory.SubFactory(LiaisonStatementFactory) + target = factory.SubFactory(LiaisonStatementFactory) + relationship_id = "refunk" diff --git a/ietf/liaisons/forms.py b/ietf/liaisons/forms.py index fa1f550d05..6ceda5ad38 100644 --- a/ietf/liaisons/forms.py +++ b/ietf/liaisons/forms.py @@ -3,39 +3,33 @@ import io -import os import operator - -from typing import Union # pyflakes:ignore - +import os from email.utils import parseaddr -from form_utils.forms import BetterModelForm +from functools import reduce +from typing import Union, Optional # pyflakes:ignore from django import forms from django.conf import settings -from django.core.exceptions import ObjectDoesNotExist, ValidationError -from django.db.models.query import QuerySet -from django.forms.utils import ErrorList -from django.db.models import Q -#from django.forms.widgets import RadioFieldRenderer +from django.core.exceptions import ValidationError from django.core.validators import validate_email +from django.db.models import Q, QuerySet +from django.forms.utils import ErrorList +from django_stubs_ext import QuerySetAny -import debug # pyflakes:ignore - +from ietf.doc.models import Document +from ietf.group.models import Group from ietf.ietfauth.utils import has_role -from ietf.name.models import DocRelationshipName -from ietf.liaisons.utils import get_person_for_user,is_authorized_individual -from ietf.liaisons.widgets import ButtonWidget,ShowAttachmentsWidget -from ietf.liaisons.models import (LiaisonStatement, - LiaisonStatementEvent,LiaisonStatementAttachment,LiaisonStatementPurposeName) from ietf.liaisons.fields import SearchableLiaisonStatementsField -from ietf.group.models import Group -from ietf.person.models import Email -from ietf.person.fields import SearchableEmailField -from ietf.doc.models import Document, DocAlias -from ietf.utils.fields import DatepickerDateField +from ietf.liaisons.models import (LiaisonStatement, + LiaisonStatementEvent, LiaisonStatementAttachment, LiaisonStatementPurposeName) +from ietf.liaisons.utils import get_person_for_user, OUTGOING_LIAISON_ROLES, \ + INCOMING_LIAISON_ROLES +from ietf.liaisons.widgets import ButtonWidget, ShowAttachmentsWidget +from ietf.name.models import DocRelationshipName +from ietf.person.models import Person +from ietf.utils.fields import DatepickerDateField, ModelMultipleChoiceField from ietf.utils.timezone import date_today, datetime_from_date, DEADLINE_TZINFO -from functools import reduce ''' NOTES: @@ -52,45 +46,105 @@ def liaison_manager_sdos(person): return Group.objects.filter(type="sdo", state="active", role__person=person, role__name="liaiman").distinct() + def flatten_choices(choices): - '''Returns a flat choice list given one with option groups defined''' + """Returns a flat choice list given one with option groups defined + + n.b., Django allows mixing grouped options and top-level options. This helper only supports + the non-mixed case where every option is in an option group. + """ flat = [] - for optgroup,options in choices: + for optgroup, options in choices: flat.extend(options) return flat + + +def choices_from_group_queryset(groups: QuerySet[Group]): + """Get choices list for internal IETF groups user is authorized to select -def get_internal_choices(user): - '''Returns the set of internal IETF groups the user has permissions for, as a list - of choices suitable for use in a select widget. If user == None, all active internal - groups are included.''' + Returns a grouped list of choices suitable for use with a ChoiceField. If user is None, + includes all groups. + """ + main = [] + areas = [] + wgs = [] + for g in groups.distinct().order_by("acronym"): + if g.acronym in ("ietf", "iesg", "iab"): + main.append((g.pk, f"The {g.acronym.upper()}")) + elif g.type_id == "area": + areas.append((g.pk, f"{g.acronym} - {g.name}")) + elif g.type_id == "wg": + wgs.append((g.pk, f"{g.acronym} - {g.name}")) choices = [] - groups = get_groups_for_person(user.person if user else None) - main = [ (g.pk, 'The {}'.format(g.acronym.upper())) for g in groups.filter(acronym__in=('ietf','iesg','iab')) ] - areas = [ (g.pk, '{} - {}'.format(g.acronym,g.name)) for g in groups.filter(type='area') ] - wgs = [ (g.pk, '{} - {}'.format(g.acronym,g.name)) for g in groups.filter(type='wg') ] - choices.append(('Main IETF Entities', main)) - choices.append(('IETF Areas', areas)) - choices.append(('IETF Working Groups', wgs )) + if len(main) > 0: + choices.append(("Main IETF Entities", main)) + if len(areas) > 0: + choices.append(("IETF Areas", areas)) + if len(wgs) > 0: + choices.append(("IETF Working Groups", wgs)) return choices -def get_groups_for_person(person): - '''Returns queryset of internal Groups the person has interesting roles in. - This is a refactor of IETFHierarchyManager.get_entities_for_person(). If Person - is None or Secretariat or Liaison Manager all internal IETF groups are returned. - ''' - if person == None or has_role(person.user, "Secretariat") or has_role(person.user, "Liaison Manager"): - # collect all internal IETF groups - queries = [Q(acronym__in=('ietf','iesg','iab')), - Q(type='area',state='active'), - Q(type='wg',state='active')] + +def all_internal_groups(): + """Get a queryset of all IETF groups suitable for LS To/From assignment""" + return Group.objects.filter( + Q(acronym__in=("ietf", "iesg", "iab")) + | Q(type="area", state="active") + | Q(type="wg", state="active") + ).distinct() + + +def internal_groups_for_person(person: Optional[Person]): + """Get a queryset of IETF groups suitable for LS To/From assignment by person""" + if person is None: + return Group.objects.none() # no person = no roles + + if has_role( + person.user, + ( + "Secretariat", + "IETF Chair", + "IAB Chair", + "Liaison Manager", + "Liaison Coordinator", + "Authorized Individual", + ), + ): + return all_internal_groups() + # Interesting roles, as Group queries + queries = [ + Q(role__person=person, role__name="chair", acronym="ietf"), + Q(role__person=person, role__name="chair", acronym="iab"), + Q(role__person=person, role__name="ad", type="area", state="active"), + Q( + role__person=person, + role__name__in=("chair", "secretary"), + type="wg", + state="active", + ), + Q( + parent__role__person=person, + parent__role__name="ad", + type="wg", + state="active", + ), + ] + if has_role(person.user, "Area Director"): + queries.append(Q(acronym__in=("ietf", "iesg"))) # AD can also choose these + return Group.objects.filter(reduce(operator.or_, queries)).distinct() + + +def external_groups_for_person(person): + """Get a queryset of external groups suitable for LS To/From assignment by person""" + filter_expr = Q(pk__in=[]) # start with no groups + # These roles can add all external sdo groups + if has_role(person.user, set(INCOMING_LIAISON_ROLES + OUTGOING_LIAISON_ROLES) - {"Liaison Manager", "Authorized Individual"}): + filter_expr |= Q(type="sdo") else: - # Interesting roles, as Group queries - queries = [Q(role__person=person,role__name='chair',acronym='ietf'), - Q(role__person=person,role__name__in=('chair','execdir'),acronym='iab'), - Q(role__person=person,role__name='ad',type='area',state='active'), - Q(role__person=person,role__name__in=('chair','secretary'),type='wg',state='active'), - Q(parent__role__person=person,parent__role__name='ad',type='wg',state='active')] - return Group.objects.filter(reduce(operator.or_,queries)).order_by('acronym').distinct() + # The person cannot add all external sdo groups; add any for which they are Liaison Manager + filter_expr |= Q(type="sdo", role__person=person, role__name__in=["auth", "liaiman"]) + return Group.objects.filter(state="active").filter(filter_expr).distinct().order_by("name") + def liaison_form_factory(request, type=None, **kwargs): """Returns appropriate Liaison entry form""" @@ -132,7 +186,7 @@ class AddCommentForm(forms.Form): # def render(self): # output = [] # for widget in self: -# output.append(format_html(force_text(widget))) +# output.append(format_html(force_str(widget))) # return mark_safe('\n'.join(output)) @@ -155,7 +209,7 @@ def get_results(self): query = self.cleaned_data.get('text') if query: q = (Q(title__icontains=query) | - Q(from_contact__address__icontains=query) | + Q(from_contact__icontains=query) | Q(to_contacts__icontains=query) | Q(other_identifiers__icontains=query) | Q(body__icontains=query) | @@ -201,10 +255,10 @@ def get_results(self): return results -class CustomModelMultipleChoiceField(forms.ModelMultipleChoiceField): +class CustomModelMultipleChoiceField(ModelMultipleChoiceField): '''If value is a QuerySet, return it as is (for use in widget.render)''' def prepare_value(self, value): - if isinstance(value, QuerySet): + if isinstance(value, QuerySetAny): return value if (hasattr(value, '__iter__') and not isinstance(value, str) and @@ -213,17 +267,12 @@ def prepare_value(self, value): return super(CustomModelMultipleChoiceField, self).prepare_value(value) -class LiaisonModelForm(BetterModelForm): +class LiaisonModelForm(forms.ModelForm): '''Specify fields which require a custom widget or that are not part of the model. ''' - from_groups = forms.ModelMultipleChoiceField(queryset=Group.objects.all(),label='Groups',required=False) - from_groups.widget.attrs["class"] = "select2-field" - from_groups.widget.attrs['data-minimum-input-length'] = 0 - from_contact = forms.EmailField() # type: Union[forms.EmailField, SearchableEmailField] + from_groups = ModelMultipleChoiceField(queryset=Group.objects.all(),label='Groups',required=False) to_contacts = forms.CharField(label="Contacts", widget=forms.Textarea(attrs={'rows':'3', }), strip=False) - to_groups = forms.ModelMultipleChoiceField(queryset=Group.objects,label='Groups',required=False) - to_groups.widget.attrs["class"] = "select2-field" - to_groups.widget.attrs['data-minimum-input-length'] = 0 + to_groups = ModelMultipleChoiceField(queryset=Group.objects,label='Groups',required=False) deadline = DatepickerDateField(date_format="yyyy-mm-dd", picker_settings={"autoclose": "1" }, label='Deadline', required=True) related_to = SearchableLiaisonStatementsField(label='Related Liaison Statement', required=False) submitted_date = DatepickerDateField(date_format="yyyy-mm-dd", picker_settings={"autoclose": "1" }, label='Submission date', required=True, initial=lambda: date_today(DEADLINE_TZINFO)) @@ -238,13 +287,6 @@ class LiaisonModelForm(BetterModelForm): class Meta: model = LiaisonStatement exclude = ('attachments','state','from_name','to_name') - fieldsets = [('From', {'fields': ['from_groups','from_contact', 'response_contacts'], 'legend': ''}), - ('To', {'fields': ['to_groups','to_contacts'], 'legend': ''}), - ('Other email addresses', {'fields': ['technical_contacts','action_holder_contacts','cc_contacts'], 'legend': ''}), - ('Purpose', {'fields':['purpose', 'deadline'], 'legend': ''}), - ('Reference', {'fields': ['other_identifiers','related_to'], 'legend': ''}), - ('Liaison Statement', {'fields': ['title', 'submitted_date', 'body', 'attachments'], 'legend': ''}), - ('Add attachment', {'fields': ['attach_title', 'attach_file', 'attach_button'], 'legend': ''})] def __init__(self, user, *args, **kwargs): super(LiaisonModelForm, self).__init__(*args, **kwargs) @@ -253,13 +295,17 @@ def __init__(self, user, *args, **kwargs): self.person = get_person_for_user(user) self.is_new = not self.instance.pk + self.fields["from_groups"].widget.attrs["class"] = "select2-field" + self.fields["from_groups"].widget.attrs["data-minimum-input-length"] = 0 self.fields["from_groups"].widget.attrs["data-placeholder"] = "Type in name to search for group" + self.fields["to_groups"].widget.attrs["class"] = "select2-field" + self.fields["to_groups"].widget.attrs["data-minimum-input-length"] = 0 self.fields["to_groups"].widget.attrs["data-placeholder"] = "Type in name to search for group" self.fields["to_contacts"].label = 'Contacts' self.fields["other_identifiers"].widget.attrs["rows"] = 2 - + # add email validators - for field in ['from_contact','to_contacts','technical_contacts','action_holder_contacts','cc_contacts']: + for field in ['to_contacts','technical_contacts','action_holder_contacts','cc_contacts']: if field in self.fields: self.fields[field].validators.append(validate_emails) @@ -278,18 +324,6 @@ def clean_to_groups(self): raise forms.ValidationError('You must specify a To Group') return to_groups - def clean_from_contact(self): - contact = self.cleaned_data.get('from_contact') - from_groups = self.cleaned_data.get('from_groups') - try: - email = Email.objects.get(address=contact) - if not email.origin: - email.origin = "liaison: %s" % (','.join([ g.acronym for g in from_groups.all() ])) - email.save() - except ObjectDoesNotExist: - raise forms.ValidationError('Email address does not exist') - return email - # Note to future person: This is the wrong place to fix the new lines # in cc_contacts and to_contacts. Those belong in the save function. # Or at least somewhere other than here. @@ -383,12 +417,12 @@ def save_attachments(self): uploaded_filename = name + extension, ) ) - if created: - DocAlias.objects.create(name=attach.name).docs.add(attach) LiaisonStatementAttachment.objects.create(statement=self.instance,document=attach) attach_file = io.open(os.path.join(settings.LIAISON_ATTACH_PATH, attach.name + extension), 'wb') attach_file.write(attached_file.read()) attach_file.close() + attached_file.seek(0) + attach.store_file(attach.uploaded_filename, attached_file) if not self.is_new: # create modified event @@ -432,42 +466,39 @@ def set_to_fields(self): assert NotImplemented class IncomingLiaisonForm(LiaisonModelForm): - def clean(self): - if 'send' in list(self.data.keys()) and self.get_post_only(): - raise forms.ValidationError('As an IETF Liaison Manager you can not send incoming liaison statements, you only can post them') - return super(IncomingLiaisonForm, self).clean() def is_approved(self): '''Incoming Liaison Statements do not required approval''' return True - def get_post_only(self): - from_groups = self.cleaned_data.get('from_groups') - if has_role(self.user, "Secretariat") or is_authorized_individual(self.user,from_groups): - return False - return True - def set_from_fields(self): - '''Set from_groups and from_contact options and initial value based on user - accessing the form.''' - if has_role(self.user, "Secretariat"): - queryset = Group.objects.filter(type="sdo", state="active").order_by('name') - else: - queryset = Group.objects.filter(type="sdo", state="active", role__person=self.person, role__name__in=("liaiman", "auth")).distinct().order_by('name') - self.fields['from_contact'].initial = self.person.role_set.filter(group=queryset[0]).first().email.address - self.fields['from_contact'].widget.attrs['disabled'] = True - self.fields['from_groups'].queryset = queryset - self.fields['from_groups'].widget.submitter = str(self.person) - + """Configure from "From" fields based on user roles""" + qs = external_groups_for_person(self.person) + self.fields["from_groups"].queryset = qs + self.fields["from_groups"].widget.submitter = str(self.person) # if there's only one possibility make it the default - if len(queryset) == 1: - self.fields['from_groups'].initial = queryset + if len(qs) == 1: + self.fields['from_groups'].initial = qs + + # Note that the IAB chair currently doesn't get to work with incoming liaison statements + + # Removing this block at the request of the IAB - as a workaround until the new liaison tool is + # create, anyone with access to the form can set any from_contact value + # + # if not ( + # has_role(self.user, "Secretariat") + # or has_role(self.user, "Liaison Coordinator") + # ): + # self.fields["from_contact"].initial = ( + # self.person.role_set.filter(group=qs[0]).first().email.formatted_email() + # ) + # self.fields["from_contact"].widget.attrs["disabled"] = True def set_to_fields(self): '''Set to_groups and to_contacts options and initial value based on user accessing the form. For incoming Liaisons, to_groups choices is the full set. ''' - self.fields['to_groups'].choices = get_internal_choices(None) + self.fields['to_groups'].choices = choices_from_group_queryset(all_internal_groups()) class OutgoingLiaisonForm(LiaisonModelForm): @@ -476,59 +507,61 @@ class OutgoingLiaisonForm(LiaisonModelForm): class Meta: model = LiaisonStatement exclude = ('attachments','state','from_name','to_name','action_holder_contacts') - # add approved field, no action_holder_contacts - fieldsets = [('From', {'fields': ['from_groups','from_contact','response_contacts','approved'], 'legend': ''}), - ('To', {'fields': ['to_groups','to_contacts'], 'legend': ''}), - ('Other email addresses', {'fields': ['technical_contacts','cc_contacts'], 'legend': ''}), - ('Purpose', {'fields':['purpose', 'deadline'], 'legend': ''}), - ('Reference', {'fields': ['other_identifiers','related_to'], 'legend': ''}), - ('Liaison Statement', {'fields': ['title', 'submitted_date', 'body', 'attachments'], 'legend': ''}), - ('Add attachment', {'fields': ['attach_title', 'attach_file', 'attach_button'], 'legend': ''})] def is_approved(self): return self.cleaned_data['approved'] def set_from_fields(self): - '''Set from_groups and from_contact options and initial value based on user - accessing the form''' - choices = get_internal_choices(self.user) - self.fields['from_groups'].choices = choices - - # set initial value if only one entry - flat_choices = flatten_choices(choices) + """Configure from "From" fields based on user roles""" + self.set_from_groups_field() + self.set_from_contact_field() + + def set_from_groups_field(self): + """Configure the from_groups field based on roles""" + grouped_choices = choices_from_group_queryset(internal_groups_for_person(self.person)) + flat_choices = flatten_choices(grouped_choices) if len(flat_choices) == 1: - self.fields['from_groups'].initial = [flat_choices[0][0]] - - if has_role(self.user, "Secretariat"): - self.fields['from_contact'] = SearchableEmailField(only_users=True) # secretariat can edit this field! - return - - if self.person.role_set.filter(name='liaiman',group__state='active'): - email = self.person.role_set.filter(name='liaiman',group__state='active').first().email.address - elif self.person.role_set.filter(name__in=('ad','chair'),group__state='active'): - email = self.person.role_set.filter(name__in=('ad','chair'),group__state='active').first().email.address + self.fields["from_groups"].choices = flat_choices + self.fields["from_groups"].initial = [flat_choices[0][0]] else: - email = self.person.email_address() + self.fields["from_groups"].choices = grouped_choices - # Non-secretariat user cannot change the from_contact field. Fill in its value. + def set_from_contact_field(self): + """Configure the from_contact field based on user roles""" + # Secretariat can set this to any valid address but gets no default + if has_role(self.user, "Secretariat"): + return + elif has_role(self.user, ["IAB Chair", "Liaison Coordinator"]): + self.fields["from_contact"].initial = "IAB Chair " + return + elif has_role(self.user, "IETF Chair"): + self.fields["from_contact"].initial = "IETF Chair " + return + # ... others have it set to the correct value and cannot change it self.fields['from_contact'].disabled = True - self.fields['from_contact'].initial = email - - def set_to_fields(self): - '''Set to_groups and to_contacts options and initial value based on user - accessing the form''' - # set options. if the user is a Liaison Manager and nothing more, reduce set to his SDOs - if has_role(self.user, "Liaison Manager") and not self.person.role_set.filter(name__in=('ad','chair'),group__state='active'): - queryset = Group.objects.filter(type="sdo", state="active", role__person=self.person, role__name="liaiman").distinct().order_by('name') + # Set up the querysets we might use - only evaluated as needed + liaison_manager_role = self.person.role_set.filter(name="liaiman", group__state="active") + chair_or_ad_role = self.person.role_set.filter( + name__in=("ad", "chair"), group__state="active" + ) + if liaison_manager_role.exists(): + from_contact_email = liaison_manager_role.first().email + elif chair_or_ad_role.exists(): + from_contact_email = chair_or_ad_role.first().email else: - # get all outgoing entities - queryset = Group.objects.filter(type="sdo", state="active").order_by('name') + from_contact_email = self.person.email() + self.fields['from_contact'].initial = from_contact_email.formatted_email() - self.fields['to_groups'].queryset = queryset + def set_to_fields(self): + """Configure the "To" fields based on user roles""" + qs = external_groups_for_person(self.person) + self.fields['to_groups'].queryset = qs # set initial if has_role(self.user, "Liaison Manager"): - self.fields['to_groups'].initial = [queryset.first()] + self.fields['to_groups'].initial = [ + qs.filter(role__person=self.person, role__name="liaiman").first() + ] class EditLiaisonForm(LiaisonModelForm): @@ -536,8 +569,7 @@ def __init__(self, *args, **kwargs): super(EditLiaisonForm, self).__init__(*args, **kwargs) self.edit = True self.fields['attachments'].initial = self.instance.liaisonstatementattachment_set.exclude(removed=True) - related = [ str(x.pk) for x in self.instance.source_of_set.all() ] - self.fields['related_to'].initial = ','.join(related) + self.fields['related_to'].initial = [ x.target for x in self.instance.source_of_set.all() ] self.fields['submitted_date'].initial = self.instance.submitted def save(self, *args, **kwargs): @@ -550,32 +582,20 @@ def save(self, *args, **kwargs): return self.instance def set_from_fields(self): - '''Set from_groups and from_contact options and initial value based on user - accessing the form.''' + """Configure from "From" fields based on user roles""" if self.instance.is_outgoing(): - self.fields['from_groups'].choices = get_internal_choices(self.user) + self.fields['from_groups'].choices = choices_from_group_queryset(internal_groups_for_person(self.person)) else: - if has_role(self.user, "Secretariat"): - queryset = Group.objects.filter(type="sdo").order_by('name') - else: - queryset = Group.objects.filter(type="sdo", role__person=self.person, role__name__in=("liaiman", "auth")).distinct().order_by('name') + self.fields["from_groups"].queryset = external_groups_for_person(self.person) + if not has_role(self.user, "Secretariat"): self.fields['from_contact'].widget.attrs['disabled'] = True - self.fields['from_groups'].queryset = queryset def set_to_fields(self): - '''Set to_groups and to_contacts options and initial value based on user - accessing the form. For incoming Liaisons, to_groups choices is the full set. - ''' + """Configure the "To" fields based on user roles""" if self.instance.is_outgoing(): - # if the user is a Liaison Manager and nothing more, reduce to set to his SDOs - if has_role(self.user, "Liaison Manager") and not self.person.role_set.filter(name__in=('ad','chair'),group__state='active'): - queryset = Group.objects.filter(type="sdo", role__person=self.person, role__name="liaiman").distinct().order_by('name') - else: - # get all outgoing entities - queryset = Group.objects.filter(type="sdo").order_by('name') - self.fields['to_groups'].queryset = queryset + self.fields['to_groups'].queryset = external_groups_for_person(self.person) else: - self.fields['to_groups'].choices = get_internal_choices(None) + self.fields['to_groups'].choices = choices_from_group_queryset(all_internal_groups()) class EditAttachmentForm(forms.Form): diff --git a/ietf/liaisons/mails.py b/ietf/liaisons/mails.py index 8708c8a078..878aada576 100644 --- a/ietf/liaisons/mails.py +++ b/ietf/liaisons/mails.py @@ -14,7 +14,10 @@ def send_liaison_by_email(request, liaison): subject = 'New Liaison Statement, "%s"' % (liaison.title) from_email = settings.LIAISON_UNIVERSAL_FROM - (to_email, cc) = gather_address_lists('liaison_statement_posted',liaison=liaison) + if liaison.is_outgoing(): + (to_email, cc) = gather_address_lists('liaison_statement_posted_outgoing',liaison=liaison) + else: + (to_email, cc) = gather_address_lists('liaison_statement_posted_incoming',liaison=liaison) bcc = ['statements@ietf.org'] body = render_to_string('liaisons/liaison_mail.txt', dict(liaison=liaison)) diff --git a/ietf/liaisons/management/.gitignore b/ietf/liaisons/management/.gitignore deleted file mode 100644 index a74b07aee4..0000000000 --- a/ietf/liaisons/management/.gitignore +++ /dev/null @@ -1 +0,0 @@ -/*.pyc diff --git a/ietf/liaisons/migrations/0001_initial.py b/ietf/liaisons/migrations/0001_initial.py index 7b58564ff1..eccd8bb331 100644 --- a/ietf/liaisons/migrations/0001_initial.py +++ b/ietf/liaisons/migrations/0001_initial.py @@ -1,7 +1,4 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-20 10:52 - +# Generated by Django 2.2.28 on 2023-03-20 19:22 from django.db import migrations, models import django.db.models.deletion @@ -34,14 +31,17 @@ class Migration(migrations.Migration): ('other_identifiers', models.TextField(blank=True, null=True)), ('body', models.TextField(blank=True)), ], + options={ + 'ordering': ['id'], + }, ), migrations.CreateModel( - name='LiaisonStatementAttachment', + name='RelatedLiaisonStatement', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('removed', models.BooleanField(default=False)), - ('document', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document')), - ('statement', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='liaisons.LiaisonStatement')), + ('relationship', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.DocRelationshipName')), + ('source', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='source_of_set', to='liaisons.LiaisonStatement')), + ('target', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='target_of_set', to='liaisons.LiaisonStatement')), ], ), migrations.CreateModel( @@ -59,21 +59,12 @@ class Migration(migrations.Migration): }, ), migrations.CreateModel( - name='LiaisonStatementGroupContacts', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('contacts', models.CharField(blank=True, max_length=255)), - ('cc_contacts', models.CharField(blank=True, max_length=255)), - ('group', ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='group.Group', unique=True)), - ], - ), - migrations.CreateModel( - name='RelatedLiaisonStatement', + name='LiaisonStatementAttachment', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('relationship', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.DocRelationshipName')), - ('source', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='source_of_set', to='liaisons.LiaisonStatement')), - ('target', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='target_of_set', to='liaisons.LiaisonStatement')), + ('removed', models.BooleanField(default=False)), + ('document', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document')), + ('statement', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='liaisons.LiaisonStatement')), ], ), migrations.AddField( @@ -111,4 +102,8 @@ class Migration(migrations.Migration): name='to_groups', field=models.ManyToManyField(blank=True, related_name='liaisonstatement_to_set', to='group.Group'), ), + migrations.AddIndex( + model_name='liaisonstatementevent', + index=models.Index(fields=['-time', '-id'], name='liaisons_li_time_3e1646_idx'), + ), ] diff --git a/ietf/liaisons/migrations/0002_alter_liaisonstatement_response_contacts.py b/ietf/liaisons/migrations/0002_alter_liaisonstatement_response_contacts.py new file mode 100644 index 0000000000..ac0a11101b --- /dev/null +++ b/ietf/liaisons/migrations/0002_alter_liaisonstatement_response_contacts.py @@ -0,0 +1,20 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +from django.db import migrations, models + + +class Migration(migrations.Migration): + + dependencies = [ + ("liaisons", "0001_initial"), + ] + + operations = [ + migrations.AlterField( + model_name="liaisonstatement", + name="response_contacts", + field=models.TextField( + blank=True, help_text="Where to send a response", max_length=1024 + ), + ), + ] diff --git a/ietf/liaisons/migrations/0002_auto_20180225_1207.py b/ietf/liaisons/migrations/0002_auto_20180225_1207.py deleted file mode 100644 index 62a9b8a13b..0000000000 --- a/ietf/liaisons/migrations/0002_auto_20180225_1207.py +++ /dev/null @@ -1,20 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-25 12:07 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('liaisons', '0001_initial'), - ] - - operations = [ - migrations.AlterModelOptions( - name='liaisonstatement', - options={'ordering': ['id']}, - ), - ] diff --git a/ietf/liaisons/migrations/0003_liaison_document2_fk.py b/ietf/liaisons/migrations/0003_liaison_document2_fk.py deleted file mode 100644 index 5fdf0f0f49..0000000000 --- a/ietf/liaisons/migrations/0003_liaison_document2_fk.py +++ /dev/null @@ -1,29 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-08 11:58 - - -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0015_1_add_fk_to_document_id'), - ('liaisons', '0002_auto_20180225_1207'), - ] - - operations = [ - migrations.AddField( - model_name='liaisonstatementattachment', - name='document2', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field=b'id'), - ), - migrations.AlterField( - model_name='liaisonstatementattachment', - name='document', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='old_liaison', to='doc.Document', to_field=b'name'), - ), - ] diff --git a/ietf/liaisons/migrations/0003_liaisonstatement_from_contact_tmp.py b/ietf/liaisons/migrations/0003_liaisonstatement_from_contact_tmp.py new file mode 100644 index 0000000000..de2ce7ff59 --- /dev/null +++ b/ietf/liaisons/migrations/0003_liaisonstatement_from_contact_tmp.py @@ -0,0 +1,22 @@ +# Copyright The IETF Trust 2025 All Rights Reserved +from django.db import migrations, models +import ietf.utils.validators + + +class Migration(migrations.Migration): + dependencies = [ + ("liaisons", "0002_alter_liaisonstatement_response_contacts"), + ] + + operations = [ + migrations.AddField( + model_name="liaisonstatement", + name="from_contact_tmp", + field=models.CharField( + blank=True, + help_text="Address of the formal sender of the statement", + max_length=512, + validators=[ietf.utils.validators.validate_mailbox_address], + ), + ), + ] diff --git a/ietf/liaisons/migrations/0004_populate_liaisonstatement_from_contact_tmp.py b/ietf/liaisons/migrations/0004_populate_liaisonstatement_from_contact_tmp.py new file mode 100644 index 0000000000..dbab326b0c --- /dev/null +++ b/ietf/liaisons/migrations/0004_populate_liaisonstatement_from_contact_tmp.py @@ -0,0 +1,60 @@ +# Copyright The IETF Trust 2025, All Rights Reserved +from itertools import islice + +from django.db import migrations + +from ietf.person.name import plain_name +from ietf.utils.mail import formataddr +from ietf.utils.validators import validate_mailbox_address + + +def forward(apps, schema_editor): + def _formatted_email(email): + """Format an email address to match Email.formatted_email()""" + person = email.person + if person: + return formataddr( + ( + # inlined Person.plain_name(), minus the caching + person.plain if person.plain else plain_name(person.name), + email.address, + ) + ) + return email.address + + def _batched(iterable, n): + """Split an iterable into lists of length <= n + + (based on itertools example code for batched(), which is added in py312) + """ + iterator = iter(iterable) + batch = list(islice(iterator, n)) # consumes first n iterations + while batch: + yield batch + batch = list(islice(iterator, n)) # consumes next n iterations + + LiaisonStatement = apps.get_model("liaisons", "LiaisonStatement") + LiaisonStatement.objects.update(from_contact_tmp="") # ensure they're all blank + for batch in _batched( + LiaisonStatement.objects.exclude(from_contact=None).select_related( + "from_contact" + ), + 100, + ): + for ls in batch: + ls.from_contact_tmp = _formatted_email(ls.from_contact) + validate_mailbox_address( + ls.from_contact_tmp + ) # be sure it's permitted before we accept it + + LiaisonStatement.objects.bulk_update(batch, fields=["from_contact_tmp"]) + + +class Migration(migrations.Migration): + dependencies = [ + ("liaisons", "0003_liaisonstatement_from_contact_tmp"), + ] + + operations = [ + migrations.RunPython(forward), + ] diff --git a/ietf/liaisons/migrations/0004_remove_liaisonstatementattachment_document.py b/ietf/liaisons/migrations/0004_remove_liaisonstatementattachment_document.py deleted file mode 100644 index 3f5868ef2b..0000000000 --- a/ietf/liaisons/migrations/0004_remove_liaisonstatementattachment_document.py +++ /dev/null @@ -1,20 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-20 09:53 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('liaisons', '0003_liaison_document2_fk'), - ] - - operations = [ - migrations.RemoveField( - model_name='liaisonstatementattachment', - name='document', - ), - ] diff --git a/ietf/liaisons/migrations/0005_rename_field_document2.py b/ietf/liaisons/migrations/0005_rename_field_document2.py deleted file mode 100644 index ddb8005735..0000000000 --- a/ietf/liaisons/migrations/0005_rename_field_document2.py +++ /dev/null @@ -1,22 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-21 05:31 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0019_rename_field_document2'), - ('liaisons', '0004_remove_liaisonstatementattachment_document'), - ] - - operations = [ - migrations.RenameField( - model_name='liaisonstatementattachment', - old_name='document2', - new_name='document', - ), - ] diff --git a/ietf/liaisons/migrations/0005_replace_liaisonstatement_from_contact.py b/ietf/liaisons/migrations/0005_replace_liaisonstatement_from_contact.py new file mode 100644 index 0000000000..e1702ae3bc --- /dev/null +++ b/ietf/liaisons/migrations/0005_replace_liaisonstatement_from_contact.py @@ -0,0 +1,20 @@ +# Copyright The IETF Trust 2025 All Rights Reserved +from django.db import migrations + + +class Migration(migrations.Migration): + dependencies = [ + ("liaisons", "0004_populate_liaisonstatement_from_contact_tmp"), + ] + + operations = [ + migrations.RemoveField( + model_name="liaisonstatement", + name="from_contact", + ), + migrations.RenameField( + model_name="liaisonstatement", + old_name="from_contact_tmp", + new_name="from_contact", + ), + ] diff --git a/ietf/liaisons/migrations/0006_document_primary_key_cleanup.py b/ietf/liaisons/migrations/0006_document_primary_key_cleanup.py deleted file mode 100644 index cae4b0ab16..0000000000 --- a/ietf/liaisons/migrations/0006_document_primary_key_cleanup.py +++ /dev/null @@ -1,23 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-06-10 03:47 - - -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('liaisons', '0005_rename_field_document2'), - ] - - operations = [ - migrations.AlterField( - model_name='liaisonstatementattachment', - name='document', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document'), - ), - ] diff --git a/ietf/liaisons/migrations/0007_auto_20201109_0439.py b/ietf/liaisons/migrations/0007_auto_20201109_0439.py deleted file mode 100644 index cd9aa967e6..0000000000 --- a/ietf/liaisons/migrations/0007_auto_20201109_0439.py +++ /dev/null @@ -1,17 +0,0 @@ -# Generated by Django 2.2.17 on 2020-11-09 04:39 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('liaisons', '0006_document_primary_key_cleanup'), - ] - - operations = [ - migrations.AddIndex( - model_name='liaisonstatementevent', - index=models.Index(fields=['-time', '-id'], name='liaisons_li_time_3e1646_idx'), - ), - ] diff --git a/ietf/liaisons/migrations/0008_purge_liaisonstatementgroupcontacts_data.py b/ietf/liaisons/migrations/0008_purge_liaisonstatementgroupcontacts_data.py deleted file mode 100644 index 4bec1a4f01..0000000000 --- a/ietf/liaisons/migrations/0008_purge_liaisonstatementgroupcontacts_data.py +++ /dev/null @@ -1,62 +0,0 @@ -# Generated by Django 2.2.17 on 2020-12-10 06:21 - -from django.db import migrations - -from ietf.person.name import plain_name - - -def forward(apps, schema_editor): - """Delete LiaisonStatementGroupContacts records""" - LiaisonStatementGroupContacts = apps.get_model('liaisons', 'LiaisonStatementGroupContacts') - LiaisonStatementGroupContacts.objects.all().delete() - - -def contacts_from_roles(roles): - """Create contacts string from Role queryset""" - emails = [] - for r in roles: - if not r.person.plain and r.person.name == r.email.address: - # Person was just a stand-in for a bare email address, so just return a bare email address - emails.append(r.email.address) - else: - # Person had a name of some sort, use that as the friendly name - person_name = r.person.plain if r.person.plain else plain_name(r.person.name) - emails.append('{} <{}>'.format(person_name,r.email.address)) - return ','.join(emails) - -def reverse(apps, schema_editor): - """Recreate LiaisonStatementGroupContacts records - - Note that this does not exactly reproduce the original contents. In particular, email addresses - in contacts or cc_contacts may have had different real names than those in the corresponding - email.person.name field. In this case, the record will be reconstructed with the name from - the Person model. The email addresses should be unchanged, though. - """ - LiaisonStatementGroupContacts = apps.get_model('liaisons', 'LiaisonStatementGroupContacts') - Group = apps.get_model('group', 'Group') - Role = apps.get_model('group', 'Role') - RoleName=apps.get_model('name', 'RoleName') - - contact_role_name = RoleName.objects.get(slug='liaison_contact') - cc_contact_role_name = RoleName.objects.get(slug='liaison_cc_contact') - - for group in Group.objects.all(): - contacts = Role.objects.filter(name=contact_role_name, group=group) - cc_contacts = Role.objects.filter(name=cc_contact_role_name, group=group) - if contacts.exists() or cc_contacts.exists(): - LiaisonStatementGroupContacts.objects.create( - group_id=group.pk, - contacts=contacts_from_roles(contacts), - cc_contacts=contacts_from_roles(cc_contacts), - ) - -class Migration(migrations.Migration): - - dependencies = [ - ('liaisons', '0007_auto_20201109_0439'), - ('group', '0041_create_liaison_contact_roles'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/liaisons/migrations/0009_delete_liaisonstatementgroupcontacts_model.py b/ietf/liaisons/migrations/0009_delete_liaisonstatementgroupcontacts_model.py deleted file mode 100644 index c4f654f1d0..0000000000 --- a/ietf/liaisons/migrations/0009_delete_liaisonstatementgroupcontacts_model.py +++ /dev/null @@ -1,16 +0,0 @@ -# Generated by Django 2.2.17 on 2020-12-10 10:05 - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('liaisons', '0008_purge_liaisonstatementgroupcontacts_data'), - ] - - operations = [ - migrations.DeleteModel( - name='LiaisonStatementGroupContacts', - ), - ] diff --git a/ietf/liaisons/models.py b/ietf/liaisons/models.py index 6302bea779..a2d79ea476 100644 --- a/ietf/liaisons/models.py +++ b/ietf/liaisons/models.py @@ -7,13 +7,14 @@ from django.db import models from django.utils.text import slugify -from ietf.person.models import Email, Person +from ietf.person.models import Person from ietf.name.models import (LiaisonStatementPurposeName, LiaisonStatementState, LiaisonStatementEventTypeName, LiaisonStatementTagName, DocRelationshipName) from ietf.doc.models import Document from ietf.group.models import Group from ietf.utils.models import ForeignKey +from ietf.utils.validators import validate_mailbox_address # maps (previous state id, new state id) to event type id STATE_EVENT_MAPPING = { @@ -29,11 +30,16 @@ class LiaisonStatement(models.Model): title = models.CharField(max_length=255) from_groups = models.ManyToManyField(Group, blank=True, related_name='liaisonstatement_from_set') - from_contact = ForeignKey(Email, blank=True, null=True) + from_contact = models.CharField( + blank=True, + max_length=512, + help_text="Address of the formal sender of the statement", + validators=(validate_mailbox_address,) + ) to_groups = models.ManyToManyField(Group, blank=True, related_name='liaisonstatement_to_set') to_contacts = models.CharField(max_length=2000, help_text="Contacts at recipient group") - response_contacts = models.CharField(blank=True, max_length=255, help_text="Where to send a response") # RFC4053 + response_contacts = models.TextField(blank=True, max_length=1024, help_text="Where to send a response") # RFC4053 technical_contacts = models.CharField(blank=True, max_length=255, help_text="Who to contact for clarification") # RFC4053 action_holder_contacts = models.CharField(blank=True, max_length=255, help_text="Who makes sure action is completed") # incoming only? cc_contacts = models.TextField(blank=True) @@ -44,7 +50,7 @@ class LiaisonStatement(models.Model): body = models.TextField(blank=True) tags = models.ManyToManyField(LiaisonStatementTagName, blank=True) - attachments = models.ManyToManyField(Document, through='LiaisonStatementAttachment', blank=True) + attachments = models.ManyToManyField(Document, through='liaisons.LiaisonStatementAttachment', blank=True) state = ForeignKey(LiaisonStatementState, default='pending') class Meta: @@ -85,7 +91,7 @@ def name(self): if self.from_groups.count(): frm = ', '.join([i.acronym or i.name for i in self.from_groups.all()]) else: - frm = self.from_contact.person.name + frm = self.from_contact if self.to_groups.count(): to = ', '.join([i.acronym or i.name for i in self.to_groups.all()]) else: diff --git a/ietf/liaisons/resources.py b/ietf/liaisons/resources.py index 8f31ea3a64..02cd159a11 100644 --- a/ietf/liaisons/resources.py +++ b/ietf/liaisons/resources.py @@ -15,12 +15,10 @@ RelatedLiaisonStatement) -from ietf.person.resources import EmailResource from ietf.group.resources import GroupResource from ietf.name.resources import LiaisonStatementPurposeNameResource, LiaisonStatementTagNameResource, LiaisonStatementStateResource from ietf.doc.resources import DocumentResource class LiaisonStatementResource(ModelResource): - from_contact = ToOneField(EmailResource, 'from_contact', null=True) purpose = ToOneField(LiaisonStatementPurposeNameResource, 'purpose') state = ToOneField(LiaisonStatementStateResource, 'state') from_groups = ToManyField(GroupResource, 'from_groups', null=True) @@ -36,6 +34,7 @@ class Meta: filtering = { "id": ALL, "title": ALL, + "from_contact": ALL, "to_contacts": ALL, "response_contacts": ALL, "technical_contacts": ALL, @@ -44,9 +43,6 @@ class Meta: "deadline": ALL, "other_identifiers": ALL, "body": ALL, - "from_name": ALL, - "to_name": ALL, - "from_contact": ALL_WITH_RELATIONS, "purpose": ALL_WITH_RELATIONS, "state": ALL_WITH_RELATIONS, "from_groups": ALL_WITH_RELATIONS, diff --git a/ietf/liaisons/tests.py b/ietf/liaisons/tests.py index f2fd5c4529..e29045443f 100644 --- a/ietf/liaisons/tests.py +++ b/ietf/liaisons/tests.py @@ -19,12 +19,13 @@ from io import StringIO from pyquery import PyQuery +from ietf.doc.storage_utils import retrieve_str from ietf.utils.test_utils import TestCase, login_testing_unauthorized from ietf.utils.mail import outbox from ietf.group.factories import GroupFactory, RoleFactory from ietf.liaisons.factories import ( LiaisonStatementFactory, - LiaisonStatementEventFactory, LiaisonStatementAttachmentFactory, ) + LiaisonStatementEventFactory, LiaisonStatementAttachmentFactory, RelatedLiaisonStatementFactory) from ietf.liaisons.models import (LiaisonStatement, LiaisonStatementPurposeName, LiaisonStatementAttachment) from ietf.person.models import Person @@ -109,65 +110,74 @@ def test_help_pages(self): self.assertEqual(self.client.get('/liaison/help/from_ietf/').status_code, 200) self.assertEqual(self.client.get('/liaison/help/to_ietf/').status_code, 200) + def test_list_other_sdo(self): + GroupFactory(type_id="sdo", state_id="conclude", acronym="third") + GroupFactory(type_id="sdo", state_id="active", acronym="second") + GroupFactory(type_id="sdo", state_id="active", acronym="first") + url = urlreverse("ietf.liaisons.views.list_other_sdo") + login_testing_unauthorized(self, "secretary", url) + r = self.client.get(url) + q = PyQuery(r.content) + self.assertEqual(len(q("h1")), 2) + first_td_elements_text = [e.text for e in q("tr").find("td:first-child a")] + self.assertEqual(first_td_elements_text, ["first", "second", "third"]) class UnitTests(TestCase): - def test_get_cc(self): - from ietf.liaisons.views import get_cc,EMAIL_ALIASES + def test_get_contacts_for_liaison_messages_for_group_primary(self): + from ietf.mailtrigger.utils import get_contacts_for_liaison_messages_for_group_primary,EMAIL_ALIASES # test IETF - cc = get_cc(Group.objects.get(acronym='ietf')) + cc = get_contacts_for_liaison_messages_for_group_primary(Group.objects.get(acronym='ietf')) self.assertTrue(EMAIL_ALIASES['IESG'] in cc) self.assertTrue(EMAIL_ALIASES['IETFCHAIR'] in cc) # test IAB - cc = get_cc(Group.objects.get(acronym='iab')) + cc = get_contacts_for_liaison_messages_for_group_primary(Group.objects.get(acronym='iab')) self.assertTrue(EMAIL_ALIASES['IAB'] in cc) self.assertTrue(EMAIL_ALIASES['IABCHAIR'] in cc) - self.assertTrue(EMAIL_ALIASES['IABEXECUTIVEDIRECTOR'] in cc) # test an Area area = Group.objects.filter(type='area').first() - cc = get_cc(area) + cc = get_contacts_for_liaison_messages_for_group_primary(area) self.assertTrue(EMAIL_ALIASES['IETFCHAIR'] in cc) self.assertTrue(contacts_from_roles([area.ad_role()]) in cc) # test a Working Group wg = Group.objects.filter(type='wg').first() - cc = get_cc(wg) + cc = get_contacts_for_liaison_messages_for_group_primary(wg) self.assertTrue(contacts_from_roles([wg.parent.ad_role()]) in cc) self.assertTrue(contacts_from_roles([wg.get_chair()]) in cc) # test an SDO sdo = RoleFactory(name_id='liaiman',group__type_id='sdo',).group - cc = get_cc(sdo) + cc = get_contacts_for_liaison_messages_for_group_primary(sdo) self.assertTrue(contacts_from_roles([sdo.role_set.filter(name='liaiman').first()]) in cc) # test a cc_contact role cc_contact_role = RoleFactory(name_id='liaison_cc_contact', group=sdo) - cc = get_cc(sdo) + cc = get_contacts_for_liaison_messages_for_group_primary(sdo) self.assertIn(contact_email_from_role(cc_contact_role), cc) - def test_get_contacts_for_group(self): - from ietf.liaisons.views import get_contacts_for_group, EMAIL_ALIASES + def test_get_contacts_for_liaison_messages_for_group_secondary(self): + from ietf.mailtrigger.utils import get_contacts_for_liaison_messages_for_group_secondary,EMAIL_ALIASES - # test explicit + # test explicit group contacts sdo = GroupFactory(type_id='sdo') contact_email = RoleFactory(name_id='liaison_contact', group=sdo).email.address - contacts = get_contacts_for_group(sdo) + contacts = get_contacts_for_liaison_messages_for_group_secondary(sdo) self.assertIsNotNone(contact_email) self.assertIn(contact_email, contacts) # test area area = Group.objects.filter(type='area').first() - contacts = get_contacts_for_group(area) + contacts = get_contacts_for_liaison_messages_for_group_secondary(area) self.assertTrue(area.ad_role().email.address in contacts) # test wg wg = Group.objects.filter(type='wg').first() - contacts = get_contacts_for_group(wg) + contacts = get_contacts_for_liaison_messages_for_group_secondary(wg) self.assertTrue(wg.get_chair().email.address in contacts) # test ietf - contacts = get_contacts_for_group(Group.objects.get(acronym='ietf')) + contacts = get_contacts_for_liaison_messages_for_group_secondary(Group.objects.get(acronym='ietf')) self.assertTrue(EMAIL_ALIASES['IETFCHAIR'] in contacts) # test iab - contacts = get_contacts_for_group(Group.objects.get(acronym='iab')) + contacts = get_contacts_for_liaison_messages_for_group_secondary(Group.objects.get(acronym='iab')) self.assertTrue(EMAIL_ALIASES['IABCHAIR'] in contacts) - self.assertTrue(EMAIL_ALIASES['IABEXECUTIVEDIRECTOR'] in contacts) # test iesg - contacts = get_contacts_for_group(Group.objects.get(acronym='iesg')) + contacts = get_contacts_for_liaison_messages_for_group_secondary(Group.objects.get(acronym='iesg')) self.assertTrue(EMAIL_ALIASES['IESG'] in contacts) def test_needs_approval(self): @@ -204,7 +214,6 @@ def test_ajax(self): self.assertEqual(r.status_code, 200) data = r.json() self.assertEqual(data["error"], False) - self.assertEqual(data["post_only"], False) self.assertTrue('cc' in data) self.assertTrue('needs_approval' in data) self.assertTrue('to_contacts' in data) @@ -364,6 +373,9 @@ def test_approval_process(self): self.assertEqual(len(q('form button[name=approved]')), 0) # check the detail page / authorized + r = self.client.post(url, dict(dead="1")) + self.assertEqual(r.status_code, 403) + mailbox_before = len(outbox) self.client.login(username="ulm-liaiman", password="ulm-liaiman+password") r = self.client.get(url) self.assertEqual(r.status_code, 200) @@ -393,6 +405,8 @@ def test_edit_liaison(self): LiaisonStatementEventFactory(statement=liaison,type_id='posted') from_group = liaison.from_groups.first() to_group = liaison.to_groups.first() + rel1 = RelatedLiaisonStatementFactory(source=liaison) + rel2 = RelatedLiaisonStatementFactory(source=liaison) url = urlreverse('ietf.liaisons.views.liaison_edit', kwargs=dict(object_id=liaison.pk)) login_testing_unauthorized(self, "secretary", url) @@ -402,10 +416,18 @@ def test_edit_liaison(self): self.assertEqual(r.status_code, 200) q = PyQuery(r.content) self.assertEqual(len(q('form input[name=from_contact]')), 1) + json_data = q('form select[name=related_to]').attr('data-pre') + try: + decoded = json.loads(json_data) + except json.JSONDecodeError as e: + self.fail('data-pre contained invalid JSON data: %s' % str(e)) + decoded_ids = [item['id'] for item in decoded] + self.assertEqual(decoded_ids, [rel1.target.id, rel2.target.id]) # edit attachments_before = liaison.attachments.count() - test_file = StringIO("hello world") + test_content = "hello world" + test_file = StringIO(test_content) test_file.name = "unnamed" r = self.client.post(url, dict(from_groups=str(from_group.pk), @@ -443,16 +465,20 @@ def test_edit_liaison(self): self.assertEqual(attachment.title, "attachment") with (Path(settings.LIAISON_ATTACH_PATH) / attachment.uploaded_filename).open() as f: written_content = f.read() + self.assertEqual(written_content, test_content) + self.assertEqual( + retrieve_str(attachment.type_id, attachment.uploaded_filename), + test_content, + ) - test_file.seek(0) - self.assertEqual(written_content, test_file.read()) def test_incoming_access(self): - '''Ensure only Secretariat, Liaison Managers, and Authorized Individuals + '''Ensure only Secretariat, Liaison Managers, Liaison Coordinators, and Authorized Individuals have access to incoming liaisons. ''' sdo = RoleFactory(name_id='liaiman',group__type_id='sdo', person__user__username='ulm-liaiman').group RoleFactory(name_id='auth',group=sdo,person__user__username='ulm-auth') + RoleFactory(name_id='liaison_coordinator', group__acronym='iab', person__user__username='liaison-coordinator') stmt = LiaisonStatementFactory(from_groups=[sdo,]) LiaisonStatementEventFactory(statement=stmt,type_id='posted') RoleFactory(name_id='chair',person__user__username='marschairman',group__acronym='mars') @@ -485,6 +511,15 @@ def test_incoming_access(self): r = self.client.get(addurl) self.assertEqual(r.status_code, 200) + # Liaison Coordinator has access + self.client.login(username="liaison-coordinator", password="liaison-coordinator+password") + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q('a.btn:contains("New incoming liaison")')), 1) + r = self.client.get(addurl) + self.assertEqual(r.status_code, 200) + # Authorized Individual has access self.client.login(username="ulm-auth", password="ulm-auth+password") r = self.client.get(url) @@ -507,9 +542,9 @@ def test_outgoing_access(self): sdo = RoleFactory(name_id='liaiman',group__type_id='sdo', person__user__username='ulm-liaiman').group RoleFactory(name_id='auth',group=sdo,person__user__username='ulm-auth') + RoleFactory(name_id='liaison_coordinator', group__acronym='iab', person__user__username='liaison-coordinator') mars = RoleFactory(name_id='chair',person__user__username='marschairman',group__acronym='mars').group RoleFactory(name_id='secr',group=mars,person__user__username='mars-secr') - RoleFactory(name_id='execdir',group=Group.objects.get(acronym='iab'),person__user__username='iab-execdir') url = urlreverse('ietf.liaisons.views.liaison_list') addurl = urlreverse('ietf.liaisons.views.liaison_add', kwargs={'type':'outgoing'}) @@ -567,17 +602,17 @@ def test_outgoing_access(self): r = self.client.get(addurl) self.assertEqual(r.status_code, 200) - # IAB Executive Director - self.assertTrue(self.client.login(username="iab-execdir", password="iab-execdir+password")) + # Liaison Manager has access + self.assertTrue(self.client.login(username="ulm-liaiman", password="ulm-liaiman+password")) r = self.client.get(url) self.assertEqual(r.status_code, 200) q = PyQuery(r.content) - self.assertEqual(len(q("a.btn:contains('New outgoing liaison')")), 1) + self.assertEqual(len(q('a.btn:contains("New outgoing liaison")')), 1) r = self.client.get(addurl) self.assertEqual(r.status_code, 200) - # Liaison Manager has access - self.assertTrue(self.client.login(username="ulm-liaiman", password="ulm-liaiman+password")) + # Liaison Coordinator has access + self.assertTrue(self.client.login(username="liaison-coordinator", password="liaison-coordinator+password")) r = self.client.get(url) self.assertEqual(r.status_code, 200) q = PyQuery(r.content) @@ -695,12 +730,13 @@ def test_add_incoming_liaison(self): # add new mailbox_before = len(outbox) - test_file = StringIO("hello world") + test_content = "hello world" + test_file = StringIO(test_content) test_file.name = "unnamed" from_groups = [ str(g.pk) for g in Group.objects.filter(type="sdo") ] to_group = Group.objects.get(acronym="mars") submitter = Person.objects.get(user__username="marschairman") - today = date_today(datetime.timezone.utc) + today = date_today(datetime.UTC) related_liaison = liaison r = self.client.post(url, dict(from_groups=from_groups, @@ -725,7 +761,7 @@ def test_add_incoming_liaison(self): l = LiaisonStatement.objects.all().order_by("-id")[0] self.assertEqual(l.from_groups.count(),2) - self.assertEqual(l.from_contact.address, submitter.email_address()) + self.assertEqual(l.from_contact, submitter.email_address()) self.assertSequenceEqual(l.to_groups.all(),[to_group]) self.assertEqual(l.technical_contacts, "technical_contact@example.com") self.assertEqual(l.action_holder_contacts, "action_holder_contacts@example.com") @@ -747,6 +783,11 @@ def test_add_incoming_liaison(self): self.assertEqual(attachment.title, "attachment") with (Path(settings.LIAISON_ATTACH_PATH) / attachment.uploaded_filename).open() as f: written_content = f.read() + self.assertEqual(written_content, test_content) + self.assertEqual( + retrieve_str(attachment.type_id, attachment.uploaded_filename), + test_content + ) test_file.seek(0) self.assertEqual(written_content, test_file.read()) @@ -755,8 +796,11 @@ def test_add_incoming_liaison(self): self.assertTrue("Liaison Statement" in outbox[-1]["Subject"]) self.assertTrue('to_contacts@' in outbox[-1]['To']) + self.assertTrue(submitter.email_address(), outbox[-1]['To']) self.assertTrue('cc@' in outbox[-1]['Cc']) + + def test_add_outgoing_liaison(self): RoleFactory(name_id='liaiman',group__type_id='sdo', person__user__username='ulm-liaiman') wg = RoleFactory(name_id='chair',person__user__username='marschairman',group__acronym='mars').group @@ -774,12 +818,13 @@ def test_add_outgoing_liaison(self): # add new mailbox_before = len(outbox) - test_file = StringIO("hello world") + test_content = "hello world" + test_file = StringIO(test_content) test_file.name = "unnamed" from_group = Group.objects.get(acronym="mars") to_group = Group.objects.filter(type="sdo")[0] submitter = Person.objects.get(user__username="marschairman") - today = date_today(datetime.timezone.utc) + today = date_today(datetime.UTC) related_liaison = liaison r = self.client.post(url, dict(from_groups=str(from_group.pk), @@ -804,7 +849,7 @@ def test_add_outgoing_liaison(self): l = LiaisonStatement.objects.all().order_by("-id")[0] self.assertSequenceEqual(l.from_groups.all(), [from_group]) - self.assertEqual(l.from_contact.address, submitter.email_address()) + self.assertEqual(l.from_contact, submitter.email_address()) self.assertSequenceEqual(l.to_groups.all(), [to_group]) self.assertEqual(l.to_contacts, "to_contacts@example.com") self.assertEqual(l.technical_contacts, "technical_contact@example.com") @@ -826,44 +871,16 @@ def test_add_outgoing_liaison(self): self.assertEqual(attachment.title, "attachment") with (Path(settings.LIAISON_ATTACH_PATH) / attachment.uploaded_filename).open() as f: written_content = f.read() - - test_file.seek(0) - self.assertEqual(written_content, test_file.read()) + self.assertEqual(written_content, test_content) + self.assertEqual( + retrieve_str(attachment.type_id, attachment.uploaded_filename), + test_content + ) self.assertEqual(len(outbox), mailbox_before + 1) self.assertTrue("Liaison Statement" in outbox[-1]["Subject"]) self.assertTrue('aread@' in outbox[-1]['To']) - - def test_add_outgoing_liaison_unapproved_post_only(self): - RoleFactory(name_id='liaiman',group__type_id='sdo', person__user__username='ulm-liaiman') - mars = RoleFactory(name_id='chair',person__user__username='marschairman',group__acronym='mars').group - RoleFactory(name_id='ad',group=mars) - - url = urlreverse('ietf.liaisons.views.liaison_add', kwargs={'type':'outgoing'}) - login_testing_unauthorized(self, "secretary", url) - - # add new - mailbox_before = len(outbox) - from_group = Group.objects.get(acronym="mars") - to_group = Group.objects.filter(type="sdo")[0] - submitter = Person.objects.get(user__username="marschairman") - today = date_today(datetime.timezone.utc) - r = self.client.post(url, - dict(from_groups=str(from_group.pk), - from_contact=submitter.email_address(), - to_groups=str(to_group.pk), - to_contacts='to_contacts@example.com', - approved="", - purpose="info", - title="title", - submitted_date=today.strftime("%Y-%m-%d"), - body="body", - post_only="1", - )) - self.assertEqual(r.status_code, 302) - l = LiaisonStatement.objects.all().order_by("-id")[0] - self.assertEqual(l.state.slug,'pending') - self.assertEqual(len(outbox), mailbox_before + 1) + self.assertTrue(submitter.email_address(), outbox[-1]['Cc']) def test_liaison_add_attachment(self): liaison = LiaisonStatementFactory(deadline=date_today(DEADLINE_TZINFO)+datetime.timedelta(days=1)) @@ -873,11 +890,12 @@ def test_liaison_add_attachment(self): # get minimum edit post data - file = StringIO('dummy file') + test_data = "dummy file" + file = StringIO(test_data) file.name = "upload.txt" post_data = dict( from_groups = ','.join([ str(x.pk) for x in liaison.from_groups.all() ]), - from_contact = liaison.from_contact.address, + from_contact = liaison.from_contact, to_groups = ','.join([ str(x.pk) for x in liaison.to_groups.all() ]), to_contacts = 'to_contacts@example.com', purpose = liaison.purpose.slug, @@ -900,28 +918,50 @@ def test_liaison_add_attachment(self): self.assertEqual(liaison.attachments.count(),1) event = liaison.liaisonstatementevent_set.order_by('id').last() self.assertTrue(event.desc.startswith('Added attachment')) + attachment = liaison.attachments.get() + self.assertEqual( + retrieve_str(attachment.type_id, attachment.uploaded_filename), + test_data + ) def test_liaison_edit_attachment(self): - - attachment = LiaisonStatementAttachmentFactory(document__name='liaiatt-1') - url = urlreverse('ietf.liaisons.views.liaison_edit_attachment', kwargs=dict(object_id=attachment.statement_id,doc_id=attachment.document_id)) + attachment = LiaisonStatementAttachmentFactory(document__name="liaiatt-1") + url = urlreverse( + "ietf.liaisons.views.liaison_edit_attachment", + kwargs=dict( + object_id=attachment.statement_id, doc_id=attachment.document_id + ), + ) login_testing_unauthorized(self, "secretary", url) r = self.client.get(url) self.assertEqual(r.status_code, 200) - post_data = dict(title='New Title') - r = self.client.post(url,post_data) + post_data = dict(title="New Title") + r = self.client.post(url, post_data) attachment = LiaisonStatementAttachment.objects.get(pk=attachment.pk) self.assertEqual(r.status_code, 302) - self.assertEqual(attachment.document.title,'New Title') - - def test_liaison_delete_attachment(self): - attachment = LiaisonStatementAttachmentFactory(document__name='liaiatt-1') - liaison = attachment.statement - url = urlreverse('ietf.liaisons.views.liaison_delete_attachment', kwargs=dict(object_id=liaison.pk,attach_id=attachment.pk)) - login_testing_unauthorized(self, "secretary", url) + self.assertEqual(attachment.document.title, "New Title") + + # ensure attempts to edit attachments not attached to this liaison statement fail + other_attachment = LiaisonStatementAttachmentFactory(document__name="liaiatt-2") + url = urlreverse( + "ietf.liaisons.views.liaison_edit_attachment", + kwargs=dict( + object_id=attachment.statement_id, doc_id=other_attachment.document_id + ), + ) r = self.client.get(url) - self.assertEqual(r.status_code, 302) - self.assertEqual(liaison.liaisonstatementattachment_set.filter(removed=False).count(),0) + self.assertEqual(r.status_code, 404) + r = self.client.post(url, dict(title="New Title")) + self.assertEqual(r.status_code, 404) + + # def test_liaison_delete_attachment(self): + # attachment = LiaisonStatementAttachmentFactory(document__name='liaiatt-1') + # liaison = attachment.statement + # url = urlreverse('ietf.liaisons.views.liaison_delete_attachment', kwargs=dict(object_id=liaison.pk,attach_id=attachment.pk)) + # login_testing_unauthorized(self, "secretary", url) + # r = self.client.get(url) + # self.assertEqual(r.status_code, 302) + # self.assertEqual(liaison.liaisonstatementattachment_set.filter(removed=False).count(),0) def test_in_response(self): '''A statement with purpose=in_response must have related statement specified''' @@ -1025,7 +1065,7 @@ def test_search(self): LiaisonStatementEventFactory(type_id='posted', statement__body="Has recently in its body",statement__from_groups=[GroupFactory(type_id='sdo',acronym='ulm'),]) # Statement 2 s2 = LiaisonStatementEventFactory(type_id='posted', statement__body="That word does not occur here", statement__title="Nor does it occur here") - s2.time=datetime.datetime(2010, 1, 1, tzinfo=datetime.timezone.utc) + s2.time=datetime.datetime(2010, 1, 1, tzinfo=datetime.UTC) s2.save() # test list only, no search filters @@ -1090,17 +1130,6 @@ def test_redirect_for_approval(self): # ------------------------------------------------- # Form validations # ------------------------------------------------- - def test_post_and_send_fail(self): - RoleFactory(name_id='liaiman',person__user__username='ulm-liaiman',group__type_id='sdo',group__acronym='ulm') - GroupFactory(type_id='wg',acronym='mars') - - url = urlreverse('ietf.liaisons.views.liaison_add', kwargs={'type':'incoming'}) - login_testing_unauthorized(self, "ulm-liaiman", url) - - r = self.client.post(url,get_liaison_post_data(),follow=True) - - self.assertEqual(r.status_code, 200) - self.assertContains(r, 'As an IETF Liaison Manager you can not send incoming liaison statements') def test_deadline_field(self): '''Required for action, comment, not info, response''' @@ -1165,4 +1194,4 @@ def test_send_liaison_deadline_reminder(self): mailbox_before = len(outbox) possibly_send_deadline_reminder(liaison) - self.assertEqual(len(outbox), mailbox_before) \ No newline at end of file + self.assertEqual(len(outbox), mailbox_before) diff --git a/ietf/liaisons/tests_forms.py b/ietf/liaisons/tests_forms.py new file mode 100644 index 0000000000..101c0c8298 --- /dev/null +++ b/ietf/liaisons/tests_forms.py @@ -0,0 +1,217 @@ +# Copyright The IETF Trust 2025, All Rights Reserved +from ietf.group.factories import GroupFactory, RoleFactory +from ietf.group.models import Group +from ietf.liaisons.forms import ( + flatten_choices, + choices_from_group_queryset, + all_internal_groups, + internal_groups_for_person, + external_groups_for_person, +) +from ietf.person.factories import PersonFactory +from ietf.person.models import Person +from ietf.utils.test_utils import TestCase + + +class HelperTests(TestCase): + @staticmethod + def _alphabetically_by_acronym(group_list): + return sorted(group_list, key=lambda item: item.acronym) + + def test_choices_from_group_queryset(self): + main_groups = list(Group.objects.filter(acronym__in=["ietf", "iab"])) + areas = GroupFactory.create_batch(2, type_id="area") + wgs = GroupFactory.create_batch(2) + + # No groups + self.assertEqual( + choices_from_group_queryset(Group.objects.none()), + [], + ) + + # Main groups only + choices = choices_from_group_queryset( + Group.objects.filter(pk__in=[g.pk for g in main_groups]) + ) + self.assertEqual(len(choices), 1, "show one optgroup, hide empty ones") + self.assertEqual(choices[0][0], "Main IETF Entities") + self.assertEqual( + [val for val, _ in choices[0][1]], # extract the choice value + [g.pk for g in self._alphabetically_by_acronym(main_groups)], + ) + + # Area groups only + choices = choices_from_group_queryset( + Group.objects.filter(pk__in=[g.pk for g in areas]) + ) + self.assertEqual(len(choices), 1, "show one optgroup, hide empty ones") + self.assertEqual(choices[0][0], "IETF Areas") + self.assertEqual( + [val for val, _ in choices[0][1]], # extract the choice value + [g.pk for g in self._alphabetically_by_acronym(areas)], + ) + + # WGs only + choices = choices_from_group_queryset( + Group.objects.filter(pk__in=[g.pk for g in wgs]) + ) + self.assertEqual(len(choices), 1, "show one optgroup, hide empty ones") + self.assertEqual(choices[0][0], "IETF Working Groups") + self.assertEqual( + [val for val, _ in choices[0][1]], # extract the choice value + [g.pk for g in self._alphabetically_by_acronym(wgs)], + ) + + # All together + choices = choices_from_group_queryset( + Group.objects.filter(pk__in=[g.pk for g in main_groups + areas + wgs]) + ) + self.assertEqual(len(choices), 3, "show all three optgroups") + self.assertEqual( + [optgroup_label for optgroup_label, _ in choices], + ["Main IETF Entities", "IETF Areas", "IETF Working Groups"], + ) + self.assertEqual( + [val for val, _ in choices[0][1]], # extract the choice value + [g.pk for g in self._alphabetically_by_acronym(main_groups)], + ) + self.assertEqual( + [val for val, _ in choices[1][1]], # extract the choice value + [g.pk for g in self._alphabetically_by_acronym(areas)], + ) + self.assertEqual( + [val for val, _ in choices[2][1]], # extract the choice value + [g.pk for g in self._alphabetically_by_acronym(wgs)], + ) + + def test_all_internal_groups(self): + # test relies on the data created in ietf.utils.test_data.make_immutable_test_data() + self.assertCountEqual( + all_internal_groups().values_list("acronym", flat=True), + {"ietf", "iab", "iesg", "farfut", "ops", "sops"}, + ) + + def test_internal_groups_for_person(self): + # test relies on the data created in ietf.utils.test_data.make_immutable_test_data() + # todo add liaison coordinator when modeled + RoleFactory( + name_id="auth", + group__type_id="sdo", + group__acronym="sdo", + person__user__username="sdo-authperson", + ) + + self.assertQuerysetEqual( + internal_groups_for_person(None), + Group.objects.none(), + msg="no Person means no groups", + ) + self.assertQuerysetEqual( + internal_groups_for_person(PersonFactory()), + Group.objects.none(), + msg="no Role means no groups", + ) + + for username in ( + "secretary", + "ietf-chair", + "iab-chair", + "sdo-authperson", + ): + returned_queryset = internal_groups_for_person( + Person.objects.get(user__username=username) + ) + self.assertCountEqual( + returned_queryset.values_list("acronym", flat=True), + {"ietf", "iab", "iesg", "farfut", "ops", "sops"}, + f"{username} should get all groups", + ) + + # "ops-ad" user is the AD of the "ops" area, which contains the "sops" wg + self.assertCountEqual( + internal_groups_for_person( + Person.objects.get(user__username="ops-ad") + ).values_list("acronym", flat=True), + {"ietf", "iesg", "ops", "sops"}, + "area director should get only their area, its wgs, and the ietf/iesg groups", + ) + + self.assertCountEqual( + internal_groups_for_person( + Person.objects.get(user__username="sopschairman"), + ).values_list("acronym", flat=True), + {"sops"}, + "wg chair should get only their wg", + ) + + def test_external_groups_for_person(self): + RoleFactory(name_id="liaison_coordinator", group__acronym="iab", person__user__username="liaison-coordinator") + the_sdo = GroupFactory(type_id="sdo", acronym="the-sdo") + liaison_manager = RoleFactory(name_id="liaiman", group=the_sdo).person + authperson = RoleFactory(name_id="auth", group=the_sdo).person + + GroupFactory(acronym="other-sdo", type_id="sdo") + for username in ( + "secretary", + "ietf-chair", + "iab-chair", + "liaison-coordinator", + "ad", + "sopschairman", + "sopssecretary", + ): + person = Person.objects.get(user__username=username) + self.assertCountEqual( + external_groups_for_person( + person, + ).values_list("acronym", flat=True), + {"the-sdo", "other-sdo"}, + f"{username} should get all SDO groups", + ) + tmp_role = RoleFactory(name_id="chair", group__type_id="wg", person=person) + self.assertCountEqual( + external_groups_for_person( + person, + ).values_list("acronym", flat=True), + {"the-sdo", "other-sdo"}, + f"{username} should still get all SDO groups when they also a liaison manager", + ) + tmp_role.delete() + + self.assertCountEqual( + external_groups_for_person(liaison_manager).values_list( + "acronym", flat=True + ), + {"the-sdo"}, + "liaison manager should get only their SDO group", + ) + self.assertCountEqual( + external_groups_for_person(authperson).values_list("acronym", flat=True), + {"the-sdo"}, + "authorized individual should get only their SDO group", + ) + + def test_flatten_choices(self): + self.assertEqual(flatten_choices([]), []) + self.assertEqual( + flatten_choices( + ( + ("group A", ()), + ("group B", (("val0", "label0"), ("val1", "label1"))), + ("group C", (("val2", "label2"),)), + ) + ), + [("val0", "label0"), ("val1", "label1"), ("val2", "label2")], + ) + + +class IncomingLiaisonFormTests(TestCase): + pass + + +class OutgoingLiaisonFormTests(TestCase): + pass + + +class EditLiaisonFormTests(TestCase): + pass diff --git a/ietf/liaisons/urls.py b/ietf/liaisons/urls.py index a4afbfef5d..498df3b965 100644 --- a/ietf/liaisons/urls.py +++ b/ietf/liaisons/urls.py @@ -26,8 +26,8 @@ url(r'^(?P\d+)/$', views.liaison_detail), url(r'^(?P\d+)/addcomment/$', views.add_comment), url(r'^(?P\d+)/edit/$', views.liaison_edit), - url(r'^(?P\d+)/edit-attachment/(?P[A-Za-z0-9._+-]+)$', views.liaison_edit_attachment), - url(r'^(?P\d+)/delete-attachment/(?P[A-Za-z0-9._+-]+)$', views.liaison_delete_attachment), + url(r'^(?P\d+)/edit-attachment/(?P[0-9]+)$', views.liaison_edit_attachment), + url(r'^(?P\d+)/delete-attachment/(?P[0-9]+)$', views.liaison_delete_attachment), url(r'^(?P\d+)/history/$', views.liaison_history), url(r'^(?P\d+)/reply/$', views.liaison_reply), url(r'^(?P\d+)/resend/$', views.liaison_resend), @@ -37,4 +37,5 @@ url(r'^add/$', views.redirect_add), url(r'^for_approval/$', views.redirect_for_approval), url(r'^for_approval/(?P\d+)/$', views.redirect_for_approval), + url(r"^list_other_sdo/$", views.list_other_sdo), ] diff --git a/ietf/liaisons/utils.py b/ietf/liaisons/utils.py index df48831917..469bbc5c87 100644 --- a/ietf/liaisons/utils.py +++ b/ietf/liaisons/utils.py @@ -4,6 +4,21 @@ from ietf.liaisons.models import LiaisonStatement from ietf.ietfauth.utils import has_role, passes_test_decorator +# Roles allowed to create and manage outgoing liaison statements. +OUTGOING_LIAISON_ROLES = [ + "Area Director", + "IAB Chair", + "IETF Chair", + "Liaison Manager", + "Liaison Coordinator", + "Secretariat", + "WG Chair", + "WG Secretary", +] + +# Roles allowed to create and manage incoming liaison statements. +INCOMING_LIAISON_ROLES = ["Authorized Individual", "Liaison Manager", "Liaison Coordinator", "Secretariat"] + can_submit_liaison_required = passes_test_decorator( lambda u, *args, **kwargs: can_add_liaison(u), "Restricted to participants who are authorized to submit liaison statements on behalf of the various IETF entities") @@ -30,13 +45,13 @@ def can_edit_liaison(user, liaison): '''Returns True if user has edit / approval authority. True if: - - user is Secretariat + - user is Secretariat or Liaison Coordinator - liaison is outgoing and user has approval authority - user is liaison manager of all SDOs involved ''' if not user.is_authenticated: return False - if has_role(user, "Secretariat"): + if has_role(user, "Secretariat") or has_role(user, "Liaison Coordinator"): return True if liaison.is_outgoing() and liaison in approvable_liaison_statements(user): @@ -59,11 +74,10 @@ def get_person_for_user(user): return None def can_add_outgoing_liaison(user): - return has_role(user, ["Area Director","WG Chair","WG Secretary","IETF Chair","IAB Chair", - "IAB Executive Director","Liaison Manager","Secretariat"]) + return has_role(user, OUTGOING_LIAISON_ROLES) def can_add_incoming_liaison(user): - return has_role(user, ["Liaison Manager","Authorized Individual","Secretariat"]) + return has_role(user, INCOMING_LIAISON_ROLES) def can_add_liaison(user): return can_add_incoming_liaison(user) or can_add_outgoing_liaison(user) diff --git a/ietf/liaisons/views.py b/ietf/liaisons/views.py index a8e80a5194..59c6ea69fc 100644 --- a/ietf/liaisons/views.py +++ b/ietf/liaisons/views.py @@ -7,19 +7,17 @@ from django.contrib import messages from django.urls import reverse as urlreverse -from django.core.exceptions import ValidationError +from django.core.exceptions import ValidationError, ObjectDoesNotExist, PermissionDenied from django.core.validators import validate_email from django.db.models import Q, Prefetch -from django.http import HttpResponse +from django.http import Http404, HttpResponse from django.shortcuts import render, get_object_or_404, redirect import debug # pyflakes:ignore -from ietf.doc.models import Document from ietf.ietfauth.utils import role_required, has_role from ietf.group.models import Group, Role -from ietf.liaisons.models import (LiaisonStatement,LiaisonStatementEvent, - LiaisonStatementAttachment) +from ietf.liaisons.models import LiaisonStatement,LiaisonStatementEvent from ietf.liaisons.utils import (get_person_for_user, can_add_outgoing_liaison, can_add_incoming_liaison, can_edit_liaison,can_submit_liaison_required, can_add_liaison) @@ -29,13 +27,6 @@ from ietf.name.models import LiaisonStatementTagName from ietf.utils.response import permission_denied -EMAIL_ALIASES = { - 'IETFCHAIR':'The IETF Chair ', - 'IESG':'The IESG ', - 'IAB':'The IAB ', - 'IABCHAIR':'The IAB Chair ', - 'IABEXECUTIVEDIRECTOR':'The IAB Executive Director '} - # ------------------------------------------------- # Helper Functions # ------------------------------------------------- @@ -57,7 +48,7 @@ def _can_take_care(liaison, user): return False if user.is_authenticated: - if has_role(user, "Secretariat"): + if has_role(user, "Secretariat") or has_role(user, "Liaison Coordinator"): return True else: return _find_person_in_emails(liaison, get_person_for_user(user)) @@ -84,8 +75,6 @@ def _find_person_in_emails(liaison, person): return True elif addr in ('iab@iab.org', 'iab-chair@iab.org') and has_role(person.user, "IAB Chair"): return True - elif addr in ('execd@iab.org', ) and has_role(person.user, "IAB Executive Director"): - return True return False @@ -97,66 +86,6 @@ def contacts_from_roles(roles): emails = [ contact_email_from_role(r) for r in roles ] return ','.join(emails) -def get_cc(group): - '''Returns list of emails to use as CC for group. Simplified refactor of IETFHierarchy - get_cc() and get_from_cc() - ''' - emails = [] - - # role based CCs - if group.acronym in ('ietf','iesg'): - emails.append(EMAIL_ALIASES['IESG']) - emails.append(EMAIL_ALIASES['IETFCHAIR']) - elif group.acronym in ('iab'): - emails.append(EMAIL_ALIASES['IAB']) - emails.append(EMAIL_ALIASES['IABCHAIR']) - emails.append(EMAIL_ALIASES['IABEXECUTIVEDIRECTOR']) - elif group.type_id == 'area': - emails.append(EMAIL_ALIASES['IETFCHAIR']) - ad_roles = group.role_set.filter(name='ad') - emails.extend([ contact_email_from_role(r) for r in ad_roles ]) - elif group.type_id == 'wg': - ad_roles = group.parent.role_set.filter(name='ad') - emails.extend([ contact_email_from_role(r) for r in ad_roles ]) - chair_roles = group.role_set.filter(name='chair') - emails.extend([ contact_email_from_role(r) for r in chair_roles ]) - if group.list_email: - emails.append('{} Discussion List <{}>'.format(group.name,group.list_email)) - elif group.type_id == 'sdo': - liaiman_roles = group.role_set.filter(name='liaiman') - emails.extend([ contact_email_from_role(r) for r in liaiman_roles ]) - - # explicit CCs - liaison_cc_roles = group.role_set.filter(name='liaison_cc_contact') - emails.extend([ contact_email_from_role(r) for r in liaison_cc_roles ]) - - return emails - -def get_contacts_for_group(group): - '''Returns default contacts for groups as a comma separated string''' - # use explicit default contacts if defined - explicit_contacts = contacts_from_roles(group.role_set.filter(name='liaison_contact')) - if explicit_contacts: - return explicit_contacts - - # otherwise construct based on group type - contacts = [] - if group.type_id == 'area': - roles = group.role_set.filter(name='ad') - contacts.append(contacts_from_roles(roles)) - elif group.type_id == 'wg': - roles = group.role_set.filter(name='chair') - contacts.append(contacts_from_roles(roles)) - elif group.acronym == 'ietf': - contacts.append(EMAIL_ALIASES['IETFCHAIR']) - elif group.acronym == 'iab': - contacts.append(EMAIL_ALIASES['IABCHAIR']) - contacts.append(EMAIL_ALIASES['IABEXECUTIVEDIRECTOR']) - elif group.acronym == 'iesg': - contacts.append(EMAIL_ALIASES['IESG']) - - return ','.join(contacts) - def get_details_tabs(stmt, selected): return [ t + (t[0].lower() == selected.lower(),) @@ -171,7 +100,7 @@ def needs_approval(group,person): user = person.user if group.acronym in ('ietf','iesg') and has_role(user, 'IETF Chair'): return False - if group.acronym == 'iab' and (has_role(user,'IAB Chair') or has_role(user,'IAB Executive Director')): + if group.acronym == 'iab' and has_role(user,'IAB Chair'): return False if group.type_id == 'area' and group.role_set.filter(name='ad',person=person): return False @@ -189,23 +118,14 @@ def normalize_sort(request): return sort, order_by -def post_only(group,person): - '''Returns true if the user is restricted to post_only (vs. post_and_send) for this - group. This is for incoming liaison statements. - - Secretariat have full access. - - Authorized Individuals have full access for the group they are associated with - - Liaison Managers can post only - ''' - if group.type_id == 'sdo' and ( not(has_role(person.user,"Secretariat") or group.role_set.filter(name='auth',person=person)) ): - return True - else: - return False # ------------------------------------------------- # Ajax Functions # ------------------------------------------------- @can_submit_liaison_required def ajax_get_liaison_info(request): + from ietf.mailtrigger.utils import get_contacts_for_liaison_messages_for_group_primary,get_contacts_for_liaison_messages_for_group_secondary + '''Returns dictionary of info to update entry form given the groups that have been selected ''' @@ -222,20 +142,18 @@ def ajax_get_liaison_info(request): cc = [] does_need_approval = [] - can_post_only = [] to_contacts = [] response_contacts = [] - result = {'response_contacts':[],'to_contacts': [], 'cc': [], 'needs_approval': False, 'post_only': False, 'full_list': []} + result = {'response_contacts':[],'to_contacts': [], 'cc': [], 'needs_approval': False, 'full_list': []} for group in from_groups: - cc.extend(get_cc(group)) + cc.extend(get_contacts_for_liaison_messages_for_group_primary(group)) does_need_approval.append(needs_approval(group,person)) - can_post_only.append(post_only(group,person)) - response_contacts.append(get_contacts_for_group(group)) + response_contacts.append(get_contacts_for_liaison_messages_for_group_secondary(group)) for group in to_groups: - cc.extend(get_cc(group)) - to_contacts.append(get_contacts_for_group(group)) + cc.extend(get_contacts_for_liaison_messages_for_group_primary(group)) + to_contacts.append(get_contacts_for_liaison_messages_for_group_secondary(group)) # if there are from_groups and any need approval if does_need_approval: @@ -246,12 +164,15 @@ def ajax_get_liaison_info(request): else: does_need_approval = True - result.update({'error': False, - 'cc': list(set(cc)), - 'response_contacts':list(set(response_contacts)), - 'to_contacts': list(set(to_contacts)), - 'needs_approval': does_need_approval, - 'post_only': any(can_post_only)}) + result.update( + { + "error": False, + "cc": list(set(cc)), + "response_contacts": list(set(response_contacts)), + "to_contacts": list(set(to_contacts)), + "needs_approval": does_need_approval, + } + ) json_result = json.dumps(result) return HttpResponse(json_result, content_type='application/json') @@ -375,23 +296,29 @@ def liaison_history(request, object_id): def liaison_delete_attachment(request, object_id, attach_id): liaison = get_object_or_404(LiaisonStatement, pk=object_id) - attach = get_object_or_404(LiaisonStatementAttachment, pk=attach_id) + if not can_edit_liaison(request.user, liaison): permission_denied(request, "You are not authorized for this action.") - - # FIXME: this view should use POST instead of GET when deleting - attach.removed = True - attach.save() - - # create event - LiaisonStatementEvent.objects.create( - type_id='modified', - by=get_person_for_user(request.user), - statement=liaison, - desc='Attachment Removed: {}'.format(attach.document.title) - ) - messages.success(request, 'Attachment Deleted') - return redirect('ietf.liaisons.views.liaison_detail', object_id=liaison.pk) + else: + permission_denied(request, "This operation is temporarily unavailable. Ask the secretariat to mark the attachment as removed using the admin.") + + # The following will be replaced with a different approach in the next generation of the liaison tool + # attach = get_object_or_404(LiaisonStatementAttachment, pk=attach_id) + + # # FIXME: this view should use POST instead of GET when deleting + # attach.removed = True + # debug.say("Got here") + # attach.save() + + # # create event + # LiaisonStatementEvent.objects.create( + # type_id='modified', + # by=get_person_for_user(request.user), + # statement=liaison, + # desc='Attachment Removed: {}'.format(attach.document.title) + # ) + # messages.success(request, 'Attachment Deleted') + # return redirect('ietf.liaisons.views.liaison_detail', object_id=liaison.pk) def liaison_detail(request, object_id): liaison = get_object_or_404(LiaisonStatement, pk=object_id) @@ -402,22 +329,28 @@ def liaison_detail(request, object_id): if request.method == 'POST': - if request.POST.get('approved'): - liaison.change_state(state_id='approved',person=person) - liaison.change_state(state_id='posted',person=person) - send_liaison_by_email(request, liaison) - messages.success(request,'Liaison Statement Approved and Posted') - elif request.POST.get('dead'): - liaison.change_state(state_id='dead',person=person) - messages.success(request,'Liaison Statement Killed') - elif request.POST.get('resurrect'): - liaison.change_state(state_id='pending',person=person) - messages.success(request,'Liaison Statement Resurrected') - elif request.POST.get('do_action_taken') and can_take_care: + if request.POST.get('do_action_taken') and can_take_care: liaison.tags.remove('required') liaison.tags.add('taken') can_take_care = False messages.success(request,'Action handled') + else: + if can_edit: + if request.POST.get('approved'): + liaison.change_state(state_id='approved',person=person) + liaison.change_state(state_id='posted',person=person) + send_liaison_by_email(request, liaison) + messages.success(request,'Liaison Statement Approved and Posted') + elif request.POST.get('dead'): + liaison.change_state(state_id='dead',person=person) + messages.success(request,'Liaison Statement Killed') + elif request.POST.get('resurrect'): + liaison.change_state(state_id='pending',person=person) + messages.success(request,'Liaison Statement Resurrected') + else: + pass + else: + raise PermissionDenied() relations_by = [i.target for i in liaison.source_of_set.filter(target__state__slug='posted')] relations_to = [i.source for i in liaison.target_of_set.filter(source__state__slug='posted')] @@ -441,7 +374,11 @@ def liaison_edit(request, object_id): def liaison_edit_attachment(request, object_id, doc_id): '''Edit the Liaison Statement attachment title''' liaison = get_object_or_404(LiaisonStatement, pk=object_id) - doc = get_object_or_404(Document, pk=doc_id) + try: + doc = liaison.attachments.get(pk=doc_id) + except ObjectDoesNotExist: + raise Http404 + if not can_edit_liaison(request.user, liaison): permission_denied(request, "You are not authorized for this action.") @@ -572,3 +509,17 @@ def liaison_resend(request, object_id): messages.success(request,'Liaison Statement resent') return redirect('ietf.liaisons.views.liaison_list') + +@role_required("Secretariat", "IAB", "Liaison Coordinator", "Liaison Manager") +def list_other_sdo(request): + def _sdo_order_key(obj:Group)-> tuple[str,str]: + state_order = { + "active" : "a", + "conclude": "b", + } + return (state_order.get(obj.state.slug,f"c{obj.state.slug}"), obj.acronym) + + sdos = sorted(list(Group.objects.filter(type="sdo")),key = _sdo_order_key) + for sdo in sdos: + sdo.liaison_managers =[r.person for r in sdo.role_set.filter(name="liaiman")] + return render(request,"liaisons/list_other_sdo.html",dict(sdos=sdos)) diff --git a/ietf/liaisons/widgets.py b/ietf/liaisons/widgets.py index d6e2fe936b..48db8af0a3 100644 --- a/ietf/liaisons/widgets.py +++ b/ietf/liaisons/widgets.py @@ -3,11 +3,12 @@ from django.urls import reverse as urlreverse -from django.db.models.query import QuerySet from django.forms.widgets import Widget from django.utils.safestring import mark_safe from django.utils.html import conditional_escape +from django_stubs_ext import QuerySetAny + class ButtonWidget(Widget): def __init__(self, *args, **kwargs): @@ -25,7 +26,9 @@ def render(self, name, value, **kwargs): html += '%s' % conditional_escape(i) required_str = 'Please fill in %s to attach a new file' % conditional_escape(self.required_label) html += '%s' % conditional_escape(required_str) - html += '' % conditional_escape(self.label) + html += ''.format( + f"id_{name}", conditional_escape(self.label) + ) return mark_safe(html) @@ -34,7 +37,7 @@ def render(self, name, value, **kwargs): html = '
' % name html += 'No files attached' html += '
' - if value and isinstance(value, QuerySet): + if value and isinstance(value, QuerySetAny): for attachment in value: html += '%s ' % (conditional_escape(attachment.document.get_href()), conditional_escape(attachment.document.title)) html += 'Edit '.format(urlreverse("ietf.liaisons.views.liaison_edit_attachment", kwargs={'object_id':attachment.statement.pk,'doc_id':attachment.document.pk})) @@ -43,4 +46,4 @@ def render(self, name, value, **kwargs): else: html += 'No files attached' html += '
' - return mark_safe(html) \ No newline at end of file + return mark_safe(html) diff --git a/ietf/mailinglists/.gitignore b/ietf/mailinglists/.gitignore deleted file mode 100644 index c7013ced9d..0000000000 --- a/ietf/mailinglists/.gitignore +++ /dev/null @@ -1,2 +0,0 @@ -/*.pyc -/settings_local.py diff --git a/ietf/mailinglists/admin.py b/ietf/mailinglists/admin.py index 90efaf9c93..081ee6477c 100644 --- a/ietf/mailinglists/admin.py +++ b/ietf/mailinglists/admin.py @@ -2,20 +2,15 @@ from django.contrib import admin -from ietf.mailinglists.models import List, Subscribed, Allowlisted +from ietf.mailinglists.models import NonWgMailingList, Allowlisted -class ListAdmin(admin.ModelAdmin): - list_display = ('id', 'name', 'description', 'advertised') - search_fields = ('name',) -admin.site.register(List, ListAdmin) -class SubscribedAdmin(admin.ModelAdmin): - list_display = ('id', 'time', 'email') - raw_id_fields = ('lists',) - search_fields = ('email',) -admin.site.register(Subscribed, SubscribedAdmin) +class NonWgMailingListAdmin(admin.ModelAdmin): + list_display = ('id', 'name', 'domain', 'description') + search_fields = ('name', 'domain') +admin.site.register(NonWgMailingList, NonWgMailingListAdmin) class AllowlistedAdmin(admin.ModelAdmin): diff --git a/ietf/mailinglists/factories.py b/ietf/mailinglists/factories.py index bc6b2b8203..3be5770d76 100644 --- a/ietf/mailinglists/factories.py +++ b/ietf/mailinglists/factories.py @@ -3,16 +3,15 @@ import factory -import random -from ietf.mailinglists.models import List +from ietf.mailinglists.models import NonWgMailingList -class ListFactory(factory.django.DjangoModelFactory): +class NonWgMailingListFactory(factory.django.DjangoModelFactory): class Meta: - model = List + model = NonWgMailingList name = factory.Sequence(lambda n: "list-name-%s" % n) + domain = factory.Sequence(lambda n: "domain-%s.org" % n) description = factory.Faker('sentence', nb_words=10) - advertised = factory.LazyAttribute(lambda obj: random.randint(0, 1)) diff --git a/ietf/mailinglists/management/commands/import_mailman_listinfo.py b/ietf/mailinglists/management/commands/import_mailman_listinfo.py deleted file mode 100644 index 43982f58b1..0000000000 --- a/ietf/mailinglists/management/commands/import_mailman_listinfo.py +++ /dev/null @@ -1,130 +0,0 @@ -# Copyright The IETF Trust 2016-2019, All Rights Reserved - -import json -import sys -import subprocess -import time -from textwrap import dedent - -import debug # pyflakes:ignore - -from pathlib import Path - -from django.conf import settings -from django.core.management.base import BaseCommand -from django.core.exceptions import MultipleObjectsReturned - - -from ietf.mailinglists.models import List, Subscribed -from ietf.utils.log import log - -mark = time.time() - -def import_mailman_listinfo(verbosity=0): - def note(msg): - if verbosity > 2: - sys.stdout.write(msg) - sys.stdout.write('\n') - def log_time(msg): - global mark - if verbosity > 1: - t = time.time() - log(msg+' (%.1fs)'% (t-mark)) - mark = t - - cmd = str(Path(settings.BASE_DIR) / "bin" / "mailman_listinfo.py") - result = subprocess.run([cmd], capture_output=True) - if result.stderr: - log("Error exporting information from mailmain") - log(result.stderr) - return - mailman_export = json.loads(result.stdout) - - names = sorted(mailman_export.keys()) - addr_max_length = Subscribed._meta.get_field('email').max_length - - subscribed = { l.name: set(l.subscribed_set.values_list('email', flat=True)) for l in List.objects.all().prefetch_related('subscribed_set') } - - for name in names: - note("List: %s" % mailman_export[name]['internal_name']) - - lists = List.objects.filter(name=mailman_export[name]['real_name']) - if lists.count() > 1: - # Arbitrary choice; we'll update the remaining item next - for item in lists[1:]: - item.delete() - mmlist, created = List.objects.get_or_create(name=mailman_export[name]['real_name']) - dirty = False - desc = mailman_export[name]['description'][:256] - if mmlist.description != desc: - mmlist.description = desc - dirty = True - if mmlist.advertised != mailman_export[name]['advertised']: - mmlist.advertised = mailman_export[name]['advertised'] - dirty = True - if dirty: - mmlist.save() - # The following calls return lowercased addresses - if mailman_export[name]['advertised']: - members = set(mailman_export[name]['members']) - if not mailman_export[name]['real_name'] in subscribed: - # 2022-7-29: lots of these going into the logs but being ignored... - # log("Note: didn't find '%s' in the dictionary of subscriptions" % mailman_export[name]['real_name']) - continue - known = subscribed[mailman_export[name]['real_name']] - log_time(" Fetched known list members from database") - to_remove = known - members - to_add = members - known - for addr in to_remove: - note(" Removing subscription: %s" % (addr)) - old = Subscribed.objects.get(email=addr) - old.lists.remove(mmlist) - if old.lists.count() == 0: - note(" Removing address with no subscriptions: %s" % (addr)) - old.delete() - if to_remove: - log(" Removed %s addresses from %s" % (len(to_remove), name)) - for addr in to_add: - if len(addr) > addr_max_length: - sys.stderr.write(" ** Email address subscribed to '%s' too long for table: <%s>\n" % (name, addr)) - continue - note(" Adding subscription: %s" % (addr)) - try: - new, created = Subscribed.objects.get_or_create(email=addr) - except MultipleObjectsReturned as e: - sys.stderr.write(" ** Error handling %s in %s: %s\n" % (addr, name, e)) - continue - new.lists.add(mmlist) - if to_add: - log(" Added %s addresses to %s" % (len(to_add), name)) - log("Completed import of list info from Mailman") - -class Command(BaseCommand): - """ - Import list information from Mailman. - - Import announced list names, descriptions, and subscribers, by calling the - appropriate Mailman functions and adding entries to the database. - - Run this from cron regularly, with sufficient permissions to access the - mailman database files. - - """ - - help = dedent(__doc__).strip() - - #option_list = BaseCommand.option_list + ( ) - - - def handle(self, *filenames, **options): - """ - - * Import announced lists, with appropriate meta-information. - - * For each list, import the members. - - """ - - verbosity = int(options.get('verbosity')) - - import_mailman_listinfo(verbosity) diff --git a/ietf/mailinglists/migrations/0001_initial.py b/ietf/mailinglists/migrations/0001_initial.py index 52c04861c9..eaf5d24e31 100644 --- a/ietf/mailinglists/migrations/0001_initial.py +++ b/ietf/mailinglists/migrations/0001_initial.py @@ -1,7 +1,4 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-20 10:52 - +# Generated by Django 2.2.28 on 2023-03-20 19:22 import django.core.validators from django.db import migrations, models @@ -32,7 +29,7 @@ class Migration(migrations.Migration): fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('time', models.DateTimeField(auto_now_add=True)), - ('email', models.CharField(max_length=64, validators=[django.core.validators.EmailValidator()])), + ('email', models.CharField(max_length=128, validators=[django.core.validators.EmailValidator()])), ('lists', models.ManyToManyField(to='mailinglists.List')), ], options={ @@ -40,7 +37,7 @@ class Migration(migrations.Migration): }, ), migrations.CreateModel( - name='Whitelisted', + name='Allowlisted', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('time', models.DateTimeField(auto_now_add=True)), @@ -48,7 +45,7 @@ class Migration(migrations.Migration): ('by', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), ], options={ - 'verbose_name_plural': 'Whitelisted', + 'verbose_name_plural': 'Allowlisted', }, ), ] diff --git a/ietf/mailinglists/migrations/0002_auto_20190703_1344.py b/ietf/mailinglists/migrations/0002_auto_20190703_1344.py deleted file mode 100644 index b55c482a28..0000000000 --- a/ietf/mailinglists/migrations/0002_auto_20190703_1344.py +++ /dev/null @@ -1,22 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# Generated by Django 1.11.22 on 2019-07-03 13:44 - - -import django.core.validators -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('mailinglists', '0001_initial'), - ] - - operations = [ - migrations.AlterField( - model_name='subscribed', - name='email', - field=models.CharField(max_length=128, validators=[django.core.validators.EmailValidator()]), - ), - ] diff --git a/ietf/mailinglists/migrations/0002_nonwgmailinglist.py b/ietf/mailinglists/migrations/0002_nonwgmailinglist.py new file mode 100644 index 0000000000..dfc941db90 --- /dev/null +++ b/ietf/mailinglists/migrations/0002_nonwgmailinglist.py @@ -0,0 +1,628 @@ +# Copyright The IETF Trust 2024, All Rights Reserved + +from django.db import migrations, models + + +def forward(apps, schema_editor): + NonWgMailingList = apps.get_model("mailinglists", "NonWgMailingList") + List = apps.get_model("mailinglists", "List") + + for l in List.objects.filter( + pk__in=[ + 10754, + 10769, + 10770, + 10768, + 10787, + 10785, + 10791, + 10786, + 10816, + 10817, + 10819, + 10818, + 10922, + 10923, + 10921, + 10940, + 10941, + 10942, + 572, + 10297, + 182, + 43, + 10704, + 10314, + 201, + 419, + 282, + 149, + 223, + 10874, + 10598, + 10639, + 10875, + 10737, + 105, + 65, + 10781, + 10771, + 10946, + 518, + 421, + 214, + 285, + 393, + 445, + 553, + 183, + 10725, + 33, + 10766, + 114, + 417, + 10789, + 10876, + 4244, + 10705, + 10706, + 10878, + 10324, + 10879, + 10642, + 10821, + 547, + 532, + 10636, + 10592, + 327, + 248, + 10697, + 288, + 346, + 10731, + 10955, + 10857, + 446, + 55, + 10799, + 10800, + 10801, + 10612, + 73, + 3, + 358, + 9640, + 10868, + 378, + 462, + 6595, + 10914, + 10915, + 197, + 63, + 558, + 10824, + 124, + 10881, + 177, + 312, + 252, + 185, + 523, + 4572, + 10618, + 206, + 68, + 10859, + 560, + 513, + 246, + 7817, + 148, + 10864, + 10589, + 10773, + 10748, + 364, + 311, + 10302, + 10272, + 10929, + 171, + 10865, + 10919, + 377, + 469, + 467, + 411, + 505, + 6318, + 10811, + 10304, + 10882, + 10845, + 568, + 10883, + 4774, + 264, + 10779, + 10884, + 10303, + 409, + 10590, + 451, + 10749, + 10765, + 486, + 519, + 10593, + 10313, + 550, + 10707, + 307, + 10861, + 10654, + 10708, + 10275, + 134, + 460, + 10911, + 10574, + 10885, + 10814, + 10676, + 10747, + 10305, + 10688, + 36, + 10844, + 10620, + 458, + 10282, + 10594, + 10752, + 389, + 296, + 10684, + 48, + 533, + 443, + 10739, + 491, + 139, + 461, + 10690, + 424, + 290, + 336, + 31, + 10709, + 382, + 10866, + 10724, + 539, + 10710, + 559, + 10609, + 74, + 10582, + 133, + 10621, + 34, + 10596, + 442, + 13, + 56, + 128, + 323, + 10285, + 80, + 315, + 3520, + 10949, + 10950, + 189, + 2599, + 10822, + 164, + 10267, + 10286, + 464, + 440, + 254, + 262, + 10943, + 465, + 75, + 179, + 162, + 457, + 10572, + 372, + 452, + 10273, + 88, + 366, + 331, + 140, + 407, + 416, + 91, + 10632, + 542, + 151, + 117, + 431, + 10628, + 10271, + 14, + 540, + 278, + 352, + 159, + 10851, + 9981, + 10694, + 10619, + 10732, + 320, + 348, + 338, + 349, + 10678, + 468, + 293, + 350, + 402, + 57, + 524, + 141, + 71, + 67, + 508, + 7828, + 10268, + 10631, + 10713, + 10889, + 345, + 78, + 342, + 190, + 10869, + 46, + 334, + 255, + 5823, + 400, + 10867, + 23, + 10666, + 10685, + 405, + 2801, + 92, + 137, + 10640, + 10656, + 104, + 123, + 10643, + 10891, + 466, + 10567, + 10318, + 526, + 30, + 222, + 194, + 10735, + 10714, + 247, + 493, + 1162, + 414, + 10648, + 10677, + 126, + 16, + 422, + 271, + 295, + 81, + 10634, + 544, + 10850, + 426, + 573, + 353, + 10829, + 538, + 10913, + 10566, + 167, + 10675, + 272, + 10673, + 10767, + 528, + 284, + 564, + 268, + 10825, + 231, + 520, + 10645, + 10872, + 515, + 10956, + 10947, + 569, + 233, + 10952, + 195, + 10938, + 2809, + 10591, + 10665, + 9639, + 10775, + 10760, + 10715, + 10716, + 10667, + 361, + 184, + 10935, + 10957, + 10944, + 94, + 449, + 525, + 1962, + 10300, + 10894, + 9156, + 10774, + 256, + 289, + 218, + 187, + 40, + 10777, + 10761, + 10670, + 249, + 10764, + 420, + 548, + 232, + 410, + 196, + 72, + 335, + 70, + 146, + 10287, + 10299, + 10311, + 10895, + 10617, + 531, + 343, + 10934, + 10933, + 10597, + 158, + 10600, + 10692, + 8630, + 556, + 324, + 11, + 10784, + 498, + 10772, + 478, + 10833, + 10691, + 391, + 10565, + 10669, + 113, + 110, + 7831, + 10855, + 10312, + 10315, + 10896, + 10672, + 10306, + 438, + 395, + 82, + 10599, + 10953, + 10858, + 10807, + 10717, + 310, + 10808, + 119, + 10595, + 10718, + 10317, + 10898, + 454, + 427, + 10583, + 10916, + 403, + 10843, + 10899, + 291, + 10812, + 10900, + 10794, + 341, + 121, + 230, + 136, + 166, + 394, + 234, + 10901, + 2466, + 10573, + 10939, + 221, + 490, + 10820, + 10873, + 10792, + 10870, + 10793, + 10904, + 181, + 10693, + 482, + 10611, + 125, + 10568, + 10788, + 211, + 10756, + 10719, + 100, + 228, + 5833, + 251, + 122, + 39, + 534, + 437, + 504, + 10613, + 439, + 306, + 10863, + 10823, + 10926, + 76, + 227, + 59, + 42, + 455, + 10927, + 10928, + 204, + 430, + 10720, + 267, + 396, + 10849, + 10308, + 281, + 10905, + 10736, + 168, + 153, + 385, + 89, + 529, + 412, + 215, + 484, + 10951, + 66, + 173, + 10633, + 10681, + 3613, + 10274, + 10750, + 367, + 387, + 10832, + 35, + 147, + 10325, + 10671, + 565, + 313, + 10871, + 10751, + 37, + 10936, + 10937, + 287, + 496, + 244, + 10841, + 10683, + 10906, + 10584, + 479, + 10856, + 163, + 10910, + 257, + 276, + 10840, + 10689, + 365, + 10847, + 99, + 77, + 435, + 213, + 15, + 10932, + 58, + 10722, + 131, + 363, + 10674, + 322, + 180, + 10917, + 10918, + 10738, + 10954, + 10581, + 208, + 337, + 4, + 571, + 10668, + 10291, + ] + ): + NonWgMailingList.objects.create(name=l.name, description=l.description) + +class Migration(migrations.Migration): + + dependencies = [ + ("mailinglists", "0001_initial"), + ] + + operations = [ + migrations.CreateModel( + name="NonWgMailingList", + fields=[ + ( + "id", + models.AutoField( + auto_created=True, + primary_key=True, + serialize=False, + verbose_name="ID", + ), + ), + ("name", models.CharField(max_length=32)), + ("description", models.CharField(max_length=256)), + ], + ), + migrations.RunPython(forward), + ] diff --git a/ietf/mailinglists/migrations/0003_allowlisted.py b/ietf/mailinglists/migrations/0003_allowlisted.py deleted file mode 100644 index a3f098d9ca..0000000000 --- a/ietf/mailinglists/migrations/0003_allowlisted.py +++ /dev/null @@ -1,22 +0,0 @@ -# Generated by Django 2.2.28 on 2022-12-05 14:26 - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0029_use_timezone_now_for_person_models'), - ('mailinglists', '0002_auto_20190703_1344'), - ] - - operations = [ - migrations.RenameModel( - old_name='Whitelisted', - new_name='Allowlisted', - ), - migrations.AlterModelOptions( - name='allowlisted', - options={'verbose_name_plural': 'Allowlisted'}, - ), - ] diff --git a/ietf/mailinglists/migrations/0003_remove_subscribed_lists_delete_list_and_more.py b/ietf/mailinglists/migrations/0003_remove_subscribed_lists_delete_list_and_more.py new file mode 100644 index 0000000000..6171136b2a --- /dev/null +++ b/ietf/mailinglists/migrations/0003_remove_subscribed_lists_delete_list_and_more.py @@ -0,0 +1,23 @@ +# Generated by Django 4.2.9 on 2024-02-02 23:04 + +from django.db import migrations + + +class Migration(migrations.Migration): + + dependencies = [ + ("mailinglists", "0002_nonwgmailinglist"), + ] + + operations = [ + migrations.RemoveField( + model_name="subscribed", + name="lists", + ), + migrations.DeleteModel( + name="List", + ), + migrations.DeleteModel( + name="Subscribed", + ), + ] diff --git a/ietf/mailinglists/migrations/0004_nonwgmailinglist_domain.py b/ietf/mailinglists/migrations/0004_nonwgmailinglist_domain.py new file mode 100644 index 0000000000..b977313a87 --- /dev/null +++ b/ietf/mailinglists/migrations/0004_nonwgmailinglist_domain.py @@ -0,0 +1,59 @@ +# Generated by Django 4.2.13 on 2024-06-05 17:51 + +from django.db import migrations, models +from django.db.models.functions import Lower + +IAB_NAMES = ["iab", "iab-stream"] +RFCED_NAMES = [ + "auth48archive", + "rfc-dist", + "rfc-editor-rfi", + "rfc-interest", + "rpat", + "rsab", +] +IRTF_NAMES = [ + "anrp-select", + "anrw-sc", + "anrw-tpc", + "crypto-panel", + "dtn-interest", + "irsg", + "irtf-announce", + "smart", + "teaching", + "travel-grants-commitee", +] + + +def forward(apps, schema_editor): + NonWgMailingList = apps.get_model("mailinglists", "NonWgMailingList") + NonWgMailingList.objects.annotate(lowername=Lower("name")).filter( + lowername__in=IAB_NAMES + ).update(domain="iab.org") + NonWgMailingList.objects.annotate(lowername=Lower("name")).filter( + lowername__in=IRTF_NAMES + ).update(domain="irtf.org") + NonWgMailingList.objects.annotate(lowername=Lower("name")).filter( + lowername__in=RFCED_NAMES + ).update(domain="rfc-editor.org") + + +def reverse(apps, schema_editor): + pass + + +class Migration(migrations.Migration): + + dependencies = [ + ("mailinglists", "0003_remove_subscribed_lists_delete_list_and_more"), + ] + + operations = [ + migrations.AddField( + model_name="nonwgmailinglist", + name="domain", + field=models.CharField(default="ietf.org", max_length=32), + ), + migrations.RunPython(forward, reverse), + ] diff --git a/ietf/mailinglists/models.py b/ietf/mailinglists/models.py index 21f3a76710..828d3823a4 100644 --- a/ietf/mailinglists/models.py +++ b/ietf/mailinglists/models.py @@ -9,25 +9,21 @@ from ietf.person.models import Person from ietf.utils.models import ForeignKey -class List(models.Model): + +# NonWgMailingList is a temporary bridging class to hold information known about mailman2 +# while decoupling from mailman2 until we integrate with mailman3 +class NonWgMailingList(models.Model): name = models.CharField(max_length=32) + domain = models.CharField(max_length=32, default="ietf.org") description = models.CharField(max_length=256) - advertised = models.BooleanField(default=True) def __str__(self): - return "" % self.name + return "" % self.name def info_url(self): - return settings.MAILING_LIST_INFO_URL % {'list_addr': self.name } - -class Subscribed(models.Model): - time = models.DateTimeField(auto_now_add=True) - email = models.CharField(max_length=128, validators=[validate_email]) - lists = models.ManyToManyField(List) - def __str__(self): - return "" % (self.email, self.time) - class Meta: - verbose_name_plural = "Subscribed" + return settings.MAILING_LIST_INFO_URL % {'list_addr': self.name.lower(), 'domain': self.domain.lower() } +# Allowlisted is unused, but is not being dropped until its human-curated content +# is archived outside this database. class Allowlisted(models.Model): time = models.DateTimeField(auto_now_add=True) email = models.CharField("Email address", max_length=64, validators=[validate_email]) diff --git a/ietf/mailinglists/resources.py b/ietf/mailinglists/resources.py index 018a8327b1..4d1713b7b6 100644 --- a/ietf/mailinglists/resources.py +++ b/ietf/mailinglists/resources.py @@ -11,7 +11,7 @@ from ietf import api from ietf.api import ToOneField # pyflakes:ignore -from ietf.mailinglists.models import Allowlisted, List, Subscribed +from ietf.mailinglists.models import Allowlisted, NonWgMailingList from ietf.person.resources import PersonResource @@ -31,34 +31,20 @@ class Meta: } api.mailinglists.register(AllowlistedResource()) -class ListResource(ModelResource): +class NonWgMailingListResource(ModelResource): class Meta: - queryset = List.objects.all() + queryset = NonWgMailingList.objects.all() serializer = api.Serializer() cache = SimpleCache() - #resource_name = 'list' + #resource_name = 'nonwgmailinglist' ordering = ['id', ] filtering = { "id": ALL, "name": ALL, + "domain": ALL, "description": ALL, - "advertised": ALL, } -api.mailinglists.register(ListResource()) +api.mailinglists.register(NonWgMailingListResource()) + -class SubscribedResource(ModelResource): - lists = ToManyField(ListResource, 'lists', null=True) - class Meta: - queryset = Subscribed.objects.all() - serializer = api.Serializer() - cache = SimpleCache() - #resource_name = 'subscribed' - ordering = ['id', ] - filtering = { - "id": ALL, - "time": ALL, - "email": ALL, - "lists": ALL_WITH_RELATIONS, - } -api.mailinglists.register(SubscribedResource()) diff --git a/ietf/mailinglists/tests.py b/ietf/mailinglists/tests.py index 0c983da80c..8c5a550dfc 100644 --- a/ietf/mailinglists/tests.py +++ b/ietf/mailinglists/tests.py @@ -9,7 +9,7 @@ import debug # pyflakes:ignore from ietf.group.factories import GroupFactory -from ietf.mailinglists.factories import ListFactory +from ietf.mailinglists.factories import NonWgMailingListFactory from ietf.utils.test_utils import TestCase @@ -32,23 +32,15 @@ def test_groups(self): def test_nonwg(self): - groups = list() - groups.append(GroupFactory(type_id='wg', acronym='mars', list_archive='https://ietf.org/mars')) - groups.append(GroupFactory(type_id='wg', acronym='ames', state_id='conclude', list_archive='https://ietf.org/ames')) - groups.append(GroupFactory(type_id='wg', acronym='newstuff', state_id='bof', list_archive='https://ietf.org/newstuff')) - groups.append(GroupFactory(type_id='rg', acronym='research', list_archive='https://irtf.org/research')) - lists = ListFactory.create_batch(7) + + lists = NonWgMailingListFactory.create_batch(7) url = urlreverse("ietf.mailinglists.views.nonwg") r = self.client.get(url) + q = PyQuery(r.content) for l in lists: - if l.advertised: self.assertContains(r, l.name) self.assertContains(r, l.description) - else: - self.assertNotContains(r, l.name, html=True) - self.assertNotContains(r, l.description, html=True) + self.assertNotEqual(q(f"a[href=\"{l.info_url()}\"]"), []) - for g in groups: - self.assertNotContains(r, g.acronym, html=True) diff --git a/ietf/mailinglists/views.py b/ietf/mailinglists/views.py index 51c31c546f..460f30e164 100644 --- a/ietf/mailinglists/views.py +++ b/ietf/mailinglists/views.py @@ -1,33 +1,25 @@ # Copyright The IETF Trust 2007-2022, All Rights Reserved -import re - from django.shortcuts import render -import debug # pyflakes:ignore +import debug # pyflakes:ignore from ietf.group.models import Group -from ietf.mailinglists.models import List +from ietf.mailinglists.models import NonWgMailingList + def groups(request): - groups = Group.objects.filter(type__features__acts_like_wg=True, list_archive__startswith='http').exclude(state__in=('bof', 'conclude')).order_by("acronym") + groups = ( + Group.objects.filter( + type__features__acts_like_wg=True, list_archive__startswith="http" + ) + .exclude(state__in=("bof", "conclude")) + .order_by("acronym") + ) + + return render(request, "mailinglists/group_archives.html", {"groups": groups}) - return render(request, "mailinglists/group_archives.html", { "groups": groups } ) def nonwg(request): - groups = Group.objects.filter(type__features__acts_like_wg=True).exclude(state__in=['bof']).order_by("acronym") - - #urls = [ g.list_archive for g in groups if '.ietf.org' in g.list_archive ] - - wg_lists = set() - for g in groups: - wg_lists.add(g.acronym) - match = re.search(r'^(https?://mailarchive.ietf.org/arch/(browse/|search/\?email-list=))(?P[^/]*)/?$', g.list_archive) - if match: - wg_lists.add(match.group('name').lower()) - - lists = List.objects.filter(advertised=True) - #debug.show('lists.count()') - lists = lists.exclude(name__in=wg_lists).order_by('name') - #debug.show('lists.count()') - return render(request, "mailinglists/nonwg.html", { "lists": lists } ) + lists = NonWgMailingList.objects.order_by("name") + return render(request, "mailinglists/nonwg.html", {"lists": lists}) diff --git a/ietf/mailtrigger/admin.py b/ietf/mailtrigger/admin.py index a60fd5b072..8c73f2ae02 100644 --- a/ietf/mailtrigger/admin.py +++ b/ietf/mailtrigger/admin.py @@ -1,9 +1,10 @@ -# Copyright The IETF Trust 2015-2019, All Rights Reserved +# Copyright The IETF Trust 2015-2025, All Rights Reserved from django.contrib import admin +from simple_history.admin import SimpleHistoryAdmin from ietf.mailtrigger.models import MailTrigger, Recipient -class RecipientAdmin(admin.ModelAdmin): +class RecipientAdmin(SimpleHistoryAdmin): list_display = [ 'slug', 'desc', 'template', 'has_code', ] def has_code(self, obj): return hasattr(obj,'gather_%s'%obj.slug) @@ -11,7 +12,7 @@ def has_code(self, obj): admin.site.register(Recipient, RecipientAdmin) -class MailTriggerAdmin(admin.ModelAdmin): +class MailTriggerAdmin(SimpleHistoryAdmin): list_display = [ 'slug', 'desc', ] filter_horizontal = [ 'to', 'cc', ] admin.site.register(MailTrigger, MailTriggerAdmin) diff --git a/ietf/mailtrigger/forms.py b/ietf/mailtrigger/forms.py index 366c429d8c..8d13c5edf3 100644 --- a/ietf/mailtrigger/forms.py +++ b/ietf/mailtrigger/forms.py @@ -11,6 +11,7 @@ class CcSelectForm(forms.Form): expansions = dict() # type: Dict[str, List[str]] cc_choices = forms.MultipleChoiceField( + required=False, label='Cc', choices=[], widget=forms.CheckboxSelectMultiple(), diff --git a/ietf/mailtrigger/migrations/0001_initial.py b/ietf/mailtrigger/migrations/0001_initial.py index 37de9a6466..c66a908aac 100644 --- a/ietf/mailtrigger/migrations/0001_initial.py +++ b/ietf/mailtrigger/migrations/0001_initial.py @@ -1,10 +1,6 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-20 10:52 - - -from typing import List, Tuple # pyflakes:ignore +# Generated by Django 2.2.28 on 2023-03-20 19:22 +from typing import List, Tuple from django.db import migrations, models @@ -12,39 +8,31 @@ class Migration(migrations.Migration): initial = True - dependencies = [ - ] # type: List[Tuple[str]] + dependencies: List[Tuple[str, str]] = [ + ] operations = [ migrations.CreateModel( - name='MailTrigger', + name='Recipient', fields=[ ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), ('desc', models.TextField(blank=True)), + ('template', models.TextField(blank=True, null=True)), ], options={ 'ordering': ['slug'], }, ), migrations.CreateModel( - name='Recipient', + name='MailTrigger', fields=[ - ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), + ('slug', models.CharField(max_length=64, primary_key=True, serialize=False)), ('desc', models.TextField(blank=True)), - ('template', models.TextField(blank=True, null=True)), + ('cc', models.ManyToManyField(blank=True, related_name='used_in_cc', to='mailtrigger.Recipient')), + ('to', models.ManyToManyField(blank=True, related_name='used_in_to', to='mailtrigger.Recipient')), ], options={ 'ordering': ['slug'], }, ), - migrations.AddField( - model_name='mailtrigger', - name='cc', - field=models.ManyToManyField(blank=True, related_name='used_in_cc', to='mailtrigger.Recipient'), - ), - migrations.AddField( - model_name='mailtrigger', - name='to', - field=models.ManyToManyField(blank=True, related_name='used_in_to', to='mailtrigger.Recipient'), - ), ] diff --git a/ietf/mailtrigger/migrations/0002_conflrev_changes.py b/ietf/mailtrigger/migrations/0002_conflrev_changes.py deleted file mode 100644 index 858d3a11d6..0000000000 --- a/ietf/mailtrigger/migrations/0002_conflrev_changes.py +++ /dev/null @@ -1,40 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.13 on 2018-05-21 12:07 - - -from django.db import migrations - -def forward(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger','MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - conflrev_ad_changed = MailTrigger.objects.create( - slug = 'conflrev_ad_changed', - desc = 'Recipients when the responsible AD for a conflict review is changed', - ) - conflrev_ad_changed.to.set(Recipient.objects.filter(slug='iesg-secretary')) - conflrev_ad_changed.cc.set(Recipient.objects.filter(slug__in=[ - 'conflict_review_steering_group', - 'conflict_review_stream_manager', - 'doc_affecteddoc_authors', - 'doc_affecteddoc_group_chairs', - 'doc_affecteddoc_notify', - 'doc_notify', - 'iesg', - ])) - - -def reverse(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger','MailTrigger') - MailTrigger.objects.filter(slug='conflrev_ad_changed').delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0001_initial'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/mailtrigger/migrations/0002_slidesubmitter.py b/ietf/mailtrigger/migrations/0002_slidesubmitter.py new file mode 100644 index 0000000000..394c7d92ce --- /dev/null +++ b/ietf/mailtrigger/migrations/0002_slidesubmitter.py @@ -0,0 +1,31 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations + +def forward(apps, schema_editor): + MailTrigger = apps.get_model("mailtrigger", "MailTrigger") + Recipient = apps.get_model("mailtrigger", "Recipient") + r = Recipient.objects.create( + slug="slides_proposer", + desc="Person who proposed slides", + template="{{ proposer.email }}" + ) + mt = MailTrigger.objects.get(slug="slides_proposed") + mt.cc.add(r) + +def reverse(apps, schema_editor): + MailTrigger = apps.get_model("mailtrigger", "MailTrigger") + Recipient = apps.get_model("mailtrigger", "Recipient") + mt = MailTrigger.objects.get(slug="slides_proposed") + r = Recipient.objects.get(slug="slides_proposer") + mt.cc.remove(r) + r.delete() + +class Migration(migrations.Migration): + dependencies = [ + ("mailtrigger", "0001_initial"), + ] + + operations = [ + migrations.RunPython(forward, reverse) + ] diff --git a/ietf/mailtrigger/migrations/0003_add_review_notify_ad.py b/ietf/mailtrigger/migrations/0003_add_review_notify_ad.py deleted file mode 100644 index 5abf175387..0000000000 --- a/ietf/mailtrigger/migrations/0003_add_review_notify_ad.py +++ /dev/null @@ -1,46 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.16 on 2018-11-02 11:34 - - -from django.db import migrations - -def forward(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger','MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - Recipient.objects.create( - slug = 'review_doc_ad', - desc = "The reviewed document's responsible area director", - template = '{% if review_req.doc.ad %}{{review_req.doc.ad.email_address}}{% endif %}' - ) - Recipient.objects.create( - slug = 'review_team_ads', - desc = "The ADs of the team reviewing the document" - ) - - review_notify_ad = MailTrigger.objects.create( - slug = 'review_notify_ad', - desc = 'Recipients when a team notifies area directors when a review with one of a certain set of results (typically results indicating problem) is submitted', - ) - review_notify_ad.to.set(Recipient.objects.filter(slug__in=['review_doc_ad','review_team_ads'])) - - -def reverse(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger','MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - MailTrigger.objects.filter(slug='review_notify_ad').delete() - Recipient.objects.filter(slug='review_doc_ad').delete() - Recipient.objects.filter(slug='review_team_ads').delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0002_conflrev_changes'), - ('review', '0004_reviewteamsettings_secr_mail_alias'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/mailtrigger/migrations/0003_ballot_approved_charter.py b/ietf/mailtrigger/migrations/0003_ballot_approved_charter.py new file mode 100644 index 0000000000..ef2e4dc444 --- /dev/null +++ b/ietf/mailtrigger/migrations/0003_ballot_approved_charter.py @@ -0,0 +1,26 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations + +def forward(apps, schema_editor): + MailTrigger = apps.get_model("mailtrigger", "MailTrigger") + Recipient = apps.get_model("mailtrigger", "Recipient") + mt = MailTrigger.objects.get(pk="ballot_approved_charter") + mt.to.remove(mt.to.first()) + mt.to.add(Recipient.objects.get(slug="group_stream_announce")) + +def reverse(apps, schema_editor): + MailTrigger = apps.get_model("mailtrigger", "MailTrigger") + Recipient = apps.get_model("mailtrigger", "Recipient") + mt = MailTrigger.objects.get(pk="ballot_approved_charter") + mt.to.remove(mt.to.first()) + mt.to.add(Recipient.objects.get(slug="ietf_announce")) + +class Migration(migrations.Migration): + dependencies = [ + ("mailtrigger", "0002_slidesubmitter"), + ] + + operations = [ + migrations.RunPython(forward, reverse) + ] diff --git a/ietf/mailtrigger/migrations/0004_ballot_rfceditornote_changed_postapproval.py b/ietf/mailtrigger/migrations/0004_ballot_rfceditornote_changed_postapproval.py deleted file mode 100644 index 50fb71a513..0000000000 --- a/ietf/mailtrigger/migrations/0004_ballot_rfceditornote_changed_postapproval.py +++ /dev/null @@ -1,30 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.16 on 2018-11-03 00:24 - - -from django.db import migrations - -def forward(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - changed = MailTrigger.objects.create( - slug = 'ballot_ednote_changed_late', - desc = 'Recipients when the RFC Editor note for a document is changed after the document has been approved', - ) - changed.to.set(Recipient.objects.filter(slug__in=['rfc_editor','iesg'])) - -def reverse(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger','MailTrigger') - MailTrigger.objects.filter(slug='ballot_ednote_changed_late').delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0003_add_review_notify_ad'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/mailtrigger/migrations/0004_slides_approved.py b/ietf/mailtrigger/migrations/0004_slides_approved.py new file mode 100644 index 0000000000..6376e80021 --- /dev/null +++ b/ietf/mailtrigger/migrations/0004_slides_approved.py @@ -0,0 +1,27 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations + +def forward(apps, schema_editor): + MailTrigger = apps.get_model("mailtrigger", "MailTrigger") + Recipient = apps.get_model("mailtrigger", "Recipient") + mt = MailTrigger.objects.create( + slug="slides_approved", + desc="Recipients when slides are approved for a given session", + ) + mt.to.add(Recipient.objects.get(slug="slides_proposer")) + mt.cc.add(Recipient.objects.get(slug="group_chairs")) + +def reverse(apps, schema_editor): + MailTrigger = apps.get_model("mailtrigger", "MailTrigger") + mt = MailTrigger.objects.get(pk="slides_approved") + mt.delete() + +class Migration(migrations.Migration): + dependencies = [ + ("mailtrigger", "0003_ballot_approved_charter"), + ] + + operations = [ + migrations.RunPython(forward, reverse) + ] diff --git a/ietf/mailtrigger/migrations/0005_rfc_recipients.py b/ietf/mailtrigger/migrations/0005_rfc_recipients.py new file mode 100644 index 0000000000..dee49d9133 --- /dev/null +++ b/ietf/mailtrigger/migrations/0005_rfc_recipients.py @@ -0,0 +1,25 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations + + +def forward(apps, schema_editor): + Recipient = apps.get_model("mailtrigger", "Recipient") + Recipient.objects.filter(slug="doc_authors").update( + template='{% if doc.type_id == "draft" or doc.type_id == "rfc" %}<{{doc.name}}@ietf.org>{% endif %}' + ) + + +def reverse(apps, schema_editor): + Recipient = apps.get_model("mailtrigger", "Recipient") + Recipient.objects.filter(slug="doc_authors").update( + template='{% if doc.type_id == "draft" %}<{{doc.name}}@ietf.org>{% endif %}' + ) + + +class Migration(migrations.Migration): + dependencies = [ + ("mailtrigger", "0004_slides_approved"), + ] + + operations = [migrations.RunPython(forward, reverse)] diff --git a/ietf/mailtrigger/migrations/0005_slides_proposed.py b/ietf/mailtrigger/migrations/0005_slides_proposed.py deleted file mode 100644 index 88af912815..0000000000 --- a/ietf/mailtrigger/migrations/0005_slides_proposed.py +++ /dev/null @@ -1,30 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-03-25 06:11 - - -from django.db import migrations - -def forward(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - changed = MailTrigger.objects.create( - slug = 'slides_proposed', - desc = 'Recipients when slides are proposed for a given session', - ) - changed.to.set(Recipient.objects.filter(slug__in=['group_chairs', 'group_responsible_directors', 'group_secretaries'])) - -def reverse(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger','MailTrigger') - MailTrigger.objects.filter(slug='slides_proposed').delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0004_ballot_rfceditornote_changed_postapproval'), - ] - - operations = [ - migrations.RunPython(forward,reverse) - ] diff --git a/ietf/mailtrigger/migrations/0006_call_for_adoption_and_last_call_issued.py b/ietf/mailtrigger/migrations/0006_call_for_adoption_and_last_call_issued.py new file mode 100644 index 0000000000..7adad150eb --- /dev/null +++ b/ietf/mailtrigger/migrations/0006_call_for_adoption_and_last_call_issued.py @@ -0,0 +1,43 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations + + +def forward(apps, schema_editor): + MailTrigger = apps.get_model("mailtrigger", "MailTrigger") + Recipient = apps.get_model("mailtrigger", "Recipient") + recipients = list( + Recipient.objects.filter( + slug__in=( + "doc_group_mail_list", + "doc_authors", + "doc_group_chairs", + "doc_shepherd", + ) + ) + ) + call_for_adoption = MailTrigger.objects.create( + slug="doc_wg_call_for_adoption_issued", + desc="Recipients when a working group call for adoption is issued", + ) + call_for_adoption.to.add(*recipients) + wg_last_call = MailTrigger.objects.create( + slug="doc_wg_last_call_issued", + desc="Recipients when a working group last call is issued", + ) + wg_last_call.to.add(*recipients) + + +def reverse(apps, schema_editor): + MailTrigger = apps.get_model("mailtrigger", "MailTrigger") + MailTrigger.objects.filter( + slug_in=("doc_wg_call_for_adoption_issued", "doc_wg_last_call_issued") + ).delete() + + +class Migration(migrations.Migration): + dependencies = [ + ("mailtrigger", "0005_rfc_recipients"), + ] + + operations = [migrations.RunPython(forward, reverse)] diff --git a/ietf/mailtrigger/migrations/0006_sub_new_wg_00.py b/ietf/mailtrigger/migrations/0006_sub_new_wg_00.py deleted file mode 100644 index 8b7936d907..0000000000 --- a/ietf/mailtrigger/migrations/0006_sub_new_wg_00.py +++ /dev/null @@ -1,39 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -from django.db import migrations - -def forward(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - Recipient.objects.create( - slug = 'new_wg_doc_list', - desc = "The email list for announcing new WG -00 submissions", - template = '' - ) - changed = MailTrigger.objects.create( - slug = 'sub_new_wg_00', - desc = 'Recipients when a new IETF WG -00 draft is announced', - ) - changed.to.set(Recipient.objects.filter(slug__in=['new_wg_doc_list'])) - - -def reverse(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger','MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - MailTrigger.objects.filter(slug='sub_new_wg_00').delete() - Recipient.objects.filter(slug='new_wg_doc_list').delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0005_slides_proposed'), - ] - - operations = [ - migrations.RunPython(forward,reverse) - ] diff --git a/ietf/mailtrigger/migrations/0007_add_review_mailtriggers.py b/ietf/mailtrigger/migrations/0007_add_review_mailtriggers.py deleted file mode 100644 index 91a963d186..0000000000 --- a/ietf/mailtrigger/migrations/0007_add_review_mailtriggers.py +++ /dev/null @@ -1,85 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -from django.db import migrations - -def forward(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger','MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - review_assignment_reviewer = Recipient.objects.create( - slug="review_assignment_reviewer", - desc="The reviewer assigned to a review assignment", - template="{% if not skip_review_reviewer %}{{review_assignment.reviewer.email_address}}{% endif %}", - ) - review_assignment_review_req_by = Recipient.objects.create( - slug="review_assignment_review_req_by", - desc="The requester of an assigned review", - template="{% if not skip_review_requested_by %}{{review_assignment.review_request.requested_by.email_address}}{% endif %}", - ) - review_req_requested_by = Recipient.objects.create( - slug="review_req_requested_by", - desc="The requester of a review", - template="{% if not skip_review_requested_by %}{{review_req.requested_by.email_address}}{% endif %}", - ) - review_req_reviewers = Recipient.objects.create( - slug="review_req_reviewers", - desc="All reviewers assigned to a review request", - template=None, - ) - review_secretaries = Recipient.objects.create( - slug="review_secretaries", - desc="The secretaries of the review team of a review request or assignment", - template=None, - ) - Recipient.objects.create( - slug="review_reviewer", - desc="A single reviewer", - template="{{reviewer.email_address}}", - ) - - review_assignment_changed = MailTrigger.objects.create( - slug="review_assignment_changed", - desc="Recipients for a change to a review assignment", - ) - review_assignment_changed.to.set([review_assignment_review_req_by, review_assignment_reviewer, - review_secretaries]) - - review_req_changed = MailTrigger.objects.create( - slug="review_req_changed", - desc="Recipients for a change to a review request", - ) - review_req_changed.to.set([review_req_requested_by, review_req_reviewers, review_secretaries]) - - review_availability_changed = MailTrigger.objects.create( - slug="review_availability_changed", - desc="Recipients for a change to a reviewer's availability", - ) - review_availability_changed.to.set( - Recipient.objects.filter(slug__in=['review_reviewer', 'group_secretaries']) - ) - - -def reverse(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger','MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - MailTrigger.objects.filter(slug__in=[ - 'review_assignment_changed', 'review_req_changed', 'review_availability_changed', - ]).delete() - Recipient.objects.filter(slug__in=[ - 'review_assignment_reviewer', 'review_assignment_review_req_by', 'review_req_requested_by', - 'review_req_reviewers', 'review_secretaries', 'review_reviewer', - ]).delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0006_sub_new_wg_00'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/mailtrigger/migrations/0007_historicalrecipient_historicalmailtrigger.py b/ietf/mailtrigger/migrations/0007_historicalrecipient_historicalmailtrigger.py new file mode 100644 index 0000000000..d23b72d737 --- /dev/null +++ b/ietf/mailtrigger/migrations/0007_historicalrecipient_historicalmailtrigger.py @@ -0,0 +1,122 @@ +# Copyright The IETF Trust 2025, All Rights Reserved +from io import StringIO + +from django.conf import settings +from django.core import management +from django.db import migrations, models +import django.db.models.deletion +import simple_history.models + +from ietf.utils.log import log + + +def forward(apps, schema_editor): + # Fill in history for existing data using the populate_history management command + captured_stdout = StringIO() + captured_stderr = StringIO() + try: + management.call_command( + "populate_history", + "mailtrigger.MailTrigger", + "mailtrigger.Recipient", + stdout=captured_stdout, + stderr=captured_stderr, + ) + except management.CommandError as err: + log( + "Failed to populate history for mailtrigger models.\n" + "\n" + f"stdout:\n{captured_stdout.getvalue() or ''}\n" + "\n" + f"stderr:\n{captured_stderr.getvalue() or ''}\n" + ) + raise RuntimeError("Failed to populate history for mailtrigger models") from err + log( + "Populated history for mailtrigger models.\n" + "\n" + f"stdout:\n{captured_stdout.getvalue() or ''}\n" + "\n" + f"stderr:\n{captured_stderr.getvalue() or ''}\n" + ) + + +def reverse(apps, schema_editor): + pass # nothing to do + + +class Migration(migrations.Migration): + dependencies = [ + migrations.swappable_dependency(settings.AUTH_USER_MODEL), + ("mailtrigger", "0006_call_for_adoption_and_last_call_issued"), + ] + + operations = [ + migrations.CreateModel( + name="HistoricalRecipient", + fields=[ + ("slug", models.CharField(db_index=True, max_length=32)), + ("desc", models.TextField(blank=True)), + ("template", models.TextField(blank=True, null=True)), + ("history_id", models.AutoField(primary_key=True, serialize=False)), + ("history_date", models.DateTimeField(db_index=True)), + ("history_change_reason", models.CharField(max_length=100, null=True)), + ( + "history_type", + models.CharField( + choices=[("+", "Created"), ("~", "Changed"), ("-", "Deleted")], + max_length=1, + ), + ), + ( + "history_user", + models.ForeignKey( + null=True, + on_delete=django.db.models.deletion.SET_NULL, + related_name="+", + to=settings.AUTH_USER_MODEL, + ), + ), + ], + options={ + "verbose_name": "historical recipient", + "verbose_name_plural": "historical recipients", + "ordering": ("-history_date", "-history_id"), + "get_latest_by": ("history_date", "history_id"), + }, + bases=(simple_history.models.HistoricalChanges, models.Model), + ), + migrations.CreateModel( + name="HistoricalMailTrigger", + fields=[ + ("slug", models.CharField(db_index=True, max_length=64)), + ("desc", models.TextField(blank=True)), + ("history_id", models.AutoField(primary_key=True, serialize=False)), + ("history_date", models.DateTimeField(db_index=True)), + ("history_change_reason", models.CharField(max_length=100, null=True)), + ( + "history_type", + models.CharField( + choices=[("+", "Created"), ("~", "Changed"), ("-", "Deleted")], + max_length=1, + ), + ), + ( + "history_user", + models.ForeignKey( + null=True, + on_delete=django.db.models.deletion.SET_NULL, + related_name="+", + to=settings.AUTH_USER_MODEL, + ), + ), + ], + options={ + "verbose_name": "historical mail trigger", + "verbose_name_plural": "historical mail triggers", + "ordering": ("-history_date", "-history_id"), + "get_latest_by": ("history_date", "history_id"), + }, + bases=(simple_history.models.HistoricalChanges, models.Model), + ), + migrations.RunPython(forward, reverse), + ] diff --git a/ietf/mailtrigger/migrations/0008_lengthen_mailtrigger_slug.py b/ietf/mailtrigger/migrations/0008_lengthen_mailtrigger_slug.py deleted file mode 100644 index df83bdb5fd..0000000000 --- a/ietf/mailtrigger/migrations/0008_lengthen_mailtrigger_slug.py +++ /dev/null @@ -1,32 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.23 on 2019-08-30 09:02 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0007_add_review_mailtriggers'), - ] - - operations = [ - migrations.AlterField( - model_name='mailtrigger', - name='slug', - field=models.CharField(max_length=64, primary_key=True, serialize=False), - ), - # The above migration will not update the ManyToMany tables, which also reference - # the mailtrigger pk as varchar(32), so manual SQL is used. - # https://code.djangoproject.com/ticket/25012 - migrations.RunSQL( - sql='ALTER TABLE `mailtrigger_mailtrigger_to` MODIFY `mailtrigger_id` varchar(64);', - reverse_sql='ALTER TABLE `mailtrigger_mailtrigger_to` MODIFY `mailtrigger_id` varchar(32);', - ), - migrations.RunSQL( - sql='ALTER TABLE `mailtrigger_mailtrigger_cc` MODIFY `mailtrigger_id` varchar(64);', - reverse_sql='ALTER TABLE `mailtrigger_mailtrigger_cc` MODIFY `mailtrigger_id` varchar(32);', - ), - ] diff --git a/ietf/mailtrigger/migrations/0008_liaison_statement_incoming_and_outgoing_posted.py b/ietf/mailtrigger/migrations/0008_liaison_statement_incoming_and_outgoing_posted.py new file mode 100644 index 0000000000..189a783a2e --- /dev/null +++ b/ietf/mailtrigger/migrations/0008_liaison_statement_incoming_and_outgoing_posted.py @@ -0,0 +1,72 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +from django.db import migrations + + +def forward(apps, schema_editor): + Mailtrigger = apps.get_model("mailtrigger", "MailTrigger") + Recipient = apps.get_model("mailtrigger", "Recipient") + recipients_to = Recipient.objects.get(pk="liaison_to_contacts") + recipients_cc = list( + Recipient.objects.filter( + slug__in=( + "liaison_cc", + "liaison_coordinators", + "liaison_response_contacts", + "liaison_technical_contacts", + ) + ) + ) + recipient_from = Recipient.objects.get(pk="liaison_from_contact") + + liaison_posted_outgoing = Mailtrigger.objects.create( + slug="liaison_statement_posted_outgoing", + desc="Recipients for a message when a new outgoing liaison statement is posted", + ) + liaison_posted_outgoing.to.add(recipients_to) + liaison_posted_outgoing.cc.add(*recipients_cc) + liaison_posted_outgoing.cc.add(recipient_from) + + liaison_posted_incoming = Mailtrigger.objects.create( + slug="liaison_statement_posted_incoming", + desc="Recipients for a message when a new incoming liaison statement is posted", + ) + liaison_posted_incoming.to.add(recipients_to) + liaison_posted_incoming.cc.add(*recipients_cc) + + Mailtrigger.objects.filter(slug=("liaison_statement_posted")).delete() + + +def reverse(apps, schema_editor): + Mailtrigger = apps.get_model("mailtrigger", "MailTrigger") + Recipient = apps.get_model("mailtrigger", "Recipient") + + Mailtrigger.objects.filter( + slug__in=( + "liaison_statement_posted_outgoing", + "liaison_statement_posted_incoming", + ) + ).delete() + + liaison_statement_posted = Mailtrigger.objects.create( + slug="liaison_statement_posted", + desc="Recipients for a message when a new liaison statement is posted", + ) + + liaison_to_contacts = Recipient.objects.get(slug="liaison_to_contacts") + recipients_ccs = Recipient.objects.filter( + slug__in=( + "liaison_cc", + "liaison_coordinators", + "liaison_response_contacts", + "liaison_technical_contacts", + ) + ) + liaison_statement_posted.to.add(liaison_to_contacts) + liaison_statement_posted.cc.add(*recipients_ccs) + + +class Migration(migrations.Migration): + dependencies = [("mailtrigger", "0007_historicalrecipient_historicalmailtrigger")] + + operations = [migrations.RunPython(forward, reverse)] diff --git a/ietf/mailtrigger/migrations/0009_custom_review_complete_mailtriggers.py b/ietf/mailtrigger/migrations/0009_custom_review_complete_mailtriggers.py deleted file mode 100644 index 496d4658fc..0000000000 --- a/ietf/mailtrigger/migrations/0009_custom_review_complete_mailtriggers.py +++ /dev/null @@ -1,51 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -from django.db import migrations - -def forward(apps, schema_editor): - ReviewTeamSettings = apps.get_model('review', 'ReviewTeamSettings') - MailTrigger = apps.get_model('mailtrigger', 'Mailtrigger') - Group = apps.get_model('group', 'Group') - GroupFeatures = apps.get_model('group', 'GroupFeatures') - - template = MailTrigger.objects.get(slug='review_completed') - template.desc = 'Default template for recipients when an review is completed - ' \ - 'customised mail triggers are used/created per team and review type.' - template.save() - - for group in Group.objects.all().only('pk', 'type', 'acronym'): - if not GroupFeatures.objects.get(type=group.type).has_reviews: - continue - try: - review_team = ReviewTeamSettings.objects.get(group=group.pk) - except ReviewTeamSettings.DoesNotExist: - continue - team_acronym = group.acronym.lower() - for review_type in review_team.review_types.all(): - slug = 'review_completed_{}_{}'.format(team_acronym, review_type.slug) - desc = 'Recipients when a {} {} review is completed'.format(team_acronym, review_type) - if MailTrigger.objects.filter(slug=slug): - # Never overwrite existing triggers - continue - mailtrigger = MailTrigger.objects.create(slug=slug, desc=desc) - mailtrigger.to.set(template.to.all()) - mailtrigger.cc.set(template.cc.all()) - - -def reverse(apps, schema_editor): - pass - - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0008_lengthen_mailtrigger_slug'), - ('review', '0014_document_primary_key_cleanup'), - ('group', '0019_rename_field_document2'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/mailtrigger/migrations/0010_add_review_reminder_mailtriggers.py b/ietf/mailtrigger/migrations/0010_add_review_reminder_mailtriggers.py deleted file mode 100644 index 384c55e8f2..0000000000 --- a/ietf/mailtrigger/migrations/0010_add_review_reminder_mailtriggers.py +++ /dev/null @@ -1,34 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -from django.db import migrations - - -def forward(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - review_reminder_overdue_assignment = MailTrigger.objects.create( - slug="review_reminder_overdue_assignment", - desc="Recipients for overdue review assignment reminders", - ) - review_reminder_overdue_assignment.to.add( - Recipient.objects.get(slug='group_secretaries') - ) - - -def reverse(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - MailTrigger.objects.filter(slug='review_reminder_overdue_assignment').delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0009_custom_review_complete_mailtriggers'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/mailtrigger/migrations/0011_ietf_last_call.py b/ietf/mailtrigger/migrations/0011_ietf_last_call.py deleted file mode 100644 index 6999d64f92..0000000000 --- a/ietf/mailtrigger/migrations/0011_ietf_last_call.py +++ /dev/null @@ -1,49 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.25 on 2019-10-04 12:12 - - -from django.db import migrations - -def forward(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger','MailTrigger') - Recipient = apps.get_model('mailtrigger','Recipient') - - ietf_last_call = Recipient.objects.create( - slug = 'ietf_last_call', - desc = 'The IETF Last Call list', - template = 'last-call@ietf.org' - ) - ietf_general = Recipient.objects.get(slug='ietf_general') - - review_completed_triggers = MailTrigger.objects.filter(slug__startswith='review_completed') - - for trigger in review_completed_triggers: - trigger.cc.remove(ietf_general) - trigger.cc.add(ietf_last_call) - -def reverse(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger','MailTrigger') - Recipient = apps.get_model('mailtrigger','Recipient') - - ietf_general = Recipient.objects.get(slug='ietf_general') - ietf_last_call = Recipient.objects.get(slug='ietf_last_call') - - review_completed_triggers = MailTrigger.objects.filter(slug__startswith='review_completed') - - for trigger in review_completed_triggers: - trigger.cc.remove(ietf_last_call) - trigger.cc.add(ietf_general) - - ietf_last_call.delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0010_add_review_reminder_mailtriggers'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/mailtrigger/migrations/0012_dont_last_call_early_reviews.py b/ietf/mailtrigger/migrations/0012_dont_last_call_early_reviews.py deleted file mode 100644 index 5302a0c891..0000000000 --- a/ietf/mailtrigger/migrations/0012_dont_last_call_early_reviews.py +++ /dev/null @@ -1,28 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.25 on 2019-10-04 13:12 - - -from django.db import migrations - -def forward(apps, shema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - for trigger in MailTrigger.objects.filter(slug__startswith='review_completed',slug__endswith='early'): - trigger.cc.remove('ietf_last_call') - - -def reverse(apps, shema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - for trigger in MailTrigger.objects.filter(slug__startswith='review_completed',slug__endswith='early'): - trigger.cc.add('ietf_last_call') - - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0011_ietf_last_call'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/mailtrigger/migrations/0013_add_irsg_ballot_saved.py b/ietf/mailtrigger/migrations/0013_add_irsg_ballot_saved.py deleted file mode 100644 index a7400e48a3..0000000000 --- a/ietf/mailtrigger/migrations/0013_add_irsg_ballot_saved.py +++ /dev/null @@ -1,52 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.25 on 2019-10-04 13:12 - - -from django.db import migrations - -def forward(apps, shema_editor): - Recipient = apps.get_model('mailtrigger','Recipient') - - irsg = Recipient.objects.create( - slug = 'irsg', - desc = 'The IRSG', - template = 'The IRSG ' - ) - - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - slug = 'irsg_ballot_saved' - desc = 'Recipients when a new IRSG ballot position with comments is saved' - irsg_ballot_saved = MailTrigger.objects.create( - slug=slug, - desc=desc - ) - irsg_ballot_saved.to.add(irsg) - irsg_ballot_saved.cc.set(Recipient.objects.filter(slug__in=['doc_affecteddoc_authors','doc_affecteddoc_group_chairs','doc_affecteddoc_notify','doc_authors','doc_group_chairs','doc_group_mail_list','doc_notify','doc_shepherd'])) - - # We cannot just change the slug of the existing ballot_saved table, - # because that will loose all the m2m entries in .to and .cc - ballot_saved = MailTrigger.objects.get(slug='ballot_saved') - iesg_ballot_saved = MailTrigger.objects.create(slug='iesg_ballot_saved') - iesg_ballot_saved.to.set(ballot_saved.to.all()) - iesg_ballot_saved.cc.set(ballot_saved.cc.all()) - iesg_ballot_saved.desc = ballot_saved.desc - iesg_ballot_saved.save() - -def reverse(apps, shema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - MailTrigger.objects.filter(slug='irsg_ballot_saved').delete() - MailTrigger.objects.filter(slug='iesg_ballot_saved').delete() - # - Recipient = apps.get_model('mailtrigger','Recipient') - Recipient.objects.filter(slug='irsg').delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0012_dont_last_call_early_reviews'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/mailtrigger/migrations/0014_add_ad_approved_conflict_review.py b/ietf/mailtrigger/migrations/0014_add_ad_approved_conflict_review.py deleted file mode 100644 index b16d2bde3f..0000000000 --- a/ietf/mailtrigger/migrations/0014_add_ad_approved_conflict_review.py +++ /dev/null @@ -1,31 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved - -from django.db import migrations - - -def forward(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - ad_approved_conflict_review = MailTrigger.objects.create( - slug='ad_approved_conflict_review', - desc='Recipients when AD approves a conflict review pending announcement', - ) - ad_approved_conflict_review.to.add( - Recipient.objects.get(pk='iesg_secretary') - ) - - -def reverse(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - MailTrigger.objects.filter(slug='ad_approved_conflict_review').delete() - - -class Migration(migrations.Migration): - dependencies = [ - ('mailtrigger', '0013_add_irsg_ballot_saved'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/mailtrigger/migrations/0015_add_ad_approved_status_change.py b/ietf/mailtrigger/migrations/0015_add_ad_approved_status_change.py deleted file mode 100644 index 8fce9cac06..0000000000 --- a/ietf/mailtrigger/migrations/0015_add_ad_approved_status_change.py +++ /dev/null @@ -1,31 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved - -from django.db import migrations - - -def forward(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - ad_approved_conflict_review = MailTrigger.objects.create( - slug='ad_approved_status_change', - desc='Recipients when AD approves a status change pending announcement', - ) - ad_approved_conflict_review.to.add( - Recipient.objects.get(pk='iesg_secretary') - ) - - -def reverse(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - MailTrigger.objects.filter(slug='ad_approved_status_change').delete() - - -class Migration(migrations.Migration): - dependencies = [ - ('mailtrigger', '0014_add_ad_approved_conflict_review'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/mailtrigger/migrations/0016_add_irsg_ballot_issued.py b/ietf/mailtrigger/migrations/0016_add_irsg_ballot_issued.py deleted file mode 100644 index 0e3accb742..0000000000 --- a/ietf/mailtrigger/migrations/0016_add_irsg_ballot_issued.py +++ /dev/null @@ -1,57 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -from django.db import migrations - - -def replace_mailtrigger(MailTrigger, old_slug, new_slug): - """Replace a MailTrigger with an equivalent using a different slug""" - # Per 0013_add_irsg_ballot_saved.py, can't just modify the existing because that - # will lose the many-to-many relations. - orig_mailtrigger = MailTrigger.objects.get(slug=old_slug) - new_mailtrigger = MailTrigger.objects.create(slug=new_slug) - new_mailtrigger.to.set(orig_mailtrigger.to.all()) - new_mailtrigger.cc.set(orig_mailtrigger.cc.all()) - new_mailtrigger.desc = orig_mailtrigger.desc - new_mailtrigger.save() - orig_mailtrigger.delete() # get rid of the obsolete MailTrigger - - -def forward(apps, schema_editor): - """Forward migration: create irsg_ballot_issued and rename ballot_issued to iesg_ballot_issued""" - # Load historical models - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - # Create the new MailTrigger - irsg_ballot_issued = MailTrigger.objects.create( - slug='irsg_ballot_issued', - desc='Recipients when a new IRSG ballot is issued', - ) - irsg_ballot_issued.to.set(Recipient.objects.filter(slug='irsg')) - irsg_ballot_issued.cc.set(Recipient.objects.filter(slug__in=[ - 'doc_stream_manager', 'doc_affecteddoc_authors', 'doc_affecteddoc_group_chairs', - 'doc_affecteddoc_notify', 'doc_authors', 'doc_group_chairs', 'doc_group_mail_list', - 'doc_notify', 'doc_shepherd' - ])) - - # Replace existing 'ballot_issued' object with an 'iesg_ballot_issued' - replace_mailtrigger(MailTrigger, 'ballot_issued', 'iesg_ballot_issued') - - -def reverse(apps, shema_editor): - """Reverse migration: rename iesg_ballot_issued to ballot_issued and remove irsg_ballot_issued""" - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - MailTrigger.objects.filter(slug='irsg_ballot_issued').delete() - replace_mailtrigger(MailTrigger, 'iesg_ballot_issued', 'ballot_issued') - - -class Migration(migrations.Migration): - dependencies = [ - ('mailtrigger', '0015_add_ad_approved_status_change'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/mailtrigger/migrations/0017_lc_to_yang_doctors.py b/ietf/mailtrigger/migrations/0017_lc_to_yang_doctors.py deleted file mode 100644 index 6685d857f5..0000000000 --- a/ietf/mailtrigger/migrations/0017_lc_to_yang_doctors.py +++ /dev/null @@ -1,38 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved - -from django.db import migrations - -def forward(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - Recipient.objects.create( - slug = 'yang_doctors_secretaries', - desc = 'Yang Doctors Secretaries', - template = '' - ) - - lc_to_yang_doctors = MailTrigger.objects.create( - slug='last_call_of_doc_with_yang_issued', - desc='Recipients when IETF LC is issued on a document with yang checks', - ) - - lc_to_yang_doctors.to.set(Recipient.objects.filter(slug='yang_doctors_secretaries')) - -def reverse(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - MailTrigger.objects.filter(slug='last_call_of_doc_with_yang_issued').delete() - Recipient.objects.filter(slug='yang_doctors_secretaries').delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0016_add_irsg_ballot_issued'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/mailtrigger/migrations/0018_interim_approve_announce.py b/ietf/mailtrigger/migrations/0018_interim_approve_announce.py deleted file mode 100644 index a3e810f0f0..0000000000 --- a/ietf/mailtrigger/migrations/0018_interim_approve_announce.py +++ /dev/null @@ -1,41 +0,0 @@ -# Copyright The IETF Trust 2020 All Rights Reserved - -from django.db import migrations - -def forward(apps,schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - interim_approved = MailTrigger.objects.get(slug='interim_approved') - interim_approved.desc = 'Recipients when an interim meeting is approved' - interim_approved.save() - interim_approved.to.set(Recipient.objects.filter(slug__in=('group_chairs','logged_in_person'))) - - interim_announce_requested = MailTrigger.objects.create( - slug='interim_announce_requested', - desc='Recipients when an interim announcement is requested', - ) - interim_announce_requested.to.set(Recipient.objects.filter(slug='iesg_secretary')) - - -def reverse(apps,schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - MailTrigger.objects.filter(slug='interim_announce_requested').delete() - - interim_approved = MailTrigger.objects.get(slug='interim_approved') - interim_approved.desc = 'Recipients when an interim meeting is approved and an announcement needs to be sent' - interim_approved.save() - interim_approved.to.set(Recipient.objects.filter(slug='iesg_secretary')) - - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0017_lc_to_yang_doctors'), - ] - - operations = [ - migrations.RunPython(forward,reverse) - ] diff --git a/ietf/mailtrigger/migrations/0019_email_iana_expert_review_state_changed.py b/ietf/mailtrigger/migrations/0019_email_iana_expert_review_state_changed.py deleted file mode 100644 index 9818b4427f..0000000000 --- a/ietf/mailtrigger/migrations/0019_email_iana_expert_review_state_changed.py +++ /dev/null @@ -1,34 +0,0 @@ -# Copyright The IETF Trust 2020 All Rights Reserved - -from django.db import migrations - -def forward(apps,schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - iana_er_state_changed = MailTrigger.objects.create( - slug='iana_expert_review_state_changed', - desc='Recipients when the IANA expert review for a document changes', - ) - - iana_er_state_changed.to.set( - Recipient.objects.filter(slug__in=[ - 'doc_ad', 'doc_authors', 'doc_group_chairs', 'doc_group_responsible_directors', 'doc_notify', 'doc_shepherd' - ]) - ) - -def reverse(apps,schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - - MailTrigger.objects.filter(slug='iana_expert_review_state_changed').delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0018_interim_approve_announce'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/mailtrigger/migrations/0020_add_ad_approval_request_mailtriggers.py b/ietf/mailtrigger/migrations/0020_add_ad_approval_request_mailtriggers.py deleted file mode 100644 index c1cfd72778..0000000000 --- a/ietf/mailtrigger/migrations/0020_add_ad_approval_request_mailtriggers.py +++ /dev/null @@ -1,58 +0,0 @@ -# Generated by Django 2.2.17 on 2020-11-23 09:54 - -from django.db import migrations - - -def forward(apps, schema_editor): - """Add new MailTrigger and Recipients, remove the one being replaced""" - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - MailTrigger.objects.create( - slug='sub_director_approval_requested', - desc='Recipients for a message requesting AD approval of a revised draft submission', - ).to.add( - Recipient.objects.create( - pk='sub_group_parent_directors', - desc="ADs for the parent group of a submission" - ) - ) - - MailTrigger.objects.create( - slug='sub_replaced_doc_chair_approval_requested', - desc='Recipients for a message requesting chair approval of a replaced WG document', - ).to.add( - Recipient.objects.get(pk='doc_group_chairs') - ) - - MailTrigger.objects.create( - slug='sub_replaced_doc_director_approval_requested', - desc='Recipients for a message requesting AD approval of a replaced WG document', - ).to.add( - Recipient.objects.create( - pk='doc_group_parent_directors', - desc='ADs for the parent group of a doc' - ) - ) - - -def reverse(apps, schema_editor): - """Remove the MailTrigger and Recipient created in the forward migration""" - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - MailTrigger.objects.filter(slug='sub_director_approval_requested').delete() - MailTrigger.objects.filter(slug='sub_replaced_doc_chair_approval_requested').delete() - MailTrigger.objects.filter(slug='sub_replaced_doc_director_approval_requested').delete() - Recipient = apps.get_model('mailtrigger', 'Recipient') - Recipient.objects.filter(pk='sub_group_parent_directors').delete() - Recipient.objects.filter(pk='doc_group_parent_directors').delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0019_email_iana_expert_review_state_changed'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/mailtrigger/migrations/0021_email_remind_action_holders.py b/ietf/mailtrigger/migrations/0021_email_remind_action_holders.py deleted file mode 100644 index 704b452a53..0000000000 --- a/ietf/mailtrigger/migrations/0021_email_remind_action_holders.py +++ /dev/null @@ -1,38 +0,0 @@ -# Copyright The IETF Trust 2020 All Rights Reserved - -from django.db import migrations - - -def forward(apps,schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - (doc_action_holders, _) = Recipient.objects.get_or_create( - slug='doc_action_holders', - desc='Action holders for a document', - template='{% for action_holder in doc.action_holders.all %}{% if doc.shepherd and action_holder == doc.shepherd.person %}{{ doc.shepherd }}{% else %}{{ action_holder.email }}{% endif %}{% if not forloop.last %},{%endif %}{% endfor %}', - ) - (doc_remind_action_holders, _) = MailTrigger.objects.get_or_create( - slug='doc_remind_action_holders', - desc='Recipients when sending a reminder email to action holders for a document', - ) - doc_remind_action_holders.to.set([doc_action_holders]) - - -def reverse(apps,schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - MailTrigger.objects.filter(slug='doc_remind_action_holders').delete() - Recipient.objects.filter(slug='doc_action_holders').delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0020_add_ad_approval_request_mailtriggers'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/mailtrigger/migrations/0022_add_doc_external_resource_change_requested.py b/ietf/mailtrigger/migrations/0022_add_doc_external_resource_change_requested.py deleted file mode 100644 index 92ace7a8d0..0000000000 --- a/ietf/mailtrigger/migrations/0022_add_doc_external_resource_change_requested.py +++ /dev/null @@ -1,37 +0,0 @@ -# Generated by Django 2.2.20 on 2021-05-13 07:20 - -from django.db import migrations - - -def forward(apps, schema_editor): - """Add new MailTrigger and Recipients""" - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - mt, created = MailTrigger.objects.get_or_create(slug='doc_external_resource_change_requested') - if created: - mt.desc='Recipients when a change to the external resources for a document is requested.' - mt.save() - for recipient_slug in [ - "doc_ad", - "doc_group_chairs", - "doc_group_delegates", - "doc_shepherd", - "doc_stream_manager" - ]: - mt.to.add(Recipient.objects.get(slug=recipient_slug)) - - -def reverse(apps, schema_editor): - pass - - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0021_email_remind_action_holders'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/mailtrigger/migrations/0023_bofreq_triggers.py b/ietf/mailtrigger/migrations/0023_bofreq_triggers.py deleted file mode 100644 index 4c3580140c..0000000000 --- a/ietf/mailtrigger/migrations/0023_bofreq_triggers.py +++ /dev/null @@ -1,50 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved -# Generated by Django 2.2.23 on 2021-05-26 07:52 - -from django.db import migrations - -def forward(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - - Recipient.objects.create(slug='bofreq_editors',desc='BOF request editors',template='') - Recipient.objects.create(slug='bofreq_previous_editors',desc='Editors of the prior version of a BOF request', - template='{% for editor in previous_editors %}{{editor.email_address}}{% if not forloop.last %},{% endif %}{% endfor %}') - - Recipient.objects.create(slug='bofreq_responsible',desc='BOF request responsible leadership',template='') - Recipient.objects.create(slug='bofreq_previous_responsible',desc='BOF request responsible leadership before change', template='') - - mt = MailTrigger.objects.create(slug='bofreq_title_changed',desc='Recipients when the title of a BOF proposal is changed.') - mt.to.set(Recipient.objects.filter(slug__in=['bofreq_responsible', 'bofreq_editors', 'doc_notify'])) - - mt = MailTrigger.objects.create(slug='bofreq_editors_changed',desc='Recipients when the editors of a BOF proposal are changed.') - mt.to.set(Recipient.objects.filter(slug__in=['bofreq_responsible', 'bofreq_editors', 'bofreq_previous_editors', 'doc_notify'])) - - mt = MailTrigger.objects.create(slug='bofreq_responsible_changed',desc='Recipients when the responsible leadership of a BOF proposal are changed.') - mt.to.set(Recipient.objects.filter(slug__in=['bofreq_responsible', 'bofreq_editors', 'bofreq_previous_responsible', 'doc_notify'])) - - mt = MailTrigger.objects.create(slug='bofreq_new_revision', desc='Recipients when a new revision of a BOF request is uploaded.') - mt.to.set(Recipient.objects.filter(slug__in=['bofreq_responsible', 'bofreq_editors', 'doc_notify'])) - - for recipient in Recipient.objects.filter(slug__in=['bofreq_responsible','bofreq_editors']): - MailTrigger.objects.get(slug='doc_state_edited').to.add(recipient) - -def reverse(apps, schema_editor): - MailTrigger = apps.get_model('mailtrigger', 'MailTrigger') - Recipient = apps.get_model('mailtrigger', 'Recipient') - for recipient in Recipient.objects.filter(slug__in=['bofreq_responsible','bofreq_editors']): - MailTrigger.objects.get(slug='doc_state_edited').to.remove(recipient) - MailTrigger.objects.filter(slug__in=('bofreq_title_changed', 'bofreq_editors_changed', 'bofreq_new_revision', 'bofreq_responsible_changed')).delete() - Recipient.objects.filter(slug__in=('bofreq_editors', 'bofreq_previous_editors')).delete() - Recipient.objects.filter(slug__in=('bofreq_responsible', 'bofreq_previous_responsible')).delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('mailtrigger', '0022_add_doc_external_resource_change_requested'), - ] - - operations = [ - migrations.RunPython(forward,reverse) - ] diff --git a/ietf/mailtrigger/models.py b/ietf/mailtrigger/models.py index a1b712c574..435729f893 100644 --- a/ietf/mailtrigger/models.py +++ b/ietf/mailtrigger/models.py @@ -1,4 +1,4 @@ -# Copyright The IETF Trust 2015-2020, All Rights Reserved +# Copyright The IETF Trust 2015-2025, All Rights Reserved # -*- coding: utf-8 -*- @@ -7,9 +7,11 @@ from email.utils import parseaddr +from simple_history.models import HistoricalRecords + from ietf.doc.utils_bofreq import bofreq_editors, bofreq_responsible from ietf.utils.mail import formataddr, get_email_addresses_from_text -from ietf.group.models import Group +from ietf.group.models import Group, Role from ietf.person.models import Email, Alias from ietf.review.models import ReviewTeamSettings @@ -36,8 +38,9 @@ def clean_duplicates(addrlist): class MailTrigger(models.Model): slug = models.CharField(max_length=64, primary_key=True) desc = models.TextField(blank=True) - to = models.ManyToManyField('Recipient', blank=True, related_name='used_in_to') - cc = models.ManyToManyField('Recipient', blank=True, related_name='used_in_cc') + to = models.ManyToManyField('mailtrigger.Recipient', blank=True, related_name='used_in_to') + cc = models.ManyToManyField('mailtrigger.Recipient', blank=True, related_name='used_in_cc') + history = HistoricalRecords() class Meta: ordering = ["slug"] @@ -49,6 +52,7 @@ class Recipient(models.Model): slug = models.CharField(max_length=32, primary_key=True) desc = models.TextField(blank=True) template = models.TextField(null=True, blank=True) + history = HistoricalRecords() class Meta: ordering = ["slug"] @@ -96,35 +100,35 @@ def gather_doc_affecteddoc_authors(self, **kwargs): addrs = [] if 'doc' in kwargs: for reldoc in kwargs['doc'].related_that_doc(('conflrev','tohist','tois','tops')): - addrs.extend(Recipient.objects.get(slug='doc_authors').gather(**{'doc':reldoc.document})) + addrs.extend(Recipient.objects.get(slug='doc_authors').gather(**{'doc':reldoc})) return addrs def gather_doc_affecteddoc_group_chairs(self, **kwargs): addrs = [] if 'doc' in kwargs: for reldoc in kwargs['doc'].related_that_doc(('conflrev','tohist','tois','tops')): - addrs.extend(Recipient.objects.get(slug='doc_group_chairs').gather(**{'doc':reldoc.document})) + addrs.extend(Recipient.objects.get(slug='doc_group_chairs').gather(**{'doc':reldoc})) return addrs def gather_doc_affecteddoc_notify(self, **kwargs): addrs = [] if 'doc' in kwargs: for reldoc in kwargs['doc'].related_that_doc(('conflrev','tohist','tois','tops')): - addrs.extend(Recipient.objects.get(slug='doc_notify').gather(**{'doc':reldoc.document})) + addrs.extend(Recipient.objects.get(slug='doc_notify').gather(**{'doc':reldoc})) return addrs def gather_conflict_review_stream_manager(self, **kwargs): addrs = [] if 'doc' in kwargs: for reldoc in kwargs['doc'].related_that_doc(('conflrev',)): - addrs.extend(Recipient.objects.get(slug='doc_stream_manager').gather(**{'doc':reldoc.document})) + addrs.extend(Recipient.objects.get(slug='doc_stream_manager').gather(**{'doc':reldoc})) return addrs def gather_conflict_review_steering_group(self,**kwargs): addrs = [] if 'doc' in kwargs: for reldoc in kwargs['doc'].related_that_doc(('conflrev',)): - if reldoc.document.stream_id=='irtf': + if reldoc.stream_id=='irtf': addrs.append('"Internet Research Steering Group" ') return addrs @@ -137,14 +141,17 @@ def gather_group_steering_group(self,**kwargs): def gather_stream_managers(self, **kwargs): addrs = [] - manager_map = dict(ise = '', - irtf = '', - ietf = '', - iab = '') + manager_map = dict( + ise = [''], + irtf = [''], + ietf = [''], + iab = [''], + editorial = Role.objects.filter(group__acronym="rsab",name_id="chair").values_list("email__address", flat=True), + ) if 'streams' in kwargs: for stream in kwargs['streams']: if stream in manager_map: - addrs.append(manager_map[stream]) + addrs.extend(manager_map[stream]) return addrs def gather_doc_stream_manager(self, **kwargs): @@ -231,7 +238,7 @@ def gather_submission_submitter(self, **kwargs): try: submitter = Alias.objects.get(name=submission.submitter).person if submitter and submitter.email(): - addrs.extend(["%s <%s>" % (submitter.name, submitter.email().address)]) + addrs.append(f"{submitter.name} <{submitter.email().address}>") except (Alias.DoesNotExist, Alias.MultipleObjectsReturned): pass return addrs diff --git a/ietf/mailtrigger/resources.py b/ietf/mailtrigger/resources.py index eb5466618a..daca055bf4 100644 --- a/ietf/mailtrigger/resources.py +++ b/ietf/mailtrigger/resources.py @@ -7,7 +7,7 @@ from ietf import api -from ietf.mailtrigger.models import Recipient, MailTrigger +from ietf.mailtrigger.models import MailTrigger, Recipient class RecipientResource(ModelResource): @@ -37,3 +37,43 @@ class Meta: } api.mailtrigger.register(MailTriggerResource()) +from ietf.utils.resources import UserResource +class HistoricalMailTriggerResource(ModelResource): + history_user = ToOneField(UserResource, 'history_user', null=True) + class Meta: + queryset = MailTrigger.history.model.objects.all() + serializer = api.Serializer() + cache = SimpleCache() + #resource_name = 'historicalmailtrigger' + ordering = ['history_id', ] + filtering = { + "slug": ALL, + "desc": ALL, + "history_id": ALL, + "history_date": ALL, + "history_change_reason": ALL, + "history_type": ALL, + "history_user": ALL_WITH_RELATIONS, + } +api.mailtrigger.register(HistoricalMailTriggerResource()) + +from ietf.utils.resources import UserResource +class HistoricalRecipientResource(ModelResource): + history_user = ToOneField(UserResource, 'history_user', null=True) + class Meta: + queryset = Recipient.history.model.objects.all() + serializer = api.Serializer() + cache = SimpleCache() + #resource_name = 'historicalrecipient' + ordering = ['history_id', ] + filtering = { + "slug": ALL, + "desc": ALL, + "template": ALL, + "history_id": ALL, + "history_date": ALL, + "history_change_reason": ALL, + "history_type": ALL, + "history_user": ALL_WITH_RELATIONS, + } +api.mailtrigger.register(HistoricalRecipientResource()) diff --git a/ietf/mailtrigger/utils.py b/ietf/mailtrigger/utils.py index 48d91ff6aa..bcdaf5e44e 100644 --- a/ietf/mailtrigger/utils.py +++ b/ietf/mailtrigger/utils.py @@ -1,45 +1,62 @@ -# Copyright The IETF Trust 2015-2019, All Rights Reserved +# Copyright The IETF Trust 2015-2023, All Rights Reserved from collections import namedtuple -import debug # pyflakes:ignore +import debug # pyflakes:ignore from ietf.mailtrigger.models import MailTrigger, Recipient from ietf.submit.models import Submission from ietf.utils.mail import excludeaddrs -class AddrLists(namedtuple('AddrLists',['to','cc'])): - __slots__ = () +EMAIL_ALIASES = { + "IETFCHAIR": "The IETF Chair ", + "IESG": "The IESG ", + "IAB": "The IAB ", + "IABCHAIR": "The IAB Chair ", +} + - def as_strings(self,compact=True): +class AddrLists(namedtuple("AddrLists", ["to", "cc"])): + __slots__ = () + def as_strings(self, compact=True): separator = ", " if compact else ",\n " to_string = separator.join(self.to) cc_string = separator.join(self.cc) - return namedtuple('AddrListsAsStrings',['to','cc'])(to=to_string,cc=cc_string) + return namedtuple("AddrListsAsStrings", ["to", "cc"])( + to=to_string, cc=cc_string + ) -def gather_address_lists(slug, skipped_recipients=None, create_from_slug_if_not_exists=None, - desc_if_not_exists=None, **kwargs): - mailtrigger = get_mailtrigger(slug, create_from_slug_if_not_exists, desc_if_not_exists) +def gather_address_lists( + slug, + skipped_recipients=None, + create_from_slug_if_not_exists=None, + desc_if_not_exists=None, + **kwargs +): + mailtrigger = get_mailtrigger( + slug, create_from_slug_if_not_exists, desc_if_not_exists + ) to = set() for recipient in mailtrigger.to.all(): to.update(recipient.gather(**kwargs)) - to.discard('') + to.discard("") if skipped_recipients: to = excludeaddrs(to, skipped_recipients) cc = set() for recipient in mailtrigger.cc.all(): cc.update(recipient.gather(**kwargs)) - cc.discard('') + cc.discard("") if skipped_recipients: cc = excludeaddrs(cc, skipped_recipients) - return AddrLists(to=sorted(list(to)),cc=sorted(list(cc))) + return AddrLists(to=sorted(list(to)), cc=sorted(list(cc))) + def get_mailtrigger(slug, create_from_slug_if_not_exists, desc_if_not_exists): try: @@ -50,77 +67,158 @@ def get_mailtrigger(slug, create_from_slug_if_not_exists, desc_if_not_exists): mailtrigger = MailTrigger.objects.create(slug=slug, desc=desc_if_not_exists) mailtrigger.to.set(template.to.all()) mailtrigger.cc.set(template.cc.all()) - if slug.startswith('review_completed') and slug.endswith('early'): - mailtrigger.cc.remove('ietf_last_call') + if slug.startswith("review_completed") and slug.endswith("early"): + mailtrigger.cc.remove("ietf_last_call") else: raise return mailtrigger -def gather_relevant_expansions(**kwargs): - - def starts_with(prefix): - return MailTrigger.objects.filter(slug__startswith=prefix).values_list('slug',flat=True) - - relevant = set() - - if 'doc' in kwargs: - - doc = kwargs['doc'] - - relevant.add('doc_state_edited') - - if not doc.type_id in ['bofreq',]: - relevant.update(['doc_telechat_details_changed','ballot_deferred','iesg_ballot_saved']) +def get_contacts_for_liaison_messages_for_group_primary(group): + from ietf.liaisons.views import contact_email_from_role + + '''Returns list of emails to use in liaison message for group + ''' + emails = [] + + # role based emails + if group.acronym in ('ietf','iesg'): + emails.append(EMAIL_ALIASES['IESG']) + emails.append(EMAIL_ALIASES['IETFCHAIR']) + elif group.acronym in ('iab'): + emails.append(EMAIL_ALIASES['IAB']) + emails.append(EMAIL_ALIASES['IABCHAIR']) + elif group.type_id == 'area': + emails.append(EMAIL_ALIASES['IETFCHAIR']) + ad_roles = group.role_set.filter(name='ad') + emails.extend([ contact_email_from_role(r) for r in ad_roles ]) + elif group.type_id == 'wg': + ad_roles = group.parent.role_set.filter(name='ad') + emails.extend([ contact_email_from_role(r) for r in ad_roles ]) + chair_roles = group.role_set.filter(name='chair') + emails.extend([ contact_email_from_role(r) for r in chair_roles ]) + if group.list_email: + emails.append('{} Discussion List <{}>'.format(group.name,group.list_email)) + elif group.type_id == 'sdo': + liaiman_roles = group.role_set.filter(name='liaiman') + emails.extend([ contact_email_from_role(r) for r in liaiman_roles ]) + + # explicit CCs + liaison_cc_roles = group.role_set.filter(name='liaison_cc_contact') + emails.extend([ contact_email_from_role(r) for r in liaison_cc_roles ]) + + return emails + + +def get_contacts_for_liaison_messages_for_group_secondary(group): + from ietf.liaisons.views import contacts_from_roles + + '''Returns default contacts for groups as a comma separated string''' + # use explicit default contacts if defined + explicit_contacts = contacts_from_roles(group.role_set.filter(name='liaison_contact')) + if explicit_contacts: + return explicit_contacts + + # otherwise construct based on group type + contacts = [] + if group.type_id == 'area': + roles = group.role_set.filter(name='ad') + contacts.append(contacts_from_roles(roles)) + elif group.type_id == 'wg': + roles = group.role_set.filter(name='chair') + contacts.append(contacts_from_roles(roles)) + elif group.acronym == 'ietf': + contacts.append(EMAIL_ALIASES['IETFCHAIR']) + elif group.acronym == 'iab': + contacts.append(EMAIL_ALIASES['IABCHAIR']) + elif group.acronym == 'iesg': + contacts.append(EMAIL_ALIASES['IESG']) + + return ','.join(contacts) - if doc.type_id in ['draft','statchg']: - relevant.update(starts_with('last_call_')) - if doc.type_id == 'draft': - relevant.update(starts_with('doc_')) - relevant.update(starts_with('resurrection_')) - relevant.update(['ipr_posted_on_doc',]) - if doc.stream_id == 'ietf': - relevant.update(['ballot_approved_ietf_stream','pubreq_iesg']) +def gather_relevant_expansions(**kwargs): + def starts_with(prefix): + return MailTrigger.objects.filter(slug__startswith=prefix).values_list( + "slug", flat=True + ) + + relevant = set() + + if "doc" in kwargs: + doc = kwargs["doc"] + + relevant.add("doc_state_edited") + + if not doc.type_id in ["bofreq", "statement", "rfc"]: + relevant.update( + ["doc_telechat_details_changed", "ballot_deferred", "iesg_ballot_saved"] + ) + + if doc.type_id in ["draft", "statchg"]: + relevant.update(starts_with("last_call_")) + + if doc.type_id == "rfc": + relevant.update( + [ + "doc_added_comment", + "doc_external_resource_change_requested", + "doc_state_edited", + "ipr_posted_on_doc", + ] + ) + + if doc.type_id == "draft": + relevant.update(starts_with("doc_")) + relevant.update(starts_with("resurrection_")) + relevant.update( + [ + "ipr_posted_on_doc", + ] + ) + if doc.stream_id == "ietf": + relevant.update(["ballot_approved_ietf_stream", "pubreq_iesg"]) else: - relevant.update(['pubreq_rfced']) - last_submission = Submission.objects.filter(name=doc.name,state='posted').order_by('-rev').first() - if last_submission and 'submission' not in kwargs: - kwargs['submission'] = last_submission - - if doc.type_id == 'conflrev': - relevant.update(['conflrev_requested','ballot_approved_conflrev']) - if doc.type_id == 'charter': - relevant.update(['charter_external_review','ballot_approved_charter']) - - if doc.type_id == 'bofreq': - relevant.update(starts_with('bofreq')) - - if 'group' in kwargs: - - relevant.update(starts_with('group_')) - relevant.update(starts_with('milestones_')) - group = kwargs['group'] + relevant.update(["pubreq_rfced"]) + last_submission = ( + Submission.objects.filter(name=doc.name, state="posted") + .order_by("-rev") + .first() + ) + if last_submission and "submission" not in kwargs: + kwargs["submission"] = last_submission + + if doc.type_id == "conflrev": + relevant.update(["conflrev_requested", "ballot_approved_conflrev"]) + if doc.type_id == "charter": + relevant.update(["charter_external_review", "ballot_approved_charter"]) + + if doc.type_id == "bofreq": + relevant.update(starts_with("bofreq")) + + if "group" in kwargs: + relevant.update(starts_with("group_")) + relevant.update(starts_with("milestones_")) + group = kwargs["group"] if group.features.acts_like_wg: - relevant.update(starts_with('session_')) + relevant.update(starts_with("session_")) if group.features.has_chartering_process: - relevant.update(['charter_external_review',]) - - if 'submission' in kwargs: + relevant.update( + [ + "charter_external_review", + ] + ) - relevant.update(starts_with('sub_')) + if "submission" in kwargs: + relevant.update(starts_with("sub_")) rule_list = [] for mailtrigger in MailTrigger.objects.filter(slug__in=relevant): - addrs = gather_address_lists(mailtrigger.slug,**kwargs) + addrs = gather_address_lists(mailtrigger.slug, **kwargs) if addrs.to or addrs.cc: - rule_list.append((mailtrigger.slug,mailtrigger.desc,addrs.to,addrs.cc)) + rule_list.append((mailtrigger.slug, mailtrigger.desc, addrs.to, addrs.cc)) return sorted(rule_list) -def get_base_submission_message_address(): - return Recipient.objects.get(slug='submission_manualpost_handling').gather()[0] def get_base_ipr_request_address(): - return Recipient.objects.get(slug='ipr_requests').gather()[0] - - + return Recipient.objects.get(slug="ipr_requests").gather()[0] diff --git a/ietf/meeting/.gitignore b/ietf/meeting/.gitignore deleted file mode 100644 index ba8bdb3afc..0000000000 --- a/ietf/meeting/.gitignore +++ /dev/null @@ -1,2 +0,0 @@ -/*.pyc -/*.swp diff --git a/ietf/meeting/admin.py b/ietf/meeting/admin.py index 2e2c7f7c22..03abf5c029 100644 --- a/ietf/meeting/admin.py +++ b/ietf/meeting/admin.py @@ -3,11 +3,14 @@ from django.contrib import admin +from django.db.models import Count from ietf.meeting.models import (Attended, Meeting, Room, Session, TimeSlot, Constraint, Schedule, SchedTimeSessAssignment, ResourceAssociation, FloorPlan, UrlResource, SessionPresentation, ImportantDate, SlideSubmission, SchedulingEvent, BusinessConstraint, - ProceedingsMaterial, MeetingHost) + ProceedingsMaterial, MeetingHost, Registration, RegistrationTicket, + AttendanceTypeName) +from ietf.utils.admin import SaferTabularInline class UrlResourceAdmin(admin.ModelAdmin): @@ -16,7 +19,7 @@ class UrlResourceAdmin(admin.ModelAdmin): raw_id_fields = ['room', ] admin.site.register(UrlResource, UrlResourceAdmin) -class UrlResourceInline(admin.TabularInline): +class UrlResourceInline(SaferTabularInline): model = UrlResource class RoomAdmin(admin.ModelAdmin): @@ -26,7 +29,7 @@ class RoomAdmin(admin.ModelAdmin): admin.site.register(Room, RoomAdmin) -class RoomInline(admin.TabularInline): +class RoomInline(SaferTabularInline): model = Room class MeetingAdmin(admin.ModelAdmin): @@ -91,12 +94,14 @@ def name_lower(self, instance): admin.site.register(Constraint, ConstraintAdmin) -class SchedulingEventInline(admin.TabularInline): +class SchedulingEventInline(SaferTabularInline): model = SchedulingEvent raw_id_fields = ["by"] class SessionAdmin(admin.ModelAdmin): - list_display = ["meeting", "name", "group_acronym", "purpose", "attendees", "requested", "current_status"] + list_display = [ + "meeting", "name", "group_acronym", "purpose", "attendees", "has_onsite_tool", "chat_room", "requested", "current_status" + ] list_filter = ["purpose", "meeting", ] raw_id_fields = ["meeting", "group", "materials", "joint_with_groups", "tombstone_for"] search_fields = ["meeting__number", "name", "group__name", "group__acronym", "purpose__name"] @@ -187,7 +192,7 @@ class ImportantDateAdmin(admin.ModelAdmin): class SlideSubmissionAdmin(admin.ModelAdmin): model = SlideSubmission list_display = ['session', 'submitter', 'title'] - raw_id_fields = ['submitter', 'session'] + raw_id_fields = ['submitter', 'session', 'doc'] admin.site.register(SlideSubmission, SlideSubmissionAdmin) @@ -211,3 +216,76 @@ class AttendedAdmin(admin.ModelAdmin): search_fields = ["person__name", "session__group__acronym", "session__meeting__number", "session__name", "session__purpose__name"] raw_id_fields= ["person", "session"] admin.site.register(Attended, AttendedAdmin) + +class MeetingFilter(admin.SimpleListFilter): + title = 'Meeting Filter' + parameter_name = 'meeting_id' + + def lookups(self, request, model_admin): + # only include meetings with registration records + meetings = Meeting.objects.filter(type='ietf').annotate(reg_count=Count('registration')).filter(reg_count__gt=0).order_by('-date') + choices = meetings.values_list('id', 'number') + return choices + + def queryset(self, request, queryset): + if self.value(): + return queryset.filter(meeting__id=self.value()) + return queryset + +class AttendanceFilter(admin.SimpleListFilter): + title = 'Attendance Type' + parameter_name = 'attendance_type' + + def lookups(self, request, model_admin): + choices = AttendanceTypeName.objects.all().values_list('slug', 'name') + return choices + + def queryset(self, request, queryset): + if self.value(): + return queryset.filter(tickets__attendance_type__slug=self.value()).distinct() + return queryset + +class RegistrationTicketInline(SaferTabularInline): + model = RegistrationTicket + +class RegistrationAdmin(admin.ModelAdmin): + model = Registration + list_filter = [AttendanceFilter, MeetingFilter] + list_display = ['meeting', 'first_name', 'last_name', 'display_attendance', 'affiliation', 'country_code', 'email', ] + search_fields = ['first_name', 'last_name', 'affiliation', 'country_code', 'email', ] + raw_id_fields = ['person'] + inlines = [RegistrationTicketInline, ] + ordering = ['-meeting__date', 'last_name'] + + def display_attendance(self, instance): + '''Only display the most significant ticket in the list. + To see all the tickets inspect the individual instance + ''' + if instance.tickets.filter(attendance_type__slug='onsite').exists(): + return 'onsite' + elif instance.tickets.filter(attendance_type__slug='remote').exists(): + return 'remote' + elif instance.tickets.filter(attendance_type__slug='hackathon_onsite').exists(): + return 'hackathon onsite' + elif instance.tickets.filter(attendance_type__slug='hackathon_remote').exists(): + return 'hackathon remote' + display_attendance.short_description = "Attendance" # type: ignore # https://github.com/python/mypy/issues/2087 + +admin.site.register(Registration, RegistrationAdmin) + +class RegistrationTicketAdmin(admin.ModelAdmin): + model = RegistrationTicket + list_filter = ['attendance_type', ] + # not available until Django 5.2, the name of a related field, using the __ notation + # list_display = ['registration__meeting', 'registration', 'attendance_type', 'ticket_type', 'registration__email'] + # list_select_related = ('registration',) + list_display = ['registration', 'attendance_type', 'ticket_type', 'display_meeting'] + search_fields = ['registration__first_name', 'registration__last_name', 'registration__email'] + raw_id_fields = ['registration'] + ordering = ['-registration__meeting__date', 'registration__last_name'] + + def display_meeting(self, instance): + return instance.registration.meeting.number + display_meeting.short_description = "Meeting" # type: ignore # https://github.com/python/mypy/issues/2087 + +admin.site.register(RegistrationTicket, RegistrationTicketAdmin) diff --git a/ietf/meeting/factories.py b/ietf/meeting/factories.py index cf3c87e7c8..fc0ce8387c 100644 --- a/ietf/meeting/factories.py +++ b/ietf/meeting/factories.py @@ -9,9 +9,10 @@ from django.core.files.base import ContentFile from django.db.models import Q +from ietf.doc.storage_utils import store_str from ietf.meeting.models import (Attended, Meeting, Session, SchedulingEvent, Schedule, TimeSlot, SessionPresentation, FloorPlan, Room, SlideSubmission, Constraint, - MeetingHost, ProceedingsMaterial) + MeetingHost, ProceedingsMaterial, Registration, RegistrationTicket) from ietf.name.models import (ConstraintName, SessionStatusName, ProceedingsMaterialTypeName, TimerangeName, SessionPurposeName) from ietf.doc.factories import ProceedingsMaterialDocFactory @@ -23,6 +24,7 @@ class MeetingFactory(factory.django.DjangoModelFactory): class Meta: model = Meeting + skip_postgeneration_save = True type_id = factory.Iterator(['ietf','interim']) @@ -103,6 +105,7 @@ def group_conflicts(obj, create, extracted, **kwargs): # pulint: disable=no-sel class SessionFactory(factory.django.DjangoModelFactory): class Meta: model = Session + skip_postgeneration_save = True meeting = factory.SubFactory(MeetingFactory) purpose_id = 'regular' @@ -110,6 +113,7 @@ class Meta: group = factory.SubFactory(GroupFactory) requested_duration = datetime.timedelta(hours=1) on_agenda = factory.lazy_attribute(lambda obj: SessionPurposeName.objects.get(pk=obj.purpose_id).on_agenda) + has_onsite_tool = factory.lazy_attribute(lambda obj: obj.purpose_id == 'regular') @factory.post_generation def status_id(obj, create, extracted, **kwargs): @@ -155,6 +159,7 @@ class Meta: class RoomFactory(factory.django.DjangoModelFactory): class Meta: model = Room + skip_postgeneration_save = True meeting = factory.SubFactory(MeetingFactory) name = factory.Faker('name') @@ -171,6 +176,7 @@ def session_types(obj, create, extracted, **kwargs): # pylint: disable=no-self-a class TimeSlotFactory(factory.django.DjangoModelFactory): class Meta: model = TimeSlot + skip_postgeneration_save = True meeting = factory.SubFactory(MeetingFactory) type_id = 'regular' @@ -224,6 +230,7 @@ class Meta: class SlideSubmissionFactory(factory.django.DjangoModelFactory): class Meta: model = SlideSubmission + skip_postgeneration_save = True session = factory.SubFactory(SessionFactory) title = factory.Faker('sentence') @@ -233,10 +240,15 @@ class Meta: make_file = factory.PostGeneration( lambda obj, create, extracted, **kwargs: open(obj.staged_filepath(),'a').close() ) + + store_submission = factory.PostGeneration( + lambda obj, create, extracted, **kwargs: store_str("staging", obj.filename, "") + ) class ConstraintFactory(factory.django.DjangoModelFactory): class Meta: model = Constraint + skip_postgeneration_save = True meeting = factory.SubFactory(MeetingFactory) source = factory.SubFactory(GroupFactory) @@ -306,3 +318,48 @@ class Meta: session = factory.SubFactory(SessionFactory) person = factory.SubFactory(PersonFactory) + + +class RegistrationFactory(factory.django.DjangoModelFactory): + """ + This will create an associated onsite week_pass ticket by default. + Methods of calling: + + RegistrationFactory() create a ticket with defaults, onsite + RegistrationFactory(with_ticket=True) same as above + RegistrationFactory(with_ticket={'attendance_type_id': 'remote'}) creates ticket with overrides + RegistrationFactory(with_ticket=False) does not create a ticket + """ + class Meta: + model = Registration + skip_postgeneration_save = True + + meeting = factory.SubFactory(MeetingFactory) + person = factory.SubFactory(PersonFactory) + email = factory.LazyAttribute(lambda obj: obj.person.email()) + first_name = factory.LazyAttribute(lambda obj: obj.person.first_name()) + last_name = factory.LazyAttribute(lambda obj: obj.person.last_name()) + affiliation = factory.Faker('company') + country_code = factory.Faker('country_code') + attended = False + checkedin = False + + @factory.post_generation + def with_ticket(self, create, extracted, **kwargs): + if not create: + return + if extracted is False: + # Explicitly disable ticket creation + return + ticket_kwargs = extracted if isinstance(extracted, dict) else {} + RegistrationTicketFactory(registration=self, **ticket_kwargs) + + +class RegistrationTicketFactory(factory.django.DjangoModelFactory): + class Meta: + model = RegistrationTicket + skip_postgeneration_save = True + + registration = factory.SubFactory(RegistrationFactory) + attendance_type_id = factory.LazyAttribute(lambda _: 'onsite') + ticket_type_id = factory.LazyAttribute(lambda _: 'week_pass') diff --git a/ietf/meeting/forms.py b/ietf/meeting/forms.py index 477a1dbb97..e5b1697f86 100644 --- a/ietf/meeting/forms.py +++ b/ietf/meeting/forms.py @@ -1,4 +1,4 @@ -# Copyright The IETF Trust 2016-2020, All Rights Reserved +# Copyright The IETF Trust 2016-2025, All Rights Reserved # -*- coding: utf-8 -*- @@ -15,27 +15,38 @@ from django.core import validators from django.core.exceptions import ValidationError from django.forms import BaseInlineFormSet +from django.template.defaultfilters import pluralize from django.utils.functional import cached_property +from django.utils.safestring import mark_safe import debug # pyflakes:ignore -from ietf.doc.models import Document, DocAlias, State, NewRevisionDocEvent +from ietf.doc.models import Document, State, NewRevisionDocEvent from ietf.group.models import Group from ietf.group.utils import groups_managed_by -from ietf.meeting.models import Session, Meeting, Schedule, countries, timezones, TimeSlot, Room +from ietf.meeting.models import (Session, Meeting, Schedule, COUNTRIES, TIMEZONES, TimeSlot, Room, + Constraint, ResourceAssociation) from ietf.meeting.helpers import get_next_interim_number, make_materials_directories from ietf.meeting.helpers import is_interim_meeting_approved, get_next_agenda_name from ietf.message.models import Message -from ietf.name.models import TimeSlotTypeName, SessionPurposeName +from ietf.name.models import TimeSlotTypeName, SessionPurposeName, TimerangeName, ConstraintName +from ietf.person.fields import SearchablePersonsField from ietf.person.models import Person -from ietf.utils.fields import DatepickerDateField, DurationField, MultiEmailField, DatepickerSplitDateTimeWidget +from ietf.utils import log +from ietf.utils.fields import ( + DatepickerDateField, + DatepickerSplitDateTimeWidget, + DurationField, + ModelMultipleChoiceField, + MultiEmailField, +) +from ietf.utils.html import clean_text_field from ietf.utils.validators import ( validate_file_size, validate_mime_type, validate_file_extension, validate_no_html_frame) -# need to insert empty option for use in ChoiceField -# countries.insert(0, ('', '-'*9 )) -countries.insert(0, ('', '-' * 9)) -timezones.insert(0, ('', '-' * 9)) +NUM_SESSION_CHOICES = (('', '--Please select'), ('1', '1'), ('2', '2')) +SESSION_TIME_RELATION_CHOICES = (('', 'No preference'),) + Constraint.TIME_RELATION_CHOICES +JOINT_FOR_SESSION_CHOICES = (('1', 'First session'), ('2', 'Second session'), ('3', 'Third session'), ) # ------------------------------------------------- # Helpers @@ -73,6 +84,27 @@ def duration_string(duration): return string +def allowed_conflicting_groups(): + return Group.objects.filter( + type__in=['wg', 'ag', 'rg', 'rag', 'program', 'edwg'], + state__in=['bof', 'proposed', 'active']) + + +def check_conflict(groups, source_group): + ''' + Takes a string which is a list of group acronyms. Checks that they are all active groups + ''' + # convert to python list (allow space or comma separated lists) + items = groups.replace(',', ' ').split() + active_groups = allowed_conflicting_groups() + for group in items: + if group == source_group.acronym: + raise forms.ValidationError("Cannot declare a conflict with the same group: %s" % group) + + if not active_groups.filter(acronym=group): + raise forms.ValidationError("Invalid or inactive group acronym: %s" % group) + + # ------------------------------------------------- # Forms # ------------------------------------------------- @@ -134,12 +166,12 @@ class InterimMeetingModelForm(forms.ModelForm): approved = forms.BooleanField(required=False) city = forms.CharField(max_length=255, required=False) city.widget.attrs['placeholder'] = "City" - country = forms.ChoiceField(choices=countries, required=False) + country = forms.ChoiceField(choices=COUNTRIES, required=False) country.widget.attrs['class'] = "select2-field" country.widget.attrs['data-max-entries'] = 1 country.widget.attrs['data-placeholder'] = "Country" country.widget.attrs['data-minimum-input-length'] = 0 - time_zone = forms.ChoiceField(choices=timezones) + time_zone = forms.ChoiceField(choices=TIMEZONES) time_zone.widget.attrs['class'] = "select2-field" time_zone.widget.attrs['data-max-entries'] = 1 time_zone.widget.attrs['data-minimum-input-length'] = 0 @@ -341,8 +373,7 @@ def save_agenda(self): # FIXME: What about agendas in html or markdown format? uploaded_filename='{}-00.txt'.format(filename)) doc.set_state(State.objects.get(type__slug=doc.type.slug, slug='active')) - DocAlias.objects.create(name=doc.name).docs.add(doc) - self.instance.sessionpresentation_set.create(document=doc, rev=doc.rev) + self.instance.presentations.create(document=doc, rev=doc.rev) NewRevisionDocEvent.objects.create( type='new_revision', by=self.user.person, @@ -356,12 +387,19 @@ def save_agenda(self): os.makedirs(directory) with io.open(path, "w", encoding='utf-8') as file: file.write(self.cleaned_data['agenda']) + doc.store_str(doc.uploaded_filename, self.cleaned_data['agenda']) class InterimAnnounceForm(forms.ModelForm): class Meta: model = Message - fields = ('to', 'frm', 'cc', 'bcc', 'reply_to', 'subject', 'body') + fields = ('to', 'cc', 'frm', 'subject', 'body') + + def __init__(self, *args, **kwargs): + super(InterimAnnounceForm, self).__init__(*args, **kwargs) + self.fields['frm'].label='From' + self.fields['frm'].widget.attrs['readonly'] = True + self.fields['to'].widget.attrs['readonly'] = True def save(self, *args, **kwargs): user = kwargs.pop('user') @@ -375,7 +413,8 @@ def save(self, *args, **kwargs): class InterimCancelForm(forms.Form): group = forms.CharField(max_length=255, required=False) date = forms.DateField(required=False) - comments = forms.CharField(required=False, widget=forms.Textarea(attrs={'placeholder': 'enter optional comments here'}), strip=False) + # max_length must match Session.agenda_note + comments = forms.CharField(max_length=512, required=False, widget=forms.Textarea(attrs={'placeholder': 'enter optional comments here'}), strip=False) def __init__(self, *args, **kwargs): super(InterimCancelForm, self).__init__(*args, **kwargs) @@ -466,6 +505,9 @@ def __init__(self, show_apply_to_all_checkbox, *args, **kwargs): class UploadMinutesForm(ApplyToAllFileUploadForm): doc_type = 'minutes' +class UploadNarrativeMinutesForm(ApplyToAllFileUploadForm): + doc_type = 'narrativeminutes' + class UploadAgendaForm(ApplyToAllFileUploadForm): doc_type = 'agenda' @@ -474,9 +516,12 @@ class UploadAgendaForm(ApplyToAllFileUploadForm): class UploadSlidesForm(ApplyToAllFileUploadForm): doc_type = 'slides' title = forms.CharField(max_length=255) + approved = forms.BooleanField(label='Auto-approve', initial=True, required=False) - def __init__(self, session, *args, **kwargs): - super().__init__(*args, **kwargs) + def __init__(self, session, show_apply_to_all_checkbox, can_manage, *args, **kwargs): + super().__init__(show_apply_to_all_checkbox, *args, **kwargs) + if not can_manage: + self.fields.pop('approved') self.session = session def clean_title(self): @@ -542,7 +587,7 @@ class SwapTimeslotsForm(forms.Form): queryset=TimeSlot.objects.none(), # default to none, fill in when we have a meeting widget=forms.TextInput, ) - rooms = forms.ModelMultipleChoiceField( + rooms = ModelMultipleChoiceField( required=True, queryset=Room.objects.none(), # default to none, fill in when we have a meeting widget=CsvModelPkInput, @@ -608,7 +653,7 @@ class TimeSlotCreateForm(forms.Form): ) duration = TimeSlotDurationField() show_location = forms.BooleanField(required=False, initial=True) - locations = forms.ModelMultipleChoiceField( + locations = ModelMultipleChoiceField( queryset=Room.objects.none(), widget=forms.CheckboxSelectMultiple, ) @@ -727,6 +772,7 @@ def __init__(self, group, *args, **kwargs): 'purpose', session_purposes[0] if len(session_purposes) > 0 else None, ) + kwargs['initial'].setdefault('has_onsite_tool', group.features.acts_like_wg) super().__init__(*args, **kwargs) self.fields['type'].widget.attrs.update({ @@ -738,12 +784,16 @@ def __init__(self, group, *args, **kwargs): self.fields['purpose'].queryset = SessionPurposeName.objects.filter(pk__in=session_purposes) if not group.features.acts_like_wg: self.fields['requested_duration'].durations = [datetime.timedelta(minutes=m) for m in range(30, 241, 30)] + # add bootstrap classes + self.fields['purpose'].widget.attrs.update({'class': 'form-select'}) + self.fields['type'].widget.attrs.update({'class': 'form-select', 'aria-label': 'session type'}) class Meta: model = Session fields = ( 'purpose', 'name', 'short', 'type', 'requested_duration', - 'on_agenda', 'remote_instructions', 'attendees', 'comments', + 'on_agenda', 'agenda_note', 'has_onsite_tool', 'chat_room', 'remote_instructions', + 'attendees', 'comments', ) labels = {'requested_duration': 'Length'} @@ -821,3 +871,296 @@ def sessiondetailsformset_factory(min_num=1, max_num=3): max_num=max_num, extra=max_num, # only creates up to max_num total ) + + +class SessionRequestStatusForm(forms.Form): + message = forms.CharField(widget=forms.Textarea(attrs={'rows': '3', 'cols': '80'}), strip=False) + + +class NameModelMultipleChoiceField(ModelMultipleChoiceField): + def label_from_instance(self, name): + return name.desc + + +class SessionRequestForm(forms.Form): + num_session = forms.ChoiceField( + choices=NUM_SESSION_CHOICES, + label="Number of sessions") + # session fields are added in __init__() + session_time_relation = forms.ChoiceField( + choices=SESSION_TIME_RELATION_CHOICES, + required=False, + label="Time between two sessions") + attendees = forms.IntegerField(label="Number of Attendees") + # FIXME: it would cleaner to have these be + # ModelMultipleChoiceField, and just customize the widgetry, that + # way validation comes for free (applies to this CharField and the + # constraints dynamically instantiated in __init__()) + joint_with_groups = forms.CharField(max_length=255, required=False) + joint_with_groups_selector = forms.ChoiceField(choices=[], required=False) # group select widget for prev field + joint_for_session = forms.ChoiceField(choices=JOINT_FOR_SESSION_CHOICES, required=False) + comments = forms.CharField( + max_length=200, + label='Special Requests', + help_text='i.e. restrictions on meeting times / days, etc. (limit 200 characters)', + required=False) + third_session = forms.BooleanField( + required=False, + help_text="Help") + resources = forms.MultipleChoiceField( + widget=forms.CheckboxSelectMultiple, + required=False, + label='Resources Requested') + bethere = SearchablePersonsField( + label="Participants who must be present", + required=False, + help_text=mark_safe('Do not include Area Directors and WG Chairs; the system already tracks their availability.')) + timeranges = NameModelMultipleChoiceField( + widget=forms.CheckboxSelectMultiple, + required=False, + label=mark_safe('Times during which this WG can not meet:
Please explain any selections in Special Requests below.'), + queryset=TimerangeName.objects.all()) + adjacent_with_wg = forms.ChoiceField( + required=False, + label=mark_safe('Plan session adjacent with another WG:
(Immediately before or after another WG, no break in between, in the same room.)')) + send_notifications = forms.BooleanField(label="Send notification emails?", required=False, initial=False) + + def __init__(self, group, meeting, data=None, *args, **kwargs): + self.hidden = kwargs.pop('hidden', False) + self.notifications_optional = kwargs.pop('notifications_optional', False) + + self.group = group + formset_class = sessiondetailsformset_factory(max_num=3 if group.features.acts_like_wg else 50) + self.session_forms = formset_class(group=self.group, meeting=meeting, data=data) + super().__init__(data=data, *args, **kwargs) + if not self.notifications_optional: + self.fields['send_notifications'].widget = forms.HiddenInput() + + # Allow additional sessions for non-wg-like groups + if not self.group.features.acts_like_wg: + self.fields['num_session'].choices = ((n, str(n)) for n in range(1, 51)) + + self._add_widget_class(self.fields['third_session'].widget, 'form-check-input') + self.fields['comments'].widget = forms.Textarea(attrs={'rows': '3', 'cols': '65'}) + + other_groups = list(allowed_conflicting_groups().exclude(pk=group.pk).values_list('acronym', 'acronym').order_by('acronym')) + self.fields['adjacent_with_wg'].choices = [('', '--No preference')] + other_groups + group_acronym_choices = [('', '--Select WG(s)')] + other_groups + self.fields['joint_with_groups_selector'].choices = group_acronym_choices + + # Set up constraints for the meeting + self._wg_field_data = [] + for constraintname in meeting.group_conflict_types.all(): + # two fields for each constraint: a CharField for the group list and a selector to add entries + constraint_field = forms.CharField(max_length=255, required=False) + constraint_field.widget.attrs['data-slug'] = constraintname.slug + constraint_field.widget.attrs['data-constraint-name'] = str(constraintname).title() + constraint_field.widget.attrs['aria-label'] = f'{constraintname.slug}_input' + self._add_widget_class(constraint_field.widget, 'wg_constraint') + self._add_widget_class(constraint_field.widget, 'form-control') + + selector_field = forms.ChoiceField(choices=group_acronym_choices, required=False) + selector_field.widget.attrs['data-slug'] = constraintname.slug # used by onchange handler + self._add_widget_class(selector_field.widget, 'wg_constraint_selector') + self._add_widget_class(selector_field.widget, 'form-control') + + cfield_id = 'constraint_{}'.format(constraintname.slug) + cselector_id = 'wg_selector_{}'.format(constraintname.slug) + # keep an eye out for field name conflicts + log.assertion('cfield_id not in self.fields') + log.assertion('cselector_id not in self.fields') + self.fields[cfield_id] = constraint_field + self.fields[cselector_id] = selector_field + self._wg_field_data.append((constraintname, cfield_id, cselector_id)) + + # Show constraints that are not actually used by the meeting so these don't get lost + self._inactive_wg_field_data = [] + inactive_cnames = ConstraintName.objects.filter( + is_group_conflict=True # Only collect group conflicts... + ).exclude( + meeting=meeting # ...that are not enabled for this meeting... + ).filter( + constraint__source=group, # ...but exist for this group... + constraint__meeting=meeting, # ... at this meeting. + ).distinct() + + for inactive_constraint_name in inactive_cnames: + field_id = 'delete_{}'.format(inactive_constraint_name.slug) + self.fields[field_id] = forms.BooleanField(required=False, label='Delete this conflict', help_text='Delete this inactive conflict?') + self._add_widget_class(self.fields[field_id].widget, 'form-control') + constraints = group.constraint_source_set.filter(meeting=meeting, name=inactive_constraint_name) + self._inactive_wg_field_data.append( + (inactive_constraint_name, + ' '.join([c.target.acronym for c in constraints]), + field_id) + ) + + self.fields['joint_with_groups_selector'].widget.attrs['onchange'] = "document.form_post.joint_with_groups.value=document.form_post.joint_with_groups.value + ' ' + this.options[this.selectedIndex].value; return 1;" + self.fields["resources"].choices = [(x.pk, x.desc) for x in ResourceAssociation.objects.filter(name__used=True).order_by('name__order')] + + if self.hidden: + # replace all the widgets to start... + for key in list(self.fields.keys()): + self.fields[key].widget = forms.HiddenInput() + # re-replace a couple special cases + self.fields['resources'].widget = forms.MultipleHiddenInput() + self.fields['timeranges'].widget = forms.MultipleHiddenInput() + # and entirely replace bethere - no need to support searching if input is hidden + self.fields['bethere'] = ModelMultipleChoiceField( + widget=forms.MultipleHiddenInput, required=False, + queryset=Person.objects.all(), + ) + + def wg_constraint_fields(self): + """Iterates over wg constraint fields + + Intended for use in the template. + """ + for cname, cfield_id, cselector_id in self._wg_field_data: + yield cname, self[cfield_id], self[cselector_id] + + def wg_constraint_count(self): + """How many wg constraints are there?""" + return len(self._wg_field_data) + + def wg_constraint_field_ids(self): + """Iterates over wg constraint field IDs""" + for cname, cfield_id, _ in self._wg_field_data: + yield cname, cfield_id + + def inactive_wg_constraints(self): + for cname, value, field_id in self._inactive_wg_field_data: + yield cname, value, self[field_id] + + def inactive_wg_constraint_count(self): + return len(self._inactive_wg_field_data) + + def inactive_wg_constraint_field_ids(self): + """Iterates over wg constraint field IDs""" + for cname, _, field_id in self._inactive_wg_field_data: + yield cname, field_id + + @staticmethod + def _add_widget_class(widget, new_class): + """Add a new class, taking care in case some already exist""" + existing_classes = widget.attrs.get('class', '').split() + widget.attrs['class'] = ' '.join(existing_classes + [new_class]) + + def _join_conflicts(self, cleaned_data, slugs): + """Concatenate constraint fields from cleaned data into a single list""" + conflicts = [] + for cname, cfield_id, _ in self._wg_field_data: + if cname.slug in slugs and cfield_id in cleaned_data: + groups = cleaned_data[cfield_id] + # convert to python list (allow space or comma separated lists) + items = groups.replace(',', ' ').split() + conflicts.extend(items) + return conflicts + + def _validate_duplicate_conflicts(self, cleaned_data): + """Validate that no WGs appear in more than one constraint that does not allow duplicates + + Raises ValidationError + """ + # Only the older constraints (conflict, conflic2, conflic3) need to be mutually exclusive. + all_conflicts = self._join_conflicts(cleaned_data, ['conflict', 'conflic2', 'conflic3']) + seen = [] + duplicated = [] + errors = [] + for c in all_conflicts: + if c not in seen: + seen.append(c) + elif c not in duplicated: # only report once + duplicated.append(c) + errors.append(forms.ValidationError('%s appears in conflicts more than once' % c)) + return errors + + def clean_joint_with_groups(self): + groups = self.cleaned_data['joint_with_groups'] + check_conflict(groups, self.group) + return groups + + def clean_comments(self): + return clean_text_field(self.cleaned_data['comments']) + + def clean_bethere(self): + bethere = self.cleaned_data["bethere"] + if bethere: + extra = set( + Person.objects.filter( + role__group=self.group, role__name__in=["chair", "ad"] + ) + & bethere + ) + if extra: + extras = ", ".join(e.name for e in extra) + raise forms.ValidationError( + ( + f"Please remove the following person{pluralize(len(extra))}, the system " + f"tracks their availability due to their role{pluralize(len(extra))}: {extras}." + ) + ) + return bethere + + def clean_send_notifications(self): + return True if not self.notifications_optional else self.cleaned_data['send_notifications'] + + def is_valid(self): + return super().is_valid() and self.session_forms.is_valid() + + def clean(self): + super(SessionRequestForm, self).clean() + self.session_forms.clean() + + data = self.cleaned_data + + # Validate the individual conflict fields + for _, cfield_id, _ in self._wg_field_data: + try: + check_conflict(data[cfield_id], self.group) + except forms.ValidationError as e: + self.add_error(cfield_id, e) + + # Skip remaining tests if individual field tests had errors, + if self.errors: + return data + + # error if conflicts contain disallowed dupes + for error in self._validate_duplicate_conflicts(data): + self.add_error(None, error) + + # Verify expected number of session entries are present + num_sessions_with_data = len(self.session_forms.forms_to_keep) + num_sessions_expected = -1 + try: + num_sessions_expected = int(data.get('num_session', '')) + except ValueError: + self.add_error('num_session', 'Invalid value for number of sessions') + if num_sessions_with_data < num_sessions_expected: + self.add_error('num_session', 'Must provide data for all sessions') + + # if default (empty) option is selected, cleaned_data won't include num_session key + if num_sessions_expected != 2 and num_sessions_expected is not None: + if data.get('session_time_relation'): + self.add_error( + 'session_time_relation', + forms.ValidationError('Time between sessions can only be used when two sessions are requested.') + ) + + joint_session = data.get('joint_for_session', '') + if joint_session != '': + joint_session = int(joint_session) + if joint_session > num_sessions_with_data: + self.add_error( + 'joint_for_session', + forms.ValidationError( + f'Session {joint_session} can not be the joint session, the session has not been requested.' + ) + ) + + return data + + @property + def media(self): + # get media for our formset + return super().media + self.session_forms.media + forms.Media(js=('ietf/js/session_form.js',)) diff --git a/ietf/meeting/helpers.py b/ietf/meeting/helpers.py index 256d10fb79..39d271ae6b 100644 --- a/ietf/meeting/helpers.py +++ b/ietf/meeting/helpers.py @@ -36,12 +36,14 @@ from ietf.utils.text import xslugify -def get_meeting(num=None,type_in=['ietf',],days=28): +def get_meeting(num=None, type_in=('ietf',), days=28): meetings = Meeting.objects - if type_in: + if type_in is not None: meetings = meetings.filter(type__in=type_in) - if num == None: - meetings = meetings.filter(date__gte=timezone.now()-datetime.timedelta(days=days)).order_by('date') + if num is None: + meetings = meetings.filter( + date__gte=timezone.now() - datetime.timedelta(days=days) + ).order_by('date') else: meetings = meetings.filter(number=num) if meetings.exists(): @@ -102,7 +104,7 @@ def preprocess_assignments_for_agenda(assignments_queryset, meeting, extra_prefe queryset=add_event_info_to_session_qs(Session.objects.all().prefetch_related( 'group', 'group__charter', 'group__charter__group', Prefetch('materials', - queryset=Document.objects.exclude(states__type=F("type"), states__slug='deleted').order_by('sessionpresentation__order').prefetch_related('states'), + queryset=Document.objects.exclude(states__type=F("type"), states__slug='deleted').order_by('presentations__order').prefetch_related('states'), to_attr='prefetched_active_materials' ) )) @@ -124,7 +126,7 @@ def preprocess_assignments_for_agenda(assignments_queryset, meeting, extra_prefe # check before blindly assigning to meeting just in case. if a.session.meeting.pk == meeting.pk: a.session.meeting = meeting - a.session.order_number = None + a.session.order_number = a.session.order_in_meeting() if a.session.group else None if a.session.group and a.session.group not in groups: groups.append(a.session.group) @@ -134,12 +136,6 @@ def preprocess_assignments_for_agenda(assignments_queryset, meeting, extra_prefe if a.session and a.session.group: sessions_for_groups[(a.session.group, a.session.type_id)].append(a) - for a in assignments: - if a.session and a.session.group: - - l = sessions_for_groups.get((a.session.group, a.session.type_id), []) - a.session.order_number = l.index(a) + 1 if a in l else 0 - timeslot_by_session_pk = {a.session_id: a.timeslot for a in assignments} for a in assignments: @@ -321,10 +317,21 @@ def _group_filter_headings(self): groups = set(self._get_group(s) for s in self.sessions if s and self._get_group(s)) - log.assertion('len(groups) == len(set(g.acronym for g in groups))') # no repeated acros + # Verify that we're not using the same acronym for more than one distinct group, accounting for + # the possibility that some groups are GroupHistory instances. This assertion will fail if a Group + # and GroupHistory for the same group have a different acronym - in that event, the filter will + # not match the meeting display, so we should be alerted that this has actually occurred. + log.assertion( + "len(set(getattr(g, 'group_id', g.id) for g in groups)) " + "== len(set(g.acronym for g in groups))" + ) group_parents = set(self._get_group_parent(g) for g in groups if self._get_group_parent(g)) - log.assertion('len(group_parents) == len(set(gp.acronym for gp in group_parents))') # no repeated acros + # See above for explanation of this assertion + log.assertion( + "len(set(getattr(gp, 'group_id', gp.id) for gp in group_parents)) " + "== len(set(gp.acronym for gp in group_parents))" + ) all_groups = groups.union(group_parents) all_groups.difference_update([g for g in all_groups if g.acronym in self.exclude_acronyms]) @@ -642,6 +649,11 @@ def read_session_file(type, num, doc): def read_agenda_file(num, doc): return read_session_file('agenda', num, doc) +# TODO-BLOBSTORE: this is _yet another_ draft derived variant created when users +# ask for drafts from the meeting agenda page. Consider whether to refactor this +# now to not call out to external binaries, and consider whether we need this extra +# format at all in the draft blobstore. if so, it would probably be stored under +# something like plainpdf/ def convert_draft_to_pdf(doc_name): inpath = os.path.join(settings.IDSUBMIT_REPOSITORY_PATH, doc_name + ".txt") outpath = os.path.join(settings.INTERNET_DRAFT_PDF_PATH, doc_name + ".pdf") @@ -830,7 +842,7 @@ def get_announcement_initial(meeting, is_change=False): desc=desc, date=meeting.date, change=change) - body = render_to_string('meeting/interim_announcement.txt', locals()) + body = render_to_string('meeting/interim_announcement.txt', locals() | {"settings": settings}) initial['body'] = body return initial @@ -883,7 +895,7 @@ def make_materials_directories(meeting): # was merged with the regular datatracker code; then in secr/proceedings/views.py # in make_directories()) saved_umask = os.umask(0) - for leaf in ('slides','agenda','minutes','id','rfc','bluesheets'): + for leaf in ('slides','agenda','minutes', 'narrativeminutes', 'id','rfc','bluesheets'): target = os.path.join(path,leaf) if not os.path.exists(target): os.makedirs(target) @@ -1092,6 +1104,7 @@ def create_interim_session_conferences(sessions): try: confs = meetecho_manager.create( group=session.group, + session_id=session.pk, description=str(session), start_time=ts.utc_start_time(), duration=ts.duration, diff --git a/ietf/meeting/management/commands/create_test_meeting.py b/ietf/meeting/management/commands/create_test_meeting.py index c857d6da8a..e48a8d66a2 100644 --- a/ietf/meeting/management/commands/create_test_meeting.py +++ b/ietf/meeting/management/commands/create_test_meeting.py @@ -14,7 +14,7 @@ # can be translated to the newly expanded Constraint objects. # # This work was done in the context of the new meeting constraints modelling: -# https://trac.ietf.org/trac/ietfdb/wiki/MeetingConstraints +# https://github.com/ietf-tools/datatracker/wiki/MeetingConstraints # Note that aside from Constraint objects, as created below, there is also # business logic that applies to all sessions, which is to be implemented # in the automatic schedule builder. diff --git a/ietf/meeting/management/commands/generate_schedule.py b/ietf/meeting/management/commands/generate_schedule.py index c0645eed65..c9279fb251 100644 --- a/ietf/meeting/management/commands/generate_schedule.py +++ b/ietf/meeting/management/commands/generate_schedule.py @@ -1,6 +1,6 @@ # Copyright The IETF Trust 2021, All Rights Reserved # For an overview of this process and context, see: -# https://trac.ietf.org/trac/ietfdb/wiki/MeetingConstraints +# https://github.com/ietf-tools/datatracker/wiki/MeetingConstraints from __future__ import absolute_import, print_function, unicode_literals import calendar @@ -25,6 +25,8 @@ from ietf.person.models import Person from ietf.meeting import models from ietf.meeting.helpers import get_person_by_email +from ietf.name.models import SessionPurposeName + # 40 runs of the optimiser for IETF 106 with cycles=160 resulted in 16 # zero-violation invocations, with a mean number of runs of 91 and @@ -72,18 +74,31 @@ def add_arguments(self, parser): 'Base schedule for generated schedule, specified as "[owner/]name"' ' (default is no base schedule; owner not required if name is unique)' )) + parser.add_argument('-p', '--purpose', + dest='purposes', + action='append', + choices=[ + spn.slug for spn in SessionPurposeName.objects.all() + if 'regular' in spn.timeslot_types # scheduler only works with "regular" timeslots + ], + default=None, + help=( + 'Limit scheduling to specified purpose ' + '(use option multiple times to specify more than one purpose; default is all purposes)' + )) - def handle(self, meeting, name, max_cycles, verbosity, base_id, *args, **kwargs): - ScheduleHandler(self.stdout, meeting, name, max_cycles, verbosity, base_id).run() + def handle(self, meeting, name, max_cycles, verbosity, base_id, purposes, *args, **kwargs): + ScheduleHandler(self.stdout, meeting, name, max_cycles, verbosity, base_id, purposes).run() class ScheduleHandler(object): def __init__(self, stdout, meeting_number, name=None, max_cycles=OPTIMISER_MAX_CYCLES, - verbosity=1, base_id=None): + verbosity=1, base_id=None, session_purposes=None): self.stdout = stdout self.verbosity = verbosity self.name = name self.max_cycles = max_cycles + self.session_purposes = session_purposes if meeting_number: try: self.meeting = models.Meeting.objects.get(type="ietf", number=meeting_number) @@ -114,6 +129,10 @@ def __init__(self, stdout, meeting_number, name=None, max_cycles=OPTIMISER_MAX_C msgs.append('Applying schedule {} as base schedule'.format(ScheduleId.from_schedule(self.base_schedule))) self.stdout.write('\n{}\n\n'.format('\n'.join(msgs))) self._load_meeting() + if len(self.schedule.sessions) == 0: + raise CommandError('No sessions found to schedule') + if len(self.schedule.timeslots) == 0: + raise CommandError('No timeslots found for schedule') def run(self): """Schedule all sessions""" @@ -194,6 +213,8 @@ def _sessions_to_schedule(self, *args, **kwargs): Extra arguments are passed to the Session constructor. """ sessions_db = self.meeting.session_set.that_can_be_scheduled().filter(type_id='regular') + if self.session_purposes is not None: + sessions_db = sessions_db.filter(purpose__slug__in=self.session_purposes) if self.base_schedule is None: fixed_sessions = models.Session.objects.none() else: diff --git a/ietf/meeting/management/commands/populate_attended.py b/ietf/meeting/management/commands/populate_attended.py deleted file mode 100644 index fac41ab78c..0000000000 --- a/ietf/meeting/management/commands/populate_attended.py +++ /dev/null @@ -1,85 +0,0 @@ -# Copyright The IETF Trust 2022, All Rights Reserved - -import debug # pyflakes: ignore - -from tqdm import tqdm - -from django.core.management.base import BaseCommand - -from ietf.meeting.models import Session -from ietf.meeting.utils import sort_sessions -from ietf.person.models import Person, Email - -import json - -class Command(BaseCommand): - - help = 'Populates the meeting Attended table based on bluesheets and registration information' - - def add_arguments(self, parser): - parser.add_argument('filename', nargs='+', type=str) - - def handle(self, *args, **options): - - issues = [] - session_cache = dict() - skipped = 0 - for filename in options['filename']: - records = json.loads(open(filename,'r').read()) - for record in tqdm(records): - user = record['sub'] - session_acronym = record['group'] - meeting_number = record['meeting'] - email = record['email'] - # In the expected dumps from MeetEcho, if there was only one session for group foo, it would just be named 'foo'. - # If there were _three_, we would see 'foo' for the first, 'foo_2' for the second, and 'foo_3' for the third. - # order below is the index into what is returned from sort_sessions -- 0 is the first session for a group at that meeting. - # There is brutal fixup below for older meetings where we had special arrangements where meetecho reported the non-existent - # group of 'plenary', mapping it into the appropriate 'ietf' group session. - # A bug in the export scripts at MeetEcho trimmed the '-t' from 'model-t'. - order = 0 - if session_acronym in ['anrw_test', 'demoanrw', 'hostspeaker']: - skipped = skipped + 1 - continue - if session_acronym=='model': - session_acronym='model-t' - if '_' in session_acronym: - session_acronym, order = session_acronym.split('_') - order = int(order)-1 - if session_acronym == 'plenary': - session_acronym = 'ietf' - if meeting_number == '111': - order = 4 - elif meeting_number == '110': - order = 3 - elif meeting_number == '109': - order = 6 - elif meeting_number == '108': - order = 13 - if session_acronym == 'ietf': - if meeting_number == '112': - order = 2 - elif meeting_number == '113': - order = 2 - if not (meeting_number, session_acronym) in session_cache: - session_cache[(meeting_number, session_acronym)] = sort_sessions([s for s in Session.objects.filter(meeting__number=meeting_number,group__acronym=session_acronym) if s.official_timeslotassignment()]) - sessions = session_cache[(meeting_number, session_acronym)] - try: - session = sessions[order] - except IndexError: - issues.append(('session not found',record)) - continue - person = None - email = Email.objects.filter(address=email).first() - if email: - person = email.person - else: - person = Person.objects.filter(user__pk=user).first() - if not person: - issues.append(('person not found',record)) - continue - obj, created = session.attended_set.get_or_create(person=person) - for issue in issues: - print(issue) - print(f'{len(issues)} issues encountered') - print(f'{skipped} records intentionally skipped') diff --git a/ietf/meeting/migrations/0001_initial.py b/ietf/meeting/migrations/0001_initial.py index 668fd0392a..d98c5619d3 100644 --- a/ietf/meeting/migrations/0001_initial.py +++ b/ietf/meeting/migrations/0001_initial.py @@ -1,15 +1,15 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-20 10:52 - +# Generated by Django 2.2.28 on 2023-03-20 19:22 import datetime import django.core.validators from django.db import migrations, models import django.db.models.deletion +import django.utils.timezone import ietf.meeting.models +import ietf.utils.fields import ietf.utils.models import ietf.utils.storage +import ietf.utils.validators class Migration(migrations.Migration): @@ -17,19 +17,20 @@ class Migration(migrations.Migration): initial = True dependencies = [ - ('group', '0001_initial'), - ('name', '0001_initial'), ('dbtemplate', '0001_initial'), + ('name', '0001_initial'), ('person', '0001_initial'), + ('group', '0001_initial'), ('doc', '0001_initial'), ] operations = [ migrations.CreateModel( - name='Constraint', + name='BusinessConstraint', fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('day', models.DateTimeField(blank=True, null=True)), + ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), + ('name', models.CharField(max_length=255)), + ('penalty', models.IntegerField(default=0, help_text='The penalty for violating this kind of constraint; for instance 10 (small penalty) or 10000 (large penalty)')), ], ), migrations.CreateModel( @@ -37,20 +38,13 @@ class Migration(migrations.Migration): fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=255)), - ('short', models.CharField(default='', max_length=2)), - ('time', models.DateTimeField(default=datetime.datetime.now)), + ('short', models.CharField(default='', max_length=3)), + ('modified', models.DateTimeField(auto_now=True)), ('order', models.SmallIntegerField()), ('image', models.ImageField(blank=True, default=None, storage=ietf.utils.storage.NoLocationMigrationFileSystemStorage(location=None), upload_to=ietf.meeting.models.floorplan_path)), ], - ), - migrations.CreateModel( - name='ImportantDate', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('date', models.DateField()), - ], options={ - 'ordering': ['-meeting', 'date'], + 'ordering': ['-id'], }, ), migrations.CreateModel( @@ -61,12 +55,12 @@ class Migration(migrations.Migration): ('date', models.DateField()), ('days', models.IntegerField(default=7, help_text='The number of days the meeting lasts', validators=[django.core.validators.MinValueValidator(1)])), ('city', models.CharField(blank=True, max_length=255)), - ('country', models.CharField(blank=True, choices=[('', ''), ('AF', 'Afghanistan'), ('AL', 'Albania'), ('DZ', 'Algeria'), ('AD', 'Andorra'), ('AO', 'Angola'), ('AI', 'Anguilla'), ('AQ', 'Antarctica'), ('AG', 'Antigua & Barbuda'), ('AR', 'Argentina'), ('AM', 'Armenia'), ('AW', 'Aruba'), ('AU', 'Australia'), ('AT', 'Austria'), ('AZ', 'Azerbaijan'), ('BS', 'Bahamas'), ('BH', 'Bahrain'), ('BD', 'Bangladesh'), ('B', 'Barbados'), ('BY', 'Belarus'), ('BE', 'Belgium'), ('BZ', 'Belize'), ('BJ', 'Benin'), ('BM', 'Bermuda'), ('BT', 'Bhutan'), ('BO', 'Bolivia'), ('BA', 'Bosnia & Herzegovina'), ('BW', 'Botswana'), ('BV', 'Bouvet Island'), ('BR', 'Brazil'), ('G', 'Britain (UK)'), ('IO', 'British Indian Ocean Territory'), ('BN', 'Brunei'), ('BG', 'Bulgaria'), ('BF', 'Burkina Faso'), ('BI', 'Burundi'), ('KH', 'Cambodia'), ('CM', 'Cameroon'), ('CA', 'Canada'), ('CV', 'Cape Verde'), ('BQ', 'Caribbean NL'), ('KY', 'Cayman Islands'), ('CF', 'Central African Rep.'), ('TD', 'Chad'), ('CL', 'Chile'), ('CN', 'China'), ('CX', 'Christmas Island'), ('CC', 'Cocos (Keeling) Islands'), ('CO', 'Colombia'), ('KM', 'Comoros'), ('CD', 'Congo (Dem. Rep.)'), ('CG', 'Congo (Rep.)'), ('CK', 'Cook Islands'), ('CR', 'Costa Rica'), ('HR', 'Croatia'), ('CU', 'Cuba'), ('CW', 'Cura\xe7ao'), ('CY', 'Cyprus'), ('CZ', 'Czech Republic'), ('CI', "C\xf4te d'Ivoire"), ('DK', 'Denmark'), ('DJ', 'Djibouti'), ('DM', 'Dominica'), ('DO', 'Dominican Republic'), ('TL', 'East Timor'), ('EC', 'Ecuador'), ('EG', 'Egypt'), ('SV', 'El Salvador'), ('GQ', 'Equatorial Guinea'), ('ER', 'Eritrea'), ('EE', 'Estonia'), ('ET', 'Ethiopia'), ('FK', 'Falkland Islands'), ('FO', 'Faroe Islands'), ('FJ', 'Fiji'), ('FI', 'Finland'), ('FR', 'France'), ('GF', 'French Guiana'), ('PF', 'French Polynesia'), ('TF', 'French Southern & Antarctic Lands'), ('GA', 'Gabon'), ('GM', 'Gambia'), ('GE', 'Georgia'), ('DE', 'Germany'), ('GH', 'Ghana'), ('GI', 'Gibraltar'), ('GR', 'Greece'), ('GL', 'Greenland'), ('GD', 'Grenada'), ('GP', 'Guadeloupe'), ('GU', 'Guam'), ('GT', 'Guatemala'), ('GG', 'Guernsey'), ('GN', 'Guinea'), ('GW', 'Guinea-Bissau'), ('GY', 'Guyana'), ('HT', 'Haiti'), ('HM', 'Heard Island & McDonald Islands'), ('HN', 'Honduras'), ('HK', 'Hong Kong'), ('HU', 'Hungary'), ('IS', 'Iceland'), ('IN', 'India'), ('ID', 'Indonesia'), ('IR', 'Iran'), ('IQ', 'Iraq'), ('IE', 'Ireland'), ('IM', 'Isle of Man'), ('IL', 'Israel'), ('IT', 'Italy'), ('JM', 'Jamaica'), ('JP', 'Japan'), ('JE', 'Jersey'), ('JO', 'Jordan'), ('KZ', 'Kazakhstan'), ('KE', 'Kenya'), ('KI', 'Kiribati'), ('KP', 'Korea (North)'), ('KR', 'Korea (South)'), ('KW', 'Kuwait'), ('KG', 'Kyrgyzstan'), ('LA', 'Laos'), ('LV', 'Latvia'), ('L', 'Lebanon'), ('LS', 'Lesotho'), ('LR', 'Liberia'), ('LY', 'Libya'), ('LI', 'Liechtenstein'), ('LT', 'Lithuania'), ('LU', 'Luxembourg'), ('MO', 'Macau'), ('MK', 'Macedonia'), ('MG', 'Madagascar'), ('MW', 'Malawi'), ('MY', 'Malaysia'), ('MV', 'Maldives'), ('ML', 'Mali'), ('MT', 'Malta'), ('MH', 'Marshall Islands'), ('MQ', 'Martinique'), ('MR', 'Mauritania'), ('MU', 'Mauritius'), ('YT', 'Mayotte'), ('MX', 'Mexico'), ('FM', 'Micronesia'), ('MD', 'Moldova'), ('MC', 'Monaco'), ('MN', 'Mongolia'), ('ME', 'Montenegro'), ('MS', 'Montserrat'), ('MA', 'Morocco'), ('MZ', 'Mozambique'), ('MM', 'Myanmar (Burma)'), ('NA', 'Namibia'), ('NR', 'Nauru'), ('NP', 'Nepal'), ('NL', 'Netherlands'), ('NC', 'New Caledonia'), ('NZ', 'New Zealand'), ('NI', 'Nicaragua'), ('NE', 'Niger'), ('NG', 'Nigeria'), ('NU', 'Niue'), ('NF', 'Norfolk Island'), ('MP', 'Northern Mariana Islands'), ('NO', 'Norway'), ('OM', 'Oman'), ('PK', 'Pakistan'), ('PW', 'Palau'), ('PS', 'Palestine'), ('PA', 'Panama'), ('PG', 'Papua New Guinea'), ('PY', 'Paraguay'), ('PE', 'Peru'), ('PH', 'Philippines'), ('PN', 'Pitcairn'), ('PL', 'Poland'), ('PT', 'Portugal'), ('PR', 'Puerto Rico'), ('QA', 'Qatar'), ('RO', 'Romania'), ('RU', 'Russia'), ('RW', 'Rwanda'), ('RE', 'R\xe9union'), ('AS', 'Samoa (American)'), ('WS', 'Samoa (western)'), ('SM', 'San Marino'), ('ST', 'Sao Tome & Principe'), ('SA', 'Saudi Arabia'), ('SN', 'Senegal'), ('RS', 'Serbia'), ('SC', 'Seychelles'), ('SL', 'Sierra Leone'), ('SG', 'Singapore'), ('SK', 'Slovakia'), ('SI', 'Slovenia'), ('S', 'Solomon Islands'), ('SO', 'Somalia'), ('ZA', 'South Africa'), ('GS', 'South Georgia & the South Sandwich Islands'), ('SS', 'South Sudan'), ('ES', 'Spain'), ('LK', 'Sri Lanka'), ('BL', 'St Barthelemy'), ('SH', 'St Helena'), ('KN', 'St Kitts & Nevis'), ('LC', 'St Lucia'), ('SX', 'St Maarten (Dutch)'), ('MF', 'St Martin (French)'), ('PM', 'St Pierre & Miquelon'), ('VC', 'St Vincent'), ('SD', 'Sudan'), ('SR', 'Suriname'), ('SJ', 'Svalbard & Jan Mayen'), ('SZ', 'Swaziland'), ('SE', 'Sweden'), ('CH', 'Switzerland'), ('SY', 'Syria'), ('TW', 'Taiwan'), ('TJ', 'Tajikistan'), ('TZ', 'Tanzania'), ('TH', 'Thailand'), ('TG', 'Togo'), ('TK', 'Tokelau'), ('TO', 'Tonga'), ('TT', 'Trinidad & Tobago'), ('TN', 'Tunisia'), ('TR', 'Turkey'), ('TM', 'Turkmenistan'), ('TC', 'Turks & Caicos Is'), ('TV', 'Tuvalu'), ('UM', 'US minor outlying islands'), ('UG', 'Uganda'), ('UA', 'Ukraine'), ('AE', 'United Arab Emirates'), ('US', 'United States'), ('UY', 'Uruguay'), ('UZ', 'Uzbekistan'), ('VU', 'Vanuatu'), ('VA', 'Vatican City'), ('VE', 'Venezuela'), ('VN', 'Vietnam'), ('VG', 'Virgin Islands (UK)'), ('VI', 'Virgin Islands (US)'), ('WF', 'Wallis & Futuna'), ('EH', 'Western Sahara'), ('YE', 'Yemen'), ('ZM', 'Zambia'), ('ZW', 'Zimbabwe'), ('AX', '\xc5land Islands')], max_length=2)), - ('time_zone', models.CharField(blank=True, choices=[('', '---------'), ('Africa/Abidjan', 'Africa/Abidjan'), ('Africa/Accra', 'Africa/Accra'), ('Africa/Addis_Ababa', 'Africa/Addis_Ababa'), ('Africa/Algiers', 'Africa/Algiers'), ('Africa/Asmara', 'Africa/Asmara'), ('Africa/Bamako', 'Africa/Bamako'), ('Africa/Bangui', 'Africa/Bangui'), ('Africa/Banjul', 'Africa/Banjul'), ('Africa/Bissau', 'Africa/Bissau'), ('Africa/Blantyre', 'Africa/Blantyre'), ('Africa/Brazzaville', 'Africa/Brazzaville'), ('Africa/Bujumbura', 'Africa/Bujumbura'), ('Africa/Cairo', 'Africa/Cairo'), ('Africa/Casablanca', 'Africa/Casablanca'), ('Africa/Ceuta', 'Africa/Ceuta'), ('Africa/Conakry', 'Africa/Conakry'), ('Africa/Dakar', 'Africa/Dakar'), ('Africa/Dar_es_Salaam', 'Africa/Dar_es_Salaam'), ('Africa/Djibouti', 'Africa/Djibouti'), ('Africa/Douala', 'Africa/Douala'), ('Africa/El_Aaiun', 'Africa/El_Aaiun'), ('Africa/Freetown', 'Africa/Freetown'), ('Africa/Gaborone', 'Africa/Gaborone'), ('Africa/Harare', 'Africa/Harare'), ('Africa/Johannesburg', 'Africa/Johannesburg'), ('Africa/Juba', 'Africa/Juba'), ('Africa/Kampala', 'Africa/Kampala'), ('Africa/Khartoum', 'Africa/Khartoum'), ('Africa/Kigali', 'Africa/Kigali'), ('Africa/Kinshasa', 'Africa/Kinshasa'), ('Africa/Lagos', 'Africa/Lagos'), ('Africa/Libreville', 'Africa/Libreville'), ('Africa/Lome', 'Africa/Lome'), ('Africa/Luanda', 'Africa/Luanda'), ('Africa/Lubumbashi', 'Africa/Lubumbashi'), ('Africa/Lusaka', 'Africa/Lusaka'), ('Africa/Malabo', 'Africa/Malabo'), ('Africa/Maputo', 'Africa/Maputo'), ('Africa/Maseru', 'Africa/Maseru'), ('Africa/Mbabane', 'Africa/Mbabane'), ('Africa/Mogadishu', 'Africa/Mogadishu'), ('Africa/Monrovia', 'Africa/Monrovia'), ('Africa/Nairobi', 'Africa/Nairobi'), ('Africa/Ndjamena', 'Africa/Ndjamena'), ('Africa/Niamey', 'Africa/Niamey'), ('Africa/Nouakchott', 'Africa/Nouakchott'), ('Africa/Ouagadougou', 'Africa/Ouagadougou'), ('Africa/Porto-Novo', 'Africa/Porto-Novo'), ('Africa/Sao_Tome', 'Africa/Sao_Tome'), ('Africa/Tripoli', 'Africa/Tripoli'), ('Africa/Tunis', 'Africa/Tunis'), ('Africa/Windhoek', 'Africa/Windhoek'), ('America/Adak', 'America/Adak'), ('America/Anchorage', 'America/Anchorage'), ('America/Anguilla', 'America/Anguilla'), ('America/Antigua', 'America/Antigua'), ('America/Araguaina', 'America/Araguaina'), ('America/Argentina/Buenos_Aires', 'America/Argentina/Buenos_Aires'), ('America/Argentina/Catamarca', 'America/Argentina/Catamarca'), ('America/Argentina/Cordoba', 'America/Argentina/Cordoba'), ('America/Argentina/Jujuy', 'America/Argentina/Jujuy'), ('America/Argentina/La_Rioja', 'America/Argentina/La_Rioja'), ('America/Argentina/Mendoza', 'America/Argentina/Mendoza'), ('America/Argentina/Rio_Gallegos', 'America/Argentina/Rio_Gallegos'), ('America/Argentina/Salta', 'America/Argentina/Salta'), ('America/Argentina/San_Juan', 'America/Argentina/San_Juan'), ('America/Argentina/San_Luis', 'America/Argentina/San_Luis'), ('America/Argentina/Tucuman', 'America/Argentina/Tucuman'), ('America/Argentina/Ushuaia', 'America/Argentina/Ushuaia'), ('America/Aruba', 'America/Aruba'), ('America/Asuncion', 'America/Asuncion'), ('America/Atikokan', 'America/Atikokan'), ('America/Bahia', 'America/Bahia'), ('America/Bahia_Banderas', 'America/Bahia_Banderas'), ('America/Barbados', 'America/Barbados'), ('America/Belem', 'America/Belem'), ('America/Belize', 'America/Belize'), ('America/Blanc-Sablon', 'America/Blanc-Sablon'), ('America/Boa_Vista', 'America/Boa_Vista'), ('America/Bogota', 'America/Bogota'), ('America/Boise', 'America/Boise'), ('America/Cambridge_Bay', 'America/Cambridge_Bay'), ('America/Campo_Grande', 'America/Campo_Grande'), ('America/Cancun', 'America/Cancun'), ('America/Caracas', 'America/Caracas'), ('America/Cayenne', 'America/Cayenne'), ('America/Cayman', 'America/Cayman'), ('America/Chicago', 'America/Chicago'), ('America/Chihuahua', 'America/Chihuahua'), ('America/Costa_Rica', 'America/Costa_Rica'), ('America/Creston', 'America/Creston'), ('America/Cuiaba', 'America/Cuiaba'), ('America/Curacao', 'America/Curacao'), ('America/Danmarkshavn', 'America/Danmarkshavn'), ('America/Dawson', 'America/Dawson'), ('America/Dawson_Creek', 'America/Dawson_Creek'), ('America/Denver', 'America/Denver'), ('America/Detroit', 'America/Detroit'), ('America/Dominica', 'America/Dominica'), ('America/Edmonton', 'America/Edmonton'), ('America/Eirunepe', 'America/Eirunepe'), ('America/El_Salvador', 'America/El_Salvador'), ('America/Fort_Nelson', 'America/Fort_Nelson'), ('America/Fortaleza', 'America/Fortaleza'), ('America/Glace_Bay', 'America/Glace_Bay'), ('America/Godtha', 'America/Godtha'), ('America/Goose_Bay', 'America/Goose_Bay'), ('America/Grand_Turk', 'America/Grand_Turk'), ('America/Grenada', 'America/Grenada'), ('America/Guadeloupe', 'America/Guadeloupe'), ('America/Guatemala', 'America/Guatemala'), ('America/Guayaquil', 'America/Guayaquil'), ('America/Guyana', 'America/Guyana'), ('America/Halifax', 'America/Halifax'), ('America/Havana', 'America/Havana'), ('America/Hermosillo', 'America/Hermosillo'), ('America/Indiana/Indianapolis', 'America/Indiana/Indianapolis'), ('America/Indiana/Knox', 'America/Indiana/Knox'), ('America/Indiana/Marengo', 'America/Indiana/Marengo'), ('America/Indiana/Petersburg', 'America/Indiana/Petersburg'), ('America/Indiana/Tell_City', 'America/Indiana/Tell_City'), ('America/Indiana/Vevay', 'America/Indiana/Vevay'), ('America/Indiana/Vincennes', 'America/Indiana/Vincennes'), ('America/Indiana/Winamac', 'America/Indiana/Winamac'), ('America/Inuvik', 'America/Inuvik'), ('America/Iqaluit', 'America/Iqaluit'), ('America/Jamaica', 'America/Jamaica'), ('America/Juneau', 'America/Juneau'), ('America/Kentucky/Louisville', 'America/Kentucky/Louisville'), ('America/Kentucky/Monticello', 'America/Kentucky/Monticello'), ('America/Kralendijk', 'America/Kralendijk'), ('America/La_Paz', 'America/La_Paz'), ('America/Lima', 'America/Lima'), ('America/Los_Angeles', 'America/Los_Angeles'), ('America/Lower_Princes', 'America/Lower_Princes'), ('America/Maceio', 'America/Maceio'), ('America/Managua', 'America/Managua'), ('America/Manaus', 'America/Manaus'), ('America/Marigot', 'America/Marigot'), ('America/Martinique', 'America/Martinique'), ('America/Matamoros', 'America/Matamoros'), ('America/Mazatlan', 'America/Mazatlan'), ('America/Menominee', 'America/Menominee'), ('America/Merida', 'America/Merida'), ('America/Metlakatla', 'America/Metlakatla'), ('America/Mexico_City', 'America/Mexico_City'), ('America/Miquelon', 'America/Miquelon'), ('America/Moncton', 'America/Moncton'), ('America/Monterrey', 'America/Monterrey'), ('America/Montevideo', 'America/Montevideo'), ('America/Montserrat', 'America/Montserrat'), ('America/Nassau', 'America/Nassau'), ('America/New_York', 'America/New_York'), ('America/Nipigon', 'America/Nipigon'), ('America/Nome', 'America/Nome'), ('America/Noronha', 'America/Noronha'), ('America/North_Dakota/Beulah', 'America/North_Dakota/Beulah'), ('America/North_Dakota/Center', 'America/North_Dakota/Center'), ('America/North_Dakota/New_Salem', 'America/North_Dakota/New_Salem'), ('America/Ojinaga', 'America/Ojinaga'), ('America/Panama', 'America/Panama'), ('America/Pangnirtung', 'America/Pangnirtung'), ('America/Paramaribo', 'America/Paramaribo'), ('America/Phoenix', 'America/Phoenix'), ('America/Port-au-Prince', 'America/Port-au-Prince'), ('America/Port_of_Spain', 'America/Port_of_Spain'), ('America/Porto_Velho', 'America/Porto_Velho'), ('America/Puerto_Rico', 'America/Puerto_Rico'), ('America/Punta_Arenas', 'America/Punta_Arenas'), ('America/Rainy_River', 'America/Rainy_River'), ('America/Rankin_Inlet', 'America/Rankin_Inlet'), ('America/Recife', 'America/Recife'), ('America/Regina', 'America/Regina'), ('America/Resolute', 'America/Resolute'), ('America/Rio_Branco', 'America/Rio_Branco'), ('America/Santarem', 'America/Santarem'), ('America/Santiago', 'America/Santiago'), ('America/Santo_Domingo', 'America/Santo_Domingo'), ('America/Sao_Paulo', 'America/Sao_Paulo'), ('America/Scoresbysund', 'America/Scoresbysund'), ('America/Sitka', 'America/Sitka'), ('America/St_Barthelemy', 'America/St_Barthelemy'), ('America/St_Johns', 'America/St_Johns'), ('America/St_Kitts', 'America/St_Kitts'), ('America/St_Lucia', 'America/St_Lucia'), ('America/St_Thomas', 'America/St_Thomas'), ('America/St_Vincent', 'America/St_Vincent'), ('America/Swift_Current', 'America/Swift_Current'), ('America/Tegucigalpa', 'America/Tegucigalpa'), ('America/Thule', 'America/Thule'), ('America/Thunder_Bay', 'America/Thunder_Bay'), ('America/Tijuana', 'America/Tijuana'), ('America/Toronto', 'America/Toronto'), ('America/Tortola', 'America/Tortola'), ('America/Vancouver', 'America/Vancouver'), ('America/Whitehorse', 'America/Whitehorse'), ('America/Winnipeg', 'America/Winnipeg'), ('America/Yakutat', 'America/Yakutat'), ('America/Yellowknife', 'America/Yellowknife'), ('Antarctica/Casey', 'Antarctica/Casey'), ('Antarctica/Davis', 'Antarctica/Davis'), ('Antarctica/DumontDUrville', 'Antarctica/DumontDUrville'), ('Antarctica/Macquarie', 'Antarctica/Macquarie'), ('Antarctica/Mawson', 'Antarctica/Mawson'), ('Antarctica/McMurdo', 'Antarctica/McMurdo'), ('Antarctica/Palmer', 'Antarctica/Palmer'), ('Antarctica/Rothera', 'Antarctica/Rothera'), ('Antarctica/Syowa', 'Antarctica/Syowa'), ('Antarctica/Troll', 'Antarctica/Troll'), ('Antarctica/Vostok', 'Antarctica/Vostok'), ('Arctic/Longyearbyen', 'Arctic/Longyearbyen'), ('Asia/Aden', 'Asia/Aden'), ('Asia/Almaty', 'Asia/Almaty'), ('Asia/Amman', 'Asia/Amman'), ('Asia/Anadyr', 'Asia/Anadyr'), ('Asia/Aqtau', 'Asia/Aqtau'), ('Asia/Aqtobe', 'Asia/Aqtobe'), ('Asia/Ashgabat', 'Asia/Ashgabat'), ('Asia/Atyrau', 'Asia/Atyrau'), ('Asia/Baghdad', 'Asia/Baghdad'), ('Asia/Bahrain', 'Asia/Bahrain'), ('Asia/Baku', 'Asia/Baku'), ('Asia/Bangkok', 'Asia/Bangkok'), ('Asia/Barnaul', 'Asia/Barnaul'), ('Asia/Beirut', 'Asia/Beirut'), ('Asia/Bishkek', 'Asia/Bishkek'), ('Asia/Brunei', 'Asia/Brunei'), ('Asia/Chita', 'Asia/Chita'), ('Asia/Choibalsan', 'Asia/Choibalsan'), ('Asia/Colombo', 'Asia/Colombo'), ('Asia/Damascus', 'Asia/Damascus'), ('Asia/Dhaka', 'Asia/Dhaka'), ('Asia/Dili', 'Asia/Dili'), ('Asia/Dubai', 'Asia/Dubai'), ('Asia/Dushanbe', 'Asia/Dushanbe'), ('Asia/Famagusta', 'Asia/Famagusta'), ('Asia/Gaza', 'Asia/Gaza'), ('Asia/Hebron', 'Asia/Hebron'), ('Asia/Ho_Chi_Minh', 'Asia/Ho_Chi_Minh'), ('Asia/Hong_Kong', 'Asia/Hong_Kong'), ('Asia/Hovd', 'Asia/Hovd'), ('Asia/Irkutsk', 'Asia/Irkutsk'), ('Asia/Jakarta', 'Asia/Jakarta'), ('Asia/Jayapura', 'Asia/Jayapura'), ('Asia/Jerusalem', 'Asia/Jerusalem'), ('Asia/Kabul', 'Asia/Kabul'), ('Asia/Kamchatka', 'Asia/Kamchatka'), ('Asia/Karachi', 'Asia/Karachi'), ('Asia/Kathmandu', 'Asia/Kathmandu'), ('Asia/Khandyga', 'Asia/Khandyga'), ('Asia/Kolkata', 'Asia/Kolkata'), ('Asia/Krasnoyarsk', 'Asia/Krasnoyarsk'), ('Asia/Kuala_Lumpur', 'Asia/Kuala_Lumpur'), ('Asia/Kuching', 'Asia/Kuching'), ('Asia/Kuwait', 'Asia/Kuwait'), ('Asia/Macau', 'Asia/Macau'), ('Asia/Magadan', 'Asia/Magadan'), ('Asia/Makassar', 'Asia/Makassar'), ('Asia/Manila', 'Asia/Manila'), ('Asia/Muscat', 'Asia/Muscat'), ('Asia/Nicosia', 'Asia/Nicosia'), ('Asia/Novokuznetsk', 'Asia/Novokuznetsk'), ('Asia/Novosibirsk', 'Asia/Novosibirsk'), ('Asia/Omsk', 'Asia/Omsk'), ('Asia/Oral', 'Asia/Oral'), ('Asia/Phnom_Penh', 'Asia/Phnom_Penh'), ('Asia/Pontianak', 'Asia/Pontianak'), ('Asia/Pyongyang', 'Asia/Pyongyang'), ('Asia/Qatar', 'Asia/Qatar'), ('Asia/Qyzylorda', 'Asia/Qyzylorda'), ('Asia/Riyadh', 'Asia/Riyadh'), ('Asia/Sakhalin', 'Asia/Sakhalin'), ('Asia/Samarkand', 'Asia/Samarkand'), ('Asia/Seoul', 'Asia/Seoul'), ('Asia/Shanghai', 'Asia/Shanghai'), ('Asia/Singapore', 'Asia/Singapore'), ('Asia/Srednekolymsk', 'Asia/Srednekolymsk'), ('Asia/Taipei', 'Asia/Taipei'), ('Asia/Tashkent', 'Asia/Tashkent'), ('Asia/Tbilisi', 'Asia/Tbilisi'), ('Asia/Tehran', 'Asia/Tehran'), ('Asia/Thimphu', 'Asia/Thimphu'), ('Asia/Tokyo', 'Asia/Tokyo'), ('Asia/Tomsk', 'Asia/Tomsk'), ('Asia/Ulaanbaatar', 'Asia/Ulaanbaatar'), ('Asia/Urumqi', 'Asia/Urumqi'), ('Asia/Ust-Nera', 'Asia/Ust-Nera'), ('Asia/Vientiane', 'Asia/Vientiane'), ('Asia/Vladivostok', 'Asia/Vladivostok'), ('Asia/Yakutsk', 'Asia/Yakutsk'), ('Asia/Yangon', 'Asia/Yangon'), ('Asia/Yekaterinburg', 'Asia/Yekaterinburg'), ('Asia/Yerevan', 'Asia/Yerevan'), ('Atlantic/Azores', 'Atlantic/Azores'), ('Atlantic/Bermuda', 'Atlantic/Bermuda'), ('Atlantic/Canary', 'Atlantic/Canary'), ('Atlantic/Cape_Verde', 'Atlantic/Cape_Verde'), ('Atlantic/Faroe', 'Atlantic/Faroe'), ('Atlantic/Madeira', 'Atlantic/Madeira'), ('Atlantic/Reykjavik', 'Atlantic/Reykjavik'), ('Atlantic/South_Georgia', 'Atlantic/South_Georgia'), ('Atlantic/St_Helena', 'Atlantic/St_Helena'), ('Atlantic/Stanley', 'Atlantic/Stanley'), ('Australia/Adelaide', 'Australia/Adelaide'), ('Australia/Brisbane', 'Australia/Brisbane'), ('Australia/Broken_Hill', 'Australia/Broken_Hill'), ('Australia/Currie', 'Australia/Currie'), ('Australia/Darwin', 'Australia/Darwin'), ('Australia/Eucla', 'Australia/Eucla'), ('Australia/Hobart', 'Australia/Hobart'), ('Australia/Lindeman', 'Australia/Lindeman'), ('Australia/Lord_Howe', 'Australia/Lord_Howe'), ('Australia/Melbourne', 'Australia/Melbourne'), ('Australia/Perth', 'Australia/Perth'), ('Australia/Sydney', 'Australia/Sydney'), ('Canada/Atlantic', 'Canada/Atlantic'), ('Canada/Central', 'Canada/Central'), ('Canada/Eastern', 'Canada/Eastern'), ('Canada/Mountain', 'Canada/Mountain'), ('Canada/Newfoundland', 'Canada/Newfoundland'), ('Canada/Pacific', 'Canada/Pacific'), ('Europe/Amsterdam', 'Europe/Amsterdam'), ('Europe/Andorra', 'Europe/Andorra'), ('Europe/Astrakhan', 'Europe/Astrakhan'), ('Europe/Athens', 'Europe/Athens'), ('Europe/Belgrade', 'Europe/Belgrade'), ('Europe/Berlin', 'Europe/Berlin'), ('Europe/Bratislava', 'Europe/Bratislava'), ('Europe/Brussels', 'Europe/Brussels'), ('Europe/Bucharest', 'Europe/Bucharest'), ('Europe/Budapest', 'Europe/Budapest'), ('Europe/Busingen', 'Europe/Busingen'), ('Europe/Chisinau', 'Europe/Chisinau'), ('Europe/Copenhagen', 'Europe/Copenhagen'), ('Europe/Dublin', 'Europe/Dublin'), ('Europe/Gibraltar', 'Europe/Gibraltar'), ('Europe/Guernsey', 'Europe/Guernsey'), ('Europe/Helsinki', 'Europe/Helsinki'), ('Europe/Isle_of_Man', 'Europe/Isle_of_Man'), ('Europe/Istanbul', 'Europe/Istanbul'), ('Europe/Jersey', 'Europe/Jersey'), ('Europe/Kaliningrad', 'Europe/Kaliningrad'), ('Europe/Kiev', 'Europe/Kiev'), ('Europe/Kirov', 'Europe/Kirov'), ('Europe/Lisbon', 'Europe/Lisbon'), ('Europe/Ljubljana', 'Europe/Ljubljana'), ('Europe/London', 'Europe/London'), ('Europe/Luxembourg', 'Europe/Luxembourg'), ('Europe/Madrid', 'Europe/Madrid'), ('Europe/Malta', 'Europe/Malta'), ('Europe/Mariehamn', 'Europe/Mariehamn'), ('Europe/Minsk', 'Europe/Minsk'), ('Europe/Monaco', 'Europe/Monaco'), ('Europe/Moscow', 'Europe/Moscow'), ('Europe/Oslo', 'Europe/Oslo'), ('Europe/Paris', 'Europe/Paris'), ('Europe/Podgorica', 'Europe/Podgorica'), ('Europe/Prague', 'Europe/Prague'), ('Europe/Riga', 'Europe/Riga'), ('Europe/Rome', 'Europe/Rome'), ('Europe/Samara', 'Europe/Samara'), ('Europe/San_Marino', 'Europe/San_Marino'), ('Europe/Sarajevo', 'Europe/Sarajevo'), ('Europe/Saratov', 'Europe/Saratov'), ('Europe/Simferopol', 'Europe/Simferopol'), ('Europe/Skopje', 'Europe/Skopje'), ('Europe/Sofia', 'Europe/Sofia'), ('Europe/Stockholm', 'Europe/Stockholm'), ('Europe/Tallinn', 'Europe/Tallinn'), ('Europe/Tirane', 'Europe/Tirane'), ('Europe/Ulyanovsk', 'Europe/Ulyanovsk'), ('Europe/Uzhgorod', 'Europe/Uzhgorod'), ('Europe/Vaduz', 'Europe/Vaduz'), ('Europe/Vatican', 'Europe/Vatican'), ('Europe/Vienna', 'Europe/Vienna'), ('Europe/Vilnius', 'Europe/Vilnius'), ('Europe/Volgograd', 'Europe/Volgograd'), ('Europe/Warsaw', 'Europe/Warsaw'), ('Europe/Zagre', 'Europe/Zagre'), ('Europe/Zaporozhye', 'Europe/Zaporozhye'), ('Europe/Zurich', 'Europe/Zurich'), ('GMT', 'GMT'), ('Indian/Antananarivo', 'Indian/Antananarivo'), ('Indian/Chagos', 'Indian/Chagos'), ('Indian/Christmas', 'Indian/Christmas'), ('Indian/Cocos', 'Indian/Cocos'), ('Indian/Comoro', 'Indian/Comoro'), ('Indian/Kerguelen', 'Indian/Kerguelen'), ('Indian/Mahe', 'Indian/Mahe'), ('Indian/Maldives', 'Indian/Maldives'), ('Indian/Mauritius', 'Indian/Mauritius'), ('Indian/Mayotte', 'Indian/Mayotte'), ('Indian/Reunion', 'Indian/Reunion'), ('Pacific/Apia', 'Pacific/Apia'), ('Pacific/Auckland', 'Pacific/Auckland'), ('Pacific/Bougainville', 'Pacific/Bougainville'), ('Pacific/Chatham', 'Pacific/Chatham'), ('Pacific/Chuuk', 'Pacific/Chuuk'), ('Pacific/Easter', 'Pacific/Easter'), ('Pacific/Efate', 'Pacific/Efate'), ('Pacific/Enderbury', 'Pacific/Enderbury'), ('Pacific/Fakaofo', 'Pacific/Fakaofo'), ('Pacific/Fiji', 'Pacific/Fiji'), ('Pacific/Funafuti', 'Pacific/Funafuti'), ('Pacific/Galapagos', 'Pacific/Galapagos'), ('Pacific/Gambier', 'Pacific/Gambier'), ('Pacific/Guadalcanal', 'Pacific/Guadalcanal'), ('Pacific/Guam', 'Pacific/Guam'), ('Pacific/Honolulu', 'Pacific/Honolulu'), ('Pacific/Kiritimati', 'Pacific/Kiritimati'), ('Pacific/Kosrae', 'Pacific/Kosrae'), ('Pacific/Kwajalein', 'Pacific/Kwajalein'), ('Pacific/Majuro', 'Pacific/Majuro'), ('Pacific/Marquesas', 'Pacific/Marquesas'), ('Pacific/Midway', 'Pacific/Midway'), ('Pacific/Nauru', 'Pacific/Nauru'), ('Pacific/Niue', 'Pacific/Niue'), ('Pacific/Norfolk', 'Pacific/Norfolk'), ('Pacific/Noumea', 'Pacific/Noumea'), ('Pacific/Pago_Pago', 'Pacific/Pago_Pago'), ('Pacific/Palau', 'Pacific/Palau'), ('Pacific/Pitcairn', 'Pacific/Pitcairn'), ('Pacific/Pohnpei', 'Pacific/Pohnpei'), ('Pacific/Port_Moresby', 'Pacific/Port_Moresby'), ('Pacific/Rarotonga', 'Pacific/Rarotonga'), ('Pacific/Saipan', 'Pacific/Saipan'), ('Pacific/Tahiti', 'Pacific/Tahiti'), ('Pacific/Tarawa', 'Pacific/Tarawa'), ('Pacific/Tongatapu', 'Pacific/Tongatapu'), ('Pacific/Wake', 'Pacific/Wake'), ('Pacific/Wallis', 'Pacific/Wallis'), ('US/Alaska', 'US/Alaska'), ('US/Arizona', 'US/Arizona'), ('US/Central', 'US/Central'), ('US/Eastern', 'US/Eastern'), ('US/Hawaii', 'US/Hawaii'), ('US/Mountain', 'US/Mountain'), ('US/Pacific', 'US/Pacific'), ('UTC', 'UTC')], max_length=255)), + ('country', models.CharField(blank=True, choices=[('', '---------'), ('AF', 'Afghanistan'), ('AL', 'Albania'), ('DZ', 'Algeria'), ('AD', 'Andorra'), ('AO', 'Angola'), ('AI', 'Anguilla'), ('AQ', 'Antarctica'), ('AG', 'Antigua & Barbuda'), ('AR', 'Argentina'), ('AM', 'Armenia'), ('AW', 'Aruba'), ('AU', 'Australia'), ('AT', 'Austria'), ('AZ', 'Azerbaijan'), ('BS', 'Bahamas'), ('BH', 'Bahrain'), ('BD', 'Bangladesh'), ('BB', 'Barbados'), ('BY', 'Belarus'), ('BE', 'Belgium'), ('BZ', 'Belize'), ('BJ', 'Benin'), ('BM', 'Bermuda'), ('BT', 'Bhutan'), ('BO', 'Bolivia'), ('BA', 'Bosnia & Herzegovina'), ('BW', 'Botswana'), ('BV', 'Bouvet Island'), ('BR', 'Brazil'), ('GB', 'Britain (UK)'), ('IO', 'British Indian Ocean Territory'), ('BN', 'Brunei'), ('BG', 'Bulgaria'), ('BF', 'Burkina Faso'), ('BI', 'Burundi'), ('KH', 'Cambodia'), ('CM', 'Cameroon'), ('CA', 'Canada'), ('CV', 'Cape Verde'), ('BQ', 'Caribbean NL'), ('KY', 'Cayman Islands'), ('CF', 'Central African Rep.'), ('TD', 'Chad'), ('CL', 'Chile'), ('CN', 'China'), ('CX', 'Christmas Island'), ('CC', 'Cocos (Keeling) Islands'), ('CO', 'Colombia'), ('KM', 'Comoros'), ('CD', 'Congo (Dem. Rep.)'), ('CG', 'Congo (Rep.)'), ('CK', 'Cook Islands'), ('CR', 'Costa Rica'), ('HR', 'Croatia'), ('CU', 'Cuba'), ('CW', 'Curaçao'), ('CY', 'Cyprus'), ('CZ', 'Czech Republic'), ('CI', "Côte d'Ivoire"), ('DK', 'Denmark'), ('DJ', 'Djibouti'), ('DM', 'Dominica'), ('DO', 'Dominican Republic'), ('TL', 'East Timor'), ('EC', 'Ecuador'), ('EG', 'Egypt'), ('SV', 'El Salvador'), ('GQ', 'Equatorial Guinea'), ('ER', 'Eritrea'), ('EE', 'Estonia'), ('SZ', 'Eswatini (Swaziland)'), ('ET', 'Ethiopia'), ('FK', 'Falkland Islands'), ('FO', 'Faroe Islands'), ('FJ', 'Fiji'), ('FI', 'Finland'), ('FR', 'France'), ('GF', 'French Guiana'), ('PF', 'French Polynesia'), ('TF', 'French Southern & Antarctic Lands'), ('GA', 'Gabon'), ('GM', 'Gambia'), ('GE', 'Georgia'), ('DE', 'Germany'), ('GH', 'Ghana'), ('GI', 'Gibraltar'), ('GR', 'Greece'), ('GL', 'Greenland'), ('GD', 'Grenada'), ('GP', 'Guadeloupe'), ('GU', 'Guam'), ('GT', 'Guatemala'), ('GG', 'Guernsey'), ('GN', 'Guinea'), ('GW', 'Guinea-Bissau'), ('GY', 'Guyana'), ('HT', 'Haiti'), ('HM', 'Heard Island & McDonald Islands'), ('HN', 'Honduras'), ('HK', 'Hong Kong'), ('HU', 'Hungary'), ('IS', 'Iceland'), ('IN', 'India'), ('ID', 'Indonesia'), ('IR', 'Iran'), ('IQ', 'Iraq'), ('IE', 'Ireland'), ('IM', 'Isle of Man'), ('IL', 'Israel'), ('IT', 'Italy'), ('JM', 'Jamaica'), ('JP', 'Japan'), ('JE', 'Jersey'), ('JO', 'Jordan'), ('KZ', 'Kazakhstan'), ('KE', 'Kenya'), ('KI', 'Kiribati'), ('KP', 'Korea (North)'), ('KR', 'Korea (South)'), ('KW', 'Kuwait'), ('KG', 'Kyrgyzstan'), ('LA', 'Laos'), ('LV', 'Latvia'), ('LB', 'Lebanon'), ('LS', 'Lesotho'), ('LR', 'Liberia'), ('LY', 'Libya'), ('LI', 'Liechtenstein'), ('LT', 'Lithuania'), ('LU', 'Luxembourg'), ('MO', 'Macau'), ('MG', 'Madagascar'), ('MW', 'Malawi'), ('MY', 'Malaysia'), ('MV', 'Maldives'), ('ML', 'Mali'), ('MT', 'Malta'), ('MH', 'Marshall Islands'), ('MQ', 'Martinique'), ('MR', 'Mauritania'), ('MU', 'Mauritius'), ('YT', 'Mayotte'), ('MX', 'Mexico'), ('FM', 'Micronesia'), ('MD', 'Moldova'), ('MC', 'Monaco'), ('MN', 'Mongolia'), ('ME', 'Montenegro'), ('MS', 'Montserrat'), ('MA', 'Morocco'), ('MZ', 'Mozambique'), ('MM', 'Myanmar (Burma)'), ('NA', 'Namibia'), ('NR', 'Nauru'), ('NP', 'Nepal'), ('NL', 'Netherlands'), ('NC', 'New Caledonia'), ('NZ', 'New Zealand'), ('NI', 'Nicaragua'), ('NE', 'Niger'), ('NG', 'Nigeria'), ('NU', 'Niue'), ('NF', 'Norfolk Island'), ('MK', 'North Macedonia'), ('MP', 'Northern Mariana Islands'), ('NO', 'Norway'), ('OM', 'Oman'), ('PK', 'Pakistan'), ('PW', 'Palau'), ('PS', 'Palestine'), ('PA', 'Panama'), ('PG', 'Papua New Guinea'), ('PY', 'Paraguay'), ('PE', 'Peru'), ('PH', 'Philippines'), ('PN', 'Pitcairn'), ('PL', 'Poland'), ('PT', 'Portugal'), ('PR', 'Puerto Rico'), ('QA', 'Qatar'), ('RO', 'Romania'), ('RU', 'Russia'), ('RW', 'Rwanda'), ('RE', 'Réunion'), ('AS', 'Samoa (American)'), ('WS', 'Samoa (western)'), ('SM', 'San Marino'), ('ST', 'Sao Tome & Principe'), ('SA', 'Saudi Arabia'), ('SN', 'Senegal'), ('RS', 'Serbia'), ('SC', 'Seychelles'), ('SL', 'Sierra Leone'), ('SG', 'Singapore'), ('SK', 'Slovakia'), ('SI', 'Slovenia'), ('SB', 'Solomon Islands'), ('SO', 'Somalia'), ('ZA', 'South Africa'), ('GS', 'South Georgia & the South Sandwich Islands'), ('SS', 'South Sudan'), ('ES', 'Spain'), ('LK', 'Sri Lanka'), ('BL', 'St Barthelemy'), ('SH', 'St Helena'), ('KN', 'St Kitts & Nevis'), ('LC', 'St Lucia'), ('SX', 'St Maarten (Dutch)'), ('MF', 'St Martin (French)'), ('PM', 'St Pierre & Miquelon'), ('VC', 'St Vincent'), ('SD', 'Sudan'), ('SR', 'Suriname'), ('SJ', 'Svalbard & Jan Mayen'), ('SE', 'Sweden'), ('CH', 'Switzerland'), ('SY', 'Syria'), ('TW', 'Taiwan'), ('TJ', 'Tajikistan'), ('TZ', 'Tanzania'), ('TH', 'Thailand'), ('TG', 'Togo'), ('TK', 'Tokelau'), ('TO', 'Tonga'), ('TT', 'Trinidad & Tobago'), ('TN', 'Tunisia'), ('TR', 'Turkey'), ('TM', 'Turkmenistan'), ('TC', 'Turks & Caicos Is'), ('TV', 'Tuvalu'), ('UM', 'US minor outlying islands'), ('UG', 'Uganda'), ('UA', 'Ukraine'), ('AE', 'United Arab Emirates'), ('US', 'United States'), ('UY', 'Uruguay'), ('UZ', 'Uzbekistan'), ('VU', 'Vanuatu'), ('VA', 'Vatican City'), ('VE', 'Venezuela'), ('VN', 'Vietnam'), ('VG', 'Virgin Islands (UK)'), ('VI', 'Virgin Islands (US)'), ('WF', 'Wallis & Futuna'), ('EH', 'Western Sahara'), ('YE', 'Yemen'), ('ZM', 'Zambia'), ('ZW', 'Zimbabwe'), ('AX', 'Åland Islands')], max_length=2)), + ('time_zone', models.CharField(choices=[('', '---------'), ('Africa/Abidjan', 'Africa/Abidjan'), ('Africa/Accra', 'Africa/Accra'), ('Africa/Addis_Ababa', 'Africa/Addis_Ababa'), ('Africa/Algiers', 'Africa/Algiers'), ('Africa/Asmara', 'Africa/Asmara'), ('Africa/Bamako', 'Africa/Bamako'), ('Africa/Bangui', 'Africa/Bangui'), ('Africa/Banjul', 'Africa/Banjul'), ('Africa/Bissau', 'Africa/Bissau'), ('Africa/Blantyre', 'Africa/Blantyre'), ('Africa/Brazzaville', 'Africa/Brazzaville'), ('Africa/Bujumbura', 'Africa/Bujumbura'), ('Africa/Cairo', 'Africa/Cairo'), ('Africa/Casablanca', 'Africa/Casablanca'), ('Africa/Ceuta', 'Africa/Ceuta'), ('Africa/Conakry', 'Africa/Conakry'), ('Africa/Dakar', 'Africa/Dakar'), ('Africa/Dar_es_Salaam', 'Africa/Dar_es_Salaam'), ('Africa/Djibouti', 'Africa/Djibouti'), ('Africa/Douala', 'Africa/Douala'), ('Africa/El_Aaiun', 'Africa/El_Aaiun'), ('Africa/Freetown', 'Africa/Freetown'), ('Africa/Gaborone', 'Africa/Gaborone'), ('Africa/Harare', 'Africa/Harare'), ('Africa/Johannesburg', 'Africa/Johannesburg'), ('Africa/Juba', 'Africa/Juba'), ('Africa/Kampala', 'Africa/Kampala'), ('Africa/Khartoum', 'Africa/Khartoum'), ('Africa/Kigali', 'Africa/Kigali'), ('Africa/Kinshasa', 'Africa/Kinshasa'), ('Africa/Lagos', 'Africa/Lagos'), ('Africa/Libreville', 'Africa/Libreville'), ('Africa/Lome', 'Africa/Lome'), ('Africa/Luanda', 'Africa/Luanda'), ('Africa/Lubumbashi', 'Africa/Lubumbashi'), ('Africa/Lusaka', 'Africa/Lusaka'), ('Africa/Malabo', 'Africa/Malabo'), ('Africa/Maputo', 'Africa/Maputo'), ('Africa/Maseru', 'Africa/Maseru'), ('Africa/Mbabane', 'Africa/Mbabane'), ('Africa/Mogadishu', 'Africa/Mogadishu'), ('Africa/Monrovia', 'Africa/Monrovia'), ('Africa/Nairobi', 'Africa/Nairobi'), ('Africa/Ndjamena', 'Africa/Ndjamena'), ('Africa/Niamey', 'Africa/Niamey'), ('Africa/Nouakchott', 'Africa/Nouakchott'), ('Africa/Ouagadougou', 'Africa/Ouagadougou'), ('Africa/Porto-Novo', 'Africa/Porto-Novo'), ('Africa/Sao_Tome', 'Africa/Sao_Tome'), ('Africa/Tripoli', 'Africa/Tripoli'), ('Africa/Tunis', 'Africa/Tunis'), ('Africa/Windhoek', 'Africa/Windhoek'), ('America/Adak', 'America/Adak'), ('America/Anchorage', 'America/Anchorage'), ('America/Anguilla', 'America/Anguilla'), ('America/Antigua', 'America/Antigua'), ('America/Araguaina', 'America/Araguaina'), ('America/Argentina/Buenos_Aires', 'America/Argentina/Buenos_Aires'), ('America/Argentina/Catamarca', 'America/Argentina/Catamarca'), ('America/Argentina/Cordoba', 'America/Argentina/Cordoba'), ('America/Argentina/Jujuy', 'America/Argentina/Jujuy'), ('America/Argentina/La_Rioja', 'America/Argentina/La_Rioja'), ('America/Argentina/Mendoza', 'America/Argentina/Mendoza'), ('America/Argentina/Rio_Gallegos', 'America/Argentina/Rio_Gallegos'), ('America/Argentina/Salta', 'America/Argentina/Salta'), ('America/Argentina/San_Juan', 'America/Argentina/San_Juan'), ('America/Argentina/San_Luis', 'America/Argentina/San_Luis'), ('America/Argentina/Tucuman', 'America/Argentina/Tucuman'), ('America/Argentina/Ushuaia', 'America/Argentina/Ushuaia'), ('America/Aruba', 'America/Aruba'), ('America/Asuncion', 'America/Asuncion'), ('America/Atikokan', 'America/Atikokan'), ('America/Bahia', 'America/Bahia'), ('America/Bahia_Banderas', 'America/Bahia_Banderas'), ('America/Barbados', 'America/Barbados'), ('America/Belem', 'America/Belem'), ('America/Belize', 'America/Belize'), ('America/Blanc-Sablon', 'America/Blanc-Sablon'), ('America/Boa_Vista', 'America/Boa_Vista'), ('America/Bogota', 'America/Bogota'), ('America/Boise', 'America/Boise'), ('America/Cambridge_Bay', 'America/Cambridge_Bay'), ('America/Campo_Grande', 'America/Campo_Grande'), ('America/Cancun', 'America/Cancun'), ('America/Caracas', 'America/Caracas'), ('America/Cayenne', 'America/Cayenne'), ('America/Cayman', 'America/Cayman'), ('America/Chicago', 'America/Chicago'), ('America/Chihuahua', 'America/Chihuahua'), ('America/Costa_Rica', 'America/Costa_Rica'), ('America/Creston', 'America/Creston'), ('America/Cuiaba', 'America/Cuiaba'), ('America/Curacao', 'America/Curacao'), ('America/Danmarkshavn', 'America/Danmarkshavn'), ('America/Dawson', 'America/Dawson'), ('America/Dawson_Creek', 'America/Dawson_Creek'), ('America/Denver', 'America/Denver'), ('America/Detroit', 'America/Detroit'), ('America/Dominica', 'America/Dominica'), ('America/Edmonton', 'America/Edmonton'), ('America/Eirunepe', 'America/Eirunepe'), ('America/El_Salvador', 'America/El_Salvador'), ('America/Fort_Nelson', 'America/Fort_Nelson'), ('America/Fortaleza', 'America/Fortaleza'), ('America/Glace_Bay', 'America/Glace_Bay'), ('America/Goose_Bay', 'America/Goose_Bay'), ('America/Grand_Turk', 'America/Grand_Turk'), ('America/Grenada', 'America/Grenada'), ('America/Guadeloupe', 'America/Guadeloupe'), ('America/Guatemala', 'America/Guatemala'), ('America/Guayaquil', 'America/Guayaquil'), ('America/Guyana', 'America/Guyana'), ('America/Halifax', 'America/Halifax'), ('America/Havana', 'America/Havana'), ('America/Hermosillo', 'America/Hermosillo'), ('America/Indiana/Indianapolis', 'America/Indiana/Indianapolis'), ('America/Indiana/Knox', 'America/Indiana/Knox'), ('America/Indiana/Marengo', 'America/Indiana/Marengo'), ('America/Indiana/Petersburg', 'America/Indiana/Petersburg'), ('America/Indiana/Tell_City', 'America/Indiana/Tell_City'), ('America/Indiana/Vevay', 'America/Indiana/Vevay'), ('America/Indiana/Vincennes', 'America/Indiana/Vincennes'), ('America/Indiana/Winamac', 'America/Indiana/Winamac'), ('America/Inuvik', 'America/Inuvik'), ('America/Iqaluit', 'America/Iqaluit'), ('America/Jamaica', 'America/Jamaica'), ('America/Juneau', 'America/Juneau'), ('America/Kentucky/Louisville', 'America/Kentucky/Louisville'), ('America/Kentucky/Monticello', 'America/Kentucky/Monticello'), ('America/La_Paz', 'America/La_Paz'), ('America/Lima', 'America/Lima'), ('America/Los_Angeles', 'America/Los_Angeles'), ('America/Maceio', 'America/Maceio'), ('America/Managua', 'America/Managua'), ('America/Manaus', 'America/Manaus'), ('America/Martinique', 'America/Martinique'), ('America/Matamoros', 'America/Matamoros'), ('America/Mazatlan', 'America/Mazatlan'), ('America/Menominee', 'America/Menominee'), ('America/Merida', 'America/Merida'), ('America/Metlakatla', 'America/Metlakatla'), ('America/Mexico_City', 'America/Mexico_City'), ('America/Miquelon', 'America/Miquelon'), ('America/Moncton', 'America/Moncton'), ('America/Monterrey', 'America/Monterrey'), ('America/Montevideo', 'America/Montevideo'), ('America/Montserrat', 'America/Montserrat'), ('America/Nassau', 'America/Nassau'), ('America/New_York', 'America/New_York'), ('America/Nipigon', 'America/Nipigon'), ('America/Nome', 'America/Nome'), ('America/Noronha', 'America/Noronha'), ('America/North_Dakota/Beulah', 'America/North_Dakota/Beulah'), ('America/North_Dakota/Center', 'America/North_Dakota/Center'), ('America/North_Dakota/New_Salem', 'America/North_Dakota/New_Salem'), ('America/Nuuk', 'America/Nuuk'), ('America/Ojinaga', 'America/Ojinaga'), ('America/Panama', 'America/Panama'), ('America/Pangnirtung', 'America/Pangnirtung'), ('America/Paramaribo', 'America/Paramaribo'), ('America/Phoenix', 'America/Phoenix'), ('America/Port-au-Prince', 'America/Port-au-Prince'), ('America/Port_of_Spain', 'America/Port_of_Spain'), ('America/Porto_Velho', 'America/Porto_Velho'), ('America/Puerto_Rico', 'America/Puerto_Rico'), ('America/Punta_Arenas', 'America/Punta_Arenas'), ('America/Rainy_River', 'America/Rainy_River'), ('America/Rankin_Inlet', 'America/Rankin_Inlet'), ('America/Recife', 'America/Recife'), ('America/Regina', 'America/Regina'), ('America/Resolute', 'America/Resolute'), ('America/Rio_Branco', 'America/Rio_Branco'), ('America/Santarem', 'America/Santarem'), ('America/Santiago', 'America/Santiago'), ('America/Santo_Domingo', 'America/Santo_Domingo'), ('America/Sao_Paulo', 'America/Sao_Paulo'), ('America/Scoresbysund', 'America/Scoresbysund'), ('America/Sitka', 'America/Sitka'), ('America/St_Johns', 'America/St_Johns'), ('America/St_Kitts', 'America/St_Kitts'), ('America/St_Lucia', 'America/St_Lucia'), ('America/St_Thomas', 'America/St_Thomas'), ('America/St_Vincent', 'America/St_Vincent'), ('America/Swift_Current', 'America/Swift_Current'), ('America/Tegucigalpa', 'America/Tegucigalpa'), ('America/Thule', 'America/Thule'), ('America/Thunder_Bay', 'America/Thunder_Bay'), ('America/Tijuana', 'America/Tijuana'), ('America/Toronto', 'America/Toronto'), ('America/Tortola', 'America/Tortola'), ('America/Vancouver', 'America/Vancouver'), ('America/Whitehorse', 'America/Whitehorse'), ('America/Winnipeg', 'America/Winnipeg'), ('America/Yakutat', 'America/Yakutat'), ('America/Yellowknife', 'America/Yellowknife'), ('Antarctica/Casey', 'Antarctica/Casey'), ('Antarctica/Davis', 'Antarctica/Davis'), ('Antarctica/DumontDUrville', 'Antarctica/DumontDUrville'), ('Antarctica/Macquarie', 'Antarctica/Macquarie'), ('Antarctica/Mawson', 'Antarctica/Mawson'), ('Antarctica/McMurdo', 'Antarctica/McMurdo'), ('Antarctica/Palmer', 'Antarctica/Palmer'), ('Antarctica/Rothera', 'Antarctica/Rothera'), ('Antarctica/Syowa', 'Antarctica/Syowa'), ('Antarctica/Troll', 'Antarctica/Troll'), ('Antarctica/Vostok', 'Antarctica/Vostok'), ('Asia/Aden', 'Asia/Aden'), ('Asia/Almaty', 'Asia/Almaty'), ('Asia/Amman', 'Asia/Amman'), ('Asia/Anadyr', 'Asia/Anadyr'), ('Asia/Aqtau', 'Asia/Aqtau'), ('Asia/Aqtobe', 'Asia/Aqtobe'), ('Asia/Ashgabat', 'Asia/Ashgabat'), ('Asia/Atyrau', 'Asia/Atyrau'), ('Asia/Baghdad', 'Asia/Baghdad'), ('Asia/Bahrain', 'Asia/Bahrain'), ('Asia/Baku', 'Asia/Baku'), ('Asia/Bangkok', 'Asia/Bangkok'), ('Asia/Barnaul', 'Asia/Barnaul'), ('Asia/Beirut', 'Asia/Beirut'), ('Asia/Bishkek', 'Asia/Bishkek'), ('Asia/Brunei', 'Asia/Brunei'), ('Asia/Chita', 'Asia/Chita'), ('Asia/Choibalsan', 'Asia/Choibalsan'), ('Asia/Colombo', 'Asia/Colombo'), ('Asia/Damascus', 'Asia/Damascus'), ('Asia/Dhaka', 'Asia/Dhaka'), ('Asia/Dili', 'Asia/Dili'), ('Asia/Dubai', 'Asia/Dubai'), ('Asia/Dushanbe', 'Asia/Dushanbe'), ('Asia/Famagusta', 'Asia/Famagusta'), ('Asia/Gaza', 'Asia/Gaza'), ('Asia/Hebron', 'Asia/Hebron'), ('Asia/Ho_Chi_Minh', 'Asia/Ho_Chi_Minh'), ('Asia/Hong_Kong', 'Asia/Hong_Kong'), ('Asia/Hovd', 'Asia/Hovd'), ('Asia/Irkutsk', 'Asia/Irkutsk'), ('Asia/Jakarta', 'Asia/Jakarta'), ('Asia/Jayapura', 'Asia/Jayapura'), ('Asia/Jerusalem', 'Asia/Jerusalem'), ('Asia/Kabul', 'Asia/Kabul'), ('Asia/Kamchatka', 'Asia/Kamchatka'), ('Asia/Karachi', 'Asia/Karachi'), ('Asia/Kathmandu', 'Asia/Kathmandu'), ('Asia/Khandyga', 'Asia/Khandyga'), ('Asia/Kolkata', 'Asia/Kolkata'), ('Asia/Krasnoyarsk', 'Asia/Krasnoyarsk'), ('Asia/Kuala_Lumpur', 'Asia/Kuala_Lumpur'), ('Asia/Kuching', 'Asia/Kuching'), ('Asia/Kuwait', 'Asia/Kuwait'), ('Asia/Macau', 'Asia/Macau'), ('Asia/Magadan', 'Asia/Magadan'), ('Asia/Makassar', 'Asia/Makassar'), ('Asia/Manila', 'Asia/Manila'), ('Asia/Muscat', 'Asia/Muscat'), ('Asia/Nicosia', 'Asia/Nicosia'), ('Asia/Novokuznetsk', 'Asia/Novokuznetsk'), ('Asia/Novosibirsk', 'Asia/Novosibirsk'), ('Asia/Omsk', 'Asia/Omsk'), ('Asia/Oral', 'Asia/Oral'), ('Asia/Phnom_Penh', 'Asia/Phnom_Penh'), ('Asia/Pontianak', 'Asia/Pontianak'), ('Asia/Pyongyang', 'Asia/Pyongyang'), ('Asia/Qatar', 'Asia/Qatar'), ('Asia/Qostanay', 'Asia/Qostanay'), ('Asia/Qyzylorda', 'Asia/Qyzylorda'), ('Asia/Riyadh', 'Asia/Riyadh'), ('Asia/Sakhalin', 'Asia/Sakhalin'), ('Asia/Samarkand', 'Asia/Samarkand'), ('Asia/Seoul', 'Asia/Seoul'), ('Asia/Shanghai', 'Asia/Shanghai'), ('Asia/Singapore', 'Asia/Singapore'), ('Asia/Srednekolymsk', 'Asia/Srednekolymsk'), ('Asia/Taipei', 'Asia/Taipei'), ('Asia/Tashkent', 'Asia/Tashkent'), ('Asia/Tbilisi', 'Asia/Tbilisi'), ('Asia/Tehran', 'Asia/Tehran'), ('Asia/Thimphu', 'Asia/Thimphu'), ('Asia/Tokyo', 'Asia/Tokyo'), ('Asia/Tomsk', 'Asia/Tomsk'), ('Asia/Ulaanbaatar', 'Asia/Ulaanbaatar'), ('Asia/Urumqi', 'Asia/Urumqi'), ('Asia/Ust-Nera', 'Asia/Ust-Nera'), ('Asia/Vientiane', 'Asia/Vientiane'), ('Asia/Vladivostok', 'Asia/Vladivostok'), ('Asia/Yakutsk', 'Asia/Yakutsk'), ('Asia/Yangon', 'Asia/Yangon'), ('Asia/Yekaterinburg', 'Asia/Yekaterinburg'), ('Asia/Yerevan', 'Asia/Yerevan'), ('Atlantic/Azores', 'Atlantic/Azores'), ('Atlantic/Bermuda', 'Atlantic/Bermuda'), ('Atlantic/Canary', 'Atlantic/Canary'), ('Atlantic/Cape_Verde', 'Atlantic/Cape_Verde'), ('Atlantic/Faroe', 'Atlantic/Faroe'), ('Atlantic/Madeira', 'Atlantic/Madeira'), ('Atlantic/Reykjavik', 'Atlantic/Reykjavik'), ('Atlantic/South_Georgia', 'Atlantic/South_Georgia'), ('Atlantic/St_Helena', 'Atlantic/St_Helena'), ('Atlantic/Stanley', 'Atlantic/Stanley'), ('Australia/Adelaide', 'Australia/Adelaide'), ('Australia/Brisbane', 'Australia/Brisbane'), ('Australia/Broken_Hill', 'Australia/Broken_Hill'), ('Australia/Darwin', 'Australia/Darwin'), ('Australia/Eucla', 'Australia/Eucla'), ('Australia/Hobart', 'Australia/Hobart'), ('Australia/Lindeman', 'Australia/Lindeman'), ('Australia/Lord_Howe', 'Australia/Lord_Howe'), ('Australia/Melbourne', 'Australia/Melbourne'), ('Australia/Perth', 'Australia/Perth'), ('Australia/Sydney', 'Australia/Sydney'), ('Europe/Amsterdam', 'Europe/Amsterdam'), ('Europe/Andorra', 'Europe/Andorra'), ('Europe/Astrakhan', 'Europe/Astrakhan'), ('Europe/Athens', 'Europe/Athens'), ('Europe/Belgrade', 'Europe/Belgrade'), ('Europe/Berlin', 'Europe/Berlin'), ('Europe/Brussels', 'Europe/Brussels'), ('Europe/Bucharest', 'Europe/Bucharest'), ('Europe/Budapest', 'Europe/Budapest'), ('Europe/Chisinau', 'Europe/Chisinau'), ('Europe/Copenhagen', 'Europe/Copenhagen'), ('Europe/Dublin', 'Europe/Dublin'), ('Europe/Gibraltar', 'Europe/Gibraltar'), ('Europe/Helsinki', 'Europe/Helsinki'), ('Europe/Istanbul', 'Europe/Istanbul'), ('Europe/Kaliningrad', 'Europe/Kaliningrad'), ('Europe/Kirov', 'Europe/Kirov'), ('Europe/Kyiv', 'Europe/Kyiv'), ('Europe/Lisbon', 'Europe/Lisbon'), ('Europe/London', 'Europe/London'), ('Europe/Luxembourg', 'Europe/Luxembourg'), ('Europe/Madrid', 'Europe/Madrid'), ('Europe/Malta', 'Europe/Malta'), ('Europe/Minsk', 'Europe/Minsk'), ('Europe/Monaco', 'Europe/Monaco'), ('Europe/Moscow', 'Europe/Moscow'), ('Europe/Oslo', 'Europe/Oslo'), ('Europe/Paris', 'Europe/Paris'), ('Europe/Prague', 'Europe/Prague'), ('Europe/Riga', 'Europe/Riga'), ('Europe/Rome', 'Europe/Rome'), ('Europe/Samara', 'Europe/Samara'), ('Europe/Saratov', 'Europe/Saratov'), ('Europe/Simferopol', 'Europe/Simferopol'), ('Europe/Sofia', 'Europe/Sofia'), ('Europe/Stockholm', 'Europe/Stockholm'), ('Europe/Tallinn', 'Europe/Tallinn'), ('Europe/Tirane', 'Europe/Tirane'), ('Europe/Ulyanovsk', 'Europe/Ulyanovsk'), ('Europe/Uzhgorod', 'Europe/Uzhgorod'), ('Europe/Vaduz', 'Europe/Vaduz'), ('Europe/Vienna', 'Europe/Vienna'), ('Europe/Vilnius', 'Europe/Vilnius'), ('Europe/Volgograd', 'Europe/Volgograd'), ('Europe/Warsaw', 'Europe/Warsaw'), ('Europe/Zaporozhye', 'Europe/Zaporozhye'), ('Europe/Zurich', 'Europe/Zurich'), ('GMT', 'GMT'), ('Indian/Antananarivo', 'Indian/Antananarivo'), ('Indian/Chagos', 'Indian/Chagos'), ('Indian/Christmas', 'Indian/Christmas'), ('Indian/Cocos', 'Indian/Cocos'), ('Indian/Comoro', 'Indian/Comoro'), ('Indian/Kerguelen', 'Indian/Kerguelen'), ('Indian/Mahe', 'Indian/Mahe'), ('Indian/Maldives', 'Indian/Maldives'), ('Indian/Mauritius', 'Indian/Mauritius'), ('Indian/Mayotte', 'Indian/Mayotte'), ('Indian/Reunion', 'Indian/Reunion'), ('Pacific/Apia', 'Pacific/Apia'), ('Pacific/Auckland', 'Pacific/Auckland'), ('Pacific/Bougainville', 'Pacific/Bougainville'), ('Pacific/Chatham', 'Pacific/Chatham'), ('Pacific/Chuuk', 'Pacific/Chuuk'), ('Pacific/Easter', 'Pacific/Easter'), ('Pacific/Efate', 'Pacific/Efate'), ('Pacific/Fakaofo', 'Pacific/Fakaofo'), ('Pacific/Fiji', 'Pacific/Fiji'), ('Pacific/Funafuti', 'Pacific/Funafuti'), ('Pacific/Galapagos', 'Pacific/Galapagos'), ('Pacific/Gambier', 'Pacific/Gambier'), ('Pacific/Guadalcanal', 'Pacific/Guadalcanal'), ('Pacific/Guam', 'Pacific/Guam'), ('Pacific/Honolulu', 'Pacific/Honolulu'), ('Pacific/Kanton', 'Pacific/Kanton'), ('Pacific/Kiritimati', 'Pacific/Kiritimati'), ('Pacific/Kosrae', 'Pacific/Kosrae'), ('Pacific/Kwajalein', 'Pacific/Kwajalein'), ('Pacific/Majuro', 'Pacific/Majuro'), ('Pacific/Marquesas', 'Pacific/Marquesas'), ('Pacific/Midway', 'Pacific/Midway'), ('Pacific/Nauru', 'Pacific/Nauru'), ('Pacific/Niue', 'Pacific/Niue'), ('Pacific/Norfolk', 'Pacific/Norfolk'), ('Pacific/Noumea', 'Pacific/Noumea'), ('Pacific/Pago_Pago', 'Pacific/Pago_Pago'), ('Pacific/Palau', 'Pacific/Palau'), ('Pacific/Pitcairn', 'Pacific/Pitcairn'), ('Pacific/Pohnpei', 'Pacific/Pohnpei'), ('Pacific/Port_Moresby', 'Pacific/Port_Moresby'), ('Pacific/Rarotonga', 'Pacific/Rarotonga'), ('Pacific/Saipan', 'Pacific/Saipan'), ('Pacific/Tahiti', 'Pacific/Tahiti'), ('Pacific/Tarawa', 'Pacific/Tarawa'), ('Pacific/Tongatapu', 'Pacific/Tongatapu'), ('Pacific/Wake', 'Pacific/Wake'), ('Pacific/Wallis', 'Pacific/Wallis'), ('UTC', 'UTC')], default='UTC', max_length=255)), ('idsubmit_cutoff_day_offset_00', models.IntegerField(blank=True, default=13, help_text='The number of days before the meeting start date when the submission of -00 drafts will be closed.')), ('idsubmit_cutoff_day_offset_01', models.IntegerField(blank=True, default=13, help_text='The number of days before the meeting start date when the submission of -01 drafts etc. will be closed.')), - ('idsubmit_cutoff_time_utc', models.DurationField(blank=True, default=datetime.timedelta(0, 86399), help_text='The time of day (UTC) after which submission will be closed. Use for example 23:59:59.')), - ('idsubmit_cutoff_warning_days', models.DurationField(blank=True, default=datetime.timedelta(21), help_text="How long before the 00 cutoff to start showing cutoff warnings. Use for example '21' or '21 days'.")), + ('idsubmit_cutoff_time_utc', models.DurationField(blank=True, default=datetime.timedelta(seconds=86399), help_text='The time of day (UTC) after which submission will be closed. Use for example 23:59:59.')), + ('idsubmit_cutoff_warning_days', models.DurationField(blank=True, default=datetime.timedelta(days=21), help_text="How long before the 00 cutoff to start showing cutoff warnings. Use for example '21' or '21 days'.")), ('submission_start_day_offset', models.IntegerField(blank=True, default=90, help_text='The number of days before the meeting start date after which meeting materials will be accepted.')), ('submission_cutoff_day_offset', models.IntegerField(blank=True, default=26, help_text='The number of days after the meeting start date in which new meeting materials will be accepted.')), ('submission_correction_day_offset', models.IntegerField(blank=True, default=50, help_text='The number of days after the meeting start date in which updates to existing meeting materials will be accepted.')), @@ -74,14 +68,18 @@ class Migration(migrations.Migration): ('venue_addr', models.TextField(blank=True)), ('break_area', models.CharField(blank=True, max_length=255)), ('reg_area', models.CharField(blank=True, max_length=255)), - ('agenda_note', models.TextField(blank=True, help_text='Text in this field will be placed at the top of the html agenda page for the meeting. HTML can be used, but will not be validated.')), + ('agenda_info_note', models.TextField(blank=True, help_text='Text in this field will be placed at the top of the html agenda page for the meeting. HTML can be used, but will not be validated.')), + ('agenda_warning_note', models.TextField(blank=True, help_text='Text in this field will be placed more prominently at the top of the html agenda page for the meeting. HTML can be used, but will not be validated.')), ('session_request_lock_message', models.CharField(blank=True, max_length=255)), ('proceedings_final', models.BooleanField(default=False, help_text='Are the proceedings for this meeting complete?')), ('acknowledgements', models.TextField(blank=True, help_text='Acknowledgements for use in meeting proceedings. Use ReStructuredText markup.')), ('show_important_dates', models.BooleanField(default=False)), + ('attendees', models.IntegerField(blank=True, default=None, help_text='Number of Attendees for backfilled meetings, leave it blank for new meetings, and then it is calculated from the registrations', null=True)), + ('group_conflict_types', models.ManyToManyField(blank=True, help_text='Types of scheduling conflict between groups to consider', limit_choices_to={'is_group_conflict': True}, to='name.ConstraintName')), + ('overview', ietf.utils.models.ForeignKey(editable=False, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='overview', to='dbtemplate.DBTemplate')), ], options={ - 'ordering': ['-date', 'id'], + 'ordering': ['-date', '-id'], }, ), migrations.CreateModel( @@ -97,7 +95,7 @@ class Migration(migrations.Migration): name='Room', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('time', models.DateTimeField(default=datetime.datetime.now)), + ('modified', models.DateTimeField(auto_now=True)), ('name', models.CharField(max_length=255)), ('functional_name', models.CharField(blank=True, max_length=255)), ('capacity', models.IntegerField(blank=True, null=True)), @@ -111,7 +109,7 @@ class Migration(migrations.Migration): ('session_types', models.ManyToManyField(blank=True, to='name.TimeSlotTypeName')), ], options={ - 'ordering': ['-meeting', 'name'], + 'ordering': ['-id'], }, ), migrations.CreateModel( @@ -128,18 +126,6 @@ class Migration(migrations.Migration): 'ordering': ['timeslot__time', 'timeslot__type__slug', 'session__group__parent__name', 'session__group__acronym', 'session__name'], }, ), - migrations.CreateModel( - name='Schedule', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('name', models.CharField(max_length=16)), - ('visible', models.BooleanField(default=True, help_text='Make this agenda available to those who know about it.')), - ('public', models.BooleanField(default=True, help_text='Make this agenda publically available.')), - ('badness', models.IntegerField(blank=True, null=True)), - ('meeting', ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='meeting.Meeting')), - ('owner', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), - ], - ), migrations.CreateModel( name='Session', fields=[ @@ -148,28 +134,24 @@ class Migration(migrations.Migration): ('short', models.CharField(blank=True, help_text="Short version of 'name' above, for use in filenames.", max_length=32)), ('attendees', models.IntegerField(blank=True, null=True)), ('agenda_note', models.CharField(blank=True, max_length=255)), - ('requested', models.DateTimeField(default=datetime.datetime.now)), ('requested_duration', models.DurationField(default=datetime.timedelta(0))), ('comments', models.TextField(blank=True)), ('scheduled', models.DateTimeField(blank=True, null=True)), ('modified', models.DateTimeField(auto_now=True)), ('remote_instructions', models.CharField(blank=True, max_length=1024)), + ('on_agenda', models.BooleanField(default=True, help_text='Is this session visible on the meeting agenda?')), ('group', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='group.Group')), + ('joint_with_groups', models.ManyToManyField(blank=True, related_name='sessions_joint_in', to='group.Group')), ], ), migrations.CreateModel( - name='SessionPresentation', + name='UrlResource', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('rev', models.CharField(blank=True, max_length=16, null=True, verbose_name='revision')), - ('order', models.PositiveSmallIntegerField(default=0)), - ('document', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document')), - ('session', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='meeting.Session')), + ('url', models.URLField(blank=True, null=True)), + ('name', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.RoomResourceName')), + ('room', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='meeting.Room')), ], - options={ - 'ordering': ('order',), - 'db_table': 'meeting_session_materials', - }, ), migrations.CreateModel( name='TimeSlot', @@ -186,18 +168,37 @@ class Migration(migrations.Migration): ('type', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.TimeSlotTypeName')), ], options={ - 'ordering': ['-time', 'id'], + 'ordering': ['-time', '-id'], }, ), migrations.CreateModel( - name='UrlResource', + name='SlideSubmission', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('url', models.URLField(blank=True, null=True)), - ('name', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.RoomResourceName')), - ('room', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='meeting.Room')), + ('time', models.DateTimeField(auto_now=True)), + ('title', models.CharField(max_length=255)), + ('filename', models.CharField(max_length=255)), + ('apply_to_all', models.BooleanField(default=False)), + ('doc', ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, to='doc.Document')), + ('session', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='meeting.Session')), + ('status', ietf.utils.models.ForeignKey(default='pending', null=True, on_delete=django.db.models.deletion.SET_NULL, to='name.SlideSubmissionStatusName')), + ('submitter', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), ], ), + migrations.CreateModel( + name='SessionPresentation', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('rev', models.CharField(blank=True, max_length=16, null=True, verbose_name='revision')), + ('order', models.PositiveSmallIntegerField(default=0)), + ('document', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document')), + ('session', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='meeting.Session')), + ], + options={ + 'db_table': 'meeting_session_materials', + 'ordering': ('order',), + }, + ), migrations.AddField( model_name='session', name='materials', @@ -210,8 +211,8 @@ class Migration(migrations.Migration): ), migrations.AddField( model_name='session', - name='requested_by', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person'), + name='purpose', + field=ietf.utils.models.ForeignKey(help_text='Purpose of the session', on_delete=django.db.models.deletion.CASCADE, to='name.SessionPurposeName'), ), migrations.AddField( model_name='session', @@ -220,14 +221,39 @@ class Migration(migrations.Migration): ), migrations.AddField( model_name='session', - name='status', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.SessionStatusName'), + name='tombstone_for', + field=models.ForeignKey(blank=True, help_text='This session is the tombstone for a session that was rescheduled', null=True, on_delete=django.db.models.deletion.CASCADE, to='meeting.Session'), ), migrations.AddField( model_name='session', name='type', field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.TimeSlotTypeName'), ), + migrations.CreateModel( + name='SchedulingEvent', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('time', models.DateTimeField(default=django.utils.timezone.now, help_text='When the event happened')), + ('by', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), + ('session', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='meeting.Session')), + ('status', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.SessionStatusName')), + ], + ), + migrations.CreateModel( + name='Schedule', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('name', models.CharField(help_text='Letters, numbers and -:_ allowed.', max_length=64, validators=[django.core.validators.RegexValidator('^[A-Za-z0-9-:_]*$')])), + ('visible', models.BooleanField(default=True, help_text='Show in the list of possible agendas for the meeting.', verbose_name='Show in agenda list')), + ('public', models.BooleanField(default=True, help_text='Allow others to see this agenda.')), + ('badness', models.IntegerField(blank=True, null=True)), + ('notes', models.TextField(blank=True)), + ('base', ietf.utils.models.ForeignKey(blank=True, help_text='Sessions scheduled in the base schedule show up in this schedule too.', limit_choices_to={'base': None}, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='derivedschedule_set', to='meeting.Schedule')), + ('meeting', ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, related_name='schedule_set', to='meeting.Meeting')), + ('origin', ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='+', to='meeting.Schedule')), + ('owner', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), + ], + ), migrations.AddField( model_name='schedtimesessassignment', name='schedule', @@ -243,59 +269,99 @@ class Migration(migrations.Migration): name='timeslot', field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='sessionassignments', to='meeting.TimeSlot'), ), - migrations.AddField( - model_name='meeting', - name='agenda', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='+', to='meeting.Schedule'), + migrations.CreateModel( + name='ProceedingsMaterial', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('document', ietf.utils.models.ForeignKey(limit_choices_to={'type_id': 'procmaterials'}, on_delete=django.db.models.deletion.CASCADE, to='doc.Document', unique=True)), + ('meeting', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='proceedings_materials', to='meeting.Meeting')), + ('type', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ProceedingsMaterialTypeName')), + ], + ), + migrations.CreateModel( + name='MeetingHost', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('name', models.CharField(max_length=255)), + ('logo', ietf.utils.fields.MissingOkImageField(height_field='logo_height', storage=ietf.utils.storage.NoLocationMigrationFileSystemStorage(location=None), upload_to=ietf.meeting.models._host_upload_path, validators=[ietf.utils.validators.MaxImageSizeValidator(400, 400), ietf.utils.validators.WrappedValidator(ietf.utils.validators.validate_file_size, True), ietf.utils.validators.WrappedValidator(ietf.utils.validators.validate_file_extension, ['.png', '.jpg', '.jpeg']), ietf.utils.validators.WrappedValidator(ietf.utils.validators.validate_mime_type, ['image/jpeg', 'image/png'], True)], width_field='logo_width')), + ('logo_width', models.PositiveIntegerField(null=True)), + ('logo_height', models.PositiveIntegerField(null=True)), + ('meeting', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='meetinghosts', to='meeting.Meeting')), + ], + options={ + 'ordering': ('pk',), + }, ), migrations.AddField( model_name='meeting', - name='overview', - field=ietf.utils.models.ForeignKey(editable=False, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='overview', to='dbtemplate.DBTemplate'), + name='schedule', + field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='+', to='meeting.Schedule'), ), migrations.AddField( model_name='meeting', name='type', field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.MeetingTypeName'), ), - migrations.AddField( - model_name='importantdate', - name='meeting', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='meeting.Meeting'), - ), - migrations.AddField( - model_name='importantdate', - name='name', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ImportantDateName'), + migrations.CreateModel( + name='ImportantDate', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('date', models.DateField()), + ('meeting', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='meeting.Meeting')), + ('name', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ImportantDateName')), + ], + options={ + 'ordering': ['-meeting_id', 'date'], + }, ), migrations.AddField( model_name='floorplan', name='meeting', field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='meeting.Meeting'), ), - migrations.AddField( - model_name='constraint', - name='meeting', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='meeting.Meeting'), + migrations.CreateModel( + name='Constraint', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('time_relation', models.CharField(blank=True, choices=[('subsequent-days', 'Schedule the sessions on subsequent days'), ('one-day-seperation', 'Leave at least one free day in between the two sessions')], max_length=200)), + ('meeting', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='meeting.Meeting')), + ('name', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ConstraintName')), + ('person', ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='person.Person')), + ('source', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='constraint_source_set', to='group.Group')), + ('target', ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, related_name='constraint_target_set', to='group.Group')), + ('timeranges', models.ManyToManyField(to='name.TimerangeName')), + ], ), - migrations.AddField( - model_name='constraint', - name='name', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ConstraintName'), + migrations.CreateModel( + name='Attended', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), + ('session', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='meeting.Session')), + ], ), - migrations.AddField( - model_name='constraint', - name='person', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='person.Person'), + migrations.AddIndex( + model_name='timeslot', + index=models.Index(fields=['-time', '-id'], name='meeting_tim_time_b802cb_idx'), ), - migrations.AddField( - model_name='constraint', - name='source', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='constraint_source_set', to='group.Group'), + migrations.AlterUniqueTogether( + name='sessionpresentation', + unique_together={('session', 'document')}, ), - migrations.AddField( - model_name='constraint', - name='target', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, related_name='constraint_target_set', to='group.Group'), + migrations.AlterUniqueTogether( + name='proceedingsmaterial', + unique_together={('meeting', 'type')}, + ), + migrations.AlterUniqueTogether( + name='meetinghost', + unique_together={('meeting', 'name')}, + ), + migrations.AddIndex( + model_name='meeting', + index=models.Index(fields=['-date', '-id'], name='meeting_mee_date_40ca21_idx'), + ), + migrations.AlterUniqueTogether( + name='attended', + unique_together={('person', 'session')}, ), ] diff --git a/ietf/meeting/migrations/0002_auto_20180225_1207.py b/ietf/meeting/migrations/0002_auto_20180225_1207.py deleted file mode 100644 index 523dd66bcf..0000000000 --- a/ietf/meeting/migrations/0002_auto_20180225_1207.py +++ /dev/null @@ -1,41 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-25 12:07 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0001_initial'), - ] - - operations = [ - migrations.AlterModelOptions( - name='floorplan', - options={'ordering': ['-id']}, - ), - migrations.AlterModelOptions( - name='importantdate', - options={'ordering': ['-meeting_id', 'date']}, - ), - migrations.AlterModelOptions( - name='meeting', - options={'ordering': ['-date', '-id']}, - ), - migrations.AlterModelOptions( - name='room', - options={'ordering': ['-id']}, - ), - migrations.AlterModelOptions( - name='timeslot', - options={'ordering': ['-time', '-id']}, - ), - migrations.AlterField( - model_name='floorplan', - name='short', - field=models.CharField(default='', max_length=3), - ), - ] diff --git a/ietf/meeting/migrations/0002_session_has_onsite_tool.py b/ietf/meeting/migrations/0002_session_has_onsite_tool.py new file mode 100644 index 0000000000..f248b1a415 --- /dev/null +++ b/ietf/meeting/migrations/0002_session_has_onsite_tool.py @@ -0,0 +1,18 @@ +# Generated by Django 2.2.28 on 2023-03-07 16:54 + +from django.db import migrations, models + + +class Migration(migrations.Migration): + + dependencies = [ + ('meeting', '0001_initial'), + ] + + operations = [ + migrations.AddField( + model_name='session', + name='has_onsite_tool', + field=models.BooleanField(default=False, help_text='Does this session use the officially supported onsite and remote tooling?'), + ), + ] diff --git a/ietf/meeting/migrations/0003_populate_session_has_onsite_tool.py b/ietf/meeting/migrations/0003_populate_session_has_onsite_tool.py new file mode 100644 index 0000000000..4499fdb3b5 --- /dev/null +++ b/ietf/meeting/migrations/0003_populate_session_has_onsite_tool.py @@ -0,0 +1,33 @@ +# Generated by Django 2.2.28 on 2023-03-07 16:54 + +from django.db import migrations + + +def forward(apps, schema_editor): + Session = apps.get_model('meeting', 'Session') + Meeting = apps.get_model('meeting', 'Meeting') + Room = apps.get_model('meeting', 'Room') + full_meetings = Meeting.objects.filter(type_id='ietf', schedule__isnull=False) + schedules = {m.schedule for m in full_meetings} | {m.schedule.base for m in full_meetings if m.schedule.base} + rooms_with_meetecho = Room.objects.filter(urlresource__name_id__in=['meetecho', 'meetecho_onsite']).distinct() + sessions_with_meetecho = Session.objects.filter( + timeslotassignments__schedule__in=schedules, + timeslotassignments__timeslot__location__in=rooms_with_meetecho, + ).distinct() + sessions_with_meetecho.update(has_onsite_tool=True) + + +def reverse(apps, schema_editor): + Session = apps.get_model('meeting', 'Session') + Session.objects.all().update(has_onsite_tool=False) + + +class Migration(migrations.Migration): + + dependencies = [ + ('meeting', '0002_session_has_onsite_tool'), + ] + + operations = [ + migrations.RunPython(forward, reverse), + ] diff --git a/ietf/meeting/migrations/0003_rename_modified_fields.py b/ietf/meeting/migrations/0003_rename_modified_fields.py deleted file mode 100644 index 4b1dbbe85f..0000000000 --- a/ietf/meeting/migrations/0003_rename_modified_fields.py +++ /dev/null @@ -1,36 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-03-02 14:33 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0002_auto_20180225_1207'), - ] - - operations = [ - migrations.RenameField( - model_name='floorplan', - old_name='time', - new_name='modified', - ), - migrations.RenameField( - model_name='room', - old_name='time', - new_name='modified', - ), - migrations.AlterField( - model_name='floorplan', - name='modified', - field=models.DateTimeField(auto_now=True), - ), - migrations.AlterField( - model_name='room', - name='modified', - field=models.DateTimeField(auto_now=True), - ), - ] diff --git a/ietf/meeting/migrations/0004_meeting_attendees.py b/ietf/meeting/migrations/0004_meeting_attendees.py deleted file mode 100644 index a929459245..0000000000 --- a/ietf/meeting/migrations/0004_meeting_attendees.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.11 on 2018-03-20 09:17 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0003_rename_modified_fields'), - ] - - operations = [ - migrations.AddField( - model_name='meeting', - name='attendees', - field=models.IntegerField(blank=True, default=None, help_text='Number of Attendees for backfilled meetings, leave it blank for new meetings, and then it is calculated from the registrations', null=True), - ), - ] diff --git a/ietf/meeting/migrations/0004_session_chat_room.py b/ietf/meeting/migrations/0004_session_chat_room.py new file mode 100644 index 0000000000..9879c8c42f --- /dev/null +++ b/ietf/meeting/migrations/0004_session_chat_room.py @@ -0,0 +1,18 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations, models + + +class Migration(migrations.Migration): + + dependencies = [ + ('meeting', '0003_populate_session_has_onsite_tool'), + ] + + operations = [ + migrations.AddField( + model_name='session', + name='chat_room', + field=models.CharField(blank=True, help_text='Name of Zulip stream, if different from group acronym', max_length=32), + ), + ] diff --git a/ietf/meeting/migrations/0005_alter_session_agenda_note.py b/ietf/meeting/migrations/0005_alter_session_agenda_note.py new file mode 100644 index 0000000000..59daeea45d --- /dev/null +++ b/ietf/meeting/migrations/0005_alter_session_agenda_note.py @@ -0,0 +1,18 @@ +# Copyright The IETF Trust 2024, All Rights Reserved + +from django.db import migrations, models + + +class Migration(migrations.Migration): + + dependencies = [ + ("meeting", "0004_session_chat_room"), + ] + + operations = [ + migrations.AlterField( + model_name="session", + name="agenda_note", + field=models.CharField(blank=True, max_length=512), + ), + ] diff --git a/ietf/meeting/migrations/0005_backfill_old_meetings.py b/ietf/meeting/migrations/0005_backfill_old_meetings.py deleted file mode 100644 index 928ed4afd8..0000000000 --- a/ietf/meeting/migrations/0005_backfill_old_meetings.py +++ /dev/null @@ -1,90 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -from django.db import migrations - - -def backfill_old_meetings(apps, schema_editor): - Meeting = apps.get_model('meeting', 'Meeting') - - for id, number, type_id, date, city, country, time_zone, continent, attendees in [ - ( 59,'59','ietf','2004-03-29','Seoul','KR','Asia/Seoul','Asia','1390' ), - ( 58,'58','ietf','2003-11-09','Minneapolis','US','America/Menominee','America','1233' ), - ( 57,'57','ietf','2003-07-13','Vienna','AT','Europe/Vienna','Europe','1304' ), - ( 56,'56','ietf','2003-03-16','San Francisco','US','America/Los_Angeles','America','1679' ), - ( 55,'55','ietf','2002-11-17','Atlanta','US','America/New_York','America','1570' ), - ( 54,'54','ietf','2002-07-14','Yokohama','JP','Asia/Tokyo','Asia','1885' ), - ( 53,'53','ietf','2002-03-17','Minneapolis','US','America/Menominee','America','1656' ), - ( 52,'52','ietf','2001-12-09','Salt Lake City','US','America/Denver','America','1691' ), - ( 51,'51','ietf','2001-08-05','London','GB','Europe/London','Europe','2226' ), - ( 50,'50','ietf','2001-03-18','Minneapolis','US','America/Menominee','America','1822' ), - ( 49,'49','ietf','2000-12-10','San Diego','US','America/Los_Angeles','America','2810' ), - ( 48,'48','ietf','2000-07-31','Pittsburgh','US','America/New_York','America','2344' ), - ( 47,'47','ietf','2000-03-26','Adelaide','AU','Australia/Adelaide','Australia','1431' ), - ( 46,'46','ietf','1999-11-07','Washington','US','America/New_York','America','2379' ), - ( 45,'45','ietf','1999-07-11','Oslo','NO','Europe/Oslo','Europe','1710' ), - ( 44,'44','ietf','1999-03-14','Minneapolis','US','America/Menominee','America','1705' ), - ( 43,'43','ietf','1998-12-07','Orlando','US','America/New_York','America','2124' ), - ( 42,'42','ietf','1998-08-24','Chicago','US','America/Chicago','America','2106' ), - ( 41,'41','ietf','1998-03-30','Los Angeles','US','America/Los_Angeles','America','1775' ), - ( 40,'40','ietf','1997-12-08','Washington','US','America/New_York','America','1897' ), - ( 39,'39','ietf','1997-08-11','Munich','DE','Europe/Berlin','Europe','1308' ), - ( 38,'38','ietf','1997-04-07','Memphis','US','America/Chicago','America','1321' ), - ( 37,'37','ietf','1996-12-09','San Jose','US','America/Los_Angeles','America','1993' ), - ( 36,'36','ietf','1996-06-24','Montreal','CA','America/New_York','America','1283' ), - ( 35,'35','ietf','1996-03-04','Los Angeles','US','America/Los_Angeles','America','1038' ), - ( 34,'34','ietf','1995-12-04','Dallas','US','America/Chicago','America','1007' ), - ( 33,'33','ietf','1995-07-17','Stockholm','SE','Europe/Stockholm','Europe','617' ), - ( 32,'32','ietf','1995-04-03','Danvers','US','America/New_York','America','983' ), - ( 31,'31','ietf','1994-12-05','San Jose','US','America/Los_Angeles','America','1079' ), - ( 30,'30','ietf','1994-07-25','Toronto','CA','America/New_York','America','710' ), - ( 29,'29','ietf','1994-03-28','Seattle','US','America/Los_Angeles','America','785' ), - ( 28,'28','ietf','1993-11-01','Houston','US','America/Chicago','America','636' ), - ( 27,'27','ietf','1993-07-12','Amsterdam','NL','Europe/Amsterdam','Europe','493' ), - ( 26,'26','ietf','1993-03-29','Columbus','US','America/New_York','America','638' ), - ( 25,'25','ietf','1992-11-16','Washington','US','America/New_York','America','633' ), - ( 24,'24','ietf','1992-07-13','Cambridge','US','America/New_York','America','677' ), - ( 23,'23','ietf','1992-03-16','San Diego','US','America/Los_Angeles','America','530' ), - ( 22,'22','ietf','1991-11-18','Santa Fe','US','America/Denver','America','372' ), - ( 21,'21','ietf','1991-07-29','Atlanta','US','America/New_York','America','387' ), - ( 20,'20','ietf','1991-03-11','St. Louis','US','America/Chicago','America','348' ), - ( 19,'19','ietf','1990-12-03','Boulder','US','America/Denver','America','292' ), - ( 18,'18','ietf','1990-07-30','Vancouver','CA','America/Los_Angeles','America','293' ), - ( 17,'17','ietf','1990-05-01','Pittsburgh','US','America/New_York','America','244' ), - ( 16,'16','ietf','1990-02-06','Tallahassee','US','America/New_York','America','196' ), - ( 15,'15','ietf','1989-10-31','Honolulu','US','Pacific/Honolulu','America','138' ), - ( 14,'14','ietf','1989-07-25','Stanford','US','America/Los_Angeles','America','217' ), - ( 13,'13','ietf','1989-04-11','Cocoa Beach','US','America/New_York','America','114' ), - ( 12,'12','ietf','1989-01-18','Austin','US','America/Chicago','America','120' ), - ( 11,'11','ietf','1988-10-17','Ann Arbor','US','America/New_York','America','114' ), - ( 10,'10','ietf','1988-06-15','Annapolis','US','America/New_York','America','112' ), - ( 9,'9','ietf','1988-03-01','San Diego','US','America/Los_Angeles','America','82' ), - ( 8,'8','ietf','1987-11-02','Boulder','US','America/Denver','America','56' ), - ( 7,'7','ietf','1987-07-27','McLean','US','America/New_York','America','101' ), - ( 6,'6','ietf','1987-04-22','Boston','US','America/New_York','America','88' ), - ( 5,'5','ietf','1987-02-04','Moffett Field','US','America/Los_Angeles','America','35' ), - ( 4,'4','ietf','1986-10-15','Menlo Park','US','America/Los_Angeles','America','35' ), - ( 3,'3','ietf','1986-07-23','Ann Arbor','US','America/New_York','America','18' ), - ( 2,'2','ietf','1986-04-08','Aberdeen','US','America/New_York','America','21' ), - ( 1,'1','ietf','1986-01-16','San Diego','US','America/Los_Angeles','America','21' ), - ]: - Meeting.objects.get_or_create(id=id, number=number, type_id=type_id, - date=date, city=city, country=country, - time_zone=time_zone); - - -def reverse(apps, schema_editor): - pass - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0004_meeting_attendees'), - ] - - operations = [ - migrations.RunPython(backfill_old_meetings, reverse) - ] - diff --git a/ietf/meeting/migrations/0006_alter_sessionpresentation_document_and_session.py b/ietf/meeting/migrations/0006_alter_sessionpresentation_document_and_session.py new file mode 100644 index 0000000000..e8d6a663f8 --- /dev/null +++ b/ietf/meeting/migrations/0006_alter_sessionpresentation_document_and_session.py @@ -0,0 +1,33 @@ +# Copyright The IETF Trust 2024, All Rights Reserved + +from django.db import migrations +import django.db.models.deletion +import ietf.utils.models + + +class Migration(migrations.Migration): + dependencies = [ + ("doc", "0021_narrativeminutes"), + ("meeting", "0005_alter_session_agenda_note"), + ] + + operations = [ + migrations.AlterField( + model_name="sessionpresentation", + name="document", + field=ietf.utils.models.ForeignKey( + on_delete=django.db.models.deletion.CASCADE, + related_name="presentations", + to="doc.document", + ), + ), + migrations.AlterField( + model_name="sessionpresentation", + name="session", + field=ietf.utils.models.ForeignKey( + on_delete=django.db.models.deletion.CASCADE, + related_name="presentations", + to="meeting.session", + ), + ), + ] diff --git a/ietf/meeting/migrations/0006_backfill_attendees.py b/ietf/meeting/migrations/0006_backfill_attendees.py deleted file mode 100644 index 212260d72d..0000000000 --- a/ietf/meeting/migrations/0006_backfill_attendees.py +++ /dev/null @@ -1,133 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -from django.db import migrations - - -def backfill_old_meetings(apps, schema_editor): - Meeting = apps.get_model('meeting', 'Meeting') - - for number, attendees in [ - ( 101,1203 ), - ( 100,1018 ), - ( 99,1235 ), - ( 98,1127 ), - ( 97,1042 ), - ( 96,1425 ), - ( 95,1043 ), - ( 94,1319 ), - ( 93,1387 ), - ( 92,1221 ), - ( 91,1109 ), - ( 90,1237 ), - ( 89,1400 ), - ( 88,1189 ), - ( 87,1435 ), - ( 86,1115 ), - ( 85,1157 ), - ( 84,1199 ), - ( 83,1395 ), - ( 82, 948 ), - ( 81,1127 ), - ( 80,1231 ), - ( 79,1208 ), - ( 78,1192 ), - ( 77,1250 ), - ( 76,1152 ), - ( 75,1124 ), - ( 74,1185 ), - ( 73, 962 ), - ( 72,1182 ), - ( 71,1174 ), - ( 70,1128 ), - ( 69,1175 ), - ( 68,1193 ), - ( 67,1245 ), - ( 66,1257 ), - ( 65,1264 ), - ( 64,1240 ), - ( 63,1450 ), - ( 62,1133 ), - ( 61,1311 ), - ( 60,1460 ), - ( 59,1390 ), - ( 58,1233 ), - ( 57,1304 ), - ( 56,1679 ), - ( 55,1570 ), - ( 54,1885 ), - ( 53,1656 ), - ( 52,1691 ), - ( 51,2226 ), - ( 50,1822 ), - ( 49,2810 ), - ( 48,2344 ), - ( 47,1431 ), - ( 46,2379 ), - ( 45,1710 ), - ( 44,1705 ), - ( 43,2124 ), - ( 42,2106 ), - ( 41,1775 ), - ( 40,1897 ), - ( 39,1308 ), - ( 38,1321 ), - ( 37,1993 ), - ( 36,1283 ), - ( 35,1038 ), - ( 34,1007 ), - ( 33,617 ), - ( 32,983 ), - ( 31,1079 ), - ( 30,710 ), - ( 29,785 ), - ( 28,636 ), - ( 27,493 ), - ( 26,638 ), - ( 25,633 ), - ( 24,677 ), - ( 23,530 ), - ( 22,372 ), - ( 21,387 ), - ( 20,348 ), - ( 19,292 ), - ( 18,293 ), - ( 17,244 ), - ( 16,196 ), - ( 15,138 ), - ( 14,217 ), - ( 13,114 ), - ( 12,120 ), - ( 11,114 ), - ( 10,112 ), - ( 9,82 ), - ( 8,56 ), - ( 7,101 ), - ( 6,88 ), - ( 5,35 ), - ( 4,35 ), - ( 3,18 ), - ( 2,21 ), - ( 1,21 ), - ]: - meeting = Meeting.objects.filter(type='ietf', - number=number).first(); - meeting.attendees = attendees - meeting.save() - - -def reverse(apps, schema_editor): - pass - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0005_backfill_old_meetings'), - ] - - operations = [ - migrations.RunPython(backfill_old_meetings, reverse) - ] - diff --git a/ietf/meeting/migrations/0007_attended_origin_attended_time.py b/ietf/meeting/migrations/0007_attended_origin_attended_time.py new file mode 100644 index 0000000000..09a8d90e07 --- /dev/null +++ b/ietf/meeting/migrations/0007_attended_origin_attended_time.py @@ -0,0 +1,26 @@ +# Copyright The IETF Trust 2024, All Rights Reserved + +from django.db import migrations, models +import django.utils.timezone + + +class Migration(migrations.Migration): + + dependencies = [ + ("meeting", "0006_alter_sessionpresentation_document_and_session"), + ] + + operations = [ + migrations.AddField( + model_name="attended", + name="origin", + field=models.CharField(default="datatracker", max_length=32), + ), + migrations.AddField( + model_name="attended", + name="time", + field=models.DateTimeField( + blank=True, default=django.utils.timezone.now, null=True + ), + ), + ] diff --git a/ietf/meeting/migrations/0007_auto_20180716_1337.py b/ietf/meeting/migrations/0007_auto_20180716_1337.py deleted file mode 100644 index a20d9dba91..0000000000 --- a/ietf/meeting/migrations/0007_auto_20180716_1337.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.14 on 2018-07-16 13:37 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0006_backfill_attendees'), - ] - - operations = [ - migrations.AlterField( - model_name='meeting', - name='time_zone', - field=models.CharField(blank=True, choices=[('', '---------'), ('Africa/Abidjan', 'Africa/Abidjan'), ('Africa/Accra', 'Africa/Accra'), ('Africa/Addis_Ababa', 'Africa/Addis_Ababa'), ('Africa/Algiers', 'Africa/Algiers'), ('Africa/Asmara', 'Africa/Asmara'), ('Africa/Bamako', 'Africa/Bamako'), ('Africa/Bangui', 'Africa/Bangui'), ('Africa/Banjul', 'Africa/Banjul'), ('Africa/Bissau', 'Africa/Bissau'), ('Africa/Blantyre', 'Africa/Blantyre'), ('Africa/Brazzaville', 'Africa/Brazzaville'), ('Africa/Bujumbura', 'Africa/Bujumbura'), ('Africa/Cairo', 'Africa/Cairo'), ('Africa/Casablanca', 'Africa/Casablanca'), ('Africa/Ceuta', 'Africa/Ceuta'), ('Africa/Conakry', 'Africa/Conakry'), ('Africa/Dakar', 'Africa/Dakar'), ('Africa/Dar_es_Salaam', 'Africa/Dar_es_Salaam'), ('Africa/Djibouti', 'Africa/Djibouti'), ('Africa/Douala', 'Africa/Douala'), ('Africa/El_Aaiun', 'Africa/El_Aaiun'), ('Africa/Freetown', 'Africa/Freetown'), ('Africa/Gaborone', 'Africa/Gaborone'), ('Africa/Harare', 'Africa/Harare'), ('Africa/Johannesburg', 'Africa/Johannesburg'), ('Africa/Juba', 'Africa/Juba'), ('Africa/Kampala', 'Africa/Kampala'), ('Africa/Khartoum', 'Africa/Khartoum'), ('Africa/Kigali', 'Africa/Kigali'), ('Africa/Kinshasa', 'Africa/Kinshasa'), ('Africa/Lagos', 'Africa/Lagos'), ('Africa/Libreville', 'Africa/Libreville'), ('Africa/Lome', 'Africa/Lome'), ('Africa/Luanda', 'Africa/Luanda'), ('Africa/Lubumbashi', 'Africa/Lubumbashi'), ('Africa/Lusaka', 'Africa/Lusaka'), ('Africa/Malabo', 'Africa/Malabo'), ('Africa/Maputo', 'Africa/Maputo'), ('Africa/Maseru', 'Africa/Maseru'), ('Africa/Mbabane', 'Africa/Mbabane'), ('Africa/Mogadishu', 'Africa/Mogadishu'), ('Africa/Monrovia', 'Africa/Monrovia'), ('Africa/Nairobi', 'Africa/Nairobi'), ('Africa/Ndjamena', 'Africa/Ndjamena'), ('Africa/Niamey', 'Africa/Niamey'), ('Africa/Nouakchott', 'Africa/Nouakchott'), ('Africa/Ouagadougou', 'Africa/Ouagadougou'), ('Africa/Porto-Novo', 'Africa/Porto-Novo'), ('Africa/Sao_Tome', 'Africa/Sao_Tome'), ('Africa/Tripoli', 'Africa/Tripoli'), ('Africa/Tunis', 'Africa/Tunis'), ('Africa/Windhoek', 'Africa/Windhoek'), ('America/Adak', 'America/Adak'), ('America/Anchorage', 'America/Anchorage'), ('America/Anguilla', 'America/Anguilla'), ('America/Antigua', 'America/Antigua'), ('America/Araguaina', 'America/Araguaina'), ('America/Argentina/Buenos_Aires', 'America/Argentina/Buenos_Aires'), ('America/Argentina/Catamarca', 'America/Argentina/Catamarca'), ('America/Argentina/Cordoba', 'America/Argentina/Cordoba'), ('America/Argentina/Jujuy', 'America/Argentina/Jujuy'), ('America/Argentina/La_Rioja', 'America/Argentina/La_Rioja'), ('America/Argentina/Mendoza', 'America/Argentina/Mendoza'), ('America/Argentina/Rio_Gallegos', 'America/Argentina/Rio_Gallegos'), ('America/Argentina/Salta', 'America/Argentina/Salta'), ('America/Argentina/San_Juan', 'America/Argentina/San_Juan'), ('America/Argentina/San_Luis', 'America/Argentina/San_Luis'), ('America/Argentina/Tucuman', 'America/Argentina/Tucuman'), ('America/Argentina/Ushuaia', 'America/Argentina/Ushuaia'), ('America/Aruba', 'America/Aruba'), ('America/Asuncion', 'America/Asuncion'), ('America/Atikokan', 'America/Atikokan'), ('America/Bahia', 'America/Bahia'), ('America/Bahia_Banderas', 'America/Bahia_Banderas'), ('America/Barbados', 'America/Barbados'), ('America/Belem', 'America/Belem'), ('America/Belize', 'America/Belize'), ('America/Blanc-Sablon', 'America/Blanc-Sablon'), ('America/Boa_Vista', 'America/Boa_Vista'), ('America/Bogota', 'America/Bogota'), ('America/Boise', 'America/Boise'), ('America/Cambridge_Bay', 'America/Cambridge_Bay'), ('America/Campo_Grande', 'America/Campo_Grande'), ('America/Cancun', 'America/Cancun'), ('America/Caracas', 'America/Caracas'), ('America/Cayenne', 'America/Cayenne'), ('America/Cayman', 'America/Cayman'), ('America/Chicago', 'America/Chicago'), ('America/Chihuahua', 'America/Chihuahua'), ('America/Costa_Rica', 'America/Costa_Rica'), ('America/Creston', 'America/Creston'), ('America/Cuiaba', 'America/Cuiaba'), ('America/Curacao', 'America/Curacao'), ('America/Danmarkshavn', 'America/Danmarkshavn'), ('America/Dawson', 'America/Dawson'), ('America/Dawson_Creek', 'America/Dawson_Creek'), ('America/Denver', 'America/Denver'), ('America/Detroit', 'America/Detroit'), ('America/Dominica', 'America/Dominica'), ('America/Edmonton', 'America/Edmonton'), ('America/Eirunepe', 'America/Eirunepe'), ('America/El_Salvador', 'America/El_Salvador'), ('America/Fort_Nelson', 'America/Fort_Nelson'), ('America/Fortaleza', 'America/Fortaleza'), ('America/Glace_Bay', 'America/Glace_Bay'), ('America/Godthab', 'America/Godthab'), ('America/Goose_Bay', 'America/Goose_Bay'), ('America/Grand_Turk', 'America/Grand_Turk'), ('America/Grenada', 'America/Grenada'), ('America/Guadeloupe', 'America/Guadeloupe'), ('America/Guatemala', 'America/Guatemala'), ('America/Guayaquil', 'America/Guayaquil'), ('America/Guyana', 'America/Guyana'), ('America/Halifax', 'America/Halifax'), ('America/Havana', 'America/Havana'), ('America/Hermosillo', 'America/Hermosillo'), ('America/Indiana/Indianapolis', 'America/Indiana/Indianapolis'), ('America/Indiana/Knox', 'America/Indiana/Knox'), ('America/Indiana/Marengo', 'America/Indiana/Marengo'), ('America/Indiana/Petersburg', 'America/Indiana/Petersburg'), ('America/Indiana/Tell_City', 'America/Indiana/Tell_City'), ('America/Indiana/Vevay', 'America/Indiana/Vevay'), ('America/Indiana/Vincennes', 'America/Indiana/Vincennes'), ('America/Indiana/Winamac', 'America/Indiana/Winamac'), ('America/Inuvik', 'America/Inuvik'), ('America/Iqaluit', 'America/Iqaluit'), ('America/Jamaica', 'America/Jamaica'), ('America/Juneau', 'America/Juneau'), ('America/Kentucky/Louisville', 'America/Kentucky/Louisville'), ('America/Kentucky/Monticello', 'America/Kentucky/Monticello'), ('America/La_Paz', 'America/La_Paz'), ('America/Lima', 'America/Lima'), ('America/Los_Angeles', 'America/Los_Angeles'), ('America/Maceio', 'America/Maceio'), ('America/Managua', 'America/Managua'), ('America/Manaus', 'America/Manaus'), ('America/Martinique', 'America/Martinique'), ('America/Matamoros', 'America/Matamoros'), ('America/Mazatlan', 'America/Mazatlan'), ('America/Menominee', 'America/Menominee'), ('America/Merida', 'America/Merida'), ('America/Metlakatla', 'America/Metlakatla'), ('America/Mexico_City', 'America/Mexico_City'), ('America/Miquelon', 'America/Miquelon'), ('America/Moncton', 'America/Moncton'), ('America/Monterrey', 'America/Monterrey'), ('America/Montevideo', 'America/Montevideo'), ('America/Montserrat', 'America/Montserrat'), ('America/Nassau', 'America/Nassau'), ('America/New_York', 'America/New_York'), ('America/Nipigon', 'America/Nipigon'), ('America/Nome', 'America/Nome'), ('America/Noronha', 'America/Noronha'), ('America/North_Dakota/Beulah', 'America/North_Dakota/Beulah'), ('America/North_Dakota/Center', 'America/North_Dakota/Center'), ('America/North_Dakota/New_Salem', 'America/North_Dakota/New_Salem'), ('America/Ojinaga', 'America/Ojinaga'), ('America/Panama', 'America/Panama'), ('America/Pangnirtung', 'America/Pangnirtung'), ('America/Paramaribo', 'America/Paramaribo'), ('America/Phoenix', 'America/Phoenix'), ('America/Port-au-Prince', 'America/Port-au-Prince'), ('America/Port_of_Spain', 'America/Port_of_Spain'), ('America/Porto_Velho', 'America/Porto_Velho'), ('America/Puerto_Rico', 'America/Puerto_Rico'), ('America/Punta_Arenas', 'America/Punta_Arenas'), ('America/Rainy_River', 'America/Rainy_River'), ('America/Rankin_Inlet', 'America/Rankin_Inlet'), ('America/Recife', 'America/Recife'), ('America/Regina', 'America/Regina'), ('America/Resolute', 'America/Resolute'), ('America/Rio_Branco', 'America/Rio_Branco'), ('America/Santarem', 'America/Santarem'), ('America/Santiago', 'America/Santiago'), ('America/Santo_Domingo', 'America/Santo_Domingo'), ('America/Sao_Paulo', 'America/Sao_Paulo'), ('America/Scoresbysund', 'America/Scoresbysund'), ('America/Sitka', 'America/Sitka'), ('America/St_Johns', 'America/St_Johns'), ('America/St_Kitts', 'America/St_Kitts'), ('America/St_Lucia', 'America/St_Lucia'), ('America/St_Thomas', 'America/St_Thomas'), ('America/St_Vincent', 'America/St_Vincent'), ('America/Swift_Current', 'America/Swift_Current'), ('America/Tegucigalpa', 'America/Tegucigalpa'), ('America/Thule', 'America/Thule'), ('America/Thunder_Bay', 'America/Thunder_Bay'), ('America/Tijuana', 'America/Tijuana'), ('America/Toronto', 'America/Toronto'), ('America/Tortola', 'America/Tortola'), ('America/Vancouver', 'America/Vancouver'), ('America/Whitehorse', 'America/Whitehorse'), ('America/Winnipeg', 'America/Winnipeg'), ('America/Yakutat', 'America/Yakutat'), ('America/Yellowknife', 'America/Yellowknife'), ('Antarctica/Casey', 'Antarctica/Casey'), ('Antarctica/Davis', 'Antarctica/Davis'), ('Antarctica/DumontDUrville', 'Antarctica/DumontDUrville'), ('Antarctica/Macquarie', 'Antarctica/Macquarie'), ('Antarctica/Mawson', 'Antarctica/Mawson'), ('Antarctica/McMurdo', 'Antarctica/McMurdo'), ('Antarctica/Palmer', 'Antarctica/Palmer'), ('Antarctica/Rothera', 'Antarctica/Rothera'), ('Antarctica/Syowa', 'Antarctica/Syowa'), ('Antarctica/Troll', 'Antarctica/Troll'), ('Antarctica/Vostok', 'Antarctica/Vostok'), ('Asia/Aden', 'Asia/Aden'), ('Asia/Almaty', 'Asia/Almaty'), ('Asia/Amman', 'Asia/Amman'), ('Asia/Anadyr', 'Asia/Anadyr'), ('Asia/Aqtau', 'Asia/Aqtau'), ('Asia/Aqtobe', 'Asia/Aqtobe'), ('Asia/Ashgabat', 'Asia/Ashgabat'), ('Asia/Atyrau', 'Asia/Atyrau'), ('Asia/Baghdad', 'Asia/Baghdad'), ('Asia/Bahrain', 'Asia/Bahrain'), ('Asia/Baku', 'Asia/Baku'), ('Asia/Bangkok', 'Asia/Bangkok'), ('Asia/Barnaul', 'Asia/Barnaul'), ('Asia/Beirut', 'Asia/Beirut'), ('Asia/Bishkek', 'Asia/Bishkek'), ('Asia/Brunei', 'Asia/Brunei'), ('Asia/Chita', 'Asia/Chita'), ('Asia/Choibalsan', 'Asia/Choibalsan'), ('Asia/Colombo', 'Asia/Colombo'), ('Asia/Damascus', 'Asia/Damascus'), ('Asia/Dhaka', 'Asia/Dhaka'), ('Asia/Dili', 'Asia/Dili'), ('Asia/Dubai', 'Asia/Dubai'), ('Asia/Dushanbe', 'Asia/Dushanbe'), ('Asia/Famagusta', 'Asia/Famagusta'), ('Asia/Gaza', 'Asia/Gaza'), ('Asia/Hebron', 'Asia/Hebron'), ('Asia/Ho_Chi_Minh', 'Asia/Ho_Chi_Minh'), ('Asia/Hong_Kong', 'Asia/Hong_Kong'), ('Asia/Hovd', 'Asia/Hovd'), ('Asia/Irkutsk', 'Asia/Irkutsk'), ('Asia/Jakarta', 'Asia/Jakarta'), ('Asia/Jayapura', 'Asia/Jayapura'), ('Asia/Jerusalem', 'Asia/Jerusalem'), ('Asia/Kabul', 'Asia/Kabul'), ('Asia/Kamchatka', 'Asia/Kamchatka'), ('Asia/Karachi', 'Asia/Karachi'), ('Asia/Kathmandu', 'Asia/Kathmandu'), ('Asia/Khandyga', 'Asia/Khandyga'), ('Asia/Kolkata', 'Asia/Kolkata'), ('Asia/Krasnoyarsk', 'Asia/Krasnoyarsk'), ('Asia/Kuala_Lumpur', 'Asia/Kuala_Lumpur'), ('Asia/Kuching', 'Asia/Kuching'), ('Asia/Kuwait', 'Asia/Kuwait'), ('Asia/Macau', 'Asia/Macau'), ('Asia/Magadan', 'Asia/Magadan'), ('Asia/Makassar', 'Asia/Makassar'), ('Asia/Manila', 'Asia/Manila'), ('Asia/Muscat', 'Asia/Muscat'), ('Asia/Nicosia', 'Asia/Nicosia'), ('Asia/Novokuznetsk', 'Asia/Novokuznetsk'), ('Asia/Novosibirsk', 'Asia/Novosibirsk'), ('Asia/Omsk', 'Asia/Omsk'), ('Asia/Oral', 'Asia/Oral'), ('Asia/Phnom_Penh', 'Asia/Phnom_Penh'), ('Asia/Pontianak', 'Asia/Pontianak'), ('Asia/Pyongyang', 'Asia/Pyongyang'), ('Asia/Qatar', 'Asia/Qatar'), ('Asia/Qyzylorda', 'Asia/Qyzylorda'), ('Asia/Riyadh', 'Asia/Riyadh'), ('Asia/Sakhalin', 'Asia/Sakhalin'), ('Asia/Samarkand', 'Asia/Samarkand'), ('Asia/Seoul', 'Asia/Seoul'), ('Asia/Shanghai', 'Asia/Shanghai'), ('Asia/Singapore', 'Asia/Singapore'), ('Asia/Srednekolymsk', 'Asia/Srednekolymsk'), ('Asia/Taipei', 'Asia/Taipei'), ('Asia/Tashkent', 'Asia/Tashkent'), ('Asia/Tbilisi', 'Asia/Tbilisi'), ('Asia/Tehran', 'Asia/Tehran'), ('Asia/Thimphu', 'Asia/Thimphu'), ('Asia/Tokyo', 'Asia/Tokyo'), ('Asia/Tomsk', 'Asia/Tomsk'), ('Asia/Ulaanbaatar', 'Asia/Ulaanbaatar'), ('Asia/Urumqi', 'Asia/Urumqi'), ('Asia/Ust-Nera', 'Asia/Ust-Nera'), ('Asia/Vientiane', 'Asia/Vientiane'), ('Asia/Vladivostok', 'Asia/Vladivostok'), ('Asia/Yakutsk', 'Asia/Yakutsk'), ('Asia/Yangon', 'Asia/Yangon'), ('Asia/Yekaterinburg', 'Asia/Yekaterinburg'), ('Asia/Yerevan', 'Asia/Yerevan'), ('Atlantic/Azores', 'Atlantic/Azores'), ('Atlantic/Bermuda', 'Atlantic/Bermuda'), ('Atlantic/Canary', 'Atlantic/Canary'), ('Atlantic/Cape_Verde', 'Atlantic/Cape_Verde'), ('Atlantic/Faroe', 'Atlantic/Faroe'), ('Atlantic/Madeira', 'Atlantic/Madeira'), ('Atlantic/Reykjavik', 'Atlantic/Reykjavik'), ('Atlantic/South_Georgia', 'Atlantic/South_Georgia'), ('Atlantic/St_Helena', 'Atlantic/St_Helena'), ('Atlantic/Stanley', 'Atlantic/Stanley'), ('Australia/Adelaide', 'Australia/Adelaide'), ('Australia/Brisbane', 'Australia/Brisbane'), ('Australia/Broken_Hill', 'Australia/Broken_Hill'), ('Australia/Currie', 'Australia/Currie'), ('Australia/Darwin', 'Australia/Darwin'), ('Australia/Eucla', 'Australia/Eucla'), ('Australia/Hobart', 'Australia/Hobart'), ('Australia/Lindeman', 'Australia/Lindeman'), ('Australia/Lord_Howe', 'Australia/Lord_Howe'), ('Australia/Melbourne', 'Australia/Melbourne'), ('Australia/Perth', 'Australia/Perth'), ('Australia/Sydney', 'Australia/Sydney'), ('Europe/Amsterdam', 'Europe/Amsterdam'), ('Europe/Andorra', 'Europe/Andorra'), ('Europe/Astrakhan', 'Europe/Astrakhan'), ('Europe/Athens', 'Europe/Athens'), ('Europe/Belgrade', 'Europe/Belgrade'), ('Europe/Berlin', 'Europe/Berlin'), ('Europe/Brussels', 'Europe/Brussels'), ('Europe/Bucharest', 'Europe/Bucharest'), ('Europe/Budapest', 'Europe/Budapest'), ('Europe/Chisinau', 'Europe/Chisinau'), ('Europe/Copenhagen', 'Europe/Copenhagen'), ('Europe/Dublin', 'Europe/Dublin'), ('Europe/Gibraltar', 'Europe/Gibraltar'), ('Europe/Helsinki', 'Europe/Helsinki'), ('Europe/Istanbul', 'Europe/Istanbul'), ('Europe/Kaliningrad', 'Europe/Kaliningrad'), ('Europe/Kiev', 'Europe/Kiev'), ('Europe/Kirov', 'Europe/Kirov'), ('Europe/Lisbon', 'Europe/Lisbon'), ('Europe/London', 'Europe/London'), ('Europe/Luxembourg', 'Europe/Luxembourg'), ('Europe/Madrid', 'Europe/Madrid'), ('Europe/Malta', 'Europe/Malta'), ('Europe/Minsk', 'Europe/Minsk'), ('Europe/Monaco', 'Europe/Monaco'), ('Europe/Moscow', 'Europe/Moscow'), ('Europe/Oslo', 'Europe/Oslo'), ('Europe/Paris', 'Europe/Paris'), ('Europe/Prague', 'Europe/Prague'), ('Europe/Riga', 'Europe/Riga'), ('Europe/Rome', 'Europe/Rome'), ('Europe/Samara', 'Europe/Samara'), ('Europe/Saratov', 'Europe/Saratov'), ('Europe/Simferopol', 'Europe/Simferopol'), ('Europe/Sofia', 'Europe/Sofia'), ('Europe/Stockholm', 'Europe/Stockholm'), ('Europe/Tallinn', 'Europe/Tallinn'), ('Europe/Tirane', 'Europe/Tirane'), ('Europe/Ulyanovsk', 'Europe/Ulyanovsk'), ('Europe/Uzhgorod', 'Europe/Uzhgorod'), ('Europe/Vaduz', 'Europe/Vaduz'), ('Europe/Vienna', 'Europe/Vienna'), ('Europe/Vilnius', 'Europe/Vilnius'), ('Europe/Volgograd', 'Europe/Volgograd'), ('Europe/Warsaw', 'Europe/Warsaw'), ('Europe/Zaporozhye', 'Europe/Zaporozhye'), ('Europe/Zurich', 'Europe/Zurich'), ('GMT', 'GMT'), ('Indian/Antananarivo', 'Indian/Antananarivo'), ('Indian/Chagos', 'Indian/Chagos'), ('Indian/Christmas', 'Indian/Christmas'), ('Indian/Cocos', 'Indian/Cocos'), ('Indian/Comoro', 'Indian/Comoro'), ('Indian/Kerguelen', 'Indian/Kerguelen'), ('Indian/Mahe', 'Indian/Mahe'), ('Indian/Maldives', 'Indian/Maldives'), ('Indian/Mauritius', 'Indian/Mauritius'), ('Indian/Mayotte', 'Indian/Mayotte'), ('Indian/Reunion', 'Indian/Reunion'), ('Pacific/Apia', 'Pacific/Apia'), ('Pacific/Auckland', 'Pacific/Auckland'), ('Pacific/Bougainville', 'Pacific/Bougainville'), ('Pacific/Chatham', 'Pacific/Chatham'), ('Pacific/Chuuk', 'Pacific/Chuuk'), ('Pacific/Easter', 'Pacific/Easter'), ('Pacific/Efate', 'Pacific/Efate'), ('Pacific/Enderbury', 'Pacific/Enderbury'), ('Pacific/Fakaofo', 'Pacific/Fakaofo'), ('Pacific/Fiji', 'Pacific/Fiji'), ('Pacific/Funafuti', 'Pacific/Funafuti'), ('Pacific/Galapagos', 'Pacific/Galapagos'), ('Pacific/Gambier', 'Pacific/Gambier'), ('Pacific/Guadalcanal', 'Pacific/Guadalcanal'), ('Pacific/Guam', 'Pacific/Guam'), ('Pacific/Honolulu', 'Pacific/Honolulu'), ('Pacific/Kiritimati', 'Pacific/Kiritimati'), ('Pacific/Kosrae', 'Pacific/Kosrae'), ('Pacific/Kwajalein', 'Pacific/Kwajalein'), ('Pacific/Majuro', 'Pacific/Majuro'), ('Pacific/Marquesas', 'Pacific/Marquesas'), ('Pacific/Midway', 'Pacific/Midway'), ('Pacific/Nauru', 'Pacific/Nauru'), ('Pacific/Niue', 'Pacific/Niue'), ('Pacific/Norfolk', 'Pacific/Norfolk'), ('Pacific/Noumea', 'Pacific/Noumea'), ('Pacific/Pago_Pago', 'Pacific/Pago_Pago'), ('Pacific/Palau', 'Pacific/Palau'), ('Pacific/Pitcairn', 'Pacific/Pitcairn'), ('Pacific/Pohnpei', 'Pacific/Pohnpei'), ('Pacific/Port_Moresby', 'Pacific/Port_Moresby'), ('Pacific/Rarotonga', 'Pacific/Rarotonga'), ('Pacific/Saipan', 'Pacific/Saipan'), ('Pacific/Tahiti', 'Pacific/Tahiti'), ('Pacific/Tarawa', 'Pacific/Tarawa'), ('Pacific/Tongatapu', 'Pacific/Tongatapu'), ('Pacific/Wake', 'Pacific/Wake'), ('Pacific/Wallis', 'Pacific/Wallis'), ('UTC', 'UTC')], max_length=255), - ), - ] diff --git a/ietf/meeting/migrations/0008_remove_schedtimesessassignment_notes.py b/ietf/meeting/migrations/0008_remove_schedtimesessassignment_notes.py new file mode 100644 index 0000000000..3c0b85fc22 --- /dev/null +++ b/ietf/meeting/migrations/0008_remove_schedtimesessassignment_notes.py @@ -0,0 +1,17 @@ +# Generated by Django 4.2.15 on 2024-08-16 13:49 + +from django.db import migrations + + +class Migration(migrations.Migration): + + dependencies = [ + ("meeting", "0007_attended_origin_attended_time"), + ] + + operations = [ + migrations.RemoveField( + model_name="schedtimesessassignment", + name="notes", + ), + ] diff --git a/ietf/meeting/migrations/0008_rename_meeting_agenda_note.py b/ietf/meeting/migrations/0008_rename_meeting_agenda_note.py deleted file mode 100644 index 884940c221..0000000000 --- a/ietf/meeting/migrations/0008_rename_meeting_agenda_note.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.15 on 2018-10-09 13:09 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0007_auto_20180716_1337'), - ] - - operations = [ - migrations.RenameField( - model_name='meeting', - old_name='agenda_note', - new_name='agenda_warning_note', - ), - ] diff --git a/ietf/meeting/migrations/0009_add_agenda_info_note.py b/ietf/meeting/migrations/0009_add_agenda_info_note.py deleted file mode 100644 index 8fc4fdb466..0000000000 --- a/ietf/meeting/migrations/0009_add_agenda_info_note.py +++ /dev/null @@ -1,26 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.16 on 2018-10-09 14:07 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0008_rename_meeting_agenda_note'), - ] - - operations = [ - migrations.AddField( - model_name='meeting', - name='agenda_info_note', - field=models.TextField(blank=True, help_text='Text in this field will be placed at the top of the html agenda page for the meeting. HTML can be used, but will not be validated.'), - ), - migrations.AlterField( - model_name='meeting', - name='agenda_warning_note', - field=models.TextField(blank=True, help_text='Text in this field will be placed more prominently at the top of the html agenda page for the meeting. HTML can be used, but will not be validated.'), - ), - ] diff --git a/ietf/meeting/migrations/0009_session_meetecho_recording_name.py b/ietf/meeting/migrations/0009_session_meetecho_recording_name.py new file mode 100644 index 0000000000..79ca4919a3 --- /dev/null +++ b/ietf/meeting/migrations/0009_session_meetecho_recording_name.py @@ -0,0 +1,20 @@ +# Copyright The IETF Trust 2024, All Rights Reserved + +from django.db import migrations, models + + +class Migration(migrations.Migration): + + dependencies = [ + ("meeting", "0008_remove_schedtimesessassignment_notes"), + ] + + operations = [ + migrations.AddField( + model_name="session", + name="meetecho_recording_name", + field=models.CharField( + blank=True, help_text="Name of the meetecho recording", max_length=64 + ), + ), + ] diff --git a/ietf/meeting/migrations/0010_alter_floorplan_image_alter_meetinghost_logo.py b/ietf/meeting/migrations/0010_alter_floorplan_image_alter_meetinghost_logo.py new file mode 100644 index 0000000000..594a1a4048 --- /dev/null +++ b/ietf/meeting/migrations/0010_alter_floorplan_image_alter_meetinghost_logo.py @@ -0,0 +1,56 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +from django.db import migrations, models +import ietf.meeting.models +import ietf.utils.fields +import ietf.utils.storage +import ietf.utils.validators + + +class Migration(migrations.Migration): + + dependencies = [ + ("meeting", "0009_session_meetecho_recording_name"), + ] + + operations = [ + migrations.AlterField( + model_name="floorplan", + name="image", + field=models.ImageField( + blank=True, + default=None, + storage=ietf.utils.storage.BlobShadowFileSystemStorage( + kind="", location=None + ), + upload_to=ietf.meeting.models.floorplan_path, + ), + ), + migrations.AlterField( + model_name="meetinghost", + name="logo", + field=ietf.utils.fields.MissingOkImageField( + height_field="logo_height", + storage=ietf.utils.storage.BlobShadowFileSystemStorage( + kind="", location=None + ), + upload_to=ietf.meeting.models._host_upload_path, + validators=[ + ietf.utils.validators.MaxImageSizeValidator(400, 400), + ietf.utils.validators.WrappedValidator( + ietf.utils.validators.validate_file_size, True + ), + ietf.utils.validators.WrappedValidator( + ietf.utils.validators.validate_file_extension, + [".png", ".jpg", ".jpeg"], + ), + ietf.utils.validators.WrappedValidator( + ietf.utils.validators.validate_mime_type, + ["image/jpeg", "image/png"], + True, + ), + ], + width_field="logo_width", + ), + ), + ] diff --git a/ietf/meeting/migrations/0010_set_ietf_103_agenda_info_note.py b/ietf/meeting/migrations/0010_set_ietf_103_agenda_info_note.py deleted file mode 100644 index bb3148f78f..0000000000 --- a/ietf/meeting/migrations/0010_set_ietf_103_agenda_info_note.py +++ /dev/null @@ -1,30 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.16 on 2018-10-09 14:23 - - -from django.db import migrations - -def forward(apps,schema_editor): - Meeting = apps.get_model('meeting','Meeting') - Meeting.objects.filter(number=103).update(agenda_info_note= - 'To see the list of unofficial side meetings, or to reserve meeting ' - 'space, please see the ' - '' - 'meeting wiki.' - ) - -def reverse(apps,schema_editor): - Meeting = apps.get_model('meeting','Meeting') - Meeting.objects.filter(number=103).update(agenda_info_note="") - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0009_add_agenda_info_note'), - ('person', '0008_auto_20181014_1448'), - ] - - operations = [ - migrations.RunPython(forward,reverse), - ] diff --git a/ietf/meeting/migrations/0011_alter_slidesubmission_doc.py b/ietf/meeting/migrations/0011_alter_slidesubmission_doc.py new file mode 100644 index 0000000000..b9cbc58e99 --- /dev/null +++ b/ietf/meeting/migrations/0011_alter_slidesubmission_doc.py @@ -0,0 +1,26 @@ +# Generated by Django 4.2.19 on 2025-03-17 09:37 + +from django.db import migrations +import django.db.models.deletion +import ietf.utils.models + + +class Migration(migrations.Migration): + + dependencies = [ + ("doc", "0025_storedobject_storedobject_unique_name_per_store"), + ("meeting", "0010_alter_floorplan_image_alter_meetinghost_logo"), + ] + + operations = [ + migrations.AlterField( + model_name="slidesubmission", + name="doc", + field=ietf.utils.models.ForeignKey( + blank=True, + null=True, + on_delete=django.db.models.deletion.SET_NULL, + to="doc.document", + ), + ), + ] diff --git a/ietf/meeting/migrations/0011_auto_20190114_0550.py b/ietf/meeting/migrations/0011_auto_20190114_0550.py deleted file mode 100644 index 50b8b71ec2..0000000000 --- a/ietf/meeting/migrations/0011_auto_20190114_0550.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.18 on 2019-01-14 05:50 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0010_set_ietf_103_agenda_info_note'), - ] - - operations = [ - migrations.AlterField( - model_name='meeting', - name='time_zone', - field=models.CharField(blank=True, choices=[('', '---------'), ('Africa/Abidjan', 'Africa/Abidjan'), ('Africa/Accra', 'Africa/Accra'), ('Africa/Addis_Ababa', 'Africa/Addis_Ababa'), ('Africa/Algiers', 'Africa/Algiers'), ('Africa/Asmara', 'Africa/Asmara'), ('Africa/Bamako', 'Africa/Bamako'), ('Africa/Bangui', 'Africa/Bangui'), ('Africa/Banjul', 'Africa/Banjul'), ('Africa/Bissau', 'Africa/Bissau'), ('Africa/Blantyre', 'Africa/Blantyre'), ('Africa/Brazzaville', 'Africa/Brazzaville'), ('Africa/Bujumbura', 'Africa/Bujumbura'), ('Africa/Cairo', 'Africa/Cairo'), ('Africa/Casablanca', 'Africa/Casablanca'), ('Africa/Ceuta', 'Africa/Ceuta'), ('Africa/Conakry', 'Africa/Conakry'), ('Africa/Dakar', 'Africa/Dakar'), ('Africa/Dar_es_Salaam', 'Africa/Dar_es_Salaam'), ('Africa/Djibouti', 'Africa/Djibouti'), ('Africa/Douala', 'Africa/Douala'), ('Africa/El_Aaiun', 'Africa/El_Aaiun'), ('Africa/Freetown', 'Africa/Freetown'), ('Africa/Gaborone', 'Africa/Gaborone'), ('Africa/Harare', 'Africa/Harare'), ('Africa/Johannesburg', 'Africa/Johannesburg'), ('Africa/Juba', 'Africa/Juba'), ('Africa/Kampala', 'Africa/Kampala'), ('Africa/Khartoum', 'Africa/Khartoum'), ('Africa/Kigali', 'Africa/Kigali'), ('Africa/Kinshasa', 'Africa/Kinshasa'), ('Africa/Lagos', 'Africa/Lagos'), ('Africa/Libreville', 'Africa/Libreville'), ('Africa/Lome', 'Africa/Lome'), ('Africa/Luanda', 'Africa/Luanda'), ('Africa/Lubumbashi', 'Africa/Lubumbashi'), ('Africa/Lusaka', 'Africa/Lusaka'), ('Africa/Malabo', 'Africa/Malabo'), ('Africa/Maputo', 'Africa/Maputo'), ('Africa/Maseru', 'Africa/Maseru'), ('Africa/Mbabane', 'Africa/Mbabane'), ('Africa/Mogadishu', 'Africa/Mogadishu'), ('Africa/Monrovia', 'Africa/Monrovia'), ('Africa/Nairobi', 'Africa/Nairobi'), ('Africa/Ndjamena', 'Africa/Ndjamena'), ('Africa/Niamey', 'Africa/Niamey'), ('Africa/Nouakchott', 'Africa/Nouakchott'), ('Africa/Ouagadougou', 'Africa/Ouagadougou'), ('Africa/Porto-Novo', 'Africa/Porto-Novo'), ('Africa/Sao_Tome', 'Africa/Sao_Tome'), ('Africa/Tripoli', 'Africa/Tripoli'), ('Africa/Tunis', 'Africa/Tunis'), ('Africa/Windhoek', 'Africa/Windhoek'), ('America/Adak', 'America/Adak'), ('America/Anchorage', 'America/Anchorage'), ('America/Anguilla', 'America/Anguilla'), ('America/Antigua', 'America/Antigua'), ('America/Araguaina', 'America/Araguaina'), ('America/Argentina/Buenos_Aires', 'America/Argentina/Buenos_Aires'), ('America/Argentina/Catamarca', 'America/Argentina/Catamarca'), ('America/Argentina/Cordoba', 'America/Argentina/Cordoba'), ('America/Argentina/Jujuy', 'America/Argentina/Jujuy'), ('America/Argentina/La_Rioja', 'America/Argentina/La_Rioja'), ('America/Argentina/Mendoza', 'America/Argentina/Mendoza'), ('America/Argentina/Rio_Gallegos', 'America/Argentina/Rio_Gallegos'), ('America/Argentina/Salta', 'America/Argentina/Salta'), ('America/Argentina/San_Juan', 'America/Argentina/San_Juan'), ('America/Argentina/San_Luis', 'America/Argentina/San_Luis'), ('America/Argentina/Tucuman', 'America/Argentina/Tucuman'), ('America/Argentina/Ushuaia', 'America/Argentina/Ushuaia'), ('America/Aruba', 'America/Aruba'), ('America/Asuncion', 'America/Asuncion'), ('America/Atikokan', 'America/Atikokan'), ('America/Bahia', 'America/Bahia'), ('America/Bahia_Banderas', 'America/Bahia_Banderas'), ('America/Barbados', 'America/Barbados'), ('America/Belem', 'America/Belem'), ('America/Belize', 'America/Belize'), ('America/Blanc-Sablon', 'America/Blanc-Sablon'), ('America/Boa_Vista', 'America/Boa_Vista'), ('America/Bogota', 'America/Bogota'), ('America/Boise', 'America/Boise'), ('America/Cambridge_Bay', 'America/Cambridge_Bay'), ('America/Campo_Grande', 'America/Campo_Grande'), ('America/Cancun', 'America/Cancun'), ('America/Caracas', 'America/Caracas'), ('America/Cayenne', 'America/Cayenne'), ('America/Cayman', 'America/Cayman'), ('America/Chicago', 'America/Chicago'), ('America/Chihuahua', 'America/Chihuahua'), ('America/Costa_Rica', 'America/Costa_Rica'), ('America/Creston', 'America/Creston'), ('America/Cuiaba', 'America/Cuiaba'), ('America/Curacao', 'America/Curacao'), ('America/Danmarkshavn', 'America/Danmarkshavn'), ('America/Dawson', 'America/Dawson'), ('America/Dawson_Creek', 'America/Dawson_Creek'), ('America/Denver', 'America/Denver'), ('America/Detroit', 'America/Detroit'), ('America/Dominica', 'America/Dominica'), ('America/Edmonton', 'America/Edmonton'), ('America/Eirunepe', 'America/Eirunepe'), ('America/El_Salvador', 'America/El_Salvador'), ('America/Fort_Nelson', 'America/Fort_Nelson'), ('America/Fortaleza', 'America/Fortaleza'), ('America/Glace_Bay', 'America/Glace_Bay'), ('America/Godthab', 'America/Godthab'), ('America/Goose_Bay', 'America/Goose_Bay'), ('America/Grand_Turk', 'America/Grand_Turk'), ('America/Grenada', 'America/Grenada'), ('America/Guadeloupe', 'America/Guadeloupe'), ('America/Guatemala', 'America/Guatemala'), ('America/Guayaquil', 'America/Guayaquil'), ('America/Guyana', 'America/Guyana'), ('America/Halifax', 'America/Halifax'), ('America/Havana', 'America/Havana'), ('America/Hermosillo', 'America/Hermosillo'), ('America/Indiana/Indianapolis', 'America/Indiana/Indianapolis'), ('America/Indiana/Knox', 'America/Indiana/Knox'), ('America/Indiana/Marengo', 'America/Indiana/Marengo'), ('America/Indiana/Petersburg', 'America/Indiana/Petersburg'), ('America/Indiana/Tell_City', 'America/Indiana/Tell_City'), ('America/Indiana/Vevay', 'America/Indiana/Vevay'), ('America/Indiana/Vincennes', 'America/Indiana/Vincennes'), ('America/Indiana/Winamac', 'America/Indiana/Winamac'), ('America/Inuvik', 'America/Inuvik'), ('America/Iqaluit', 'America/Iqaluit'), ('America/Jamaica', 'America/Jamaica'), ('America/Juneau', 'America/Juneau'), ('America/Kentucky/Louisville', 'America/Kentucky/Louisville'), ('America/Kentucky/Monticello', 'America/Kentucky/Monticello'), ('America/La_Paz', 'America/La_Paz'), ('America/Lima', 'America/Lima'), ('America/Los_Angeles', 'America/Los_Angeles'), ('America/Maceio', 'America/Maceio'), ('America/Managua', 'America/Managua'), ('America/Manaus', 'America/Manaus'), ('America/Martinique', 'America/Martinique'), ('America/Matamoros', 'America/Matamoros'), ('America/Mazatlan', 'America/Mazatlan'), ('America/Menominee', 'America/Menominee'), ('America/Merida', 'America/Merida'), ('America/Metlakatla', 'America/Metlakatla'), ('America/Mexico_City', 'America/Mexico_City'), ('America/Miquelon', 'America/Miquelon'), ('America/Moncton', 'America/Moncton'), ('America/Monterrey', 'America/Monterrey'), ('America/Montevideo', 'America/Montevideo'), ('America/Montserrat', 'America/Montserrat'), ('America/Nassau', 'America/Nassau'), ('America/New_York', 'America/New_York'), ('America/Nipigon', 'America/Nipigon'), ('America/Nome', 'America/Nome'), ('America/Noronha', 'America/Noronha'), ('America/North_Dakota/Beulah', 'America/North_Dakota/Beulah'), ('America/North_Dakota/Center', 'America/North_Dakota/Center'), ('America/North_Dakota/New_Salem', 'America/North_Dakota/New_Salem'), ('America/Ojinaga', 'America/Ojinaga'), ('America/Panama', 'America/Panama'), ('America/Pangnirtung', 'America/Pangnirtung'), ('America/Paramaribo', 'America/Paramaribo'), ('America/Phoenix', 'America/Phoenix'), ('America/Port-au-Prince', 'America/Port-au-Prince'), ('America/Port_of_Spain', 'America/Port_of_Spain'), ('America/Porto_Velho', 'America/Porto_Velho'), ('America/Puerto_Rico', 'America/Puerto_Rico'), ('America/Punta_Arenas', 'America/Punta_Arenas'), ('America/Rainy_River', 'America/Rainy_River'), ('America/Rankin_Inlet', 'America/Rankin_Inlet'), ('America/Recife', 'America/Recife'), ('America/Regina', 'America/Regina'), ('America/Resolute', 'America/Resolute'), ('America/Rio_Branco', 'America/Rio_Branco'), ('America/Santarem', 'America/Santarem'), ('America/Santiago', 'America/Santiago'), ('America/Santo_Domingo', 'America/Santo_Domingo'), ('America/Sao_Paulo', 'America/Sao_Paulo'), ('America/Scoresbysund', 'America/Scoresbysund'), ('America/Sitka', 'America/Sitka'), ('America/St_Johns', 'America/St_Johns'), ('America/St_Kitts', 'America/St_Kitts'), ('America/St_Lucia', 'America/St_Lucia'), ('America/St_Thomas', 'America/St_Thomas'), ('America/St_Vincent', 'America/St_Vincent'), ('America/Swift_Current', 'America/Swift_Current'), ('America/Tegucigalpa', 'America/Tegucigalpa'), ('America/Thule', 'America/Thule'), ('America/Thunder_Bay', 'America/Thunder_Bay'), ('America/Tijuana', 'America/Tijuana'), ('America/Toronto', 'America/Toronto'), ('America/Tortola', 'America/Tortola'), ('America/Vancouver', 'America/Vancouver'), ('America/Whitehorse', 'America/Whitehorse'), ('America/Winnipeg', 'America/Winnipeg'), ('America/Yakutat', 'America/Yakutat'), ('America/Yellowknife', 'America/Yellowknife'), ('Antarctica/Casey', 'Antarctica/Casey'), ('Antarctica/Davis', 'Antarctica/Davis'), ('Antarctica/DumontDUrville', 'Antarctica/DumontDUrville'), ('Antarctica/Macquarie', 'Antarctica/Macquarie'), ('Antarctica/Mawson', 'Antarctica/Mawson'), ('Antarctica/McMurdo', 'Antarctica/McMurdo'), ('Antarctica/Palmer', 'Antarctica/Palmer'), ('Antarctica/Rothera', 'Antarctica/Rothera'), ('Antarctica/Syowa', 'Antarctica/Syowa'), ('Antarctica/Troll', 'Antarctica/Troll'), ('Antarctica/Vostok', 'Antarctica/Vostok'), ('Asia/Aden', 'Asia/Aden'), ('Asia/Almaty', 'Asia/Almaty'), ('Asia/Amman', 'Asia/Amman'), ('Asia/Anadyr', 'Asia/Anadyr'), ('Asia/Aqtau', 'Asia/Aqtau'), ('Asia/Aqtobe', 'Asia/Aqtobe'), ('Asia/Ashgabat', 'Asia/Ashgabat'), ('Asia/Atyrau', 'Asia/Atyrau'), ('Asia/Baghdad', 'Asia/Baghdad'), ('Asia/Bahrain', 'Asia/Bahrain'), ('Asia/Baku', 'Asia/Baku'), ('Asia/Bangkok', 'Asia/Bangkok'), ('Asia/Barnaul', 'Asia/Barnaul'), ('Asia/Beirut', 'Asia/Beirut'), ('Asia/Bishkek', 'Asia/Bishkek'), ('Asia/Brunei', 'Asia/Brunei'), ('Asia/Chita', 'Asia/Chita'), ('Asia/Choibalsan', 'Asia/Choibalsan'), ('Asia/Colombo', 'Asia/Colombo'), ('Asia/Damascus', 'Asia/Damascus'), ('Asia/Dhaka', 'Asia/Dhaka'), ('Asia/Dili', 'Asia/Dili'), ('Asia/Dubai', 'Asia/Dubai'), ('Asia/Dushanbe', 'Asia/Dushanbe'), ('Asia/Famagusta', 'Asia/Famagusta'), ('Asia/Gaza', 'Asia/Gaza'), ('Asia/Hebron', 'Asia/Hebron'), ('Asia/Ho_Chi_Minh', 'Asia/Ho_Chi_Minh'), ('Asia/Hong_Kong', 'Asia/Hong_Kong'), ('Asia/Hovd', 'Asia/Hovd'), ('Asia/Irkutsk', 'Asia/Irkutsk'), ('Asia/Jakarta', 'Asia/Jakarta'), ('Asia/Jayapura', 'Asia/Jayapura'), ('Asia/Jerusalem', 'Asia/Jerusalem'), ('Asia/Kabul', 'Asia/Kabul'), ('Asia/Kamchatka', 'Asia/Kamchatka'), ('Asia/Karachi', 'Asia/Karachi'), ('Asia/Kathmandu', 'Asia/Kathmandu'), ('Asia/Khandyga', 'Asia/Khandyga'), ('Asia/Kolkata', 'Asia/Kolkata'), ('Asia/Krasnoyarsk', 'Asia/Krasnoyarsk'), ('Asia/Kuala_Lumpur', 'Asia/Kuala_Lumpur'), ('Asia/Kuching', 'Asia/Kuching'), ('Asia/Kuwait', 'Asia/Kuwait'), ('Asia/Macau', 'Asia/Macau'), ('Asia/Magadan', 'Asia/Magadan'), ('Asia/Makassar', 'Asia/Makassar'), ('Asia/Manila', 'Asia/Manila'), ('Asia/Muscat', 'Asia/Muscat'), ('Asia/Nicosia', 'Asia/Nicosia'), ('Asia/Novokuznetsk', 'Asia/Novokuznetsk'), ('Asia/Novosibirsk', 'Asia/Novosibirsk'), ('Asia/Omsk', 'Asia/Omsk'), ('Asia/Oral', 'Asia/Oral'), ('Asia/Phnom_Penh', 'Asia/Phnom_Penh'), ('Asia/Pontianak', 'Asia/Pontianak'), ('Asia/Pyongyang', 'Asia/Pyongyang'), ('Asia/Qatar', 'Asia/Qatar'), ('Asia/Qostanay', 'Asia/Qostanay'), ('Asia/Qyzylorda', 'Asia/Qyzylorda'), ('Asia/Riyadh', 'Asia/Riyadh'), ('Asia/Sakhalin', 'Asia/Sakhalin'), ('Asia/Samarkand', 'Asia/Samarkand'), ('Asia/Seoul', 'Asia/Seoul'), ('Asia/Shanghai', 'Asia/Shanghai'), ('Asia/Singapore', 'Asia/Singapore'), ('Asia/Srednekolymsk', 'Asia/Srednekolymsk'), ('Asia/Taipei', 'Asia/Taipei'), ('Asia/Tashkent', 'Asia/Tashkent'), ('Asia/Tbilisi', 'Asia/Tbilisi'), ('Asia/Tehran', 'Asia/Tehran'), ('Asia/Thimphu', 'Asia/Thimphu'), ('Asia/Tokyo', 'Asia/Tokyo'), ('Asia/Tomsk', 'Asia/Tomsk'), ('Asia/Ulaanbaatar', 'Asia/Ulaanbaatar'), ('Asia/Urumqi', 'Asia/Urumqi'), ('Asia/Ust-Nera', 'Asia/Ust-Nera'), ('Asia/Vientiane', 'Asia/Vientiane'), ('Asia/Vladivostok', 'Asia/Vladivostok'), ('Asia/Yakutsk', 'Asia/Yakutsk'), ('Asia/Yangon', 'Asia/Yangon'), ('Asia/Yekaterinburg', 'Asia/Yekaterinburg'), ('Asia/Yerevan', 'Asia/Yerevan'), ('Atlantic/Azores', 'Atlantic/Azores'), ('Atlantic/Bermuda', 'Atlantic/Bermuda'), ('Atlantic/Canary', 'Atlantic/Canary'), ('Atlantic/Cape_Verde', 'Atlantic/Cape_Verde'), ('Atlantic/Faroe', 'Atlantic/Faroe'), ('Atlantic/Madeira', 'Atlantic/Madeira'), ('Atlantic/Reykjavik', 'Atlantic/Reykjavik'), ('Atlantic/South_Georgia', 'Atlantic/South_Georgia'), ('Atlantic/St_Helena', 'Atlantic/St_Helena'), ('Atlantic/Stanley', 'Atlantic/Stanley'), ('Australia/Adelaide', 'Australia/Adelaide'), ('Australia/Brisbane', 'Australia/Brisbane'), ('Australia/Broken_Hill', 'Australia/Broken_Hill'), ('Australia/Currie', 'Australia/Currie'), ('Australia/Darwin', 'Australia/Darwin'), ('Australia/Eucla', 'Australia/Eucla'), ('Australia/Hobart', 'Australia/Hobart'), ('Australia/Lindeman', 'Australia/Lindeman'), ('Australia/Lord_Howe', 'Australia/Lord_Howe'), ('Australia/Melbourne', 'Australia/Melbourne'), ('Australia/Perth', 'Australia/Perth'), ('Australia/Sydney', 'Australia/Sydney'), ('Europe/Amsterdam', 'Europe/Amsterdam'), ('Europe/Andorra', 'Europe/Andorra'), ('Europe/Astrakhan', 'Europe/Astrakhan'), ('Europe/Athens', 'Europe/Athens'), ('Europe/Belgrade', 'Europe/Belgrade'), ('Europe/Berlin', 'Europe/Berlin'), ('Europe/Brussels', 'Europe/Brussels'), ('Europe/Bucharest', 'Europe/Bucharest'), ('Europe/Budapest', 'Europe/Budapest'), ('Europe/Chisinau', 'Europe/Chisinau'), ('Europe/Copenhagen', 'Europe/Copenhagen'), ('Europe/Dublin', 'Europe/Dublin'), ('Europe/Gibraltar', 'Europe/Gibraltar'), ('Europe/Helsinki', 'Europe/Helsinki'), ('Europe/Istanbul', 'Europe/Istanbul'), ('Europe/Kaliningrad', 'Europe/Kaliningrad'), ('Europe/Kiev', 'Europe/Kiev'), ('Europe/Kirov', 'Europe/Kirov'), ('Europe/Lisbon', 'Europe/Lisbon'), ('Europe/London', 'Europe/London'), ('Europe/Luxembourg', 'Europe/Luxembourg'), ('Europe/Madrid', 'Europe/Madrid'), ('Europe/Malta', 'Europe/Malta'), ('Europe/Minsk', 'Europe/Minsk'), ('Europe/Monaco', 'Europe/Monaco'), ('Europe/Moscow', 'Europe/Moscow'), ('Europe/Oslo', 'Europe/Oslo'), ('Europe/Paris', 'Europe/Paris'), ('Europe/Prague', 'Europe/Prague'), ('Europe/Riga', 'Europe/Riga'), ('Europe/Rome', 'Europe/Rome'), ('Europe/Samara', 'Europe/Samara'), ('Europe/Saratov', 'Europe/Saratov'), ('Europe/Simferopol', 'Europe/Simferopol'), ('Europe/Sofia', 'Europe/Sofia'), ('Europe/Stockholm', 'Europe/Stockholm'), ('Europe/Tallinn', 'Europe/Tallinn'), ('Europe/Tirane', 'Europe/Tirane'), ('Europe/Ulyanovsk', 'Europe/Ulyanovsk'), ('Europe/Uzhgorod', 'Europe/Uzhgorod'), ('Europe/Vaduz', 'Europe/Vaduz'), ('Europe/Vienna', 'Europe/Vienna'), ('Europe/Vilnius', 'Europe/Vilnius'), ('Europe/Volgograd', 'Europe/Volgograd'), ('Europe/Warsaw', 'Europe/Warsaw'), ('Europe/Zaporozhye', 'Europe/Zaporozhye'), ('Europe/Zurich', 'Europe/Zurich'), ('GMT', 'GMT'), ('Indian/Antananarivo', 'Indian/Antananarivo'), ('Indian/Chagos', 'Indian/Chagos'), ('Indian/Christmas', 'Indian/Christmas'), ('Indian/Cocos', 'Indian/Cocos'), ('Indian/Comoro', 'Indian/Comoro'), ('Indian/Kerguelen', 'Indian/Kerguelen'), ('Indian/Mahe', 'Indian/Mahe'), ('Indian/Maldives', 'Indian/Maldives'), ('Indian/Mauritius', 'Indian/Mauritius'), ('Indian/Mayotte', 'Indian/Mayotte'), ('Indian/Reunion', 'Indian/Reunion'), ('Pacific/Apia', 'Pacific/Apia'), ('Pacific/Auckland', 'Pacific/Auckland'), ('Pacific/Bougainville', 'Pacific/Bougainville'), ('Pacific/Chatham', 'Pacific/Chatham'), ('Pacific/Chuuk', 'Pacific/Chuuk'), ('Pacific/Easter', 'Pacific/Easter'), ('Pacific/Efate', 'Pacific/Efate'), ('Pacific/Enderbury', 'Pacific/Enderbury'), ('Pacific/Fakaofo', 'Pacific/Fakaofo'), ('Pacific/Fiji', 'Pacific/Fiji'), ('Pacific/Funafuti', 'Pacific/Funafuti'), ('Pacific/Galapagos', 'Pacific/Galapagos'), ('Pacific/Gambier', 'Pacific/Gambier'), ('Pacific/Guadalcanal', 'Pacific/Guadalcanal'), ('Pacific/Guam', 'Pacific/Guam'), ('Pacific/Honolulu', 'Pacific/Honolulu'), ('Pacific/Kiritimati', 'Pacific/Kiritimati'), ('Pacific/Kosrae', 'Pacific/Kosrae'), ('Pacific/Kwajalein', 'Pacific/Kwajalein'), ('Pacific/Majuro', 'Pacific/Majuro'), ('Pacific/Marquesas', 'Pacific/Marquesas'), ('Pacific/Midway', 'Pacific/Midway'), ('Pacific/Nauru', 'Pacific/Nauru'), ('Pacific/Niue', 'Pacific/Niue'), ('Pacific/Norfolk', 'Pacific/Norfolk'), ('Pacific/Noumea', 'Pacific/Noumea'), ('Pacific/Pago_Pago', 'Pacific/Pago_Pago'), ('Pacific/Palau', 'Pacific/Palau'), ('Pacific/Pitcairn', 'Pacific/Pitcairn'), ('Pacific/Pohnpei', 'Pacific/Pohnpei'), ('Pacific/Port_Moresby', 'Pacific/Port_Moresby'), ('Pacific/Rarotonga', 'Pacific/Rarotonga'), ('Pacific/Saipan', 'Pacific/Saipan'), ('Pacific/Tahiti', 'Pacific/Tahiti'), ('Pacific/Tarawa', 'Pacific/Tarawa'), ('Pacific/Tongatapu', 'Pacific/Tongatapu'), ('Pacific/Wake', 'Pacific/Wake'), ('Pacific/Wallis', 'Pacific/Wallis'), ('UTC', 'UTC')], max_length=255), - ), - ] diff --git a/ietf/meeting/migrations/0012_add_slide_submissions.py b/ietf/meeting/migrations/0012_add_slide_submissions.py deleted file mode 100644 index e1a798d59b..0000000000 --- a/ietf/meeting/migrations/0012_add_slide_submissions.py +++ /dev/null @@ -1,30 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-03-23 07:41 - - -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0009_auto_20190118_0725'), - ('meeting', '0011_auto_20190114_0550'), - ] - - operations = [ - migrations.CreateModel( - name='SlideSubmission', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('title', models.CharField(max_length=255)), - ('filename', models.CharField(max_length=255)), - ('apply_to_all', models.BooleanField(default=False)), - ('session', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='meeting.Session')), - ('submitter', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), - ], - ), - ] diff --git a/ietf/meeting/migrations/0012_registration_registrationticket.py b/ietf/meeting/migrations/0012_registration_registrationticket.py new file mode 100644 index 0000000000..c555f52e8b --- /dev/null +++ b/ietf/meeting/migrations/0012_registration_registrationticket.py @@ -0,0 +1,90 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +from django.db import migrations, models +import django.db.models.deletion +import ietf.utils.models + + +class Migration(migrations.Migration): + + dependencies = [ + ("name", "0017_populate_new_reg_names"), + ("person", "0004_alter_person_photo_alter_person_photo_thumb"), + ("meeting", "0011_alter_slidesubmission_doc"), + ] + + operations = [ + migrations.CreateModel( + name="Registration", + fields=[ + ( + "id", + models.AutoField( + auto_created=True, + primary_key=True, + serialize=False, + verbose_name="ID", + ), + ), + ("first_name", models.CharField(max_length=255)), + ("last_name", models.CharField(max_length=255)), + ("affiliation", models.CharField(blank=True, max_length=255)), + ("country_code", models.CharField(max_length=2)), + ("email", models.EmailField(blank=True, max_length=254, null=True)), + ("attended", models.BooleanField(default=False)), + ("checkedin", models.BooleanField(default=False)), + ( + "meeting", + ietf.utils.models.ForeignKey( + on_delete=django.db.models.deletion.CASCADE, + to="meeting.meeting", + ), + ), + ( + "person", + ietf.utils.models.ForeignKey( + blank=True, + null=True, + on_delete=django.db.models.deletion.PROTECT, + to="person.person", + ), + ), + ], + ), + migrations.CreateModel( + name="RegistrationTicket", + fields=[ + ( + "id", + models.AutoField( + auto_created=True, + primary_key=True, + serialize=False, + verbose_name="ID", + ), + ), + ( + "attendance_type", + ietf.utils.models.ForeignKey( + on_delete=django.db.models.deletion.PROTECT, + to="name.attendancetypename", + ), + ), + ( + "registration", + ietf.utils.models.ForeignKey( + on_delete=django.db.models.deletion.CASCADE, + related_name="tickets", + to="meeting.registration", + ), + ), + ( + "ticket_type", + ietf.utils.models.ForeignKey( + on_delete=django.db.models.deletion.PROTECT, + to="name.registrationtickettypename", + ), + ), + ], + ), + ] diff --git a/ietf/meeting/migrations/0013_correct_reg_checkedin.py b/ietf/meeting/migrations/0013_correct_reg_checkedin.py new file mode 100644 index 0000000000..88b3efceac --- /dev/null +++ b/ietf/meeting/migrations/0013_correct_reg_checkedin.py @@ -0,0 +1,41 @@ +# Generated by Django 4.2.21 on 2025-05-20 22:28 + +''' +The original migration had a flaw. If a participant had both a remote and onsite +registration, which is rare but does occur, which registration the checkedin state +came from was indeterminate. If it came from the remote registration it would be +False which might be wrong. This migration finds all registrations with onsite tickets +and checkedin is False, and checks if it is correct, and fixes if needed. +''' + +from django.db import migrations +import datetime + + +def forward(apps, schema_editor): + Registration = apps.get_model('meeting', 'Registration') + MeetingRegistration = apps.get_model('stats', 'MeetingRegistration') + today = datetime.date.today() + for reg in Registration.objects.filter(tickets__attendance_type__slug='onsite', checkedin=False, meeting__date__lt=today).order_by('meeting__number'): + # get original MeetingRegistration + mregs = MeetingRegistration.objects.filter(meeting=reg.meeting, email=reg.email, reg_type='onsite') + mregs_checkedin = [mr.checkedin for mr in mregs] + if any(mregs_checkedin): + reg.checkedin = True + reg.save() + print(f'updating {reg.meeting}:{reg.email}:{reg.pk}') + + +def reverse(apps, schema_editor): + pass + + +class Migration(migrations.Migration): + + dependencies = [ + ("meeting", "0012_registration_registrationticket"), + ] + + operations = [ + migrations.RunPython(forward, reverse), + ] diff --git a/ietf/meeting/migrations/0013_make_separate_break_sessobj.py b/ietf/meeting/migrations/0013_make_separate_break_sessobj.py deleted file mode 100644 index ceea5179d2..0000000000 --- a/ietf/meeting/migrations/0013_make_separate_break_sessobj.py +++ /dev/null @@ -1,43 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-03-23 06:11 - - -import datetime -from django.db import migrations - - -def copy_session(session): - session.pk = None - session.save() - return session - - -def forward(apps, schema_editor): - Meeting = apps.get_model('meeting', 'Meeting') - today = datetime.datetime.today() - meetings = Meeting.objects.filter(date__gt=today, type='ietf') - for meeting in meetings: - sessions = meeting.session_set.filter(type__in=['break', 'reg']) - for session in sessions: - assignments = session.timeslotassignments.filter(schedule=meeting.agenda) - if assignments.count() > 1: - ids = [ a.id for a in assignments ] - first_assignment = session.timeslotassignments.get(id=ids[0]) - original_session = first_assignment.session - for assignment in session.timeslotassignments.filter(id__in=ids[1:]): - assignment.session = copy_session(original_session) - assignment.save() - -def backward(apps, schema_editor): - return - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0012_add_slide_submissions'), - ] - - operations = [ - migrations.RunPython(forward, backward), - ] diff --git a/ietf/meeting/migrations/0014_alter_floorplan_image.py b/ietf/meeting/migrations/0014_alter_floorplan_image.py new file mode 100644 index 0000000000..e125625edc --- /dev/null +++ b/ietf/meeting/migrations/0014_alter_floorplan_image.py @@ -0,0 +1,25 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +from django.db import migrations, models +import ietf.meeting.models +import ietf.utils.storage + + +class Migration(migrations.Migration): + dependencies = [ + ("meeting", "0013_correct_reg_checkedin"), + ] + + operations = [ + migrations.AlterField( + model_name="floorplan", + name="image", + field=models.ImageField( + default=None, + storage=ietf.utils.storage.BlobShadowFileSystemStorage( + kind="", location=None + ), + upload_to=ietf.meeting.models.floorplan_path, + ), + ), + ] diff --git a/ietf/meeting/migrations/0014_auto_20190426_0305.py b/ietf/meeting/migrations/0014_auto_20190426_0305.py deleted file mode 100644 index 8f4b6b0b3b..0000000000 --- a/ietf/meeting/migrations/0014_auto_20190426_0305.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-04-26 03:05 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0013_make_separate_break_sessobj'), - ] - - operations = [ - migrations.AlterField( - model_name='meeting', - name='country', - field=models.CharField(blank=True, choices=[('', ''), ('AF', 'Afghanistan'), ('AL', 'Albania'), ('DZ', 'Algeria'), ('AD', 'Andorra'), ('AO', 'Angola'), ('AI', 'Anguilla'), ('AQ', 'Antarctica'), ('AG', 'Antigua & Barbuda'), ('AR', 'Argentina'), ('AM', 'Armenia'), ('AW', 'Aruba'), ('AU', 'Australia'), ('AT', 'Austria'), ('AZ', 'Azerbaijan'), ('BS', 'Bahamas'), ('BH', 'Bahrain'), ('BD', 'Bangladesh'), ('BB', 'Barbados'), ('BY', 'Belarus'), ('BE', 'Belgium'), ('BZ', 'Belize'), ('BJ', 'Benin'), ('BM', 'Bermuda'), ('BT', 'Bhutan'), ('BO', 'Bolivia'), ('BA', 'Bosnia & Herzegovina'), ('BW', 'Botswana'), ('BV', 'Bouvet Island'), ('BR', 'Brazil'), ('GB', 'Britain (UK)'), ('IO', 'British Indian Ocean Territory'), ('BN', 'Brunei'), ('BG', 'Bulgaria'), ('BF', 'Burkina Faso'), ('BI', 'Burundi'), ('KH', 'Cambodia'), ('CM', 'Cameroon'), ('CA', 'Canada'), ('CV', 'Cape Verde'), ('BQ', 'Caribbean NL'), ('KY', 'Cayman Islands'), ('CF', 'Central African Rep.'), ('TD', 'Chad'), ('CL', 'Chile'), ('CN', 'China'), ('CX', 'Christmas Island'), ('CC', 'Cocos (Keeling) Islands'), ('CO', 'Colombia'), ('KM', 'Comoros'), ('CD', 'Congo (Dem. Rep.)'), ('CG', 'Congo (Rep.)'), ('CK', 'Cook Islands'), ('CR', 'Costa Rica'), ('HR', 'Croatia'), ('CU', 'Cuba'), ('CW', 'Cura\xe7ao'), ('CY', 'Cyprus'), ('CZ', 'Czech Republic'), ('CI', "C\xf4te d'Ivoire"), ('DK', 'Denmark'), ('DJ', 'Djibouti'), ('DM', 'Dominica'), ('DO', 'Dominican Republic'), ('TL', 'East Timor'), ('EC', 'Ecuador'), ('EG', 'Egypt'), ('SV', 'El Salvador'), ('GQ', 'Equatorial Guinea'), ('ER', 'Eritrea'), ('EE', 'Estonia'), ('SZ', 'Eswatini (Swaziland)'), ('ET', 'Ethiopia'), ('FK', 'Falkland Islands'), ('FO', 'Faroe Islands'), ('FJ', 'Fiji'), ('FI', 'Finland'), ('FR', 'France'), ('GF', 'French Guiana'), ('PF', 'French Polynesia'), ('TF', 'French Southern & Antarctic Lands'), ('GA', 'Gabon'), ('GM', 'Gambia'), ('GE', 'Georgia'), ('DE', 'Germany'), ('GH', 'Ghana'), ('GI', 'Gibraltar'), ('GR', 'Greece'), ('GL', 'Greenland'), ('GD', 'Grenada'), ('GP', 'Guadeloupe'), ('GU', 'Guam'), ('GT', 'Guatemala'), ('GG', 'Guernsey'), ('GN', 'Guinea'), ('GW', 'Guinea-Bissau'), ('GY', 'Guyana'), ('HT', 'Haiti'), ('HM', 'Heard Island & McDonald Islands'), ('HN', 'Honduras'), ('HK', 'Hong Kong'), ('HU', 'Hungary'), ('IS', 'Iceland'), ('IN', 'India'), ('ID', 'Indonesia'), ('IR', 'Iran'), ('IQ', 'Iraq'), ('IE', 'Ireland'), ('IM', 'Isle of Man'), ('IL', 'Israel'), ('IT', 'Italy'), ('JM', 'Jamaica'), ('JP', 'Japan'), ('JE', 'Jersey'), ('JO', 'Jordan'), ('KZ', 'Kazakhstan'), ('KE', 'Kenya'), ('KI', 'Kiribati'), ('KP', 'Korea (North)'), ('KR', 'Korea (South)'), ('KW', 'Kuwait'), ('KG', 'Kyrgyzstan'), ('LA', 'Laos'), ('LV', 'Latvia'), ('LB', 'Lebanon'), ('LS', 'Lesotho'), ('LR', 'Liberia'), ('LY', 'Libya'), ('LI', 'Liechtenstein'), ('LT', 'Lithuania'), ('LU', 'Luxembourg'), ('MO', 'Macau'), ('MG', 'Madagascar'), ('MW', 'Malawi'), ('MY', 'Malaysia'), ('MV', 'Maldives'), ('ML', 'Mali'), ('MT', 'Malta'), ('MH', 'Marshall Islands'), ('MQ', 'Martinique'), ('MR', 'Mauritania'), ('MU', 'Mauritius'), ('YT', 'Mayotte'), ('MX', 'Mexico'), ('FM', 'Micronesia'), ('MD', 'Moldova'), ('MC', 'Monaco'), ('MN', 'Mongolia'), ('ME', 'Montenegro'), ('MS', 'Montserrat'), ('MA', 'Morocco'), ('MZ', 'Mozambique'), ('MM', 'Myanmar (Burma)'), ('NA', 'Namibia'), ('NR', 'Nauru'), ('NP', 'Nepal'), ('NL', 'Netherlands'), ('NC', 'New Caledonia'), ('NZ', 'New Zealand'), ('NI', 'Nicaragua'), ('NE', 'Niger'), ('NG', 'Nigeria'), ('NU', 'Niue'), ('NF', 'Norfolk Island'), ('MK', 'North Macedonia'), ('MP', 'Northern Mariana Islands'), ('NO', 'Norway'), ('OM', 'Oman'), ('PK', 'Pakistan'), ('PW', 'Palau'), ('PS', 'Palestine'), ('PA', 'Panama'), ('PG', 'Papua New Guinea'), ('PY', 'Paraguay'), ('PE', 'Peru'), ('PH', 'Philippines'), ('PN', 'Pitcairn'), ('PL', 'Poland'), ('PT', 'Portugal'), ('PR', 'Puerto Rico'), ('QA', 'Qatar'), ('RO', 'Romania'), ('RU', 'Russia'), ('RW', 'Rwanda'), ('RE', 'R\xe9union'), ('AS', 'Samoa (American)'), ('WS', 'Samoa (western)'), ('SM', 'San Marino'), ('ST', 'Sao Tome & Principe'), ('SA', 'Saudi Arabia'), ('SN', 'Senegal'), ('RS', 'Serbia'), ('SC', 'Seychelles'), ('SL', 'Sierra Leone'), ('SG', 'Singapore'), ('SK', 'Slovakia'), ('SI', 'Slovenia'), ('SB', 'Solomon Islands'), ('SO', 'Somalia'), ('ZA', 'South Africa'), ('GS', 'South Georgia & the South Sandwich Islands'), ('SS', 'South Sudan'), ('ES', 'Spain'), ('LK', 'Sri Lanka'), ('BL', 'St Barthelemy'), ('SH', 'St Helena'), ('KN', 'St Kitts & Nevis'), ('LC', 'St Lucia'), ('SX', 'St Maarten (Dutch)'), ('MF', 'St Martin (French)'), ('PM', 'St Pierre & Miquelon'), ('VC', 'St Vincent'), ('SD', 'Sudan'), ('SR', 'Suriname'), ('SJ', 'Svalbard & Jan Mayen'), ('SE', 'Sweden'), ('CH', 'Switzerland'), ('SY', 'Syria'), ('TW', 'Taiwan'), ('TJ', 'Tajikistan'), ('TZ', 'Tanzania'), ('TH', 'Thailand'), ('TG', 'Togo'), ('TK', 'Tokelau'), ('TO', 'Tonga'), ('TT', 'Trinidad & Tobago'), ('TN', 'Tunisia'), ('TR', 'Turkey'), ('TM', 'Turkmenistan'), ('TC', 'Turks & Caicos Is'), ('TV', 'Tuvalu'), ('UM', 'US minor outlying islands'), ('UG', 'Uganda'), ('UA', 'Ukraine'), ('AE', 'United Arab Emirates'), ('US', 'United States'), ('UY', 'Uruguay'), ('UZ', 'Uzbekistan'), ('VU', 'Vanuatu'), ('VA', 'Vatican City'), ('VE', 'Venezuela'), ('VN', 'Vietnam'), ('VG', 'Virgin Islands (UK)'), ('VI', 'Virgin Islands (US)'), ('WF', 'Wallis & Futuna'), ('EH', 'Western Sahara'), ('YE', 'Yemen'), ('ZM', 'Zambia'), ('ZW', 'Zimbabwe'), ('AX', '\xc5land Islands')], max_length=2), - ), - ] diff --git a/ietf/meeting/migrations/0015_alter_meeting_time_zone.py b/ietf/meeting/migrations/0015_alter_meeting_time_zone.py new file mode 100644 index 0000000000..2a4b7859ee --- /dev/null +++ b/ietf/meeting/migrations/0015_alter_meeting_time_zone.py @@ -0,0 +1,451 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +from django.db import migrations, models + + +def forward(apps, schema_editor): + """Migrate 'GMT' meeting time zones to 'UTC'""" + Meeting = apps.get_model("meeting", "Meeting") + Meeting.objects.filter(time_zone="GMT").update(time_zone="UTC") + + +def reverse(apps, schema_editor): + pass # nothing to do + + +class Migration(migrations.Migration): + + dependencies = [ + ("meeting", "0014_alter_floorplan_image"), + ] + + operations = [ + migrations.RunPython(forward, reverse), + migrations.AlterField( + model_name="meeting", + name="time_zone", + field=models.CharField( + choices=[ + ("", "---------"), + ("Africa/Abidjan", "Africa/Abidjan"), + ("Africa/Accra", "Africa/Accra"), + ("Africa/Addis_Ababa", "Africa/Addis_Ababa"), + ("Africa/Algiers", "Africa/Algiers"), + ("Africa/Asmara", "Africa/Asmara"), + ("Africa/Bamako", "Africa/Bamako"), + ("Africa/Bangui", "Africa/Bangui"), + ("Africa/Banjul", "Africa/Banjul"), + ("Africa/Bissau", "Africa/Bissau"), + ("Africa/Blantyre", "Africa/Blantyre"), + ("Africa/Brazzaville", "Africa/Brazzaville"), + ("Africa/Bujumbura", "Africa/Bujumbura"), + ("Africa/Cairo", "Africa/Cairo"), + ("Africa/Casablanca", "Africa/Casablanca"), + ("Africa/Ceuta", "Africa/Ceuta"), + ("Africa/Conakry", "Africa/Conakry"), + ("Africa/Dakar", "Africa/Dakar"), + ("Africa/Dar_es_Salaam", "Africa/Dar_es_Salaam"), + ("Africa/Djibouti", "Africa/Djibouti"), + ("Africa/Douala", "Africa/Douala"), + ("Africa/El_Aaiun", "Africa/El_Aaiun"), + ("Africa/Freetown", "Africa/Freetown"), + ("Africa/Gaborone", "Africa/Gaborone"), + ("Africa/Harare", "Africa/Harare"), + ("Africa/Johannesburg", "Africa/Johannesburg"), + ("Africa/Juba", "Africa/Juba"), + ("Africa/Kampala", "Africa/Kampala"), + ("Africa/Khartoum", "Africa/Khartoum"), + ("Africa/Kigali", "Africa/Kigali"), + ("Africa/Kinshasa", "Africa/Kinshasa"), + ("Africa/Lagos", "Africa/Lagos"), + ("Africa/Libreville", "Africa/Libreville"), + ("Africa/Lome", "Africa/Lome"), + ("Africa/Luanda", "Africa/Luanda"), + ("Africa/Lubumbashi", "Africa/Lubumbashi"), + ("Africa/Lusaka", "Africa/Lusaka"), + ("Africa/Malabo", "Africa/Malabo"), + ("Africa/Maputo", "Africa/Maputo"), + ("Africa/Maseru", "Africa/Maseru"), + ("Africa/Mbabane", "Africa/Mbabane"), + ("Africa/Mogadishu", "Africa/Mogadishu"), + ("Africa/Monrovia", "Africa/Monrovia"), + ("Africa/Nairobi", "Africa/Nairobi"), + ("Africa/Ndjamena", "Africa/Ndjamena"), + ("Africa/Niamey", "Africa/Niamey"), + ("Africa/Nouakchott", "Africa/Nouakchott"), + ("Africa/Ouagadougou", "Africa/Ouagadougou"), + ("Africa/Porto-Novo", "Africa/Porto-Novo"), + ("Africa/Sao_Tome", "Africa/Sao_Tome"), + ("Africa/Tripoli", "Africa/Tripoli"), + ("Africa/Tunis", "Africa/Tunis"), + ("Africa/Windhoek", "Africa/Windhoek"), + ("America/Adak", "America/Adak"), + ("America/Anchorage", "America/Anchorage"), + ("America/Anguilla", "America/Anguilla"), + ("America/Antigua", "America/Antigua"), + ("America/Araguaina", "America/Araguaina"), + ( + "America/Argentina/Buenos_Aires", + "America/Argentina/Buenos_Aires", + ), + ("America/Argentina/Catamarca", "America/Argentina/Catamarca"), + ("America/Argentina/Cordoba", "America/Argentina/Cordoba"), + ("America/Argentina/Jujuy", "America/Argentina/Jujuy"), + ("America/Argentina/La_Rioja", "America/Argentina/La_Rioja"), + ("America/Argentina/Mendoza", "America/Argentina/Mendoza"), + ( + "America/Argentina/Rio_Gallegos", + "America/Argentina/Rio_Gallegos", + ), + ("America/Argentina/Salta", "America/Argentina/Salta"), + ("America/Argentina/San_Juan", "America/Argentina/San_Juan"), + ("America/Argentina/San_Luis", "America/Argentina/San_Luis"), + ("America/Argentina/Tucuman", "America/Argentina/Tucuman"), + ("America/Argentina/Ushuaia", "America/Argentina/Ushuaia"), + ("America/Aruba", "America/Aruba"), + ("America/Asuncion", "America/Asuncion"), + ("America/Atikokan", "America/Atikokan"), + ("America/Bahia", "America/Bahia"), + ("America/Bahia_Banderas", "America/Bahia_Banderas"), + ("America/Barbados", "America/Barbados"), + ("America/Belem", "America/Belem"), + ("America/Belize", "America/Belize"), + ("America/Blanc-Sablon", "America/Blanc-Sablon"), + ("America/Boa_Vista", "America/Boa_Vista"), + ("America/Bogota", "America/Bogota"), + ("America/Boise", "America/Boise"), + ("America/Cambridge_Bay", "America/Cambridge_Bay"), + ("America/Campo_Grande", "America/Campo_Grande"), + ("America/Cancun", "America/Cancun"), + ("America/Caracas", "America/Caracas"), + ("America/Cayenne", "America/Cayenne"), + ("America/Cayman", "America/Cayman"), + ("America/Chicago", "America/Chicago"), + ("America/Chihuahua", "America/Chihuahua"), + ("America/Costa_Rica", "America/Costa_Rica"), + ("America/Creston", "America/Creston"), + ("America/Cuiaba", "America/Cuiaba"), + ("America/Curacao", "America/Curacao"), + ("America/Danmarkshavn", "America/Danmarkshavn"), + ("America/Dawson", "America/Dawson"), + ("America/Dawson_Creek", "America/Dawson_Creek"), + ("America/Denver", "America/Denver"), + ("America/Detroit", "America/Detroit"), + ("America/Dominica", "America/Dominica"), + ("America/Edmonton", "America/Edmonton"), + ("America/Eirunepe", "America/Eirunepe"), + ("America/El_Salvador", "America/El_Salvador"), + ("America/Fort_Nelson", "America/Fort_Nelson"), + ("America/Fortaleza", "America/Fortaleza"), + ("America/Glace_Bay", "America/Glace_Bay"), + ("America/Goose_Bay", "America/Goose_Bay"), + ("America/Grand_Turk", "America/Grand_Turk"), + ("America/Grenada", "America/Grenada"), + ("America/Guadeloupe", "America/Guadeloupe"), + ("America/Guatemala", "America/Guatemala"), + ("America/Guayaquil", "America/Guayaquil"), + ("America/Guyana", "America/Guyana"), + ("America/Halifax", "America/Halifax"), + ("America/Havana", "America/Havana"), + ("America/Hermosillo", "America/Hermosillo"), + ("America/Indiana/Indianapolis", "America/Indiana/Indianapolis"), + ("America/Indiana/Knox", "America/Indiana/Knox"), + ("America/Indiana/Marengo", "America/Indiana/Marengo"), + ("America/Indiana/Petersburg", "America/Indiana/Petersburg"), + ("America/Indiana/Tell_City", "America/Indiana/Tell_City"), + ("America/Indiana/Vevay", "America/Indiana/Vevay"), + ("America/Indiana/Vincennes", "America/Indiana/Vincennes"), + ("America/Indiana/Winamac", "America/Indiana/Winamac"), + ("America/Inuvik", "America/Inuvik"), + ("America/Iqaluit", "America/Iqaluit"), + ("America/Jamaica", "America/Jamaica"), + ("America/Juneau", "America/Juneau"), + ("America/Kentucky/Louisville", "America/Kentucky/Louisville"), + ("America/Kentucky/Monticello", "America/Kentucky/Monticello"), + ("America/La_Paz", "America/La_Paz"), + ("America/Lima", "America/Lima"), + ("America/Los_Angeles", "America/Los_Angeles"), + ("America/Maceio", "America/Maceio"), + ("America/Managua", "America/Managua"), + ("America/Manaus", "America/Manaus"), + ("America/Martinique", "America/Martinique"), + ("America/Matamoros", "America/Matamoros"), + ("America/Mazatlan", "America/Mazatlan"), + ("America/Menominee", "America/Menominee"), + ("America/Merida", "America/Merida"), + ("America/Metlakatla", "America/Metlakatla"), + ("America/Mexico_City", "America/Mexico_City"), + ("America/Miquelon", "America/Miquelon"), + ("America/Moncton", "America/Moncton"), + ("America/Monterrey", "America/Monterrey"), + ("America/Montevideo", "America/Montevideo"), + ("America/Montserrat", "America/Montserrat"), + ("America/Nassau", "America/Nassau"), + ("America/New_York", "America/New_York"), + ("America/Nipigon", "America/Nipigon"), + ("America/Nome", "America/Nome"), + ("America/Noronha", "America/Noronha"), + ("America/North_Dakota/Beulah", "America/North_Dakota/Beulah"), + ("America/North_Dakota/Center", "America/North_Dakota/Center"), + ( + "America/North_Dakota/New_Salem", + "America/North_Dakota/New_Salem", + ), + ("America/Nuuk", "America/Nuuk"), + ("America/Ojinaga", "America/Ojinaga"), + ("America/Panama", "America/Panama"), + ("America/Pangnirtung", "America/Pangnirtung"), + ("America/Paramaribo", "America/Paramaribo"), + ("America/Phoenix", "America/Phoenix"), + ("America/Port-au-Prince", "America/Port-au-Prince"), + ("America/Port_of_Spain", "America/Port_of_Spain"), + ("America/Porto_Velho", "America/Porto_Velho"), + ("America/Puerto_Rico", "America/Puerto_Rico"), + ("America/Punta_Arenas", "America/Punta_Arenas"), + ("America/Rainy_River", "America/Rainy_River"), + ("America/Rankin_Inlet", "America/Rankin_Inlet"), + ("America/Recife", "America/Recife"), + ("America/Regina", "America/Regina"), + ("America/Resolute", "America/Resolute"), + ("America/Rio_Branco", "America/Rio_Branco"), + ("America/Santarem", "America/Santarem"), + ("America/Santiago", "America/Santiago"), + ("America/Santo_Domingo", "America/Santo_Domingo"), + ("America/Sao_Paulo", "America/Sao_Paulo"), + ("America/Scoresbysund", "America/Scoresbysund"), + ("America/Sitka", "America/Sitka"), + ("America/St_Johns", "America/St_Johns"), + ("America/St_Kitts", "America/St_Kitts"), + ("America/St_Lucia", "America/St_Lucia"), + ("America/St_Thomas", "America/St_Thomas"), + ("America/St_Vincent", "America/St_Vincent"), + ("America/Swift_Current", "America/Swift_Current"), + ("America/Tegucigalpa", "America/Tegucigalpa"), + ("America/Thule", "America/Thule"), + ("America/Thunder_Bay", "America/Thunder_Bay"), + ("America/Tijuana", "America/Tijuana"), + ("America/Toronto", "America/Toronto"), + ("America/Tortola", "America/Tortola"), + ("America/Vancouver", "America/Vancouver"), + ("America/Whitehorse", "America/Whitehorse"), + ("America/Winnipeg", "America/Winnipeg"), + ("America/Yakutat", "America/Yakutat"), + ("America/Yellowknife", "America/Yellowknife"), + ("Antarctica/Casey", "Antarctica/Casey"), + ("Antarctica/Davis", "Antarctica/Davis"), + ("Antarctica/DumontDUrville", "Antarctica/DumontDUrville"), + ("Antarctica/Macquarie", "Antarctica/Macquarie"), + ("Antarctica/Mawson", "Antarctica/Mawson"), + ("Antarctica/McMurdo", "Antarctica/McMurdo"), + ("Antarctica/Palmer", "Antarctica/Palmer"), + ("Antarctica/Rothera", "Antarctica/Rothera"), + ("Antarctica/Syowa", "Antarctica/Syowa"), + ("Antarctica/Troll", "Antarctica/Troll"), + ("Antarctica/Vostok", "Antarctica/Vostok"), + ("Asia/Aden", "Asia/Aden"), + ("Asia/Almaty", "Asia/Almaty"), + ("Asia/Amman", "Asia/Amman"), + ("Asia/Anadyr", "Asia/Anadyr"), + ("Asia/Aqtau", "Asia/Aqtau"), + ("Asia/Aqtobe", "Asia/Aqtobe"), + ("Asia/Ashgabat", "Asia/Ashgabat"), + ("Asia/Atyrau", "Asia/Atyrau"), + ("Asia/Baghdad", "Asia/Baghdad"), + ("Asia/Bahrain", "Asia/Bahrain"), + ("Asia/Baku", "Asia/Baku"), + ("Asia/Bangkok", "Asia/Bangkok"), + ("Asia/Barnaul", "Asia/Barnaul"), + ("Asia/Beirut", "Asia/Beirut"), + ("Asia/Bishkek", "Asia/Bishkek"), + ("Asia/Brunei", "Asia/Brunei"), + ("Asia/Chita", "Asia/Chita"), + ("Asia/Choibalsan", "Asia/Choibalsan"), + ("Asia/Colombo", "Asia/Colombo"), + ("Asia/Damascus", "Asia/Damascus"), + ("Asia/Dhaka", "Asia/Dhaka"), + ("Asia/Dili", "Asia/Dili"), + ("Asia/Dubai", "Asia/Dubai"), + ("Asia/Dushanbe", "Asia/Dushanbe"), + ("Asia/Famagusta", "Asia/Famagusta"), + ("Asia/Gaza", "Asia/Gaza"), + ("Asia/Hebron", "Asia/Hebron"), + ("Asia/Ho_Chi_Minh", "Asia/Ho_Chi_Minh"), + ("Asia/Hong_Kong", "Asia/Hong_Kong"), + ("Asia/Hovd", "Asia/Hovd"), + ("Asia/Irkutsk", "Asia/Irkutsk"), + ("Asia/Jakarta", "Asia/Jakarta"), + ("Asia/Jayapura", "Asia/Jayapura"), + ("Asia/Jerusalem", "Asia/Jerusalem"), + ("Asia/Kabul", "Asia/Kabul"), + ("Asia/Kamchatka", "Asia/Kamchatka"), + ("Asia/Karachi", "Asia/Karachi"), + ("Asia/Kathmandu", "Asia/Kathmandu"), + ("Asia/Khandyga", "Asia/Khandyga"), + ("Asia/Kolkata", "Asia/Kolkata"), + ("Asia/Krasnoyarsk", "Asia/Krasnoyarsk"), + ("Asia/Kuala_Lumpur", "Asia/Kuala_Lumpur"), + ("Asia/Kuching", "Asia/Kuching"), + ("Asia/Kuwait", "Asia/Kuwait"), + ("Asia/Macau", "Asia/Macau"), + ("Asia/Magadan", "Asia/Magadan"), + ("Asia/Makassar", "Asia/Makassar"), + ("Asia/Manila", "Asia/Manila"), + ("Asia/Muscat", "Asia/Muscat"), + ("Asia/Nicosia", "Asia/Nicosia"), + ("Asia/Novokuznetsk", "Asia/Novokuznetsk"), + ("Asia/Novosibirsk", "Asia/Novosibirsk"), + ("Asia/Omsk", "Asia/Omsk"), + ("Asia/Oral", "Asia/Oral"), + ("Asia/Phnom_Penh", "Asia/Phnom_Penh"), + ("Asia/Pontianak", "Asia/Pontianak"), + ("Asia/Pyongyang", "Asia/Pyongyang"), + ("Asia/Qatar", "Asia/Qatar"), + ("Asia/Qostanay", "Asia/Qostanay"), + ("Asia/Qyzylorda", "Asia/Qyzylorda"), + ("Asia/Riyadh", "Asia/Riyadh"), + ("Asia/Sakhalin", "Asia/Sakhalin"), + ("Asia/Samarkand", "Asia/Samarkand"), + ("Asia/Seoul", "Asia/Seoul"), + ("Asia/Shanghai", "Asia/Shanghai"), + ("Asia/Singapore", "Asia/Singapore"), + ("Asia/Srednekolymsk", "Asia/Srednekolymsk"), + ("Asia/Taipei", "Asia/Taipei"), + ("Asia/Tashkent", "Asia/Tashkent"), + ("Asia/Tbilisi", "Asia/Tbilisi"), + ("Asia/Tehran", "Asia/Tehran"), + ("Asia/Thimphu", "Asia/Thimphu"), + ("Asia/Tokyo", "Asia/Tokyo"), + ("Asia/Tomsk", "Asia/Tomsk"), + ("Asia/Ulaanbaatar", "Asia/Ulaanbaatar"), + ("Asia/Urumqi", "Asia/Urumqi"), + ("Asia/Ust-Nera", "Asia/Ust-Nera"), + ("Asia/Vientiane", "Asia/Vientiane"), + ("Asia/Vladivostok", "Asia/Vladivostok"), + ("Asia/Yakutsk", "Asia/Yakutsk"), + ("Asia/Yangon", "Asia/Yangon"), + ("Asia/Yekaterinburg", "Asia/Yekaterinburg"), + ("Asia/Yerevan", "Asia/Yerevan"), + ("Atlantic/Azores", "Atlantic/Azores"), + ("Atlantic/Bermuda", "Atlantic/Bermuda"), + ("Atlantic/Canary", "Atlantic/Canary"), + ("Atlantic/Cape_Verde", "Atlantic/Cape_Verde"), + ("Atlantic/Faroe", "Atlantic/Faroe"), + ("Atlantic/Madeira", "Atlantic/Madeira"), + ("Atlantic/Reykjavik", "Atlantic/Reykjavik"), + ("Atlantic/South_Georgia", "Atlantic/South_Georgia"), + ("Atlantic/St_Helena", "Atlantic/St_Helena"), + ("Atlantic/Stanley", "Atlantic/Stanley"), + ("Australia/Adelaide", "Australia/Adelaide"), + ("Australia/Brisbane", "Australia/Brisbane"), + ("Australia/Broken_Hill", "Australia/Broken_Hill"), + ("Australia/Darwin", "Australia/Darwin"), + ("Australia/Eucla", "Australia/Eucla"), + ("Australia/Hobart", "Australia/Hobart"), + ("Australia/Lindeman", "Australia/Lindeman"), + ("Australia/Lord_Howe", "Australia/Lord_Howe"), + ("Australia/Melbourne", "Australia/Melbourne"), + ("Australia/Perth", "Australia/Perth"), + ("Australia/Sydney", "Australia/Sydney"), + ("Europe/Amsterdam", "Europe/Amsterdam"), + ("Europe/Andorra", "Europe/Andorra"), + ("Europe/Astrakhan", "Europe/Astrakhan"), + ("Europe/Athens", "Europe/Athens"), + ("Europe/Belgrade", "Europe/Belgrade"), + ("Europe/Berlin", "Europe/Berlin"), + ("Europe/Brussels", "Europe/Brussels"), + ("Europe/Bucharest", "Europe/Bucharest"), + ("Europe/Budapest", "Europe/Budapest"), + ("Europe/Chisinau", "Europe/Chisinau"), + ("Europe/Copenhagen", "Europe/Copenhagen"), + ("Europe/Dublin", "Europe/Dublin"), + ("Europe/Gibraltar", "Europe/Gibraltar"), + ("Europe/Helsinki", "Europe/Helsinki"), + ("Europe/Istanbul", "Europe/Istanbul"), + ("Europe/Kaliningrad", "Europe/Kaliningrad"), + ("Europe/Kirov", "Europe/Kirov"), + ("Europe/Kyiv", "Europe/Kyiv"), + ("Europe/Lisbon", "Europe/Lisbon"), + ("Europe/London", "Europe/London"), + ("Europe/Luxembourg", "Europe/Luxembourg"), + ("Europe/Madrid", "Europe/Madrid"), + ("Europe/Malta", "Europe/Malta"), + ("Europe/Minsk", "Europe/Minsk"), + ("Europe/Monaco", "Europe/Monaco"), + ("Europe/Moscow", "Europe/Moscow"), + ("Europe/Oslo", "Europe/Oslo"), + ("Europe/Paris", "Europe/Paris"), + ("Europe/Prague", "Europe/Prague"), + ("Europe/Riga", "Europe/Riga"), + ("Europe/Rome", "Europe/Rome"), + ("Europe/Samara", "Europe/Samara"), + ("Europe/Saratov", "Europe/Saratov"), + ("Europe/Simferopol", "Europe/Simferopol"), + ("Europe/Sofia", "Europe/Sofia"), + ("Europe/Stockholm", "Europe/Stockholm"), + ("Europe/Tallinn", "Europe/Tallinn"), + ("Europe/Tirane", "Europe/Tirane"), + ("Europe/Ulyanovsk", "Europe/Ulyanovsk"), + ("Europe/Uzhgorod", "Europe/Uzhgorod"), + ("Europe/Vaduz", "Europe/Vaduz"), + ("Europe/Vienna", "Europe/Vienna"), + ("Europe/Vilnius", "Europe/Vilnius"), + ("Europe/Volgograd", "Europe/Volgograd"), + ("Europe/Warsaw", "Europe/Warsaw"), + ("Europe/Zaporozhye", "Europe/Zaporozhye"), + ("Europe/Zurich", "Europe/Zurich"), + ("Indian/Antananarivo", "Indian/Antananarivo"), + ("Indian/Chagos", "Indian/Chagos"), + ("Indian/Christmas", "Indian/Christmas"), + ("Indian/Cocos", "Indian/Cocos"), + ("Indian/Comoro", "Indian/Comoro"), + ("Indian/Kerguelen", "Indian/Kerguelen"), + ("Indian/Mahe", "Indian/Mahe"), + ("Indian/Maldives", "Indian/Maldives"), + ("Indian/Mauritius", "Indian/Mauritius"), + ("Indian/Mayotte", "Indian/Mayotte"), + ("Indian/Reunion", "Indian/Reunion"), + ("Pacific/Apia", "Pacific/Apia"), + ("Pacific/Auckland", "Pacific/Auckland"), + ("Pacific/Bougainville", "Pacific/Bougainville"), + ("Pacific/Chatham", "Pacific/Chatham"), + ("Pacific/Chuuk", "Pacific/Chuuk"), + ("Pacific/Easter", "Pacific/Easter"), + ("Pacific/Efate", "Pacific/Efate"), + ("Pacific/Fakaofo", "Pacific/Fakaofo"), + ("Pacific/Fiji", "Pacific/Fiji"), + ("Pacific/Funafuti", "Pacific/Funafuti"), + ("Pacific/Galapagos", "Pacific/Galapagos"), + ("Pacific/Gambier", "Pacific/Gambier"), + ("Pacific/Guadalcanal", "Pacific/Guadalcanal"), + ("Pacific/Guam", "Pacific/Guam"), + ("Pacific/Honolulu", "Pacific/Honolulu"), + ("Pacific/Kanton", "Pacific/Kanton"), + ("Pacific/Kiritimati", "Pacific/Kiritimati"), + ("Pacific/Kosrae", "Pacific/Kosrae"), + ("Pacific/Kwajalein", "Pacific/Kwajalein"), + ("Pacific/Majuro", "Pacific/Majuro"), + ("Pacific/Marquesas", "Pacific/Marquesas"), + ("Pacific/Midway", "Pacific/Midway"), + ("Pacific/Nauru", "Pacific/Nauru"), + ("Pacific/Niue", "Pacific/Niue"), + ("Pacific/Norfolk", "Pacific/Norfolk"), + ("Pacific/Noumea", "Pacific/Noumea"), + ("Pacific/Pago_Pago", "Pacific/Pago_Pago"), + ("Pacific/Palau", "Pacific/Palau"), + ("Pacific/Pitcairn", "Pacific/Pitcairn"), + ("Pacific/Pohnpei", "Pacific/Pohnpei"), + ("Pacific/Port_Moresby", "Pacific/Port_Moresby"), + ("Pacific/Rarotonga", "Pacific/Rarotonga"), + ("Pacific/Saipan", "Pacific/Saipan"), + ("Pacific/Tahiti", "Pacific/Tahiti"), + ("Pacific/Tarawa", "Pacific/Tarawa"), + ("Pacific/Tongatapu", "Pacific/Tongatapu"), + ("Pacific/Wake", "Pacific/Wake"), + ("Pacific/Wallis", "Pacific/Wallis"), + ("UTC", "UTC"), + ], + default="UTC", + max_length=255, + ), + ), + ] diff --git a/ietf/meeting/migrations/0015_sessionpresentation_document2_fk.py b/ietf/meeting/migrations/0015_sessionpresentation_document2_fk.py deleted file mode 100644 index 257a5060cb..0000000000 --- a/ietf/meeting/migrations/0015_sessionpresentation_document2_fk.py +++ /dev/null @@ -1,29 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-08 11:58 - - -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0015_1_add_fk_to_document_id'), - ('meeting', '0014_auto_20190426_0305'), - ] - - operations = [ - migrations.AddField( - model_name='sessionpresentation', - name='document2', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field=b'id'), - ), - migrations.AlterField( - model_name='sessionpresentation', - name='document', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='old_sesspres', to='doc.Document', to_field=b'name'), - ), - ] diff --git a/ietf/meeting/migrations/0016_alter_meeting_country_alter_meeting_time_zone.py b/ietf/meeting/migrations/0016_alter_meeting_country_alter_meeting_time_zone.py new file mode 100644 index 0000000000..8c467ea156 --- /dev/null +++ b/ietf/meeting/migrations/0016_alter_meeting_country_alter_meeting_time_zone.py @@ -0,0 +1,694 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +from django.db import migrations, models + + +class Migration(migrations.Migration): + + dependencies = [ + ("meeting", "0015_alter_meeting_time_zone"), + ] + + operations = [ + migrations.AlterField( + model_name="meeting", + name="country", + field=models.CharField( + blank=True, + choices=[ + ("", "---------"), + ("AF", "Afghanistan"), + ("AL", "Albania"), + ("DZ", "Algeria"), + ("AD", "Andorra"), + ("AO", "Angola"), + ("AI", "Anguilla"), + ("AQ", "Antarctica"), + ("AG", "Antigua & Barbuda"), + ("AR", "Argentina"), + ("AM", "Armenia"), + ("AW", "Aruba"), + ("AU", "Australia"), + ("AT", "Austria"), + ("AZ", "Azerbaijan"), + ("BS", "Bahamas"), + ("BH", "Bahrain"), + ("BD", "Bangladesh"), + ("BB", "Barbados"), + ("BY", "Belarus"), + ("BE", "Belgium"), + ("BZ", "Belize"), + ("BJ", "Benin"), + ("BM", "Bermuda"), + ("BT", "Bhutan"), + ("BO", "Bolivia"), + ("BA", "Bosnia & Herzegovina"), + ("BW", "Botswana"), + ("BV", "Bouvet Island"), + ("BR", "Brazil"), + ("GB", "Britain (UK)"), + ("IO", "British Indian Ocean Territory"), + ("BN", "Brunei"), + ("BG", "Bulgaria"), + ("BF", "Burkina Faso"), + ("BI", "Burundi"), + ("KH", "Cambodia"), + ("CM", "Cameroon"), + ("CA", "Canada"), + ("CV", "Cape Verde"), + ("BQ", "Caribbean NL"), + ("KY", "Cayman Islands"), + ("CF", "Central African Rep."), + ("TD", "Chad"), + ("CL", "Chile"), + ("CN", "China"), + ("CX", "Christmas Island"), + ("CC", "Cocos (Keeling) Islands"), + ("CO", "Colombia"), + ("KM", "Comoros"), + ("CD", "Congo (Dem. Rep.)"), + ("CG", "Congo (Rep.)"), + ("CK", "Cook Islands"), + ("CR", "Costa Rica"), + ("HR", "Croatia"), + ("CU", "Cuba"), + ("CW", "Curaçao"), + ("CY", "Cyprus"), + ("CZ", "Czech Republic"), + ("CI", "Côte d'Ivoire"), + ("DK", "Denmark"), + ("DJ", "Djibouti"), + ("DM", "Dominica"), + ("DO", "Dominican Republic"), + ("TL", "East Timor"), + ("EC", "Ecuador"), + ("EG", "Egypt"), + ("SV", "El Salvador"), + ("GQ", "Equatorial Guinea"), + ("ER", "Eritrea"), + ("EE", "Estonia"), + ("SZ", "Eswatini (Swaziland)"), + ("ET", "Ethiopia"), + ("FK", "Falkland Islands"), + ("FO", "Faroe Islands"), + ("FJ", "Fiji"), + ("FI", "Finland"), + ("FR", "France"), + ("GF", "French Guiana"), + ("PF", "French Polynesia"), + ("TF", "French S. Terr."), + ("GA", "Gabon"), + ("GM", "Gambia"), + ("GE", "Georgia"), + ("DE", "Germany"), + ("GH", "Ghana"), + ("GI", "Gibraltar"), + ("GR", "Greece"), + ("GL", "Greenland"), + ("GD", "Grenada"), + ("GP", "Guadeloupe"), + ("GU", "Guam"), + ("GT", "Guatemala"), + ("GG", "Guernsey"), + ("GN", "Guinea"), + ("GW", "Guinea-Bissau"), + ("GY", "Guyana"), + ("HT", "Haiti"), + ("HM", "Heard Island & McDonald Islands"), + ("HN", "Honduras"), + ("HK", "Hong Kong"), + ("HU", "Hungary"), + ("IS", "Iceland"), + ("IN", "India"), + ("ID", "Indonesia"), + ("IR", "Iran"), + ("IQ", "Iraq"), + ("IE", "Ireland"), + ("IM", "Isle of Man"), + ("IL", "Israel"), + ("IT", "Italy"), + ("JM", "Jamaica"), + ("JP", "Japan"), + ("JE", "Jersey"), + ("JO", "Jordan"), + ("KZ", "Kazakhstan"), + ("KE", "Kenya"), + ("KI", "Kiribati"), + ("KP", "Korea (North)"), + ("KR", "Korea (South)"), + ("KW", "Kuwait"), + ("KG", "Kyrgyzstan"), + ("LA", "Laos"), + ("LV", "Latvia"), + ("LB", "Lebanon"), + ("LS", "Lesotho"), + ("LR", "Liberia"), + ("LY", "Libya"), + ("LI", "Liechtenstein"), + ("LT", "Lithuania"), + ("LU", "Luxembourg"), + ("MO", "Macau"), + ("MG", "Madagascar"), + ("MW", "Malawi"), + ("MY", "Malaysia"), + ("MV", "Maldives"), + ("ML", "Mali"), + ("MT", "Malta"), + ("MH", "Marshall Islands"), + ("MQ", "Martinique"), + ("MR", "Mauritania"), + ("MU", "Mauritius"), + ("YT", "Mayotte"), + ("MX", "Mexico"), + ("FM", "Micronesia"), + ("MD", "Moldova"), + ("MC", "Monaco"), + ("MN", "Mongolia"), + ("ME", "Montenegro"), + ("MS", "Montserrat"), + ("MA", "Morocco"), + ("MZ", "Mozambique"), + ("MM", "Myanmar (Burma)"), + ("NA", "Namibia"), + ("NR", "Nauru"), + ("NP", "Nepal"), + ("NL", "Netherlands"), + ("NC", "New Caledonia"), + ("NZ", "New Zealand"), + ("NI", "Nicaragua"), + ("NE", "Niger"), + ("NG", "Nigeria"), + ("NU", "Niue"), + ("NF", "Norfolk Island"), + ("MK", "North Macedonia"), + ("MP", "Northern Mariana Islands"), + ("NO", "Norway"), + ("OM", "Oman"), + ("PK", "Pakistan"), + ("PW", "Palau"), + ("PS", "Palestine"), + ("PA", "Panama"), + ("PG", "Papua New Guinea"), + ("PY", "Paraguay"), + ("PE", "Peru"), + ("PH", "Philippines"), + ("PN", "Pitcairn"), + ("PL", "Poland"), + ("PT", "Portugal"), + ("PR", "Puerto Rico"), + ("QA", "Qatar"), + ("RO", "Romania"), + ("RU", "Russia"), + ("RW", "Rwanda"), + ("RE", "Réunion"), + ("AS", "Samoa (American)"), + ("WS", "Samoa (western)"), + ("SM", "San Marino"), + ("ST", "Sao Tome & Principe"), + ("SA", "Saudi Arabia"), + ("SN", "Senegal"), + ("RS", "Serbia"), + ("SC", "Seychelles"), + ("SL", "Sierra Leone"), + ("SG", "Singapore"), + ("SK", "Slovakia"), + ("SI", "Slovenia"), + ("SB", "Solomon Islands"), + ("SO", "Somalia"), + ("ZA", "South Africa"), + ("GS", "South Georgia & the South Sandwich Islands"), + ("SS", "South Sudan"), + ("ES", "Spain"), + ("LK", "Sri Lanka"), + ("BL", "St Barthelemy"), + ("SH", "St Helena"), + ("KN", "St Kitts & Nevis"), + ("LC", "St Lucia"), + ("SX", "St Maarten (Dutch)"), + ("MF", "St Martin (French)"), + ("PM", "St Pierre & Miquelon"), + ("VC", "St Vincent"), + ("SD", "Sudan"), + ("SR", "Suriname"), + ("SJ", "Svalbard & Jan Mayen"), + ("SE", "Sweden"), + ("CH", "Switzerland"), + ("SY", "Syria"), + ("TW", "Taiwan"), + ("TJ", "Tajikistan"), + ("TZ", "Tanzania"), + ("TH", "Thailand"), + ("TG", "Togo"), + ("TK", "Tokelau"), + ("TO", "Tonga"), + ("TT", "Trinidad & Tobago"), + ("TN", "Tunisia"), + ("TR", "Turkey"), + ("TM", "Turkmenistan"), + ("TC", "Turks & Caicos Is"), + ("TV", "Tuvalu"), + ("UM", "US minor outlying islands"), + ("UG", "Uganda"), + ("UA", "Ukraine"), + ("AE", "United Arab Emirates"), + ("US", "United States"), + ("UY", "Uruguay"), + ("UZ", "Uzbekistan"), + ("VU", "Vanuatu"), + ("VA", "Vatican City"), + ("VE", "Venezuela"), + ("VN", "Vietnam"), + ("VG", "Virgin Islands (UK)"), + ("VI", "Virgin Islands (US)"), + ("WF", "Wallis & Futuna"), + ("EH", "Western Sahara"), + ("YE", "Yemen"), + ("ZM", "Zambia"), + ("ZW", "Zimbabwe"), + ("AX", "Åland Islands"), + ], + max_length=2, + ), + ), + migrations.AlterField( + model_name="meeting", + name="time_zone", + field=models.CharField( + choices=[ + ("", "---------"), + ("Africa/Abidjan", "Africa/Abidjan"), + ("Africa/Accra", "Africa/Accra"), + ("Africa/Addis_Ababa", "Africa/Addis_Ababa"), + ("Africa/Algiers", "Africa/Algiers"), + ("Africa/Asmara", "Africa/Asmara"), + ("Africa/Bamako", "Africa/Bamako"), + ("Africa/Bangui", "Africa/Bangui"), + ("Africa/Banjul", "Africa/Banjul"), + ("Africa/Bissau", "Africa/Bissau"), + ("Africa/Blantyre", "Africa/Blantyre"), + ("Africa/Brazzaville", "Africa/Brazzaville"), + ("Africa/Bujumbura", "Africa/Bujumbura"), + ("Africa/Cairo", "Africa/Cairo"), + ("Africa/Casablanca", "Africa/Casablanca"), + ("Africa/Ceuta", "Africa/Ceuta"), + ("Africa/Conakry", "Africa/Conakry"), + ("Africa/Dakar", "Africa/Dakar"), + ("Africa/Dar_es_Salaam", "Africa/Dar_es_Salaam"), + ("Africa/Djibouti", "Africa/Djibouti"), + ("Africa/Douala", "Africa/Douala"), + ("Africa/El_Aaiun", "Africa/El_Aaiun"), + ("Africa/Freetown", "Africa/Freetown"), + ("Africa/Gaborone", "Africa/Gaborone"), + ("Africa/Harare", "Africa/Harare"), + ("Africa/Johannesburg", "Africa/Johannesburg"), + ("Africa/Juba", "Africa/Juba"), + ("Africa/Kampala", "Africa/Kampala"), + ("Africa/Khartoum", "Africa/Khartoum"), + ("Africa/Kigali", "Africa/Kigali"), + ("Africa/Kinshasa", "Africa/Kinshasa"), + ("Africa/Lagos", "Africa/Lagos"), + ("Africa/Libreville", "Africa/Libreville"), + ("Africa/Lome", "Africa/Lome"), + ("Africa/Luanda", "Africa/Luanda"), + ("Africa/Lubumbashi", "Africa/Lubumbashi"), + ("Africa/Lusaka", "Africa/Lusaka"), + ("Africa/Malabo", "Africa/Malabo"), + ("Africa/Maputo", "Africa/Maputo"), + ("Africa/Maseru", "Africa/Maseru"), + ("Africa/Mbabane", "Africa/Mbabane"), + ("Africa/Mogadishu", "Africa/Mogadishu"), + ("Africa/Monrovia", "Africa/Monrovia"), + ("Africa/Nairobi", "Africa/Nairobi"), + ("Africa/Ndjamena", "Africa/Ndjamena"), + ("Africa/Niamey", "Africa/Niamey"), + ("Africa/Nouakchott", "Africa/Nouakchott"), + ("Africa/Ouagadougou", "Africa/Ouagadougou"), + ("Africa/Porto-Novo", "Africa/Porto-Novo"), + ("Africa/Sao_Tome", "Africa/Sao_Tome"), + ("Africa/Tripoli", "Africa/Tripoli"), + ("Africa/Tunis", "Africa/Tunis"), + ("Africa/Windhoek", "Africa/Windhoek"), + ("America/Adak", "America/Adak"), + ("America/Anchorage", "America/Anchorage"), + ("America/Anguilla", "America/Anguilla"), + ("America/Antigua", "America/Antigua"), + ("America/Araguaina", "America/Araguaina"), + ( + "America/Argentina/Buenos_Aires", + "America/Argentina/Buenos_Aires", + ), + ("America/Argentina/Catamarca", "America/Argentina/Catamarca"), + ("America/Argentina/Cordoba", "America/Argentina/Cordoba"), + ("America/Argentina/Jujuy", "America/Argentina/Jujuy"), + ("America/Argentina/La_Rioja", "America/Argentina/La_Rioja"), + ("America/Argentina/Mendoza", "America/Argentina/Mendoza"), + ( + "America/Argentina/Rio_Gallegos", + "America/Argentina/Rio_Gallegos", + ), + ("America/Argentina/Salta", "America/Argentina/Salta"), + ("America/Argentina/San_Juan", "America/Argentina/San_Juan"), + ("America/Argentina/San_Luis", "America/Argentina/San_Luis"), + ("America/Argentina/Tucuman", "America/Argentina/Tucuman"), + ("America/Argentina/Ushuaia", "America/Argentina/Ushuaia"), + ("America/Aruba", "America/Aruba"), + ("America/Asuncion", "America/Asuncion"), + ("America/Atikokan", "America/Atikokan"), + ("America/Bahia", "America/Bahia"), + ("America/Bahia_Banderas", "America/Bahia_Banderas"), + ("America/Barbados", "America/Barbados"), + ("America/Belem", "America/Belem"), + ("America/Belize", "America/Belize"), + ("America/Blanc-Sablon", "America/Blanc-Sablon"), + ("America/Boa_Vista", "America/Boa_Vista"), + ("America/Bogota", "America/Bogota"), + ("America/Boise", "America/Boise"), + ("America/Cambridge_Bay", "America/Cambridge_Bay"), + ("America/Campo_Grande", "America/Campo_Grande"), + ("America/Cancun", "America/Cancun"), + ("America/Caracas", "America/Caracas"), + ("America/Cayenne", "America/Cayenne"), + ("America/Cayman", "America/Cayman"), + ("America/Chicago", "America/Chicago"), + ("America/Chihuahua", "America/Chihuahua"), + ("America/Ciudad_Juarez", "America/Ciudad_Juarez"), + ("America/Costa_Rica", "America/Costa_Rica"), + ("America/Coyhaique", "America/Coyhaique"), + ("America/Creston", "America/Creston"), + ("America/Cuiaba", "America/Cuiaba"), + ("America/Curacao", "America/Curacao"), + ("America/Danmarkshavn", "America/Danmarkshavn"), + ("America/Dawson", "America/Dawson"), + ("America/Dawson_Creek", "America/Dawson_Creek"), + ("America/Denver", "America/Denver"), + ("America/Detroit", "America/Detroit"), + ("America/Dominica", "America/Dominica"), + ("America/Edmonton", "America/Edmonton"), + ("America/Eirunepe", "America/Eirunepe"), + ("America/El_Salvador", "America/El_Salvador"), + ("America/Fort_Nelson", "America/Fort_Nelson"), + ("America/Fortaleza", "America/Fortaleza"), + ("America/Glace_Bay", "America/Glace_Bay"), + ("America/Goose_Bay", "America/Goose_Bay"), + ("America/Grand_Turk", "America/Grand_Turk"), + ("America/Grenada", "America/Grenada"), + ("America/Guadeloupe", "America/Guadeloupe"), + ("America/Guatemala", "America/Guatemala"), + ("America/Guayaquil", "America/Guayaquil"), + ("America/Guyana", "America/Guyana"), + ("America/Halifax", "America/Halifax"), + ("America/Havana", "America/Havana"), + ("America/Hermosillo", "America/Hermosillo"), + ("America/Indiana/Indianapolis", "America/Indiana/Indianapolis"), + ("America/Indiana/Knox", "America/Indiana/Knox"), + ("America/Indiana/Marengo", "America/Indiana/Marengo"), + ("America/Indiana/Petersburg", "America/Indiana/Petersburg"), + ("America/Indiana/Tell_City", "America/Indiana/Tell_City"), + ("America/Indiana/Vevay", "America/Indiana/Vevay"), + ("America/Indiana/Vincennes", "America/Indiana/Vincennes"), + ("America/Indiana/Winamac", "America/Indiana/Winamac"), + ("America/Inuvik", "America/Inuvik"), + ("America/Iqaluit", "America/Iqaluit"), + ("America/Jamaica", "America/Jamaica"), + ("America/Juneau", "America/Juneau"), + ("America/Kentucky/Louisville", "America/Kentucky/Louisville"), + ("America/Kentucky/Monticello", "America/Kentucky/Monticello"), + ("America/La_Paz", "America/La_Paz"), + ("America/Lima", "America/Lima"), + ("America/Los_Angeles", "America/Los_Angeles"), + ("America/Maceio", "America/Maceio"), + ("America/Managua", "America/Managua"), + ("America/Manaus", "America/Manaus"), + ("America/Martinique", "America/Martinique"), + ("America/Matamoros", "America/Matamoros"), + ("America/Mazatlan", "America/Mazatlan"), + ("America/Menominee", "America/Menominee"), + ("America/Merida", "America/Merida"), + ("America/Metlakatla", "America/Metlakatla"), + ("America/Mexico_City", "America/Mexico_City"), + ("America/Miquelon", "America/Miquelon"), + ("America/Moncton", "America/Moncton"), + ("America/Monterrey", "America/Monterrey"), + ("America/Montevideo", "America/Montevideo"), + ("America/Montserrat", "America/Montserrat"), + ("America/Nassau", "America/Nassau"), + ("America/New_York", "America/New_York"), + ("America/Nome", "America/Nome"), + ("America/Noronha", "America/Noronha"), + ("America/North_Dakota/Beulah", "America/North_Dakota/Beulah"), + ("America/North_Dakota/Center", "America/North_Dakota/Center"), + ( + "America/North_Dakota/New_Salem", + "America/North_Dakota/New_Salem", + ), + ("America/Nuuk", "America/Nuuk"), + ("America/Ojinaga", "America/Ojinaga"), + ("America/Panama", "America/Panama"), + ("America/Paramaribo", "America/Paramaribo"), + ("America/Phoenix", "America/Phoenix"), + ("America/Port-au-Prince", "America/Port-au-Prince"), + ("America/Port_of_Spain", "America/Port_of_Spain"), + ("America/Porto_Velho", "America/Porto_Velho"), + ("America/Puerto_Rico", "America/Puerto_Rico"), + ("America/Punta_Arenas", "America/Punta_Arenas"), + ("America/Rankin_Inlet", "America/Rankin_Inlet"), + ("America/Recife", "America/Recife"), + ("America/Regina", "America/Regina"), + ("America/Resolute", "America/Resolute"), + ("America/Rio_Branco", "America/Rio_Branco"), + ("America/Santarem", "America/Santarem"), + ("America/Santiago", "America/Santiago"), + ("America/Santo_Domingo", "America/Santo_Domingo"), + ("America/Sao_Paulo", "America/Sao_Paulo"), + ("America/Scoresbysund", "America/Scoresbysund"), + ("America/Sitka", "America/Sitka"), + ("America/St_Johns", "America/St_Johns"), + ("America/St_Kitts", "America/St_Kitts"), + ("America/St_Lucia", "America/St_Lucia"), + ("America/St_Thomas", "America/St_Thomas"), + ("America/St_Vincent", "America/St_Vincent"), + ("America/Swift_Current", "America/Swift_Current"), + ("America/Tegucigalpa", "America/Tegucigalpa"), + ("America/Thule", "America/Thule"), + ("America/Tijuana", "America/Tijuana"), + ("America/Toronto", "America/Toronto"), + ("America/Tortola", "America/Tortola"), + ("America/Vancouver", "America/Vancouver"), + ("America/Whitehorse", "America/Whitehorse"), + ("America/Winnipeg", "America/Winnipeg"), + ("America/Yakutat", "America/Yakutat"), + ("Antarctica/Casey", "Antarctica/Casey"), + ("Antarctica/Davis", "Antarctica/Davis"), + ("Antarctica/DumontDUrville", "Antarctica/DumontDUrville"), + ("Antarctica/Macquarie", "Antarctica/Macquarie"), + ("Antarctica/Mawson", "Antarctica/Mawson"), + ("Antarctica/McMurdo", "Antarctica/McMurdo"), + ("Antarctica/Palmer", "Antarctica/Palmer"), + ("Antarctica/Rothera", "Antarctica/Rothera"), + ("Antarctica/Syowa", "Antarctica/Syowa"), + ("Antarctica/Troll", "Antarctica/Troll"), + ("Antarctica/Vostok", "Antarctica/Vostok"), + ("Asia/Aden", "Asia/Aden"), + ("Asia/Almaty", "Asia/Almaty"), + ("Asia/Amman", "Asia/Amman"), + ("Asia/Anadyr", "Asia/Anadyr"), + ("Asia/Aqtau", "Asia/Aqtau"), + ("Asia/Aqtobe", "Asia/Aqtobe"), + ("Asia/Ashgabat", "Asia/Ashgabat"), + ("Asia/Atyrau", "Asia/Atyrau"), + ("Asia/Baghdad", "Asia/Baghdad"), + ("Asia/Bahrain", "Asia/Bahrain"), + ("Asia/Baku", "Asia/Baku"), + ("Asia/Bangkok", "Asia/Bangkok"), + ("Asia/Barnaul", "Asia/Barnaul"), + ("Asia/Beirut", "Asia/Beirut"), + ("Asia/Bishkek", "Asia/Bishkek"), + ("Asia/Brunei", "Asia/Brunei"), + ("Asia/Chita", "Asia/Chita"), + ("Asia/Colombo", "Asia/Colombo"), + ("Asia/Damascus", "Asia/Damascus"), + ("Asia/Dhaka", "Asia/Dhaka"), + ("Asia/Dili", "Asia/Dili"), + ("Asia/Dubai", "Asia/Dubai"), + ("Asia/Dushanbe", "Asia/Dushanbe"), + ("Asia/Famagusta", "Asia/Famagusta"), + ("Asia/Gaza", "Asia/Gaza"), + ("Asia/Hebron", "Asia/Hebron"), + ("Asia/Ho_Chi_Minh", "Asia/Ho_Chi_Minh"), + ("Asia/Hong_Kong", "Asia/Hong_Kong"), + ("Asia/Hovd", "Asia/Hovd"), + ("Asia/Irkutsk", "Asia/Irkutsk"), + ("Asia/Jakarta", "Asia/Jakarta"), + ("Asia/Jayapura", "Asia/Jayapura"), + ("Asia/Jerusalem", "Asia/Jerusalem"), + ("Asia/Kabul", "Asia/Kabul"), + ("Asia/Kamchatka", "Asia/Kamchatka"), + ("Asia/Karachi", "Asia/Karachi"), + ("Asia/Kathmandu", "Asia/Kathmandu"), + ("Asia/Khandyga", "Asia/Khandyga"), + ("Asia/Kolkata", "Asia/Kolkata"), + ("Asia/Krasnoyarsk", "Asia/Krasnoyarsk"), + ("Asia/Kuala_Lumpur", "Asia/Kuala_Lumpur"), + ("Asia/Kuching", "Asia/Kuching"), + ("Asia/Kuwait", "Asia/Kuwait"), + ("Asia/Macau", "Asia/Macau"), + ("Asia/Magadan", "Asia/Magadan"), + ("Asia/Makassar", "Asia/Makassar"), + ("Asia/Manila", "Asia/Manila"), + ("Asia/Muscat", "Asia/Muscat"), + ("Asia/Nicosia", "Asia/Nicosia"), + ("Asia/Novokuznetsk", "Asia/Novokuznetsk"), + ("Asia/Novosibirsk", "Asia/Novosibirsk"), + ("Asia/Omsk", "Asia/Omsk"), + ("Asia/Oral", "Asia/Oral"), + ("Asia/Phnom_Penh", "Asia/Phnom_Penh"), + ("Asia/Pontianak", "Asia/Pontianak"), + ("Asia/Pyongyang", "Asia/Pyongyang"), + ("Asia/Qatar", "Asia/Qatar"), + ("Asia/Qostanay", "Asia/Qostanay"), + ("Asia/Qyzylorda", "Asia/Qyzylorda"), + ("Asia/Riyadh", "Asia/Riyadh"), + ("Asia/Sakhalin", "Asia/Sakhalin"), + ("Asia/Samarkand", "Asia/Samarkand"), + ("Asia/Seoul", "Asia/Seoul"), + ("Asia/Shanghai", "Asia/Shanghai"), + ("Asia/Singapore", "Asia/Singapore"), + ("Asia/Srednekolymsk", "Asia/Srednekolymsk"), + ("Asia/Taipei", "Asia/Taipei"), + ("Asia/Tashkent", "Asia/Tashkent"), + ("Asia/Tbilisi", "Asia/Tbilisi"), + ("Asia/Tehran", "Asia/Tehran"), + ("Asia/Thimphu", "Asia/Thimphu"), + ("Asia/Tokyo", "Asia/Tokyo"), + ("Asia/Tomsk", "Asia/Tomsk"), + ("Asia/Ulaanbaatar", "Asia/Ulaanbaatar"), + ("Asia/Urumqi", "Asia/Urumqi"), + ("Asia/Ust-Nera", "Asia/Ust-Nera"), + ("Asia/Vientiane", "Asia/Vientiane"), + ("Asia/Vladivostok", "Asia/Vladivostok"), + ("Asia/Yakutsk", "Asia/Yakutsk"), + ("Asia/Yangon", "Asia/Yangon"), + ("Asia/Yekaterinburg", "Asia/Yekaterinburg"), + ("Asia/Yerevan", "Asia/Yerevan"), + ("Atlantic/Azores", "Atlantic/Azores"), + ("Atlantic/Bermuda", "Atlantic/Bermuda"), + ("Atlantic/Canary", "Atlantic/Canary"), + ("Atlantic/Cape_Verde", "Atlantic/Cape_Verde"), + ("Atlantic/Faroe", "Atlantic/Faroe"), + ("Atlantic/Madeira", "Atlantic/Madeira"), + ("Atlantic/Reykjavik", "Atlantic/Reykjavik"), + ("Atlantic/South_Georgia", "Atlantic/South_Georgia"), + ("Atlantic/St_Helena", "Atlantic/St_Helena"), + ("Atlantic/Stanley", "Atlantic/Stanley"), + ("Australia/Adelaide", "Australia/Adelaide"), + ("Australia/Brisbane", "Australia/Brisbane"), + ("Australia/Broken_Hill", "Australia/Broken_Hill"), + ("Australia/Darwin", "Australia/Darwin"), + ("Australia/Eucla", "Australia/Eucla"), + ("Australia/Hobart", "Australia/Hobart"), + ("Australia/Lindeman", "Australia/Lindeman"), + ("Australia/Lord_Howe", "Australia/Lord_Howe"), + ("Australia/Melbourne", "Australia/Melbourne"), + ("Australia/Perth", "Australia/Perth"), + ("Australia/Sydney", "Australia/Sydney"), + ("Europe/Amsterdam", "Europe/Amsterdam"), + ("Europe/Andorra", "Europe/Andorra"), + ("Europe/Astrakhan", "Europe/Astrakhan"), + ("Europe/Athens", "Europe/Athens"), + ("Europe/Belgrade", "Europe/Belgrade"), + ("Europe/Berlin", "Europe/Berlin"), + ("Europe/Brussels", "Europe/Brussels"), + ("Europe/Bucharest", "Europe/Bucharest"), + ("Europe/Budapest", "Europe/Budapest"), + ("Europe/Chisinau", "Europe/Chisinau"), + ("Europe/Copenhagen", "Europe/Copenhagen"), + ("Europe/Dublin", "Europe/Dublin"), + ("Europe/Gibraltar", "Europe/Gibraltar"), + ("Europe/Helsinki", "Europe/Helsinki"), + ("Europe/Istanbul", "Europe/Istanbul"), + ("Europe/Kaliningrad", "Europe/Kaliningrad"), + ("Europe/Kirov", "Europe/Kirov"), + ("Europe/Kyiv", "Europe/Kyiv"), + ("Europe/Lisbon", "Europe/Lisbon"), + ("Europe/London", "Europe/London"), + ("Europe/Luxembourg", "Europe/Luxembourg"), + ("Europe/Madrid", "Europe/Madrid"), + ("Europe/Malta", "Europe/Malta"), + ("Europe/Minsk", "Europe/Minsk"), + ("Europe/Monaco", "Europe/Monaco"), + ("Europe/Moscow", "Europe/Moscow"), + ("Europe/Oslo", "Europe/Oslo"), + ("Europe/Paris", "Europe/Paris"), + ("Europe/Prague", "Europe/Prague"), + ("Europe/Riga", "Europe/Riga"), + ("Europe/Rome", "Europe/Rome"), + ("Europe/Samara", "Europe/Samara"), + ("Europe/Saratov", "Europe/Saratov"), + ("Europe/Simferopol", "Europe/Simferopol"), + ("Europe/Sofia", "Europe/Sofia"), + ("Europe/Stockholm", "Europe/Stockholm"), + ("Europe/Tallinn", "Europe/Tallinn"), + ("Europe/Tirane", "Europe/Tirane"), + ("Europe/Ulyanovsk", "Europe/Ulyanovsk"), + ("Europe/Vaduz", "Europe/Vaduz"), + ("Europe/Vienna", "Europe/Vienna"), + ("Europe/Vilnius", "Europe/Vilnius"), + ("Europe/Volgograd", "Europe/Volgograd"), + ("Europe/Warsaw", "Europe/Warsaw"), + ("Europe/Zurich", "Europe/Zurich"), + ("Indian/Antananarivo", "Indian/Antananarivo"), + ("Indian/Chagos", "Indian/Chagos"), + ("Indian/Christmas", "Indian/Christmas"), + ("Indian/Cocos", "Indian/Cocos"), + ("Indian/Comoro", "Indian/Comoro"), + ("Indian/Kerguelen", "Indian/Kerguelen"), + ("Indian/Mahe", "Indian/Mahe"), + ("Indian/Maldives", "Indian/Maldives"), + ("Indian/Mauritius", "Indian/Mauritius"), + ("Indian/Mayotte", "Indian/Mayotte"), + ("Indian/Reunion", "Indian/Reunion"), + ("Pacific/Apia", "Pacific/Apia"), + ("Pacific/Auckland", "Pacific/Auckland"), + ("Pacific/Bougainville", "Pacific/Bougainville"), + ("Pacific/Chatham", "Pacific/Chatham"), + ("Pacific/Chuuk", "Pacific/Chuuk"), + ("Pacific/Easter", "Pacific/Easter"), + ("Pacific/Efate", "Pacific/Efate"), + ("Pacific/Fakaofo", "Pacific/Fakaofo"), + ("Pacific/Fiji", "Pacific/Fiji"), + ("Pacific/Funafuti", "Pacific/Funafuti"), + ("Pacific/Galapagos", "Pacific/Galapagos"), + ("Pacific/Gambier", "Pacific/Gambier"), + ("Pacific/Guadalcanal", "Pacific/Guadalcanal"), + ("Pacific/Guam", "Pacific/Guam"), + ("Pacific/Honolulu", "Pacific/Honolulu"), + ("Pacific/Kanton", "Pacific/Kanton"), + ("Pacific/Kiritimati", "Pacific/Kiritimati"), + ("Pacific/Kosrae", "Pacific/Kosrae"), + ("Pacific/Kwajalein", "Pacific/Kwajalein"), + ("Pacific/Majuro", "Pacific/Majuro"), + ("Pacific/Marquesas", "Pacific/Marquesas"), + ("Pacific/Midway", "Pacific/Midway"), + ("Pacific/Nauru", "Pacific/Nauru"), + ("Pacific/Niue", "Pacific/Niue"), + ("Pacific/Norfolk", "Pacific/Norfolk"), + ("Pacific/Noumea", "Pacific/Noumea"), + ("Pacific/Pago_Pago", "Pacific/Pago_Pago"), + ("Pacific/Palau", "Pacific/Palau"), + ("Pacific/Pitcairn", "Pacific/Pitcairn"), + ("Pacific/Pohnpei", "Pacific/Pohnpei"), + ("Pacific/Port_Moresby", "Pacific/Port_Moresby"), + ("Pacific/Rarotonga", "Pacific/Rarotonga"), + ("Pacific/Saipan", "Pacific/Saipan"), + ("Pacific/Tahiti", "Pacific/Tahiti"), + ("Pacific/Tarawa", "Pacific/Tarawa"), + ("Pacific/Tongatapu", "Pacific/Tongatapu"), + ("Pacific/Wake", "Pacific/Wake"), + ("Pacific/Wallis", "Pacific/Wallis"), + ("UTC", "UTC"), + ], + default="UTC", + max_length=255, + ), + ), + ] diff --git a/ietf/meeting/migrations/0016_remove_sessionpresentation_document.py b/ietf/meeting/migrations/0016_remove_sessionpresentation_document.py deleted file mode 100644 index e93baaa33b..0000000000 --- a/ietf/meeting/migrations/0016_remove_sessionpresentation_document.py +++ /dev/null @@ -1,34 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-21 03:57 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0018_remove_old_document_field'), - ('meeting', '0015_sessionpresentation_document2_fk'), - ] - - operations = [ - - # We need to get rid of the current through table uniqueness - # constraint before we can remove the document column: The table - # has "UNIQUE KEY `session_id` (`session_id`,`document_id`)" - migrations.RunSQL( - "ALTER TABLE `meeting_session_materials` DROP INDEX `session_id`;", - "CREATE UNIQUE INDEX `session_id` ON `meeting_session_materials` (`session_id`, `document_id`);" - ), - ## This doesn't work: - # migrations.RemoveIndex( - # model_name='sessionpresentation', - # name='session_id' - # ), - migrations.RemoveField( - model_name='sessionpresentation', - name='document', - ), - ] diff --git a/ietf/meeting/migrations/0017_rename_field_document2.py b/ietf/meeting/migrations/0017_rename_field_document2.py deleted file mode 100644 index 5965d8d20e..0000000000 --- a/ietf/meeting/migrations/0017_rename_field_document2.py +++ /dev/null @@ -1,26 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-21 05:31 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0019_rename_field_document2'), - ('meeting', '0016_remove_sessionpresentation_document'), - ] - - operations = [ - migrations.RenameField( - model_name='sessionpresentation', - old_name='document2', - new_name='document', - ), - migrations.AlterUniqueTogether( - name='sessionpresentation', - unique_together=set([('session', 'document')]), - ), - ] diff --git a/ietf/meeting/migrations/0018_document_primary_key_cleanup.py b/ietf/meeting/migrations/0018_document_primary_key_cleanup.py deleted file mode 100644 index 749caa443b..0000000000 --- a/ietf/meeting/migrations/0018_document_primary_key_cleanup.py +++ /dev/null @@ -1,23 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-06-10 03:47 - - -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0017_rename_field_document2'), - ] - - operations = [ - migrations.AlterField( - model_name='sessionpresentation', - name='document', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document'), - ), - ] diff --git a/ietf/meeting/migrations/0019_slidesubmission_time.py b/ietf/meeting/migrations/0019_slidesubmission_time.py deleted file mode 100644 index 3dc0e2c523..0000000000 --- a/ietf/meeting/migrations/0019_slidesubmission_time.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.22 on 2019-07-21 14:03 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0018_document_primary_key_cleanup'), - ] - - operations = [ - migrations.AddField( - model_name='slidesubmission', - name='time', - field=models.DateTimeField(auto_now=True), - ), - ] diff --git a/ietf/meeting/migrations/0020_remove_future_break_sessions.py b/ietf/meeting/migrations/0020_remove_future_break_sessions.py deleted file mode 100644 index 5d2569b1cc..0000000000 --- a/ietf/meeting/migrations/0020_remove_future_break_sessions.py +++ /dev/null @@ -1,32 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.22 on 2019-07-22 14:56 - - -import datetime -from django.db import migrations - - -def forward(apps, schema_editor): - Meeting = apps.get_model('meeting', 'Meeting') - today = datetime.datetime.today() - meetings = Meeting.objects.filter(date__gt=today, type='ietf') - for meeting in meetings: - meeting.agenda.assignments.all().delete() - meeting.session_set.all().delete() - meeting.timeslot_set.all().delete() - - -def backward(apps, schema_editor): - return - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0019_slidesubmission_time'), - ] - - operations = [ - migrations.RunPython(forward, backward), - ] diff --git a/ietf/meeting/migrations/0021_rename_meeting_agenda_to_schedule.py b/ietf/meeting/migrations/0021_rename_meeting_agenda_to_schedule.py deleted file mode 100644 index 5399968cb7..0000000000 --- a/ietf/meeting/migrations/0021_rename_meeting_agenda_to_schedule.py +++ /dev/null @@ -1,28 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.26 on 2019-11-18 04:01 - - -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0020_remove_future_break_sessions'), - ] - - operations = [ - migrations.RenameField( - model_name='meeting', - old_name='agenda', - new_name='schedule', - ), - migrations.AlterField( - model_name='schedule', - name='meeting', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, related_name='schedule_set', to='meeting.Meeting'), - ), - ] diff --git a/ietf/meeting/migrations/0022_schedulingevent.py b/ietf/meeting/migrations/0022_schedulingevent.py deleted file mode 100644 index 22602a06b4..0000000000 --- a/ietf/meeting/migrations/0022_schedulingevent.py +++ /dev/null @@ -1,31 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.26 on 2019-11-19 02:41 - - -import datetime -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0009_auto_20190118_0725'), - ('name', '0007_fix_m2m_slug_id_length'), - ('meeting', '0021_rename_meeting_agenda_to_schedule'), - ] - - operations = [ - migrations.CreateModel( - name='SchedulingEvent', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('time', models.DateTimeField(default=datetime.datetime.now, help_text='When the event happened')), - ('by', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), - ('session', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='meeting.Session')), - ('status', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.SessionStatusName')), - ], - ), - ] diff --git a/ietf/meeting/migrations/0023_create_scheduling_events.py b/ietf/meeting/migrations/0023_create_scheduling_events.py deleted file mode 100644 index 5adcdd59e1..0000000000 --- a/ietf/meeting/migrations/0023_create_scheduling_events.py +++ /dev/null @@ -1,67 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.26 on 2019-11-19 02:42 - - -from django.db import migrations - -import datetime - -def create_scheduling_events(apps, schema_editor): - Session = apps.get_model('meeting', 'Session') - SchedulingEvent = apps.get_model('meeting', 'SchedulingEvent') - Person = apps.get_model('person', 'Person') - SessionStatusName = apps.get_model('name', 'SessionStatusName') - - system_person = Person.objects.get(name='(System)') - session_status_names = { n.slug: n for n in SessionStatusName.objects.all() } - - epoch_time = datetime.datetime(1970, 1, 1, 0, 0, 0) - - for s in Session.objects.select_related('requested_by').filter(schedulingevent=None).iterator(): - # temporarily fix up weird timestamps for the migration - if s.requested == epoch_time: - s.requested = s.modified - - requested_event = SchedulingEvent() - requested_event.session = s - requested_event.time = s.requested - requested_event.by = s.requested_by - requested_event.status = session_status_names[s.status_id if s.status_id == 'apprw' or (s.status_id == 'notmeet' and not s.scheduled) else 'schedw'] - requested_event.save() - - scheduled_event = None - if s.status_id != requested_event.status_id: - if s.scheduled or s.status_id in ['sched', 'scheda']: - scheduled_event = SchedulingEvent() - scheduled_event.session = s - if s.scheduled: - scheduled_event.time = s.scheduled - else: - # we don't know when this happened - scheduled_event.time = s.modified - scheduled_event.by = system_person # we don't know who did it - scheduled_event.status = session_status_names[s.status_id if s.status_id == 'scheda' else 'sched'] - scheduled_event.save() - - final_event = None - if s.status_id not in ['apprw', 'schedw', 'notmeet', 'sched', 'scheda']: - final_event = SchedulingEvent() - final_event.session = s - final_event.time = s.modified - final_event.by = system_person # we don't know who did it - final_event.status = session_status_names[s.status_id] - final_event.save() - -def noop(apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0022_schedulingevent'), - ] - - operations = [ - migrations.RunPython(create_scheduling_events, noop), - ] diff --git a/ietf/meeting/migrations/0024_auto_20191204_1731.py b/ietf/meeting/migrations/0024_auto_20191204_1731.py deleted file mode 100644 index 828611bb98..0000000000 --- a/ietf/meeting/migrations/0024_auto_20191204_1731.py +++ /dev/null @@ -1,28 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.26 on 2019-12-04 17:31 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0023_create_scheduling_events'), - ] - - operations = [ - migrations.RemoveField( - model_name='session', - name='requested', - ), - migrations.RemoveField( - model_name='session', - name='requested_by', - ), - migrations.RemoveField( - model_name='session', - name='status', - ), - ] diff --git a/ietf/meeting/migrations/0025_rename_type_session_to_regular.py b/ietf/meeting/migrations/0025_rename_type_session_to_regular.py deleted file mode 100644 index 1eeac57b5a..0000000000 --- a/ietf/meeting/migrations/0025_rename_type_session_to_regular.py +++ /dev/null @@ -1,38 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.26 on 2019-12-06 11:13 - - -from django.db import migrations - -def rename_session_to_regular(apps, schema_editor): - Session = apps.get_model('meeting', 'Session') - TimeSlot = apps.get_model('meeting', 'TimeSlot') - Room = apps.get_model('meeting', 'Room') - TimeSlotTypeName = apps.get_model('name', 'TimeSlotTypeName') - - TimeSlotTypeName.objects.create( - slug='regular', - name='Regular', - used=True, - order=0, - ) - - Session.objects.filter(type='session').update(type='regular') - TimeSlot.objects.filter(type='session').update(type='regular') - Room.session_types.through.objects.filter(timeslottypename='session').update(timeslottypename='regular') - - TimeSlotTypeName.objects.filter(slug='session').delete() - -def noop(apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0024_auto_20191204_1731'), - ] - - operations = [ - migrations.RunPython(rename_session_to_regular, noop), - ] diff --git a/ietf/meeting/migrations/0026_cancel_107_sessions.py b/ietf/meeting/migrations/0026_cancel_107_sessions.py deleted file mode 100644 index 754cd7c49e..0000000000 --- a/ietf/meeting/migrations/0026_cancel_107_sessions.py +++ /dev/null @@ -1,40 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.29 on 2020-03-18 16:18 -from __future__ import unicode_literals - -from django.db import migrations - - -def cancel_sessions(apps, schema_editor): - Session = apps.get_model('meeting', 'Session') - SchedulingEvent = apps.get_model('meeting', 'SchedulingEvent') - SessionStatusName = apps.get_model('name', 'SessionStatusName') - Person = apps.get_model('person', 'Person') - excludes = ['txauth','dispatch','add','raw','masque','wpack','drip','gendispatch','privacypass', 'ript', 'secdispatch', 'webtrans'] - canceled = SessionStatusName.objects.get(slug='canceled') - person = Person.objects.get(name='Ryan Cross') - sessions = Session.objects.filter(meeting__number=107,group__type__in=['wg','rg','ag']).exclude(group__acronym__in=excludes) - for session in sessions: - SchedulingEvent.objects.create( - session = session, - status = canceled, - by = person) - - -def reverse(apps, schema_editor): - SchedulingEvent = apps.get_model('meeting', 'SchedulingEvent') - Person = apps.get_model('person', 'Person') - person = Person.objects.get(name='Ryan Cross') - SchedulingEvent.objects.filter(session__meeting__number=107, by=person).delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0025_rename_type_session_to_regular'), - ] - - operations = [ - migrations.RunPython(cancel_sessions, reverse), - ] diff --git a/ietf/meeting/migrations/0027_add_constraint_options_and_joint_groups.py b/ietf/meeting/migrations/0027_add_constraint_options_and_joint_groups.py deleted file mode 100644 index 7a64fbe658..0000000000 --- a/ietf/meeting/migrations/0027_add_constraint_options_and_joint_groups.py +++ /dev/null @@ -1,52 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.27 on 2020-02-11 04:47 -from __future__ import unicode_literals - -from django.db import migrations, models - - -def forward(apps, schema_editor): - ConstraintName = apps.get_model("name", "ConstraintName") - ConstraintName.objects.create(slug="timerange", desc="", penalty=100000, - name="Can't meet within timerange") - ConstraintName.objects.create(slug="time_relation", desc="", penalty=1000, - name="Preference for time between sessions") - ConstraintName.objects.create(slug="wg_adjacent", desc="", penalty=10000, - name="Request for adjacent scheduling with another WG") - - -def reverse(apps, schema_editor): - ConstraintName = apps.get_model("name", "ConstraintName") - ConstraintName.objects.filter(slug__in=["timerange", "time_relation", "wg_adjacent"]).delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0010_timerangename'), - ('meeting', '0026_cancel_107_sessions'), - ] - - operations = [ - migrations.RemoveField( - model_name='constraint', - name='day', - ), - migrations.AddField( - model_name='constraint', - name='time_relation', - field=models.CharField(blank=True, choices=[('subsequent-days', 'Schedule the sessions on subsequent days'), ('one-day-seperation', 'Leave at least one free day in between the two sessions')], max_length=200), - ), - migrations.AddField( - model_name='constraint', - name='timeranges', - field=models.ManyToManyField(to='name.TimerangeName'), - ), - migrations.AddField( - model_name='session', - name='joint_with_groups', - field=models.ManyToManyField(related_name='sessions_joint_in', to='group.Group'), - ), - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/meeting/migrations/0028_auto_20200501_0139.py b/ietf/meeting/migrations/0028_auto_20200501_0139.py deleted file mode 100644 index 70c46c3d26..0000000000 --- a/ietf/meeting/migrations/0028_auto_20200501_0139.py +++ /dev/null @@ -1,26 +0,0 @@ -# -*- coding: utf-8 -*- -# Generated by Django 1.11.29 on 2020-05-01 01:39 -from __future__ import unicode_literals - -import django.core.validators -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0027_add_constraint_options_and_joint_groups'), - ] - - operations = [ - migrations.AlterField( - model_name='meeting', - name='time_zone', - field=models.CharField(blank=True, choices=[('', '---------'), ('Africa/Abidjan', 'Africa/Abidjan'), ('Africa/Accra', 'Africa/Accra'), ('Africa/Addis_Ababa', 'Africa/Addis_Ababa'), ('Africa/Algiers', 'Africa/Algiers'), ('Africa/Asmara', 'Africa/Asmara'), ('Africa/Bamako', 'Africa/Bamako'), ('Africa/Bangui', 'Africa/Bangui'), ('Africa/Banjul', 'Africa/Banjul'), ('Africa/Bissau', 'Africa/Bissau'), ('Africa/Blantyre', 'Africa/Blantyre'), ('Africa/Brazzaville', 'Africa/Brazzaville'), ('Africa/Bujumbura', 'Africa/Bujumbura'), ('Africa/Cairo', 'Africa/Cairo'), ('Africa/Casablanca', 'Africa/Casablanca'), ('Africa/Ceuta', 'Africa/Ceuta'), ('Africa/Conakry', 'Africa/Conakry'), ('Africa/Dakar', 'Africa/Dakar'), ('Africa/Dar_es_Salaam', 'Africa/Dar_es_Salaam'), ('Africa/Djibouti', 'Africa/Djibouti'), ('Africa/Douala', 'Africa/Douala'), ('Africa/El_Aaiun', 'Africa/El_Aaiun'), ('Africa/Freetown', 'Africa/Freetown'), ('Africa/Gaborone', 'Africa/Gaborone'), ('Africa/Harare', 'Africa/Harare'), ('Africa/Johannesburg', 'Africa/Johannesburg'), ('Africa/Juba', 'Africa/Juba'), ('Africa/Kampala', 'Africa/Kampala'), ('Africa/Khartoum', 'Africa/Khartoum'), ('Africa/Kigali', 'Africa/Kigali'), ('Africa/Kinshasa', 'Africa/Kinshasa'), ('Africa/Lagos', 'Africa/Lagos'), ('Africa/Libreville', 'Africa/Libreville'), ('Africa/Lome', 'Africa/Lome'), ('Africa/Luanda', 'Africa/Luanda'), ('Africa/Lubumbashi', 'Africa/Lubumbashi'), ('Africa/Lusaka', 'Africa/Lusaka'), ('Africa/Malabo', 'Africa/Malabo'), ('Africa/Maputo', 'Africa/Maputo'), ('Africa/Maseru', 'Africa/Maseru'), ('Africa/Mbabane', 'Africa/Mbabane'), ('Africa/Mogadishu', 'Africa/Mogadishu'), ('Africa/Monrovia', 'Africa/Monrovia'), ('Africa/Nairobi', 'Africa/Nairobi'), ('Africa/Ndjamena', 'Africa/Ndjamena'), ('Africa/Niamey', 'Africa/Niamey'), ('Africa/Nouakchott', 'Africa/Nouakchott'), ('Africa/Ouagadougou', 'Africa/Ouagadougou'), ('Africa/Porto-Novo', 'Africa/Porto-Novo'), ('Africa/Sao_Tome', 'Africa/Sao_Tome'), ('Africa/Tripoli', 'Africa/Tripoli'), ('Africa/Tunis', 'Africa/Tunis'), ('Africa/Windhoek', 'Africa/Windhoek'), ('America/Adak', 'America/Adak'), ('America/Anchorage', 'America/Anchorage'), ('America/Anguilla', 'America/Anguilla'), ('America/Antigua', 'America/Antigua'), ('America/Araguaina', 'America/Araguaina'), ('America/Argentina/Buenos_Aires', 'America/Argentina/Buenos_Aires'), ('America/Argentina/Catamarca', 'America/Argentina/Catamarca'), ('America/Argentina/Cordoba', 'America/Argentina/Cordoba'), ('America/Argentina/Jujuy', 'America/Argentina/Jujuy'), ('America/Argentina/La_Rioja', 'America/Argentina/La_Rioja'), ('America/Argentina/Mendoza', 'America/Argentina/Mendoza'), ('America/Argentina/Rio_Gallegos', 'America/Argentina/Rio_Gallegos'), ('America/Argentina/Salta', 'America/Argentina/Salta'), ('America/Argentina/San_Juan', 'America/Argentina/San_Juan'), ('America/Argentina/San_Luis', 'America/Argentina/San_Luis'), ('America/Argentina/Tucuman', 'America/Argentina/Tucuman'), ('America/Argentina/Ushuaia', 'America/Argentina/Ushuaia'), ('America/Aruba', 'America/Aruba'), ('America/Asuncion', 'America/Asuncion'), ('America/Atikokan', 'America/Atikokan'), ('America/Bahia', 'America/Bahia'), ('America/Bahia_Banderas', 'America/Bahia_Banderas'), ('America/Barbados', 'America/Barbados'), ('America/Belem', 'America/Belem'), ('America/Belize', 'America/Belize'), ('America/Blanc-Sablon', 'America/Blanc-Sablon'), ('America/Boa_Vista', 'America/Boa_Vista'), ('America/Bogota', 'America/Bogota'), ('America/Boise', 'America/Boise'), ('America/Cambridge_Bay', 'America/Cambridge_Bay'), ('America/Campo_Grande', 'America/Campo_Grande'), ('America/Cancun', 'America/Cancun'), ('America/Caracas', 'America/Caracas'), ('America/Cayenne', 'America/Cayenne'), ('America/Cayman', 'America/Cayman'), ('America/Chicago', 'America/Chicago'), ('America/Chihuahua', 'America/Chihuahua'), ('America/Costa_Rica', 'America/Costa_Rica'), ('America/Creston', 'America/Creston'), ('America/Cuiaba', 'America/Cuiaba'), ('America/Curacao', 'America/Curacao'), ('America/Danmarkshavn', 'America/Danmarkshavn'), ('America/Dawson', 'America/Dawson'), ('America/Dawson_Creek', 'America/Dawson_Creek'), ('America/Denver', 'America/Denver'), ('America/Detroit', 'America/Detroit'), ('America/Dominica', 'America/Dominica'), ('America/Edmonton', 'America/Edmonton'), ('America/Eirunepe', 'America/Eirunepe'), ('America/El_Salvador', 'America/El_Salvador'), ('America/Fort_Nelson', 'America/Fort_Nelson'), ('America/Fortaleza', 'America/Fortaleza'), ('America/Glace_Bay', 'America/Glace_Bay'), ('America/Goose_Bay', 'America/Goose_Bay'), ('America/Grand_Turk', 'America/Grand_Turk'), ('America/Grenada', 'America/Grenada'), ('America/Guadeloupe', 'America/Guadeloupe'), ('America/Guatemala', 'America/Guatemala'), ('America/Guayaquil', 'America/Guayaquil'), ('America/Guyana', 'America/Guyana'), ('America/Halifax', 'America/Halifax'), ('America/Havana', 'America/Havana'), ('America/Hermosillo', 'America/Hermosillo'), ('America/Indiana/Indianapolis', 'America/Indiana/Indianapolis'), ('America/Indiana/Knox', 'America/Indiana/Knox'), ('America/Indiana/Marengo', 'America/Indiana/Marengo'), ('America/Indiana/Petersburg', 'America/Indiana/Petersburg'), ('America/Indiana/Tell_City', 'America/Indiana/Tell_City'), ('America/Indiana/Vevay', 'America/Indiana/Vevay'), ('America/Indiana/Vincennes', 'America/Indiana/Vincennes'), ('America/Indiana/Winamac', 'America/Indiana/Winamac'), ('America/Inuvik', 'America/Inuvik'), ('America/Iqaluit', 'America/Iqaluit'), ('America/Jamaica', 'America/Jamaica'), ('America/Juneau', 'America/Juneau'), ('America/Kentucky/Louisville', 'America/Kentucky/Louisville'), ('America/Kentucky/Monticello', 'America/Kentucky/Monticello'), ('America/La_Paz', 'America/La_Paz'), ('America/Lima', 'America/Lima'), ('America/Los_Angeles', 'America/Los_Angeles'), ('America/Maceio', 'America/Maceio'), ('America/Managua', 'America/Managua'), ('America/Manaus', 'America/Manaus'), ('America/Martinique', 'America/Martinique'), ('America/Matamoros', 'America/Matamoros'), ('America/Mazatlan', 'America/Mazatlan'), ('America/Menominee', 'America/Menominee'), ('America/Merida', 'America/Merida'), ('America/Metlakatla', 'America/Metlakatla'), ('America/Mexico_City', 'America/Mexico_City'), ('America/Miquelon', 'America/Miquelon'), ('America/Moncton', 'America/Moncton'), ('America/Monterrey', 'America/Monterrey'), ('America/Montevideo', 'America/Montevideo'), ('America/Montserrat', 'America/Montserrat'), ('America/Nassau', 'America/Nassau'), ('America/New_York', 'America/New_York'), ('America/Nipigon', 'America/Nipigon'), ('America/Nome', 'America/Nome'), ('America/Noronha', 'America/Noronha'), ('America/North_Dakota/Beulah', 'America/North_Dakota/Beulah'), ('America/North_Dakota/Center', 'America/North_Dakota/Center'), ('America/North_Dakota/New_Salem', 'America/North_Dakota/New_Salem'), ('America/Nuuk', 'America/Nuuk'), ('America/Ojinaga', 'America/Ojinaga'), ('America/Panama', 'America/Panama'), ('America/Pangnirtung', 'America/Pangnirtung'), ('America/Paramaribo', 'America/Paramaribo'), ('America/Phoenix', 'America/Phoenix'), ('America/Port-au-Prince', 'America/Port-au-Prince'), ('America/Port_of_Spain', 'America/Port_of_Spain'), ('America/Porto_Velho', 'America/Porto_Velho'), ('America/Puerto_Rico', 'America/Puerto_Rico'), ('America/Punta_Arenas', 'America/Punta_Arenas'), ('America/Rainy_River', 'America/Rainy_River'), ('America/Rankin_Inlet', 'America/Rankin_Inlet'), ('America/Recife', 'America/Recife'), ('America/Regina', 'America/Regina'), ('America/Resolute', 'America/Resolute'), ('America/Rio_Branco', 'America/Rio_Branco'), ('America/Santarem', 'America/Santarem'), ('America/Santiago', 'America/Santiago'), ('America/Santo_Domingo', 'America/Santo_Domingo'), ('America/Sao_Paulo', 'America/Sao_Paulo'), ('America/Scoresbysund', 'America/Scoresbysund'), ('America/Sitka', 'America/Sitka'), ('America/St_Johns', 'America/St_Johns'), ('America/St_Kitts', 'America/St_Kitts'), ('America/St_Lucia', 'America/St_Lucia'), ('America/St_Thomas', 'America/St_Thomas'), ('America/St_Vincent', 'America/St_Vincent'), ('America/Swift_Current', 'America/Swift_Current'), ('America/Tegucigalpa', 'America/Tegucigalpa'), ('America/Thule', 'America/Thule'), ('America/Thunder_Bay', 'America/Thunder_Bay'), ('America/Tijuana', 'America/Tijuana'), ('America/Toronto', 'America/Toronto'), ('America/Tortola', 'America/Tortola'), ('America/Vancouver', 'America/Vancouver'), ('America/Whitehorse', 'America/Whitehorse'), ('America/Winnipeg', 'America/Winnipeg'), ('America/Yakutat', 'America/Yakutat'), ('America/Yellowknife', 'America/Yellowknife'), ('Antarctica/Casey', 'Antarctica/Casey'), ('Antarctica/Davis', 'Antarctica/Davis'), ('Antarctica/DumontDUrville', 'Antarctica/DumontDUrville'), ('Antarctica/Macquarie', 'Antarctica/Macquarie'), ('Antarctica/Mawson', 'Antarctica/Mawson'), ('Antarctica/McMurdo', 'Antarctica/McMurdo'), ('Antarctica/Palmer', 'Antarctica/Palmer'), ('Antarctica/Rothera', 'Antarctica/Rothera'), ('Antarctica/Syowa', 'Antarctica/Syowa'), ('Antarctica/Troll', 'Antarctica/Troll'), ('Antarctica/Vostok', 'Antarctica/Vostok'), ('Asia/Aden', 'Asia/Aden'), ('Asia/Almaty', 'Asia/Almaty'), ('Asia/Amman', 'Asia/Amman'), ('Asia/Anadyr', 'Asia/Anadyr'), ('Asia/Aqtau', 'Asia/Aqtau'), ('Asia/Aqtobe', 'Asia/Aqtobe'), ('Asia/Ashgabat', 'Asia/Ashgabat'), ('Asia/Atyrau', 'Asia/Atyrau'), ('Asia/Baghdad', 'Asia/Baghdad'), ('Asia/Bahrain', 'Asia/Bahrain'), ('Asia/Baku', 'Asia/Baku'), ('Asia/Bangkok', 'Asia/Bangkok'), ('Asia/Barnaul', 'Asia/Barnaul'), ('Asia/Beirut', 'Asia/Beirut'), ('Asia/Bishkek', 'Asia/Bishkek'), ('Asia/Brunei', 'Asia/Brunei'), ('Asia/Chita', 'Asia/Chita'), ('Asia/Choibalsan', 'Asia/Choibalsan'), ('Asia/Colombo', 'Asia/Colombo'), ('Asia/Damascus', 'Asia/Damascus'), ('Asia/Dhaka', 'Asia/Dhaka'), ('Asia/Dili', 'Asia/Dili'), ('Asia/Dubai', 'Asia/Dubai'), ('Asia/Dushanbe', 'Asia/Dushanbe'), ('Asia/Famagusta', 'Asia/Famagusta'), ('Asia/Gaza', 'Asia/Gaza'), ('Asia/Hebron', 'Asia/Hebron'), ('Asia/Ho_Chi_Minh', 'Asia/Ho_Chi_Minh'), ('Asia/Hong_Kong', 'Asia/Hong_Kong'), ('Asia/Hovd', 'Asia/Hovd'), ('Asia/Irkutsk', 'Asia/Irkutsk'), ('Asia/Jakarta', 'Asia/Jakarta'), ('Asia/Jayapura', 'Asia/Jayapura'), ('Asia/Jerusalem', 'Asia/Jerusalem'), ('Asia/Kabul', 'Asia/Kabul'), ('Asia/Kamchatka', 'Asia/Kamchatka'), ('Asia/Karachi', 'Asia/Karachi'), ('Asia/Kathmandu', 'Asia/Kathmandu'), ('Asia/Khandyga', 'Asia/Khandyga'), ('Asia/Kolkata', 'Asia/Kolkata'), ('Asia/Krasnoyarsk', 'Asia/Krasnoyarsk'), ('Asia/Kuala_Lumpur', 'Asia/Kuala_Lumpur'), ('Asia/Kuching', 'Asia/Kuching'), ('Asia/Kuwait', 'Asia/Kuwait'), ('Asia/Macau', 'Asia/Macau'), ('Asia/Magadan', 'Asia/Magadan'), ('Asia/Makassar', 'Asia/Makassar'), ('Asia/Manila', 'Asia/Manila'), ('Asia/Muscat', 'Asia/Muscat'), ('Asia/Nicosia', 'Asia/Nicosia'), ('Asia/Novokuznetsk', 'Asia/Novokuznetsk'), ('Asia/Novosibirsk', 'Asia/Novosibirsk'), ('Asia/Omsk', 'Asia/Omsk'), ('Asia/Oral', 'Asia/Oral'), ('Asia/Phnom_Penh', 'Asia/Phnom_Penh'), ('Asia/Pontianak', 'Asia/Pontianak'), ('Asia/Pyongyang', 'Asia/Pyongyang'), ('Asia/Qatar', 'Asia/Qatar'), ('Asia/Qostanay', 'Asia/Qostanay'), ('Asia/Qyzylorda', 'Asia/Qyzylorda'), ('Asia/Riyadh', 'Asia/Riyadh'), ('Asia/Sakhalin', 'Asia/Sakhalin'), ('Asia/Samarkand', 'Asia/Samarkand'), ('Asia/Seoul', 'Asia/Seoul'), ('Asia/Shanghai', 'Asia/Shanghai'), ('Asia/Singapore', 'Asia/Singapore'), ('Asia/Srednekolymsk', 'Asia/Srednekolymsk'), ('Asia/Taipei', 'Asia/Taipei'), ('Asia/Tashkent', 'Asia/Tashkent'), ('Asia/Tbilisi', 'Asia/Tbilisi'), ('Asia/Tehran', 'Asia/Tehran'), ('Asia/Thimphu', 'Asia/Thimphu'), ('Asia/Tokyo', 'Asia/Tokyo'), ('Asia/Tomsk', 'Asia/Tomsk'), ('Asia/Ulaanbaatar', 'Asia/Ulaanbaatar'), ('Asia/Urumqi', 'Asia/Urumqi'), ('Asia/Ust-Nera', 'Asia/Ust-Nera'), ('Asia/Vientiane', 'Asia/Vientiane'), ('Asia/Vladivostok', 'Asia/Vladivostok'), ('Asia/Yakutsk', 'Asia/Yakutsk'), ('Asia/Yangon', 'Asia/Yangon'), ('Asia/Yekaterinburg', 'Asia/Yekaterinburg'), ('Asia/Yerevan', 'Asia/Yerevan'), ('Atlantic/Azores', 'Atlantic/Azores'), ('Atlantic/Bermuda', 'Atlantic/Bermuda'), ('Atlantic/Canary', 'Atlantic/Canary'), ('Atlantic/Cape_Verde', 'Atlantic/Cape_Verde'), ('Atlantic/Faroe', 'Atlantic/Faroe'), ('Atlantic/Madeira', 'Atlantic/Madeira'), ('Atlantic/Reykjavik', 'Atlantic/Reykjavik'), ('Atlantic/South_Georgia', 'Atlantic/South_Georgia'), ('Atlantic/St_Helena', 'Atlantic/St_Helena'), ('Atlantic/Stanley', 'Atlantic/Stanley'), ('Australia/Adelaide', 'Australia/Adelaide'), ('Australia/Brisbane', 'Australia/Brisbane'), ('Australia/Broken_Hill', 'Australia/Broken_Hill'), ('Australia/Currie', 'Australia/Currie'), ('Australia/Darwin', 'Australia/Darwin'), ('Australia/Eucla', 'Australia/Eucla'), ('Australia/Hobart', 'Australia/Hobart'), ('Australia/Lindeman', 'Australia/Lindeman'), ('Australia/Lord_Howe', 'Australia/Lord_Howe'), ('Australia/Melbourne', 'Australia/Melbourne'), ('Australia/Perth', 'Australia/Perth'), ('Australia/Sydney', 'Australia/Sydney'), ('Europe/Amsterdam', 'Europe/Amsterdam'), ('Europe/Andorra', 'Europe/Andorra'), ('Europe/Astrakhan', 'Europe/Astrakhan'), ('Europe/Athens', 'Europe/Athens'), ('Europe/Belgrade', 'Europe/Belgrade'), ('Europe/Berlin', 'Europe/Berlin'), ('Europe/Brussels', 'Europe/Brussels'), ('Europe/Bucharest', 'Europe/Bucharest'), ('Europe/Budapest', 'Europe/Budapest'), ('Europe/Chisinau', 'Europe/Chisinau'), ('Europe/Copenhagen', 'Europe/Copenhagen'), ('Europe/Dublin', 'Europe/Dublin'), ('Europe/Gibraltar', 'Europe/Gibraltar'), ('Europe/Helsinki', 'Europe/Helsinki'), ('Europe/Istanbul', 'Europe/Istanbul'), ('Europe/Kaliningrad', 'Europe/Kaliningrad'), ('Europe/Kiev', 'Europe/Kiev'), ('Europe/Kirov', 'Europe/Kirov'), ('Europe/Lisbon', 'Europe/Lisbon'), ('Europe/London', 'Europe/London'), ('Europe/Luxembourg', 'Europe/Luxembourg'), ('Europe/Madrid', 'Europe/Madrid'), ('Europe/Malta', 'Europe/Malta'), ('Europe/Minsk', 'Europe/Minsk'), ('Europe/Monaco', 'Europe/Monaco'), ('Europe/Moscow', 'Europe/Moscow'), ('Europe/Oslo', 'Europe/Oslo'), ('Europe/Paris', 'Europe/Paris'), ('Europe/Prague', 'Europe/Prague'), ('Europe/Riga', 'Europe/Riga'), ('Europe/Rome', 'Europe/Rome'), ('Europe/Samara', 'Europe/Samara'), ('Europe/Saratov', 'Europe/Saratov'), ('Europe/Simferopol', 'Europe/Simferopol'), ('Europe/Sofia', 'Europe/Sofia'), ('Europe/Stockholm', 'Europe/Stockholm'), ('Europe/Tallinn', 'Europe/Tallinn'), ('Europe/Tirane', 'Europe/Tirane'), ('Europe/Ulyanovsk', 'Europe/Ulyanovsk'), ('Europe/Uzhgorod', 'Europe/Uzhgorod'), ('Europe/Vaduz', 'Europe/Vaduz'), ('Europe/Vienna', 'Europe/Vienna'), ('Europe/Vilnius', 'Europe/Vilnius'), ('Europe/Volgograd', 'Europe/Volgograd'), ('Europe/Warsaw', 'Europe/Warsaw'), ('Europe/Zaporozhye', 'Europe/Zaporozhye'), ('Europe/Zurich', 'Europe/Zurich'), ('GMT', 'GMT'), ('Indian/Antananarivo', 'Indian/Antananarivo'), ('Indian/Chagos', 'Indian/Chagos'), ('Indian/Christmas', 'Indian/Christmas'), ('Indian/Cocos', 'Indian/Cocos'), ('Indian/Comoro', 'Indian/Comoro'), ('Indian/Kerguelen', 'Indian/Kerguelen'), ('Indian/Mahe', 'Indian/Mahe'), ('Indian/Maldives', 'Indian/Maldives'), ('Indian/Mauritius', 'Indian/Mauritius'), ('Indian/Mayotte', 'Indian/Mayotte'), ('Indian/Reunion', 'Indian/Reunion'), ('Pacific/Apia', 'Pacific/Apia'), ('Pacific/Auckland', 'Pacific/Auckland'), ('Pacific/Bougainville', 'Pacific/Bougainville'), ('Pacific/Chatham', 'Pacific/Chatham'), ('Pacific/Chuuk', 'Pacific/Chuuk'), ('Pacific/Easter', 'Pacific/Easter'), ('Pacific/Efate', 'Pacific/Efate'), ('Pacific/Enderbury', 'Pacific/Enderbury'), ('Pacific/Fakaofo', 'Pacific/Fakaofo'), ('Pacific/Fiji', 'Pacific/Fiji'), ('Pacific/Funafuti', 'Pacific/Funafuti'), ('Pacific/Galapagos', 'Pacific/Galapagos'), ('Pacific/Gambier', 'Pacific/Gambier'), ('Pacific/Guadalcanal', 'Pacific/Guadalcanal'), ('Pacific/Guam', 'Pacific/Guam'), ('Pacific/Honolulu', 'Pacific/Honolulu'), ('Pacific/Kiritimati', 'Pacific/Kiritimati'), ('Pacific/Kosrae', 'Pacific/Kosrae'), ('Pacific/Kwajalein', 'Pacific/Kwajalein'), ('Pacific/Majuro', 'Pacific/Majuro'), ('Pacific/Marquesas', 'Pacific/Marquesas'), ('Pacific/Midway', 'Pacific/Midway'), ('Pacific/Nauru', 'Pacific/Nauru'), ('Pacific/Niue', 'Pacific/Niue'), ('Pacific/Norfolk', 'Pacific/Norfolk'), ('Pacific/Noumea', 'Pacific/Noumea'), ('Pacific/Pago_Pago', 'Pacific/Pago_Pago'), ('Pacific/Palau', 'Pacific/Palau'), ('Pacific/Pitcairn', 'Pacific/Pitcairn'), ('Pacific/Pohnpei', 'Pacific/Pohnpei'), ('Pacific/Port_Moresby', 'Pacific/Port_Moresby'), ('Pacific/Rarotonga', 'Pacific/Rarotonga'), ('Pacific/Saipan', 'Pacific/Saipan'), ('Pacific/Tahiti', 'Pacific/Tahiti'), ('Pacific/Tarawa', 'Pacific/Tarawa'), ('Pacific/Tongatapu', 'Pacific/Tongatapu'), ('Pacific/Wake', 'Pacific/Wake'), ('Pacific/Wallis', 'Pacific/Wallis'), ('UTC', 'UTC')], max_length=255), - ), - migrations.AlterField( - model_name='schedule', - name='name', - field=models.CharField(help_text='Letters, numbers and -:_ allowed.', max_length=16, validators=[django.core.validators.RegexValidator('^[A-Za-z0-9-:_]*$')]), - ), - ] diff --git a/ietf/meeting/migrations/0029_businessconstraint.py b/ietf/meeting/migrations/0029_businessconstraint.py deleted file mode 100644 index 7e1f13ee76..0000000000 --- a/ietf/meeting/migrations/0029_businessconstraint.py +++ /dev/null @@ -1,93 +0,0 @@ -# -*- coding: utf-8 -*- -# Generated by Django 1.11.27 on 2020-05-29 02:52 -from __future__ import unicode_literals - -from django.db import migrations, models - - -def forward(apps, schema_editor): - BusinessConstraint = apps.get_model("meeting", "BusinessConstraint") - BusinessConstraint.objects.create( - slug="bof_overlapping_prg", - name="BOFs cannot conflict with PRGs", - penalty=100000, - ) - BusinessConstraint.objects.create( - slug="bof_overlapping_bof", - name="BOFs cannot conflict with any other BOFs", - penalty=100000, - ) - BusinessConstraint.objects.create( - slug="bof_overlapping_area_wg", - name="BOFs cannot conflict with any other WGs in their area", - penalty=100000, - ) - BusinessConstraint.objects.create( - slug="bof_overlapping_area_meeting", - name="BOFs cannot conflict with any area-wide meetings (of any area)", - penalty=10000, - ) - BusinessConstraint.objects.create( - slug="area_overlapping_in_area", - name="Area meetings cannot conflict with anything else in their area", - penalty=10000, - ) - BusinessConstraint.objects.create( - slug="area_overlapping_other_area", - name="Area meetings cannot conflict with other area meetings", - penalty=100000, - ) - BusinessConstraint.objects.create( - slug="session_overlap_ad", - name="WGs overseen by the same Area Director should not conflict", - penalty=100, - ) - BusinessConstraint.objects.create( - slug="sessions_out_of_order", - name="Sessions should be scheduled in requested order", - penalty=100000, - ) - BusinessConstraint.objects.create( - slug="session_requires_trim", - name="Sessions should be scheduled according to requested duration and attendees", - penalty=100000, - ) - - ConstraintName = apps.get_model("name", "ConstraintName") - ConstraintName.objects.filter(slug='conflict').update(penalty=100000) - ConstraintName.objects.filter(slug='conflic2').update(penalty=10000) - ConstraintName.objects.filter(slug='conflic3').update(penalty=100000) - ConstraintName.objects.filter(slug='bethere').update(penalty=10000) - ConstraintName.objects.filter(slug='timerange').update(penalty=1000000) - ConstraintName.objects.filter(slug='time_relation').update(penalty=1000) - ConstraintName.objects.filter(slug='wg_adjacent').update(penalty=1000) - - -def reverse(apps, schema_editor): - ConstraintName = apps.get_model("name", "ConstraintName") - ConstraintName.objects.filter(slug='conflict').update(penalty=100000) - ConstraintName.objects.filter(slug='conflic2').update(penalty=10000) - ConstraintName.objects.filter(slug='conflic3').update(penalty=1000) - ConstraintName.objects.filter(slug='bethere').update(penalty=200000) - ConstraintName.objects.filter(slug='timerange').update(penalty=100000) - ConstraintName.objects.filter(slug='time_relation').update(penalty=100000) - ConstraintName.objects.filter(slug='wg_adjacent').update(penalty=100000) - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0028_auto_20200501_0139'), - ] - - operations = [ - migrations.CreateModel( - name='BusinessConstraint', - fields=[ - ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), - ('name', models.CharField(max_length=255)), - ('penalty', models.IntegerField(default=0, help_text='The penalty for violating this kind of constraint; for instance 10 (small penalty) or 10000 (large penalty)')), - ], - ), - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/meeting/migrations/0030_allow_empty_joint_with_sessions.py b/ietf/meeting/migrations/0030_allow_empty_joint_with_sessions.py deleted file mode 100644 index d8ab7cff24..0000000000 --- a/ietf/meeting/migrations/0030_allow_empty_joint_with_sessions.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.2.14 on 2020-07-20 14:11 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0029_businessconstraint'), - ] - - operations = [ - migrations.AlterField( - model_name='session', - name='joint_with_groups', - field=models.ManyToManyField(blank=True, related_name='sessions_joint_in', to='group.Group'), - ), - ] diff --git a/ietf/meeting/migrations/0031_auto_20200803_1153.py b/ietf/meeting/migrations/0031_auto_20200803_1153.py deleted file mode 100644 index 8a5ac54aa7..0000000000 --- a/ietf/meeting/migrations/0031_auto_20200803_1153.py +++ /dev/null @@ -1,22 +0,0 @@ -# Generated by Django 2.2.14 on 2020-08-03 11:53 - -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0018_slidesubmissionstatusname'), - ('doc', '0035_populate_docextresources'), - ('meeting', '0030_allow_empty_joint_with_sessions'), - ] - - operations = [ - migrations.AddField( - model_name='slidesubmission', - name='doc', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, to='doc.Document'), - ), - ] diff --git a/ietf/meeting/migrations/0032_auto_20200824_1642.py b/ietf/meeting/migrations/0032_auto_20200824_1642.py deleted file mode 100644 index 31155da6a4..0000000000 --- a/ietf/meeting/migrations/0032_auto_20200824_1642.py +++ /dev/null @@ -1,20 +0,0 @@ -# Generated by Django 2.2.14 on 2020-08-03 11:53 - -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0031_auto_20200803_1153'), - ] - - operations = [ - migrations.AddField( - model_name='slidesubmission', - name='status', - field=ietf.utils.models.ForeignKey(null=True, default='pending', on_delete=django.db.models.deletion.SET_NULL, to='name.SlideSubmissionStatusName'), - ), - ] diff --git a/ietf/meeting/migrations/0033_session_tombstone_for.py b/ietf/meeting/migrations/0033_session_tombstone_for.py deleted file mode 100644 index dc74615cbf..0000000000 --- a/ietf/meeting/migrations/0033_session_tombstone_for.py +++ /dev/null @@ -1,19 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved - -from django.db import migrations, models -import django.db.models.deletion - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0032_auto_20200824_1642'), - ] - - operations = [ - migrations.AddField( - model_name='session', - name='tombstone_for', - field=models.ForeignKey(blank=True, help_text='This session is the tombstone for a session that was rescheduled', null=True, on_delete=django.db.models.deletion.CASCADE, to='meeting.Session'), - ), - ] diff --git a/ietf/meeting/migrations/0034_schedule_notes.py b/ietf/meeting/migrations/0034_schedule_notes.py deleted file mode 100644 index f851a6b9ff..0000000000 --- a/ietf/meeting/migrations/0034_schedule_notes.py +++ /dev/null @@ -1,28 +0,0 @@ -# Generated by Django 2.0.13 on 2020-07-01 02:45 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0033_session_tombstone_for'), - ] - - operations = [ - migrations.AddField( - model_name='schedule', - name='notes', - field=models.TextField(blank=True), - ), - migrations.AlterField( - model_name='schedule', - name='public', - field=models.BooleanField(default=True, help_text='Allow others to see this agenda.'), - ), - migrations.AlterField( - model_name='schedule', - name='visible', - field=models.BooleanField(default=True, help_text='Show in the list of possible agendas for the meeting.', verbose_name='Show in agenda list'), - ), - ] diff --git a/ietf/meeting/migrations/0035_add_session_origin.py b/ietf/meeting/migrations/0035_add_session_origin.py deleted file mode 100644 index ff782d52cd..0000000000 --- a/ietf/meeting/migrations/0035_add_session_origin.py +++ /dev/null @@ -1,20 +0,0 @@ -# Generated by Django 2.0.13 on 2020-08-04 06:22 - -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0034_schedule_notes'), - ] - - operations = [ - migrations.AddField( - model_name='schedule', - name='origin', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, to='meeting.Schedule'), - ), - ] diff --git a/ietf/meeting/migrations/0036_add_schedule_base.py b/ietf/meeting/migrations/0036_add_schedule_base.py deleted file mode 100644 index 436fa5cf01..0000000000 --- a/ietf/meeting/migrations/0036_add_schedule_base.py +++ /dev/null @@ -1,25 +0,0 @@ -# Generated by Django 2.0.13 on 2020-08-07 09:30 - -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0035_add_session_origin'), - ] - - operations = [ - migrations.AddField( - model_name='schedule', - name='base', - field=ietf.utils.models.ForeignKey(blank=True, help_text='Sessions scheduled in the base show up in this schedule.', null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='derivedschedule_set', to='meeting.Schedule'), - ), - migrations.AlterField( - model_name='schedule', - name='origin', - field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='+', to='meeting.Schedule'), - ), - ] diff --git a/ietf/meeting/migrations/0037_auto_20200908_0334.py b/ietf/meeting/migrations/0037_auto_20200908_0334.py deleted file mode 100644 index 1ecd8fc9cc..0000000000 --- a/ietf/meeting/migrations/0037_auto_20200908_0334.py +++ /dev/null @@ -1,20 +0,0 @@ -# Generated by Django 2.2.16 on 2020-09-08 03:34 - -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0036_add_schedule_base'), - ] - - operations = [ - migrations.AlterField( - model_name='schedule', - name='base', - field=ietf.utils.models.ForeignKey(blank=True, help_text='Sessions scheduled in the base schedule show up in this schedule too.', limit_choices_to={'base': None}, null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='derivedschedule_set', to='meeting.Schedule'), - ), - ] diff --git a/ietf/meeting/migrations/0038_auto_20201005_1123.py b/ietf/meeting/migrations/0038_auto_20201005_1123.py deleted file mode 100644 index faff10a0a2..0000000000 --- a/ietf/meeting/migrations/0038_auto_20201005_1123.py +++ /dev/null @@ -1,19 +0,0 @@ -# Generated by Django 2.2.16 on 2020-10-05 11:23 - -import django.core.validators -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0037_auto_20200908_0334'), - ] - - operations = [ - migrations.AlterField( - model_name='schedule', - name='name', - field=models.CharField(help_text='Letters, numbers and -:_ allowed.', max_length=64, validators=[django.core.validators.RegexValidator('^[A-Za-z0-9-:_]*$')]), - ), - ] diff --git a/ietf/meeting/migrations/0039_auto_20201109_0439.py b/ietf/meeting/migrations/0039_auto_20201109_0439.py deleted file mode 100644 index 9790a158a1..0000000000 --- a/ietf/meeting/migrations/0039_auto_20201109_0439.py +++ /dev/null @@ -1,21 +0,0 @@ -# Generated by Django 2.2.17 on 2020-11-09 04:39 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0038_auto_20201005_1123'), - ] - - operations = [ - migrations.AddIndex( - model_name='meeting', - index=models.Index(fields=['-date', '-id'], name='meeting_mee_date_40ca21_idx'), - ), - migrations.AddIndex( - model_name='timeslot', - index=models.Index(fields=['-time', '-id'], name='meeting_tim_time_b802cb_idx'), - ), - ] diff --git a/ietf/meeting/migrations/0040_auto_20210130_1027.py b/ietf/meeting/migrations/0040_auto_20210130_1027.py deleted file mode 100644 index 84794371ba..0000000000 --- a/ietf/meeting/migrations/0040_auto_20210130_1027.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.2.17 on 2021-01-30 10:27 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0039_auto_20201109_0439'), - ] - - operations = [ - migrations.AlterField( - model_name='meeting', - name='time_zone', - field=models.CharField(blank=True, choices=[('', '---------'), ('Africa/Abidjan', 'Africa/Abidjan'), ('Africa/Accra', 'Africa/Accra'), ('Africa/Addis_Ababa', 'Africa/Addis_Ababa'), ('Africa/Algiers', 'Africa/Algiers'), ('Africa/Asmara', 'Africa/Asmara'), ('Africa/Bamako', 'Africa/Bamako'), ('Africa/Bangui', 'Africa/Bangui'), ('Africa/Banjul', 'Africa/Banjul'), ('Africa/Bissau', 'Africa/Bissau'), ('Africa/Blantyre', 'Africa/Blantyre'), ('Africa/Brazzaville', 'Africa/Brazzaville'), ('Africa/Bujumbura', 'Africa/Bujumbura'), ('Africa/Cairo', 'Africa/Cairo'), ('Africa/Casablanca', 'Africa/Casablanca'), ('Africa/Ceuta', 'Africa/Ceuta'), ('Africa/Conakry', 'Africa/Conakry'), ('Africa/Dakar', 'Africa/Dakar'), ('Africa/Dar_es_Salaam', 'Africa/Dar_es_Salaam'), ('Africa/Djibouti', 'Africa/Djibouti'), ('Africa/Douala', 'Africa/Douala'), ('Africa/El_Aaiun', 'Africa/El_Aaiun'), ('Africa/Freetown', 'Africa/Freetown'), ('Africa/Gaborone', 'Africa/Gaborone'), ('Africa/Harare', 'Africa/Harare'), ('Africa/Johannesburg', 'Africa/Johannesburg'), ('Africa/Juba', 'Africa/Juba'), ('Africa/Kampala', 'Africa/Kampala'), ('Africa/Khartoum', 'Africa/Khartoum'), ('Africa/Kigali', 'Africa/Kigali'), ('Africa/Kinshasa', 'Africa/Kinshasa'), ('Africa/Lagos', 'Africa/Lagos'), ('Africa/Libreville', 'Africa/Libreville'), ('Africa/Lome', 'Africa/Lome'), ('Africa/Luanda', 'Africa/Luanda'), ('Africa/Lubumbashi', 'Africa/Lubumbashi'), ('Africa/Lusaka', 'Africa/Lusaka'), ('Africa/Malabo', 'Africa/Malabo'), ('Africa/Maputo', 'Africa/Maputo'), ('Africa/Maseru', 'Africa/Maseru'), ('Africa/Mbabane', 'Africa/Mbabane'), ('Africa/Mogadishu', 'Africa/Mogadishu'), ('Africa/Monrovia', 'Africa/Monrovia'), ('Africa/Nairobi', 'Africa/Nairobi'), ('Africa/Ndjamena', 'Africa/Ndjamena'), ('Africa/Niamey', 'Africa/Niamey'), ('Africa/Nouakchott', 'Africa/Nouakchott'), ('Africa/Ouagadougou', 'Africa/Ouagadougou'), ('Africa/Porto-Novo', 'Africa/Porto-Novo'), ('Africa/Sao_Tome', 'Africa/Sao_Tome'), ('Africa/Tripoli', 'Africa/Tripoli'), ('Africa/Tunis', 'Africa/Tunis'), ('Africa/Windhoek', 'Africa/Windhoek'), ('America/Adak', 'America/Adak'), ('America/Anchorage', 'America/Anchorage'), ('America/Anguilla', 'America/Anguilla'), ('America/Antigua', 'America/Antigua'), ('America/Araguaina', 'America/Araguaina'), ('America/Argentina/Buenos_Aires', 'America/Argentina/Buenos_Aires'), ('America/Argentina/Catamarca', 'America/Argentina/Catamarca'), ('America/Argentina/Cordoba', 'America/Argentina/Cordoba'), ('America/Argentina/Jujuy', 'America/Argentina/Jujuy'), ('America/Argentina/La_Rioja', 'America/Argentina/La_Rioja'), ('America/Argentina/Mendoza', 'America/Argentina/Mendoza'), ('America/Argentina/Rio_Gallegos', 'America/Argentina/Rio_Gallegos'), ('America/Argentina/Salta', 'America/Argentina/Salta'), ('America/Argentina/San_Juan', 'America/Argentina/San_Juan'), ('America/Argentina/San_Luis', 'America/Argentina/San_Luis'), ('America/Argentina/Tucuman', 'America/Argentina/Tucuman'), ('America/Argentina/Ushuaia', 'America/Argentina/Ushuaia'), ('America/Aruba', 'America/Aruba'), ('America/Asuncion', 'America/Asuncion'), ('America/Atikokan', 'America/Atikokan'), ('America/Bahia', 'America/Bahia'), ('America/Bahia_Banderas', 'America/Bahia_Banderas'), ('America/Barbados', 'America/Barbados'), ('America/Belem', 'America/Belem'), ('America/Belize', 'America/Belize'), ('America/Blanc-Sablon', 'America/Blanc-Sablon'), ('America/Boa_Vista', 'America/Boa_Vista'), ('America/Bogota', 'America/Bogota'), ('America/Boise', 'America/Boise'), ('America/Cambridge_Bay', 'America/Cambridge_Bay'), ('America/Campo_Grande', 'America/Campo_Grande'), ('America/Cancun', 'America/Cancun'), ('America/Caracas', 'America/Caracas'), ('America/Cayenne', 'America/Cayenne'), ('America/Cayman', 'America/Cayman'), ('America/Chicago', 'America/Chicago'), ('America/Chihuahua', 'America/Chihuahua'), ('America/Costa_Rica', 'America/Costa_Rica'), ('America/Creston', 'America/Creston'), ('America/Cuiaba', 'America/Cuiaba'), ('America/Curacao', 'America/Curacao'), ('America/Danmarkshavn', 'America/Danmarkshavn'), ('America/Dawson', 'America/Dawson'), ('America/Dawson_Creek', 'America/Dawson_Creek'), ('America/Denver', 'America/Denver'), ('America/Detroit', 'America/Detroit'), ('America/Dominica', 'America/Dominica'), ('America/Edmonton', 'America/Edmonton'), ('America/Eirunepe', 'America/Eirunepe'), ('America/El_Salvador', 'America/El_Salvador'), ('America/Fort_Nelson', 'America/Fort_Nelson'), ('America/Fortaleza', 'America/Fortaleza'), ('America/Glace_Bay', 'America/Glace_Bay'), ('America/Goose_Bay', 'America/Goose_Bay'), ('America/Grand_Turk', 'America/Grand_Turk'), ('America/Grenada', 'America/Grenada'), ('America/Guadeloupe', 'America/Guadeloupe'), ('America/Guatemala', 'America/Guatemala'), ('America/Guayaquil', 'America/Guayaquil'), ('America/Guyana', 'America/Guyana'), ('America/Halifax', 'America/Halifax'), ('America/Havana', 'America/Havana'), ('America/Hermosillo', 'America/Hermosillo'), ('America/Indiana/Indianapolis', 'America/Indiana/Indianapolis'), ('America/Indiana/Knox', 'America/Indiana/Knox'), ('America/Indiana/Marengo', 'America/Indiana/Marengo'), ('America/Indiana/Petersburg', 'America/Indiana/Petersburg'), ('America/Indiana/Tell_City', 'America/Indiana/Tell_City'), ('America/Indiana/Vevay', 'America/Indiana/Vevay'), ('America/Indiana/Vincennes', 'America/Indiana/Vincennes'), ('America/Indiana/Winamac', 'America/Indiana/Winamac'), ('America/Inuvik', 'America/Inuvik'), ('America/Iqaluit', 'America/Iqaluit'), ('America/Jamaica', 'America/Jamaica'), ('America/Juneau', 'America/Juneau'), ('America/Kentucky/Louisville', 'America/Kentucky/Louisville'), ('America/Kentucky/Monticello', 'America/Kentucky/Monticello'), ('America/La_Paz', 'America/La_Paz'), ('America/Lima', 'America/Lima'), ('America/Los_Angeles', 'America/Los_Angeles'), ('America/Maceio', 'America/Maceio'), ('America/Managua', 'America/Managua'), ('America/Manaus', 'America/Manaus'), ('America/Martinique', 'America/Martinique'), ('America/Matamoros', 'America/Matamoros'), ('America/Mazatlan', 'America/Mazatlan'), ('America/Menominee', 'America/Menominee'), ('America/Merida', 'America/Merida'), ('America/Metlakatla', 'America/Metlakatla'), ('America/Mexico_City', 'America/Mexico_City'), ('America/Miquelon', 'America/Miquelon'), ('America/Moncton', 'America/Moncton'), ('America/Monterrey', 'America/Monterrey'), ('America/Montevideo', 'America/Montevideo'), ('America/Montserrat', 'America/Montserrat'), ('America/Nassau', 'America/Nassau'), ('America/New_York', 'America/New_York'), ('America/Nipigon', 'America/Nipigon'), ('America/Nome', 'America/Nome'), ('America/Noronha', 'America/Noronha'), ('America/North_Dakota/Beulah', 'America/North_Dakota/Beulah'), ('America/North_Dakota/Center', 'America/North_Dakota/Center'), ('America/North_Dakota/New_Salem', 'America/North_Dakota/New_Salem'), ('America/Nuuk', 'America/Nuuk'), ('America/Ojinaga', 'America/Ojinaga'), ('America/Panama', 'America/Panama'), ('America/Pangnirtung', 'America/Pangnirtung'), ('America/Paramaribo', 'America/Paramaribo'), ('America/Phoenix', 'America/Phoenix'), ('America/Port-au-Prince', 'America/Port-au-Prince'), ('America/Port_of_Spain', 'America/Port_of_Spain'), ('America/Porto_Velho', 'America/Porto_Velho'), ('America/Puerto_Rico', 'America/Puerto_Rico'), ('America/Punta_Arenas', 'America/Punta_Arenas'), ('America/Rainy_River', 'America/Rainy_River'), ('America/Rankin_Inlet', 'America/Rankin_Inlet'), ('America/Recife', 'America/Recife'), ('America/Regina', 'America/Regina'), ('America/Resolute', 'America/Resolute'), ('America/Rio_Branco', 'America/Rio_Branco'), ('America/Santarem', 'America/Santarem'), ('America/Santiago', 'America/Santiago'), ('America/Santo_Domingo', 'America/Santo_Domingo'), ('America/Sao_Paulo', 'America/Sao_Paulo'), ('America/Scoresbysund', 'America/Scoresbysund'), ('America/Sitka', 'America/Sitka'), ('America/St_Johns', 'America/St_Johns'), ('America/St_Kitts', 'America/St_Kitts'), ('America/St_Lucia', 'America/St_Lucia'), ('America/St_Thomas', 'America/St_Thomas'), ('America/St_Vincent', 'America/St_Vincent'), ('America/Swift_Current', 'America/Swift_Current'), ('America/Tegucigalpa', 'America/Tegucigalpa'), ('America/Thule', 'America/Thule'), ('America/Thunder_Bay', 'America/Thunder_Bay'), ('America/Tijuana', 'America/Tijuana'), ('America/Toronto', 'America/Toronto'), ('America/Tortola', 'America/Tortola'), ('America/Vancouver', 'America/Vancouver'), ('America/Whitehorse', 'America/Whitehorse'), ('America/Winnipeg', 'America/Winnipeg'), ('America/Yakutat', 'America/Yakutat'), ('America/Yellowknife', 'America/Yellowknife'), ('Antarctica/Casey', 'Antarctica/Casey'), ('Antarctica/Davis', 'Antarctica/Davis'), ('Antarctica/DumontDUrville', 'Antarctica/DumontDUrville'), ('Antarctica/Macquarie', 'Antarctica/Macquarie'), ('Antarctica/Mawson', 'Antarctica/Mawson'), ('Antarctica/McMurdo', 'Antarctica/McMurdo'), ('Antarctica/Palmer', 'Antarctica/Palmer'), ('Antarctica/Rothera', 'Antarctica/Rothera'), ('Antarctica/Syowa', 'Antarctica/Syowa'), ('Antarctica/Troll', 'Antarctica/Troll'), ('Antarctica/Vostok', 'Antarctica/Vostok'), ('Asia/Aden', 'Asia/Aden'), ('Asia/Almaty', 'Asia/Almaty'), ('Asia/Amman', 'Asia/Amman'), ('Asia/Anadyr', 'Asia/Anadyr'), ('Asia/Aqtau', 'Asia/Aqtau'), ('Asia/Aqtobe', 'Asia/Aqtobe'), ('Asia/Ashgabat', 'Asia/Ashgabat'), ('Asia/Atyrau', 'Asia/Atyrau'), ('Asia/Baghdad', 'Asia/Baghdad'), ('Asia/Bahrain', 'Asia/Bahrain'), ('Asia/Baku', 'Asia/Baku'), ('Asia/Bangkok', 'Asia/Bangkok'), ('Asia/Barnaul', 'Asia/Barnaul'), ('Asia/Beirut', 'Asia/Beirut'), ('Asia/Bishkek', 'Asia/Bishkek'), ('Asia/Brunei', 'Asia/Brunei'), ('Asia/Chita', 'Asia/Chita'), ('Asia/Choibalsan', 'Asia/Choibalsan'), ('Asia/Colombo', 'Asia/Colombo'), ('Asia/Damascus', 'Asia/Damascus'), ('Asia/Dhaka', 'Asia/Dhaka'), ('Asia/Dili', 'Asia/Dili'), ('Asia/Dubai', 'Asia/Dubai'), ('Asia/Dushanbe', 'Asia/Dushanbe'), ('Asia/Famagusta', 'Asia/Famagusta'), ('Asia/Gaza', 'Asia/Gaza'), ('Asia/Hebron', 'Asia/Hebron'), ('Asia/Ho_Chi_Minh', 'Asia/Ho_Chi_Minh'), ('Asia/Hong_Kong', 'Asia/Hong_Kong'), ('Asia/Hovd', 'Asia/Hovd'), ('Asia/Irkutsk', 'Asia/Irkutsk'), ('Asia/Jakarta', 'Asia/Jakarta'), ('Asia/Jayapura', 'Asia/Jayapura'), ('Asia/Jerusalem', 'Asia/Jerusalem'), ('Asia/Kabul', 'Asia/Kabul'), ('Asia/Kamchatka', 'Asia/Kamchatka'), ('Asia/Karachi', 'Asia/Karachi'), ('Asia/Kathmandu', 'Asia/Kathmandu'), ('Asia/Khandyga', 'Asia/Khandyga'), ('Asia/Kolkata', 'Asia/Kolkata'), ('Asia/Krasnoyarsk', 'Asia/Krasnoyarsk'), ('Asia/Kuala_Lumpur', 'Asia/Kuala_Lumpur'), ('Asia/Kuching', 'Asia/Kuching'), ('Asia/Kuwait', 'Asia/Kuwait'), ('Asia/Macau', 'Asia/Macau'), ('Asia/Magadan', 'Asia/Magadan'), ('Asia/Makassar', 'Asia/Makassar'), ('Asia/Manila', 'Asia/Manila'), ('Asia/Muscat', 'Asia/Muscat'), ('Asia/Nicosia', 'Asia/Nicosia'), ('Asia/Novokuznetsk', 'Asia/Novokuznetsk'), ('Asia/Novosibirsk', 'Asia/Novosibirsk'), ('Asia/Omsk', 'Asia/Omsk'), ('Asia/Oral', 'Asia/Oral'), ('Asia/Phnom_Penh', 'Asia/Phnom_Penh'), ('Asia/Pontianak', 'Asia/Pontianak'), ('Asia/Pyongyang', 'Asia/Pyongyang'), ('Asia/Qatar', 'Asia/Qatar'), ('Asia/Qostanay', 'Asia/Qostanay'), ('Asia/Qyzylorda', 'Asia/Qyzylorda'), ('Asia/Riyadh', 'Asia/Riyadh'), ('Asia/Sakhalin', 'Asia/Sakhalin'), ('Asia/Samarkand', 'Asia/Samarkand'), ('Asia/Seoul', 'Asia/Seoul'), ('Asia/Shanghai', 'Asia/Shanghai'), ('Asia/Singapore', 'Asia/Singapore'), ('Asia/Srednekolymsk', 'Asia/Srednekolymsk'), ('Asia/Taipei', 'Asia/Taipei'), ('Asia/Tashkent', 'Asia/Tashkent'), ('Asia/Tbilisi', 'Asia/Tbilisi'), ('Asia/Tehran', 'Asia/Tehran'), ('Asia/Thimphu', 'Asia/Thimphu'), ('Asia/Tokyo', 'Asia/Tokyo'), ('Asia/Tomsk', 'Asia/Tomsk'), ('Asia/Ulaanbaatar', 'Asia/Ulaanbaatar'), ('Asia/Urumqi', 'Asia/Urumqi'), ('Asia/Ust-Nera', 'Asia/Ust-Nera'), ('Asia/Vientiane', 'Asia/Vientiane'), ('Asia/Vladivostok', 'Asia/Vladivostok'), ('Asia/Yakutsk', 'Asia/Yakutsk'), ('Asia/Yangon', 'Asia/Yangon'), ('Asia/Yekaterinburg', 'Asia/Yekaterinburg'), ('Asia/Yerevan', 'Asia/Yerevan'), ('Atlantic/Azores', 'Atlantic/Azores'), ('Atlantic/Bermuda', 'Atlantic/Bermuda'), ('Atlantic/Canary', 'Atlantic/Canary'), ('Atlantic/Cape_Verde', 'Atlantic/Cape_Verde'), ('Atlantic/Faroe', 'Atlantic/Faroe'), ('Atlantic/Madeira', 'Atlantic/Madeira'), ('Atlantic/Reykjavik', 'Atlantic/Reykjavik'), ('Atlantic/South_Georgia', 'Atlantic/South_Georgia'), ('Atlantic/St_Helena', 'Atlantic/St_Helena'), ('Atlantic/Stanley', 'Atlantic/Stanley'), ('Australia/Adelaide', 'Australia/Adelaide'), ('Australia/Brisbane', 'Australia/Brisbane'), ('Australia/Broken_Hill', 'Australia/Broken_Hill'), ('Australia/Darwin', 'Australia/Darwin'), ('Australia/Eucla', 'Australia/Eucla'), ('Australia/Hobart', 'Australia/Hobart'), ('Australia/Lindeman', 'Australia/Lindeman'), ('Australia/Lord_Howe', 'Australia/Lord_Howe'), ('Australia/Melbourne', 'Australia/Melbourne'), ('Australia/Perth', 'Australia/Perth'), ('Australia/Sydney', 'Australia/Sydney'), ('Europe/Amsterdam', 'Europe/Amsterdam'), ('Europe/Andorra', 'Europe/Andorra'), ('Europe/Astrakhan', 'Europe/Astrakhan'), ('Europe/Athens', 'Europe/Athens'), ('Europe/Belgrade', 'Europe/Belgrade'), ('Europe/Berlin', 'Europe/Berlin'), ('Europe/Brussels', 'Europe/Brussels'), ('Europe/Bucharest', 'Europe/Bucharest'), ('Europe/Budapest', 'Europe/Budapest'), ('Europe/Chisinau', 'Europe/Chisinau'), ('Europe/Copenhagen', 'Europe/Copenhagen'), ('Europe/Dublin', 'Europe/Dublin'), ('Europe/Gibraltar', 'Europe/Gibraltar'), ('Europe/Helsinki', 'Europe/Helsinki'), ('Europe/Istanbul', 'Europe/Istanbul'), ('Europe/Kaliningrad', 'Europe/Kaliningrad'), ('Europe/Kiev', 'Europe/Kiev'), ('Europe/Kirov', 'Europe/Kirov'), ('Europe/Lisbon', 'Europe/Lisbon'), ('Europe/London', 'Europe/London'), ('Europe/Luxembourg', 'Europe/Luxembourg'), ('Europe/Madrid', 'Europe/Madrid'), ('Europe/Malta', 'Europe/Malta'), ('Europe/Minsk', 'Europe/Minsk'), ('Europe/Monaco', 'Europe/Monaco'), ('Europe/Moscow', 'Europe/Moscow'), ('Europe/Oslo', 'Europe/Oslo'), ('Europe/Paris', 'Europe/Paris'), ('Europe/Prague', 'Europe/Prague'), ('Europe/Riga', 'Europe/Riga'), ('Europe/Rome', 'Europe/Rome'), ('Europe/Samara', 'Europe/Samara'), ('Europe/Saratov', 'Europe/Saratov'), ('Europe/Simferopol', 'Europe/Simferopol'), ('Europe/Sofia', 'Europe/Sofia'), ('Europe/Stockholm', 'Europe/Stockholm'), ('Europe/Tallinn', 'Europe/Tallinn'), ('Europe/Tirane', 'Europe/Tirane'), ('Europe/Ulyanovsk', 'Europe/Ulyanovsk'), ('Europe/Uzhgorod', 'Europe/Uzhgorod'), ('Europe/Vaduz', 'Europe/Vaduz'), ('Europe/Vienna', 'Europe/Vienna'), ('Europe/Vilnius', 'Europe/Vilnius'), ('Europe/Volgograd', 'Europe/Volgograd'), ('Europe/Warsaw', 'Europe/Warsaw'), ('Europe/Zaporozhye', 'Europe/Zaporozhye'), ('Europe/Zurich', 'Europe/Zurich'), ('GMT', 'GMT'), ('Indian/Antananarivo', 'Indian/Antananarivo'), ('Indian/Chagos', 'Indian/Chagos'), ('Indian/Christmas', 'Indian/Christmas'), ('Indian/Cocos', 'Indian/Cocos'), ('Indian/Comoro', 'Indian/Comoro'), ('Indian/Kerguelen', 'Indian/Kerguelen'), ('Indian/Mahe', 'Indian/Mahe'), ('Indian/Maldives', 'Indian/Maldives'), ('Indian/Mauritius', 'Indian/Mauritius'), ('Indian/Mayotte', 'Indian/Mayotte'), ('Indian/Reunion', 'Indian/Reunion'), ('Pacific/Apia', 'Pacific/Apia'), ('Pacific/Auckland', 'Pacific/Auckland'), ('Pacific/Bougainville', 'Pacific/Bougainville'), ('Pacific/Chatham', 'Pacific/Chatham'), ('Pacific/Chuuk', 'Pacific/Chuuk'), ('Pacific/Easter', 'Pacific/Easter'), ('Pacific/Efate', 'Pacific/Efate'), ('Pacific/Enderbury', 'Pacific/Enderbury'), ('Pacific/Fakaofo', 'Pacific/Fakaofo'), ('Pacific/Fiji', 'Pacific/Fiji'), ('Pacific/Funafuti', 'Pacific/Funafuti'), ('Pacific/Galapagos', 'Pacific/Galapagos'), ('Pacific/Gambier', 'Pacific/Gambier'), ('Pacific/Guadalcanal', 'Pacific/Guadalcanal'), ('Pacific/Guam', 'Pacific/Guam'), ('Pacific/Honolulu', 'Pacific/Honolulu'), ('Pacific/Kiritimati', 'Pacific/Kiritimati'), ('Pacific/Kosrae', 'Pacific/Kosrae'), ('Pacific/Kwajalein', 'Pacific/Kwajalein'), ('Pacific/Majuro', 'Pacific/Majuro'), ('Pacific/Marquesas', 'Pacific/Marquesas'), ('Pacific/Midway', 'Pacific/Midway'), ('Pacific/Nauru', 'Pacific/Nauru'), ('Pacific/Niue', 'Pacific/Niue'), ('Pacific/Norfolk', 'Pacific/Norfolk'), ('Pacific/Noumea', 'Pacific/Noumea'), ('Pacific/Pago_Pago', 'Pacific/Pago_Pago'), ('Pacific/Palau', 'Pacific/Palau'), ('Pacific/Pitcairn', 'Pacific/Pitcairn'), ('Pacific/Pohnpei', 'Pacific/Pohnpei'), ('Pacific/Port_Moresby', 'Pacific/Port_Moresby'), ('Pacific/Rarotonga', 'Pacific/Rarotonga'), ('Pacific/Saipan', 'Pacific/Saipan'), ('Pacific/Tahiti', 'Pacific/Tahiti'), ('Pacific/Tarawa', 'Pacific/Tarawa'), ('Pacific/Tongatapu', 'Pacific/Tongatapu'), ('Pacific/Wake', 'Pacific/Wake'), ('Pacific/Wallis', 'Pacific/Wallis'), ('UTC', 'UTC')], max_length=255), - ), - ] diff --git a/ietf/meeting/migrations/0041_assign_correct_constraintnames.py b/ietf/meeting/migrations/0041_assign_correct_constraintnames.py deleted file mode 100644 index a5f5cf1301..0000000000 --- a/ietf/meeting/migrations/0041_assign_correct_constraintnames.py +++ /dev/null @@ -1,47 +0,0 @@ -# Generated by Django 2.2.20 on 2021-05-13 07:51 - -from django.db import migrations - - -replacement_slugs = ( - ('conflict', 'chair_conflict'), - ('conflic2', 'tech_overlap'), - ('conflic3', 'key_participant'), -) - - -def affected(constraint_qs): - """Filter constraints, keeping only those to be updated""" - # The constraints were renamed in the UI in commit 16699 on 2019-09-03. - # This was between meetings 105 and 106. Assuming this migration and - # the new conflict types are in place before meeting 111, these - # are the meetings for which the UI disagreed with the constraint - # type actually created. - affected_meetings = ['106', '107', '108', '109', '110'] - return constraint_qs.filter(meeting__number__in=affected_meetings) - - -def forward(apps, schema_editor): - Constraint = apps.get_model('meeting', 'Constraint') - affected_constraints = affected(Constraint.objects.all()) - for old, new in replacement_slugs: - affected_constraints.filter(name_id=old).update(name_id=new) - - -def reverse(apps, schema_editor): - Constraint = apps.get_model('meeting', 'Constraint') - affected_constraints = affected(Constraint.objects.all()) - for old, new in replacement_slugs: - affected_constraints.filter(name_id=new).update(name_id=old) - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0040_auto_20210130_1027'), - ('name', '0026_add_conflict_constraintnames'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/meeting/migrations/0042_meeting_group_conflict_types.py b/ietf/meeting/migrations/0042_meeting_group_conflict_types.py deleted file mode 100644 index 2d3320f456..0000000000 --- a/ietf/meeting/migrations/0042_meeting_group_conflict_types.py +++ /dev/null @@ -1,19 +0,0 @@ -# Generated by Django 2.2.20 on 2021-05-20 12:28 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0026_add_conflict_constraintnames'), - ('meeting', '0041_assign_correct_constraintnames'), - ] - - operations = [ - migrations.AddField( - model_name='meeting', - name='group_conflict_types', - field=models.ManyToManyField(blank=True, limit_choices_to={'is_group_conflict': True}, help_text='Types of scheduling conflict between groups to consider', to='name.ConstraintName'), - ), - ] diff --git a/ietf/meeting/migrations/0043_populate_meeting_group_conflict_types.py b/ietf/meeting/migrations/0043_populate_meeting_group_conflict_types.py deleted file mode 100644 index 8092635641..0000000000 --- a/ietf/meeting/migrations/0043_populate_meeting_group_conflict_types.py +++ /dev/null @@ -1,42 +0,0 @@ -# Generated by Django 2.2.20 on 2021-05-20 12:30 - -from django.db import migrations -from django.db.models import IntegerField -from django.db.models.functions import Cast - - -def forward(apps, schema_editor): - Meeting = apps.get_model('meeting', 'Meeting') - ConstraintName = apps.get_model('name', 'ConstraintName') - - # old for pre-106 - old_constraints = ConstraintName.objects.filter(slug__in=['conflict', 'conflic2', 'conflic3']) - new_constraints = ConstraintName.objects.filter(slug__in=['chair_conflict', 'tech_overlap', 'key_participant']) - - # get meetings with numeric 'number' field to avoid lexicographic ordering - ietf_meetings = Meeting.objects.filter( - type='ietf' - ).annotate( - number_as_int=Cast('number', output_field=IntegerField()) - ) - - for mtg in ietf_meetings.filter(number_as_int__lt=106): - for cn in old_constraints: - mtg.group_conflict_types.add(cn) - for mtg in ietf_meetings.filter(number_as_int__gte=106): - for cn in new_constraints: - mtg.group_conflict_types.add(cn) - -def reverse(apps, schema_editor): - pass - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0042_meeting_group_conflict_types'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/meeting/migrations/0044_again_assign_correct_constraintnames.py b/ietf/meeting/migrations/0044_again_assign_correct_constraintnames.py deleted file mode 100644 index ed2c07a3b8..0000000000 --- a/ietf/meeting/migrations/0044_again_assign_correct_constraintnames.py +++ /dev/null @@ -1,47 +0,0 @@ -# Generated by Django 2.2.20 on 2021-05-13 07:51 - -from django.db import migrations - - -replacement_slugs = ( - ('conflict', 'chair_conflict'), - ('conflic2', 'tech_overlap'), - ('conflic3', 'key_participant'), -) - - -def affected(constraint_qs): - """Filter constraints, keeping only those to be updated""" - # The constraints were renamed in the UI in commit 16699 on 2019-09-03. - # This was between meetings 105 and 106. Assuming this migration and - # the new conflict types are in place before meeting 111, these - # are the meetings for which the UI disagreed with the constraint - # type actually created. - affected_meetings = ['111'] - return constraint_qs.filter(meeting__number__in=affected_meetings) - - -def forward(apps, schema_editor): - Constraint = apps.get_model('meeting', 'Constraint') - affected_constraints = affected(Constraint.objects.all()) - for old, new in replacement_slugs: - affected_constraints.filter(name_id=old).update(name_id=new) - - -def reverse(apps, schema_editor): - Constraint = apps.get_model('meeting', 'Constraint') - affected_constraints = affected(Constraint.objects.all()) - for old, new in replacement_slugs: - affected_constraints.filter(name_id=new).update(name_id=old) - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0043_populate_meeting_group_conflict_types'), - ('name', '0027_add_bofrequest'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/meeting/migrations/0045_proceedingsmaterial.py b/ietf/meeting/migrations/0045_proceedingsmaterial.py deleted file mode 100644 index 1ee553f4cd..0000000000 --- a/ietf/meeting/migrations/0045_proceedingsmaterial.py +++ /dev/null @@ -1,30 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -# Generated by Django 2.2.24 on 2021-07-26 17:09 - -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0029_proceedingsmaterialtypename'), - ('meeting', '0044_again_assign_correct_constraintnames'), - ] - - operations = [ - migrations.CreateModel( - name='ProceedingsMaterial', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('meeting', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='proceedings_materials', to='meeting.Meeting')), - ('type', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ProceedingsMaterialTypeName')), - ('document', ietf.utils.models.ForeignKey(limit_choices_to={'type_id': 'procmaterials'}, on_delete=django.db.models.deletion.CASCADE, to='doc.Document', unique=True)), - ], - options={ - 'unique_together': {('meeting', 'type')}, - }, - ), - ] diff --git a/ietf/meeting/migrations/0046_meetinghost.py b/ietf/meeting/migrations/0046_meetinghost.py deleted file mode 100644 index 02392913b3..0000000000 --- a/ietf/meeting/migrations/0046_meetinghost.py +++ /dev/null @@ -1,34 +0,0 @@ -# Generated by Django 2.2.24 on 2021-08-26 10:18 - -from django.db import migrations, models -import django.db.models.deletion -import ietf.meeting.models -import ietf.utils.fields -import ietf.utils.models -import ietf.utils.storage -import ietf.utils.validators - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0045_proceedingsmaterial'), - ] - - operations = [ - migrations.CreateModel( - name='MeetingHost', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('name', models.CharField(max_length=255)), - ('logo', ietf.utils.fields.MissingOkImageField(height_field='logo_height', storage=ietf.utils.storage.NoLocationMigrationFileSystemStorage(location=None), upload_to=ietf.meeting.models._host_upload_path, validators=[ietf.utils.validators.MaxImageSizeValidator(1600, 1600), ietf.utils.validators.WrappedValidator(ietf.utils.validators.validate_file_size, True), ietf.utils.validators.WrappedValidator(ietf.utils.validators.validate_file_extension, ['.png', '.jpg', '.jpeg']), ietf.utils.validators.WrappedValidator(ietf.utils.validators.validate_mime_type, ['image/jpeg', 'image/png'], True)], width_field='logo_width')), - ('logo_width', models.PositiveIntegerField(null=True)), - ('logo_height', models.PositiveIntegerField(null=True)), - ('meeting', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='meetinghosts', to='meeting.Meeting')), - ], - options={ - 'ordering': ('pk',), - 'unique_together': {('meeting', 'name')}, - }, - ), - ] diff --git a/ietf/meeting/migrations/0047_auto_20210906_0702.py b/ietf/meeting/migrations/0047_auto_20210906_0702.py deleted file mode 100644 index d44b64b9da..0000000000 --- a/ietf/meeting/migrations/0047_auto_20210906_0702.py +++ /dev/null @@ -1,22 +0,0 @@ -# Generated by Django 2.2.24 on 2021-09-06 07:02 - -from django.db import migrations -import ietf.meeting.models -import ietf.utils.fields -import ietf.utils.storage -import ietf.utils.validators - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0046_meetinghost'), - ] - - operations = [ - migrations.AlterField( - model_name='meetinghost', - name='logo', - field=ietf.utils.fields.MissingOkImageField(height_field='logo_height', storage=ietf.utils.storage.NoLocationMigrationFileSystemStorage(location=None), upload_to=ietf.meeting.models._host_upload_path, validators=[ietf.utils.validators.MaxImageSizeValidator(400, 400), ietf.utils.validators.WrappedValidator(ietf.utils.validators.validate_file_size, True), ietf.utils.validators.WrappedValidator(ietf.utils.validators.validate_file_extension, ['.png', '.jpg', '.jpeg']), ietf.utils.validators.WrappedValidator(ietf.utils.validators.validate_mime_type, ['image/jpeg', 'image/png'], True)], width_field='logo_width'), - ), - ] diff --git a/ietf/meeting/migrations/0048_auto_20211008_0907.py b/ietf/meeting/migrations/0048_auto_20211008_0907.py deleted file mode 100644 index 5f8f844360..0000000000 --- a/ietf/meeting/migrations/0048_auto_20211008_0907.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.2.24 on 2021-10-08 09:07 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0047_auto_20210906_0702'), - ] - - operations = [ - migrations.AlterField( - model_name='meeting', - name='time_zone', - field=models.CharField(blank=True, choices=[('', '---------'), ('Africa/Abidjan', 'Africa/Abidjan'), ('Africa/Accra', 'Africa/Accra'), ('Africa/Addis_Ababa', 'Africa/Addis_Ababa'), ('Africa/Algiers', 'Africa/Algiers'), ('Africa/Asmara', 'Africa/Asmara'), ('Africa/Bamako', 'Africa/Bamako'), ('Africa/Bangui', 'Africa/Bangui'), ('Africa/Banjul', 'Africa/Banjul'), ('Africa/Bissau', 'Africa/Bissau'), ('Africa/Blantyre', 'Africa/Blantyre'), ('Africa/Brazzaville', 'Africa/Brazzaville'), ('Africa/Bujumbura', 'Africa/Bujumbura'), ('Africa/Cairo', 'Africa/Cairo'), ('Africa/Casablanca', 'Africa/Casablanca'), ('Africa/Ceuta', 'Africa/Ceuta'), ('Africa/Conakry', 'Africa/Conakry'), ('Africa/Dakar', 'Africa/Dakar'), ('Africa/Dar_es_Salaam', 'Africa/Dar_es_Salaam'), ('Africa/Djibouti', 'Africa/Djibouti'), ('Africa/Douala', 'Africa/Douala'), ('Africa/El_Aaiun', 'Africa/El_Aaiun'), ('Africa/Freetown', 'Africa/Freetown'), ('Africa/Gaborone', 'Africa/Gaborone'), ('Africa/Harare', 'Africa/Harare'), ('Africa/Johannesburg', 'Africa/Johannesburg'), ('Africa/Juba', 'Africa/Juba'), ('Africa/Kampala', 'Africa/Kampala'), ('Africa/Khartoum', 'Africa/Khartoum'), ('Africa/Kigali', 'Africa/Kigali'), ('Africa/Kinshasa', 'Africa/Kinshasa'), ('Africa/Lagos', 'Africa/Lagos'), ('Africa/Libreville', 'Africa/Libreville'), ('Africa/Lome', 'Africa/Lome'), ('Africa/Luanda', 'Africa/Luanda'), ('Africa/Lubumbashi', 'Africa/Lubumbashi'), ('Africa/Lusaka', 'Africa/Lusaka'), ('Africa/Malabo', 'Africa/Malabo'), ('Africa/Maputo', 'Africa/Maputo'), ('Africa/Maseru', 'Africa/Maseru'), ('Africa/Mbabane', 'Africa/Mbabane'), ('Africa/Mogadishu', 'Africa/Mogadishu'), ('Africa/Monrovia', 'Africa/Monrovia'), ('Africa/Nairobi', 'Africa/Nairobi'), ('Africa/Ndjamena', 'Africa/Ndjamena'), ('Africa/Niamey', 'Africa/Niamey'), ('Africa/Nouakchott', 'Africa/Nouakchott'), ('Africa/Ouagadougou', 'Africa/Ouagadougou'), ('Africa/Porto-Novo', 'Africa/Porto-Novo'), ('Africa/Sao_Tome', 'Africa/Sao_Tome'), ('Africa/Tripoli', 'Africa/Tripoli'), ('Africa/Tunis', 'Africa/Tunis'), ('Africa/Windhoek', 'Africa/Windhoek'), ('America/Adak', 'America/Adak'), ('America/Anchorage', 'America/Anchorage'), ('America/Anguilla', 'America/Anguilla'), ('America/Antigua', 'America/Antigua'), ('America/Araguaina', 'America/Araguaina'), ('America/Argentina/Buenos_Aires', 'America/Argentina/Buenos_Aires'), ('America/Argentina/Catamarca', 'America/Argentina/Catamarca'), ('America/Argentina/Cordoba', 'America/Argentina/Cordoba'), ('America/Argentina/Jujuy', 'America/Argentina/Jujuy'), ('America/Argentina/La_Rioja', 'America/Argentina/La_Rioja'), ('America/Argentina/Mendoza', 'America/Argentina/Mendoza'), ('America/Argentina/Rio_Gallegos', 'America/Argentina/Rio_Gallegos'), ('America/Argentina/Salta', 'America/Argentina/Salta'), ('America/Argentina/San_Juan', 'America/Argentina/San_Juan'), ('America/Argentina/San_Luis', 'America/Argentina/San_Luis'), ('America/Argentina/Tucuman', 'America/Argentina/Tucuman'), ('America/Argentina/Ushuaia', 'America/Argentina/Ushuaia'), ('America/Aruba', 'America/Aruba'), ('America/Asuncion', 'America/Asuncion'), ('America/Atikokan', 'America/Atikokan'), ('America/Bahia', 'America/Bahia'), ('America/Bahia_Banderas', 'America/Bahia_Banderas'), ('America/Barbados', 'America/Barbados'), ('America/Belem', 'America/Belem'), ('America/Belize', 'America/Belize'), ('America/Blanc-Sablon', 'America/Blanc-Sablon'), ('America/Boa_Vista', 'America/Boa_Vista'), ('America/Bogota', 'America/Bogota'), ('America/Boise', 'America/Boise'), ('America/Cambridge_Bay', 'America/Cambridge_Bay'), ('America/Campo_Grande', 'America/Campo_Grande'), ('America/Cancun', 'America/Cancun'), ('America/Caracas', 'America/Caracas'), ('America/Cayenne', 'America/Cayenne'), ('America/Cayman', 'America/Cayman'), ('America/Chicago', 'America/Chicago'), ('America/Chihuahua', 'America/Chihuahua'), ('America/Costa_Rica', 'America/Costa_Rica'), ('America/Creston', 'America/Creston'), ('America/Cuiaba', 'America/Cuiaba'), ('America/Curacao', 'America/Curacao'), ('America/Danmarkshavn', 'America/Danmarkshavn'), ('America/Dawson', 'America/Dawson'), ('America/Dawson_Creek', 'America/Dawson_Creek'), ('America/Denver', 'America/Denver'), ('America/Detroit', 'America/Detroit'), ('America/Dominica', 'America/Dominica'), ('America/Edmonton', 'America/Edmonton'), ('America/Eirunepe', 'America/Eirunepe'), ('America/El_Salvador', 'America/El_Salvador'), ('America/Fort_Nelson', 'America/Fort_Nelson'), ('America/Fortaleza', 'America/Fortaleza'), ('America/Glace_Bay', 'America/Glace_Bay'), ('America/Goose_Bay', 'America/Goose_Bay'), ('America/Grand_Turk', 'America/Grand_Turk'), ('America/Grenada', 'America/Grenada'), ('America/Guadeloupe', 'America/Guadeloupe'), ('America/Guatemala', 'America/Guatemala'), ('America/Guayaquil', 'America/Guayaquil'), ('America/Guyana', 'America/Guyana'), ('America/Halifax', 'America/Halifax'), ('America/Havana', 'America/Havana'), ('America/Hermosillo', 'America/Hermosillo'), ('America/Indiana/Indianapolis', 'America/Indiana/Indianapolis'), ('America/Indiana/Knox', 'America/Indiana/Knox'), ('America/Indiana/Marengo', 'America/Indiana/Marengo'), ('America/Indiana/Petersburg', 'America/Indiana/Petersburg'), ('America/Indiana/Tell_City', 'America/Indiana/Tell_City'), ('America/Indiana/Vevay', 'America/Indiana/Vevay'), ('America/Indiana/Vincennes', 'America/Indiana/Vincennes'), ('America/Indiana/Winamac', 'America/Indiana/Winamac'), ('America/Inuvik', 'America/Inuvik'), ('America/Iqaluit', 'America/Iqaluit'), ('America/Jamaica', 'America/Jamaica'), ('America/Juneau', 'America/Juneau'), ('America/Kentucky/Louisville', 'America/Kentucky/Louisville'), ('America/Kentucky/Monticello', 'America/Kentucky/Monticello'), ('America/La_Paz', 'America/La_Paz'), ('America/Lima', 'America/Lima'), ('America/Los_Angeles', 'America/Los_Angeles'), ('America/Maceio', 'America/Maceio'), ('America/Managua', 'America/Managua'), ('America/Manaus', 'America/Manaus'), ('America/Martinique', 'America/Martinique'), ('America/Matamoros', 'America/Matamoros'), ('America/Mazatlan', 'America/Mazatlan'), ('America/Menominee', 'America/Menominee'), ('America/Merida', 'America/Merida'), ('America/Metlakatla', 'America/Metlakatla'), ('America/Mexico_City', 'America/Mexico_City'), ('America/Miquelon', 'America/Miquelon'), ('America/Moncton', 'America/Moncton'), ('America/Monterrey', 'America/Monterrey'), ('America/Montevideo', 'America/Montevideo'), ('America/Montserrat', 'America/Montserrat'), ('America/Nassau', 'America/Nassau'), ('America/New_York', 'America/New_York'), ('America/Nipigon', 'America/Nipigon'), ('America/Nome', 'America/Nome'), ('America/Noronha', 'America/Noronha'), ('America/North_Dakota/Beulah', 'America/North_Dakota/Beulah'), ('America/North_Dakota/Center', 'America/North_Dakota/Center'), ('America/North_Dakota/New_Salem', 'America/North_Dakota/New_Salem'), ('America/Nuuk', 'America/Nuuk'), ('America/Ojinaga', 'America/Ojinaga'), ('America/Panama', 'America/Panama'), ('America/Pangnirtung', 'America/Pangnirtung'), ('America/Paramaribo', 'America/Paramaribo'), ('America/Phoenix', 'America/Phoenix'), ('America/Port-au-Prince', 'America/Port-au-Prince'), ('America/Port_of_Spain', 'America/Port_of_Spain'), ('America/Porto_Velho', 'America/Porto_Velho'), ('America/Puerto_Rico', 'America/Puerto_Rico'), ('America/Punta_Arenas', 'America/Punta_Arenas'), ('America/Rainy_River', 'America/Rainy_River'), ('America/Rankin_Inlet', 'America/Rankin_Inlet'), ('America/Recife', 'America/Recife'), ('America/Regina', 'America/Regina'), ('America/Resolute', 'America/Resolute'), ('America/Rio_Branco', 'America/Rio_Branco'), ('America/Santarem', 'America/Santarem'), ('America/Santiago', 'America/Santiago'), ('America/Santo_Domingo', 'America/Santo_Domingo'), ('America/Sao_Paulo', 'America/Sao_Paulo'), ('America/Scoresbysund', 'America/Scoresbysund'), ('America/Sitka', 'America/Sitka'), ('America/St_Johns', 'America/St_Johns'), ('America/St_Kitts', 'America/St_Kitts'), ('America/St_Lucia', 'America/St_Lucia'), ('America/St_Thomas', 'America/St_Thomas'), ('America/St_Vincent', 'America/St_Vincent'), ('America/Swift_Current', 'America/Swift_Current'), ('America/Tegucigalpa', 'America/Tegucigalpa'), ('America/Thule', 'America/Thule'), ('America/Thunder_Bay', 'America/Thunder_Bay'), ('America/Tijuana', 'America/Tijuana'), ('America/Toronto', 'America/Toronto'), ('America/Tortola', 'America/Tortola'), ('America/Vancouver', 'America/Vancouver'), ('America/Whitehorse', 'America/Whitehorse'), ('America/Winnipeg', 'America/Winnipeg'), ('America/Yakutat', 'America/Yakutat'), ('America/Yellowknife', 'America/Yellowknife'), ('Antarctica/Casey', 'Antarctica/Casey'), ('Antarctica/Davis', 'Antarctica/Davis'), ('Antarctica/DumontDUrville', 'Antarctica/DumontDUrville'), ('Antarctica/Macquarie', 'Antarctica/Macquarie'), ('Antarctica/Mawson', 'Antarctica/Mawson'), ('Antarctica/McMurdo', 'Antarctica/McMurdo'), ('Antarctica/Palmer', 'Antarctica/Palmer'), ('Antarctica/Rothera', 'Antarctica/Rothera'), ('Antarctica/Syowa', 'Antarctica/Syowa'), ('Antarctica/Troll', 'Antarctica/Troll'), ('Antarctica/Vostok', 'Antarctica/Vostok'), ('Asia/Aden', 'Asia/Aden'), ('Asia/Almaty', 'Asia/Almaty'), ('Asia/Amman', 'Asia/Amman'), ('Asia/Anadyr', 'Asia/Anadyr'), ('Asia/Aqtau', 'Asia/Aqtau'), ('Asia/Aqtobe', 'Asia/Aqtobe'), ('Asia/Ashgabat', 'Asia/Ashgabat'), ('Asia/Atyrau', 'Asia/Atyrau'), ('Asia/Baghdad', 'Asia/Baghdad'), ('Asia/Bahrain', 'Asia/Bahrain'), ('Asia/Baku', 'Asia/Baku'), ('Asia/Bangkok', 'Asia/Bangkok'), ('Asia/Barnaul', 'Asia/Barnaul'), ('Asia/Beirut', 'Asia/Beirut'), ('Asia/Bishkek', 'Asia/Bishkek'), ('Asia/Brunei', 'Asia/Brunei'), ('Asia/Chita', 'Asia/Chita'), ('Asia/Choibalsan', 'Asia/Choibalsan'), ('Asia/Colombo', 'Asia/Colombo'), ('Asia/Damascus', 'Asia/Damascus'), ('Asia/Dhaka', 'Asia/Dhaka'), ('Asia/Dili', 'Asia/Dili'), ('Asia/Dubai', 'Asia/Dubai'), ('Asia/Dushanbe', 'Asia/Dushanbe'), ('Asia/Famagusta', 'Asia/Famagusta'), ('Asia/Gaza', 'Asia/Gaza'), ('Asia/Hebron', 'Asia/Hebron'), ('Asia/Ho_Chi_Minh', 'Asia/Ho_Chi_Minh'), ('Asia/Hong_Kong', 'Asia/Hong_Kong'), ('Asia/Hovd', 'Asia/Hovd'), ('Asia/Irkutsk', 'Asia/Irkutsk'), ('Asia/Jakarta', 'Asia/Jakarta'), ('Asia/Jayapura', 'Asia/Jayapura'), ('Asia/Jerusalem', 'Asia/Jerusalem'), ('Asia/Kabul', 'Asia/Kabul'), ('Asia/Kamchatka', 'Asia/Kamchatka'), ('Asia/Karachi', 'Asia/Karachi'), ('Asia/Kathmandu', 'Asia/Kathmandu'), ('Asia/Khandyga', 'Asia/Khandyga'), ('Asia/Kolkata', 'Asia/Kolkata'), ('Asia/Krasnoyarsk', 'Asia/Krasnoyarsk'), ('Asia/Kuala_Lumpur', 'Asia/Kuala_Lumpur'), ('Asia/Kuching', 'Asia/Kuching'), ('Asia/Kuwait', 'Asia/Kuwait'), ('Asia/Macau', 'Asia/Macau'), ('Asia/Magadan', 'Asia/Magadan'), ('Asia/Makassar', 'Asia/Makassar'), ('Asia/Manila', 'Asia/Manila'), ('Asia/Muscat', 'Asia/Muscat'), ('Asia/Nicosia', 'Asia/Nicosia'), ('Asia/Novokuznetsk', 'Asia/Novokuznetsk'), ('Asia/Novosibirsk', 'Asia/Novosibirsk'), ('Asia/Omsk', 'Asia/Omsk'), ('Asia/Oral', 'Asia/Oral'), ('Asia/Phnom_Penh', 'Asia/Phnom_Penh'), ('Asia/Pontianak', 'Asia/Pontianak'), ('Asia/Pyongyang', 'Asia/Pyongyang'), ('Asia/Qatar', 'Asia/Qatar'), ('Asia/Qostanay', 'Asia/Qostanay'), ('Asia/Qyzylorda', 'Asia/Qyzylorda'), ('Asia/Riyadh', 'Asia/Riyadh'), ('Asia/Sakhalin', 'Asia/Sakhalin'), ('Asia/Samarkand', 'Asia/Samarkand'), ('Asia/Seoul', 'Asia/Seoul'), ('Asia/Shanghai', 'Asia/Shanghai'), ('Asia/Singapore', 'Asia/Singapore'), ('Asia/Srednekolymsk', 'Asia/Srednekolymsk'), ('Asia/Taipei', 'Asia/Taipei'), ('Asia/Tashkent', 'Asia/Tashkent'), ('Asia/Tbilisi', 'Asia/Tbilisi'), ('Asia/Tehran', 'Asia/Tehran'), ('Asia/Thimphu', 'Asia/Thimphu'), ('Asia/Tokyo', 'Asia/Tokyo'), ('Asia/Tomsk', 'Asia/Tomsk'), ('Asia/Ulaanbaatar', 'Asia/Ulaanbaatar'), ('Asia/Urumqi', 'Asia/Urumqi'), ('Asia/Ust-Nera', 'Asia/Ust-Nera'), ('Asia/Vientiane', 'Asia/Vientiane'), ('Asia/Vladivostok', 'Asia/Vladivostok'), ('Asia/Yakutsk', 'Asia/Yakutsk'), ('Asia/Yangon', 'Asia/Yangon'), ('Asia/Yekaterinburg', 'Asia/Yekaterinburg'), ('Asia/Yerevan', 'Asia/Yerevan'), ('Atlantic/Azores', 'Atlantic/Azores'), ('Atlantic/Bermuda', 'Atlantic/Bermuda'), ('Atlantic/Canary', 'Atlantic/Canary'), ('Atlantic/Cape_Verde', 'Atlantic/Cape_Verde'), ('Atlantic/Faroe', 'Atlantic/Faroe'), ('Atlantic/Madeira', 'Atlantic/Madeira'), ('Atlantic/Reykjavik', 'Atlantic/Reykjavik'), ('Atlantic/South_Georgia', 'Atlantic/South_Georgia'), ('Atlantic/St_Helena', 'Atlantic/St_Helena'), ('Atlantic/Stanley', 'Atlantic/Stanley'), ('Australia/Adelaide', 'Australia/Adelaide'), ('Australia/Brisbane', 'Australia/Brisbane'), ('Australia/Broken_Hill', 'Australia/Broken_Hill'), ('Australia/Darwin', 'Australia/Darwin'), ('Australia/Eucla', 'Australia/Eucla'), ('Australia/Hobart', 'Australia/Hobart'), ('Australia/Lindeman', 'Australia/Lindeman'), ('Australia/Lord_Howe', 'Australia/Lord_Howe'), ('Australia/Melbourne', 'Australia/Melbourne'), ('Australia/Perth', 'Australia/Perth'), ('Australia/Sydney', 'Australia/Sydney'), ('Europe/Amsterdam', 'Europe/Amsterdam'), ('Europe/Andorra', 'Europe/Andorra'), ('Europe/Astrakhan', 'Europe/Astrakhan'), ('Europe/Athens', 'Europe/Athens'), ('Europe/Belgrade', 'Europe/Belgrade'), ('Europe/Berlin', 'Europe/Berlin'), ('Europe/Brussels', 'Europe/Brussels'), ('Europe/Bucharest', 'Europe/Bucharest'), ('Europe/Budapest', 'Europe/Budapest'), ('Europe/Chisinau', 'Europe/Chisinau'), ('Europe/Copenhagen', 'Europe/Copenhagen'), ('Europe/Dublin', 'Europe/Dublin'), ('Europe/Gibraltar', 'Europe/Gibraltar'), ('Europe/Helsinki', 'Europe/Helsinki'), ('Europe/Istanbul', 'Europe/Istanbul'), ('Europe/Kaliningrad', 'Europe/Kaliningrad'), ('Europe/Kiev', 'Europe/Kiev'), ('Europe/Kirov', 'Europe/Kirov'), ('Europe/Lisbon', 'Europe/Lisbon'), ('Europe/London', 'Europe/London'), ('Europe/Luxembourg', 'Europe/Luxembourg'), ('Europe/Madrid', 'Europe/Madrid'), ('Europe/Malta', 'Europe/Malta'), ('Europe/Minsk', 'Europe/Minsk'), ('Europe/Monaco', 'Europe/Monaco'), ('Europe/Moscow', 'Europe/Moscow'), ('Europe/Oslo', 'Europe/Oslo'), ('Europe/Paris', 'Europe/Paris'), ('Europe/Prague', 'Europe/Prague'), ('Europe/Riga', 'Europe/Riga'), ('Europe/Rome', 'Europe/Rome'), ('Europe/Samara', 'Europe/Samara'), ('Europe/Saratov', 'Europe/Saratov'), ('Europe/Simferopol', 'Europe/Simferopol'), ('Europe/Sofia', 'Europe/Sofia'), ('Europe/Stockholm', 'Europe/Stockholm'), ('Europe/Tallinn', 'Europe/Tallinn'), ('Europe/Tirane', 'Europe/Tirane'), ('Europe/Ulyanovsk', 'Europe/Ulyanovsk'), ('Europe/Uzhgorod', 'Europe/Uzhgorod'), ('Europe/Vaduz', 'Europe/Vaduz'), ('Europe/Vienna', 'Europe/Vienna'), ('Europe/Vilnius', 'Europe/Vilnius'), ('Europe/Volgograd', 'Europe/Volgograd'), ('Europe/Warsaw', 'Europe/Warsaw'), ('Europe/Zaporozhye', 'Europe/Zaporozhye'), ('Europe/Zurich', 'Europe/Zurich'), ('GMT', 'GMT'), ('Indian/Antananarivo', 'Indian/Antananarivo'), ('Indian/Chagos', 'Indian/Chagos'), ('Indian/Christmas', 'Indian/Christmas'), ('Indian/Cocos', 'Indian/Cocos'), ('Indian/Comoro', 'Indian/Comoro'), ('Indian/Kerguelen', 'Indian/Kerguelen'), ('Indian/Mahe', 'Indian/Mahe'), ('Indian/Maldives', 'Indian/Maldives'), ('Indian/Mauritius', 'Indian/Mauritius'), ('Indian/Mayotte', 'Indian/Mayotte'), ('Indian/Reunion', 'Indian/Reunion'), ('Pacific/Apia', 'Pacific/Apia'), ('Pacific/Auckland', 'Pacific/Auckland'), ('Pacific/Bougainville', 'Pacific/Bougainville'), ('Pacific/Chatham', 'Pacific/Chatham'), ('Pacific/Chuuk', 'Pacific/Chuuk'), ('Pacific/Easter', 'Pacific/Easter'), ('Pacific/Efate', 'Pacific/Efate'), ('Pacific/Fakaofo', 'Pacific/Fakaofo'), ('Pacific/Fiji', 'Pacific/Fiji'), ('Pacific/Funafuti', 'Pacific/Funafuti'), ('Pacific/Galapagos', 'Pacific/Galapagos'), ('Pacific/Gambier', 'Pacific/Gambier'), ('Pacific/Guadalcanal', 'Pacific/Guadalcanal'), ('Pacific/Guam', 'Pacific/Guam'), ('Pacific/Honolulu', 'Pacific/Honolulu'), ('Pacific/Kanton', 'Pacific/Kanton'), ('Pacific/Kiritimati', 'Pacific/Kiritimati'), ('Pacific/Kosrae', 'Pacific/Kosrae'), ('Pacific/Kwajalein', 'Pacific/Kwajalein'), ('Pacific/Majuro', 'Pacific/Majuro'), ('Pacific/Marquesas', 'Pacific/Marquesas'), ('Pacific/Midway', 'Pacific/Midway'), ('Pacific/Nauru', 'Pacific/Nauru'), ('Pacific/Niue', 'Pacific/Niue'), ('Pacific/Norfolk', 'Pacific/Norfolk'), ('Pacific/Noumea', 'Pacific/Noumea'), ('Pacific/Pago_Pago', 'Pacific/Pago_Pago'), ('Pacific/Palau', 'Pacific/Palau'), ('Pacific/Pitcairn', 'Pacific/Pitcairn'), ('Pacific/Pohnpei', 'Pacific/Pohnpei'), ('Pacific/Port_Moresby', 'Pacific/Port_Moresby'), ('Pacific/Rarotonga', 'Pacific/Rarotonga'), ('Pacific/Saipan', 'Pacific/Saipan'), ('Pacific/Tahiti', 'Pacific/Tahiti'), ('Pacific/Tarawa', 'Pacific/Tarawa'), ('Pacific/Tongatapu', 'Pacific/Tongatapu'), ('Pacific/Wake', 'Pacific/Wake'), ('Pacific/Wallis', 'Pacific/Wallis'), ('UTC', 'UTC')], max_length=255), - ), - ] diff --git a/ietf/meeting/migrations/0049_session_purpose.py b/ietf/meeting/migrations/0049_session_purpose.py deleted file mode 100644 index 50e8305bf1..0000000000 --- a/ietf/meeting/migrations/0049_session_purpose.py +++ /dev/null @@ -1,22 +0,0 @@ -# Generated by Django 2.2.24 on 2021-09-16 18:04 - -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0035_populate_sessionpurposename'), - ('meeting', '0048_auto_20211008_0907'), - ] - - operations = [ - migrations.AddField( - model_name='session', - name='purpose', - field=ietf.utils.models.ForeignKey(default='none', help_text='Purpose of the session', on_delete=django.db.models.deletion.CASCADE, to='name.SessionPurposeName'), - preserve_default=False, - ), - ] diff --git a/ietf/meeting/migrations/0050_session_on_agenda.py b/ietf/meeting/migrations/0050_session_on_agenda.py deleted file mode 100644 index 15f1885ae7..0000000000 --- a/ietf/meeting/migrations/0050_session_on_agenda.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.2.24 on 2021-10-22 06:58 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0049_session_purpose'), - ] - - operations = [ - migrations.AddField( - model_name='session', - name='on_agenda', - field=models.BooleanField(default=True, help_text='Is this session visible on the meeting agenda?'), - ), - ] diff --git a/ietf/meeting/migrations/0051_populate_session_on_agenda.py b/ietf/meeting/migrations/0051_populate_session_on_agenda.py deleted file mode 100644 index b63942072a..0000000000 --- a/ietf/meeting/migrations/0051_populate_session_on_agenda.py +++ /dev/null @@ -1,39 +0,0 @@ -# Generated by Django 2.2.24 on 2021-10-22 06:58 - -from django.db import migrations, models - - -def forward(apps, schema_editor): - Session = apps.get_model('meeting', 'Session') - SchedTimeSessAssignment = apps.get_model('meeting', 'SchedTimeSessAssignment') - # find official assignments that are to private timeslots and fill in session.on_agenda - private_assignments = SchedTimeSessAssignment.objects.filter( - models.Q( - schedule=models.F('session__meeting__schedule') - ) | models.Q( - schedule=models.F('session__meeting__schedule__base') - ), - timeslot__type__private=True, - ) - for pa in private_assignments: - pa.session.on_agenda = False - pa.session.save() - # Also update any sessions to match their purpose's default setting (this intentionally - # overrides the timeslot settings above, but that is unlikely to matter because the - # purposes will roll out at the same time as the on_agenda field) - Session.objects.filter(purpose__on_agenda=False).update(on_agenda=False) - Session.objects.filter(purpose__on_agenda=True).update(on_agenda=True) - -def reverse(apps, schema_editor): - Session = apps.get_model('meeting', 'Session') - Session.objects.update(on_agenda=True) # restore all to default on_agenda=True state - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0050_session_on_agenda'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/meeting/migrations/0052_auto_20220503_1815.py b/ietf/meeting/migrations/0052_auto_20220503_1815.py deleted file mode 100644 index 8a01c23ab9..0000000000 --- a/ietf/meeting/migrations/0052_auto_20220503_1815.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.2.28 on 2022-05-03 18:15 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0051_populate_session_on_agenda'), - ] - - operations = [ - migrations.AlterField( - model_name='meeting', - name='country', - field=models.CharField(blank=True, choices=[('', '---------'), ('AF', 'Afghanistan'), ('AL', 'Albania'), ('DZ', 'Algeria'), ('AD', 'Andorra'), ('AO', 'Angola'), ('AI', 'Anguilla'), ('AQ', 'Antarctica'), ('AG', 'Antigua & Barbuda'), ('AR', 'Argentina'), ('AM', 'Armenia'), ('AW', 'Aruba'), ('AU', 'Australia'), ('AT', 'Austria'), ('AZ', 'Azerbaijan'), ('BS', 'Bahamas'), ('BH', 'Bahrain'), ('BD', 'Bangladesh'), ('BB', 'Barbados'), ('BY', 'Belarus'), ('BE', 'Belgium'), ('BZ', 'Belize'), ('BJ', 'Benin'), ('BM', 'Bermuda'), ('BT', 'Bhutan'), ('BO', 'Bolivia'), ('BA', 'Bosnia & Herzegovina'), ('BW', 'Botswana'), ('BV', 'Bouvet Island'), ('BR', 'Brazil'), ('GB', 'Britain (UK)'), ('IO', 'British Indian Ocean Territory'), ('BN', 'Brunei'), ('BG', 'Bulgaria'), ('BF', 'Burkina Faso'), ('BI', 'Burundi'), ('KH', 'Cambodia'), ('CM', 'Cameroon'), ('CA', 'Canada'), ('CV', 'Cape Verde'), ('BQ', 'Caribbean NL'), ('KY', 'Cayman Islands'), ('CF', 'Central African Rep.'), ('TD', 'Chad'), ('CL', 'Chile'), ('CN', 'China'), ('CX', 'Christmas Island'), ('CC', 'Cocos (Keeling) Islands'), ('CO', 'Colombia'), ('KM', 'Comoros'), ('CD', 'Congo (Dem. Rep.)'), ('CG', 'Congo (Rep.)'), ('CK', 'Cook Islands'), ('CR', 'Costa Rica'), ('HR', 'Croatia'), ('CU', 'Cuba'), ('CW', 'Curaçao'), ('CY', 'Cyprus'), ('CZ', 'Czech Republic'), ('CI', "Côte d'Ivoire"), ('DK', 'Denmark'), ('DJ', 'Djibouti'), ('DM', 'Dominica'), ('DO', 'Dominican Republic'), ('TL', 'East Timor'), ('EC', 'Ecuador'), ('EG', 'Egypt'), ('SV', 'El Salvador'), ('GQ', 'Equatorial Guinea'), ('ER', 'Eritrea'), ('EE', 'Estonia'), ('SZ', 'Eswatini (Swaziland)'), ('ET', 'Ethiopia'), ('FK', 'Falkland Islands'), ('FO', 'Faroe Islands'), ('FJ', 'Fiji'), ('FI', 'Finland'), ('FR', 'France'), ('GF', 'French Guiana'), ('PF', 'French Polynesia'), ('TF', 'French Southern & Antarctic Lands'), ('GA', 'Gabon'), ('GM', 'Gambia'), ('GE', 'Georgia'), ('DE', 'Germany'), ('GH', 'Ghana'), ('GI', 'Gibraltar'), ('GR', 'Greece'), ('GL', 'Greenland'), ('GD', 'Grenada'), ('GP', 'Guadeloupe'), ('GU', 'Guam'), ('GT', 'Guatemala'), ('GG', 'Guernsey'), ('GN', 'Guinea'), ('GW', 'Guinea-Bissau'), ('GY', 'Guyana'), ('HT', 'Haiti'), ('HM', 'Heard Island & McDonald Islands'), ('HN', 'Honduras'), ('HK', 'Hong Kong'), ('HU', 'Hungary'), ('IS', 'Iceland'), ('IN', 'India'), ('ID', 'Indonesia'), ('IR', 'Iran'), ('IQ', 'Iraq'), ('IE', 'Ireland'), ('IM', 'Isle of Man'), ('IL', 'Israel'), ('IT', 'Italy'), ('JM', 'Jamaica'), ('JP', 'Japan'), ('JE', 'Jersey'), ('JO', 'Jordan'), ('KZ', 'Kazakhstan'), ('KE', 'Kenya'), ('KI', 'Kiribati'), ('KP', 'Korea (North)'), ('KR', 'Korea (South)'), ('KW', 'Kuwait'), ('KG', 'Kyrgyzstan'), ('LA', 'Laos'), ('LV', 'Latvia'), ('LB', 'Lebanon'), ('LS', 'Lesotho'), ('LR', 'Liberia'), ('LY', 'Libya'), ('LI', 'Liechtenstein'), ('LT', 'Lithuania'), ('LU', 'Luxembourg'), ('MO', 'Macau'), ('MG', 'Madagascar'), ('MW', 'Malawi'), ('MY', 'Malaysia'), ('MV', 'Maldives'), ('ML', 'Mali'), ('MT', 'Malta'), ('MH', 'Marshall Islands'), ('MQ', 'Martinique'), ('MR', 'Mauritania'), ('MU', 'Mauritius'), ('YT', 'Mayotte'), ('MX', 'Mexico'), ('FM', 'Micronesia'), ('MD', 'Moldova'), ('MC', 'Monaco'), ('MN', 'Mongolia'), ('ME', 'Montenegro'), ('MS', 'Montserrat'), ('MA', 'Morocco'), ('MZ', 'Mozambique'), ('MM', 'Myanmar (Burma)'), ('NA', 'Namibia'), ('NR', 'Nauru'), ('NP', 'Nepal'), ('NL', 'Netherlands'), ('NC', 'New Caledonia'), ('NZ', 'New Zealand'), ('NI', 'Nicaragua'), ('NE', 'Niger'), ('NG', 'Nigeria'), ('NU', 'Niue'), ('NF', 'Norfolk Island'), ('MK', 'North Macedonia'), ('MP', 'Northern Mariana Islands'), ('NO', 'Norway'), ('OM', 'Oman'), ('PK', 'Pakistan'), ('PW', 'Palau'), ('PS', 'Palestine'), ('PA', 'Panama'), ('PG', 'Papua New Guinea'), ('PY', 'Paraguay'), ('PE', 'Peru'), ('PH', 'Philippines'), ('PN', 'Pitcairn'), ('PL', 'Poland'), ('PT', 'Portugal'), ('PR', 'Puerto Rico'), ('QA', 'Qatar'), ('RO', 'Romania'), ('RU', 'Russia'), ('RW', 'Rwanda'), ('RE', 'Réunion'), ('AS', 'Samoa (American)'), ('WS', 'Samoa (western)'), ('SM', 'San Marino'), ('ST', 'Sao Tome & Principe'), ('SA', 'Saudi Arabia'), ('SN', 'Senegal'), ('RS', 'Serbia'), ('SC', 'Seychelles'), ('SL', 'Sierra Leone'), ('SG', 'Singapore'), ('SK', 'Slovakia'), ('SI', 'Slovenia'), ('SB', 'Solomon Islands'), ('SO', 'Somalia'), ('ZA', 'South Africa'), ('GS', 'South Georgia & the South Sandwich Islands'), ('SS', 'South Sudan'), ('ES', 'Spain'), ('LK', 'Sri Lanka'), ('BL', 'St Barthelemy'), ('SH', 'St Helena'), ('KN', 'St Kitts & Nevis'), ('LC', 'St Lucia'), ('SX', 'St Maarten (Dutch)'), ('MF', 'St Martin (French)'), ('PM', 'St Pierre & Miquelon'), ('VC', 'St Vincent'), ('SD', 'Sudan'), ('SR', 'Suriname'), ('SJ', 'Svalbard & Jan Mayen'), ('SE', 'Sweden'), ('CH', 'Switzerland'), ('SY', 'Syria'), ('TW', 'Taiwan'), ('TJ', 'Tajikistan'), ('TZ', 'Tanzania'), ('TH', 'Thailand'), ('TG', 'Togo'), ('TK', 'Tokelau'), ('TO', 'Tonga'), ('TT', 'Trinidad & Tobago'), ('TN', 'Tunisia'), ('TR', 'Turkey'), ('TM', 'Turkmenistan'), ('TC', 'Turks & Caicos Is'), ('TV', 'Tuvalu'), ('UM', 'US minor outlying islands'), ('UG', 'Uganda'), ('UA', 'Ukraine'), ('AE', 'United Arab Emirates'), ('US', 'United States'), ('UY', 'Uruguay'), ('UZ', 'Uzbekistan'), ('VU', 'Vanuatu'), ('VA', 'Vatican City'), ('VE', 'Venezuela'), ('VN', 'Vietnam'), ('VG', 'Virgin Islands (UK)'), ('VI', 'Virgin Islands (US)'), ('WF', 'Wallis & Futuna'), ('EH', 'Western Sahara'), ('YE', 'Yemen'), ('ZM', 'Zambia'), ('ZW', 'Zimbabwe'), ('AX', 'Åland Islands')], max_length=2), - ), - ] diff --git a/ietf/meeting/migrations/0053_attended.py b/ietf/meeting/migrations/0053_attended.py deleted file mode 100644 index 110101cf15..0000000000 --- a/ietf/meeting/migrations/0053_attended.py +++ /dev/null @@ -1,28 +0,0 @@ -# Copyright The IETF Trust 2022, All Rights Reserved -# Generated by Django 2.2.28 on 2022-06-17 08:40 - -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0023_auto_20220615_1006'), - ('meeting', '0052_auto_20220503_1815'), - ] - - operations = [ - migrations.CreateModel( - name='Attended', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), - ('session', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='meeting.Session')), - ], - options={ - 'unique_together': {('person', 'session')}, - }, - ), - ] diff --git a/ietf/meeting/migrations/0054_pytz_2002_2.py b/ietf/meeting/migrations/0054_pytz_2002_2.py deleted file mode 100644 index bb9ad62a54..0000000000 --- a/ietf/meeting/migrations/0054_pytz_2002_2.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.2.28 on 2022-08-12 09:24 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0053_attended'), - ] - - operations = [ - migrations.AlterField( - model_name='meeting', - name='time_zone', - field=models.CharField(blank=True, choices=[('', '---------'), ('Africa/Abidjan', 'Africa/Abidjan'), ('Africa/Accra', 'Africa/Accra'), ('Africa/Addis_Ababa', 'Africa/Addis_Ababa'), ('Africa/Algiers', 'Africa/Algiers'), ('Africa/Asmara', 'Africa/Asmara'), ('Africa/Bamako', 'Africa/Bamako'), ('Africa/Bangui', 'Africa/Bangui'), ('Africa/Banjul', 'Africa/Banjul'), ('Africa/Bissau', 'Africa/Bissau'), ('Africa/Blantyre', 'Africa/Blantyre'), ('Africa/Brazzaville', 'Africa/Brazzaville'), ('Africa/Bujumbura', 'Africa/Bujumbura'), ('Africa/Cairo', 'Africa/Cairo'), ('Africa/Casablanca', 'Africa/Casablanca'), ('Africa/Ceuta', 'Africa/Ceuta'), ('Africa/Conakry', 'Africa/Conakry'), ('Africa/Dakar', 'Africa/Dakar'), ('Africa/Dar_es_Salaam', 'Africa/Dar_es_Salaam'), ('Africa/Djibouti', 'Africa/Djibouti'), ('Africa/Douala', 'Africa/Douala'), ('Africa/El_Aaiun', 'Africa/El_Aaiun'), ('Africa/Freetown', 'Africa/Freetown'), ('Africa/Gaborone', 'Africa/Gaborone'), ('Africa/Harare', 'Africa/Harare'), ('Africa/Johannesburg', 'Africa/Johannesburg'), ('Africa/Juba', 'Africa/Juba'), ('Africa/Kampala', 'Africa/Kampala'), ('Africa/Khartoum', 'Africa/Khartoum'), ('Africa/Kigali', 'Africa/Kigali'), ('Africa/Kinshasa', 'Africa/Kinshasa'), ('Africa/Lagos', 'Africa/Lagos'), ('Africa/Libreville', 'Africa/Libreville'), ('Africa/Lome', 'Africa/Lome'), ('Africa/Luanda', 'Africa/Luanda'), ('Africa/Lubumbashi', 'Africa/Lubumbashi'), ('Africa/Lusaka', 'Africa/Lusaka'), ('Africa/Malabo', 'Africa/Malabo'), ('Africa/Maputo', 'Africa/Maputo'), ('Africa/Maseru', 'Africa/Maseru'), ('Africa/Mbabane', 'Africa/Mbabane'), ('Africa/Mogadishu', 'Africa/Mogadishu'), ('Africa/Monrovia', 'Africa/Monrovia'), ('Africa/Nairobi', 'Africa/Nairobi'), ('Africa/Ndjamena', 'Africa/Ndjamena'), ('Africa/Niamey', 'Africa/Niamey'), ('Africa/Nouakchott', 'Africa/Nouakchott'), ('Africa/Ouagadougou', 'Africa/Ouagadougou'), ('Africa/Porto-Novo', 'Africa/Porto-Novo'), ('Africa/Sao_Tome', 'Africa/Sao_Tome'), ('Africa/Tripoli', 'Africa/Tripoli'), ('Africa/Tunis', 'Africa/Tunis'), ('Africa/Windhoek', 'Africa/Windhoek'), ('America/Adak', 'America/Adak'), ('America/Anchorage', 'America/Anchorage'), ('America/Anguilla', 'America/Anguilla'), ('America/Antigua', 'America/Antigua'), ('America/Araguaina', 'America/Araguaina'), ('America/Argentina/Buenos_Aires', 'America/Argentina/Buenos_Aires'), ('America/Argentina/Catamarca', 'America/Argentina/Catamarca'), ('America/Argentina/Cordoba', 'America/Argentina/Cordoba'), ('America/Argentina/Jujuy', 'America/Argentina/Jujuy'), ('America/Argentina/La_Rioja', 'America/Argentina/La_Rioja'), ('America/Argentina/Mendoza', 'America/Argentina/Mendoza'), ('America/Argentina/Rio_Gallegos', 'America/Argentina/Rio_Gallegos'), ('America/Argentina/Salta', 'America/Argentina/Salta'), ('America/Argentina/San_Juan', 'America/Argentina/San_Juan'), ('America/Argentina/San_Luis', 'America/Argentina/San_Luis'), ('America/Argentina/Tucuman', 'America/Argentina/Tucuman'), ('America/Argentina/Ushuaia', 'America/Argentina/Ushuaia'), ('America/Aruba', 'America/Aruba'), ('America/Asuncion', 'America/Asuncion'), ('America/Atikokan', 'America/Atikokan'), ('America/Bahia', 'America/Bahia'), ('America/Bahia_Banderas', 'America/Bahia_Banderas'), ('America/Barbados', 'America/Barbados'), ('America/Belem', 'America/Belem'), ('America/Belize', 'America/Belize'), ('America/Blanc-Sablon', 'America/Blanc-Sablon'), ('America/Boa_Vista', 'America/Boa_Vista'), ('America/Bogota', 'America/Bogota'), ('America/Boise', 'America/Boise'), ('America/Cambridge_Bay', 'America/Cambridge_Bay'), ('America/Campo_Grande', 'America/Campo_Grande'), ('America/Cancun', 'America/Cancun'), ('America/Caracas', 'America/Caracas'), ('America/Cayenne', 'America/Cayenne'), ('America/Cayman', 'America/Cayman'), ('America/Chicago', 'America/Chicago'), ('America/Chihuahua', 'America/Chihuahua'), ('America/Costa_Rica', 'America/Costa_Rica'), ('America/Creston', 'America/Creston'), ('America/Cuiaba', 'America/Cuiaba'), ('America/Curacao', 'America/Curacao'), ('America/Danmarkshavn', 'America/Danmarkshavn'), ('America/Dawson', 'America/Dawson'), ('America/Dawson_Creek', 'America/Dawson_Creek'), ('America/Denver', 'America/Denver'), ('America/Detroit', 'America/Detroit'), ('America/Dominica', 'America/Dominica'), ('America/Edmonton', 'America/Edmonton'), ('America/Eirunepe', 'America/Eirunepe'), ('America/El_Salvador', 'America/El_Salvador'), ('America/Fort_Nelson', 'America/Fort_Nelson'), ('America/Fortaleza', 'America/Fortaleza'), ('America/Glace_Bay', 'America/Glace_Bay'), ('America/Goose_Bay', 'America/Goose_Bay'), ('America/Grand_Turk', 'America/Grand_Turk'), ('America/Grenada', 'America/Grenada'), ('America/Guadeloupe', 'America/Guadeloupe'), ('America/Guatemala', 'America/Guatemala'), ('America/Guayaquil', 'America/Guayaquil'), ('America/Guyana', 'America/Guyana'), ('America/Halifax', 'America/Halifax'), ('America/Havana', 'America/Havana'), ('America/Hermosillo', 'America/Hermosillo'), ('America/Indiana/Indianapolis', 'America/Indiana/Indianapolis'), ('America/Indiana/Knox', 'America/Indiana/Knox'), ('America/Indiana/Marengo', 'America/Indiana/Marengo'), ('America/Indiana/Petersburg', 'America/Indiana/Petersburg'), ('America/Indiana/Tell_City', 'America/Indiana/Tell_City'), ('America/Indiana/Vevay', 'America/Indiana/Vevay'), ('America/Indiana/Vincennes', 'America/Indiana/Vincennes'), ('America/Indiana/Winamac', 'America/Indiana/Winamac'), ('America/Inuvik', 'America/Inuvik'), ('America/Iqaluit', 'America/Iqaluit'), ('America/Jamaica', 'America/Jamaica'), ('America/Juneau', 'America/Juneau'), ('America/Kentucky/Louisville', 'America/Kentucky/Louisville'), ('America/Kentucky/Monticello', 'America/Kentucky/Monticello'), ('America/La_Paz', 'America/La_Paz'), ('America/Lima', 'America/Lima'), ('America/Los_Angeles', 'America/Los_Angeles'), ('America/Maceio', 'America/Maceio'), ('America/Managua', 'America/Managua'), ('America/Manaus', 'America/Manaus'), ('America/Martinique', 'America/Martinique'), ('America/Matamoros', 'America/Matamoros'), ('America/Mazatlan', 'America/Mazatlan'), ('America/Menominee', 'America/Menominee'), ('America/Merida', 'America/Merida'), ('America/Metlakatla', 'America/Metlakatla'), ('America/Mexico_City', 'America/Mexico_City'), ('America/Miquelon', 'America/Miquelon'), ('America/Moncton', 'America/Moncton'), ('America/Monterrey', 'America/Monterrey'), ('America/Montevideo', 'America/Montevideo'), ('America/Montreal', 'America/Montreal'), ('America/Montserrat', 'America/Montserrat'), ('America/Nassau', 'America/Nassau'), ('America/New_York', 'America/New_York'), ('America/Nipigon', 'America/Nipigon'), ('America/Nome', 'America/Nome'), ('America/Noronha', 'America/Noronha'), ('America/North_Dakota/Beulah', 'America/North_Dakota/Beulah'), ('America/North_Dakota/Center', 'America/North_Dakota/Center'), ('America/North_Dakota/New_Salem', 'America/North_Dakota/New_Salem'), ('America/Nuuk', 'America/Nuuk'), ('America/Ojinaga', 'America/Ojinaga'), ('America/Panama', 'America/Panama'), ('America/Pangnirtung', 'America/Pangnirtung'), ('America/Paramaribo', 'America/Paramaribo'), ('America/Phoenix', 'America/Phoenix'), ('America/Port-au-Prince', 'America/Port-au-Prince'), ('America/Port_of_Spain', 'America/Port_of_Spain'), ('America/Porto_Velho', 'America/Porto_Velho'), ('America/Puerto_Rico', 'America/Puerto_Rico'), ('America/Punta_Arenas', 'America/Punta_Arenas'), ('America/Rainy_River', 'America/Rainy_River'), ('America/Rankin_Inlet', 'America/Rankin_Inlet'), ('America/Recife', 'America/Recife'), ('America/Regina', 'America/Regina'), ('America/Resolute', 'America/Resolute'), ('America/Rio_Branco', 'America/Rio_Branco'), ('America/Santarem', 'America/Santarem'), ('America/Santiago', 'America/Santiago'), ('America/Santo_Domingo', 'America/Santo_Domingo'), ('America/Sao_Paulo', 'America/Sao_Paulo'), ('America/Scoresbysund', 'America/Scoresbysund'), ('America/Sitka', 'America/Sitka'), ('America/St_Johns', 'America/St_Johns'), ('America/St_Kitts', 'America/St_Kitts'), ('America/St_Lucia', 'America/St_Lucia'), ('America/St_Thomas', 'America/St_Thomas'), ('America/St_Vincent', 'America/St_Vincent'), ('America/Swift_Current', 'America/Swift_Current'), ('America/Tegucigalpa', 'America/Tegucigalpa'), ('America/Thule', 'America/Thule'), ('America/Thunder_Bay', 'America/Thunder_Bay'), ('America/Tijuana', 'America/Tijuana'), ('America/Toronto', 'America/Toronto'), ('America/Tortola', 'America/Tortola'), ('America/Vancouver', 'America/Vancouver'), ('America/Whitehorse', 'America/Whitehorse'), ('America/Winnipeg', 'America/Winnipeg'), ('America/Yakutat', 'America/Yakutat'), ('America/Yellowknife', 'America/Yellowknife'), ('Antarctica/Casey', 'Antarctica/Casey'), ('Antarctica/Davis', 'Antarctica/Davis'), ('Antarctica/DumontDUrville', 'Antarctica/DumontDUrville'), ('Antarctica/Macquarie', 'Antarctica/Macquarie'), ('Antarctica/Mawson', 'Antarctica/Mawson'), ('Antarctica/McMurdo', 'Antarctica/McMurdo'), ('Antarctica/Palmer', 'Antarctica/Palmer'), ('Antarctica/Rothera', 'Antarctica/Rothera'), ('Antarctica/Syowa', 'Antarctica/Syowa'), ('Antarctica/Troll', 'Antarctica/Troll'), ('Antarctica/Vostok', 'Antarctica/Vostok'), ('Asia/Aden', 'Asia/Aden'), ('Asia/Almaty', 'Asia/Almaty'), ('Asia/Amman', 'Asia/Amman'), ('Asia/Anadyr', 'Asia/Anadyr'), ('Asia/Aqtau', 'Asia/Aqtau'), ('Asia/Aqtobe', 'Asia/Aqtobe'), ('Asia/Ashgabat', 'Asia/Ashgabat'), ('Asia/Atyrau', 'Asia/Atyrau'), ('Asia/Baghdad', 'Asia/Baghdad'), ('Asia/Bahrain', 'Asia/Bahrain'), ('Asia/Baku', 'Asia/Baku'), ('Asia/Bangkok', 'Asia/Bangkok'), ('Asia/Barnaul', 'Asia/Barnaul'), ('Asia/Beirut', 'Asia/Beirut'), ('Asia/Bishkek', 'Asia/Bishkek'), ('Asia/Brunei', 'Asia/Brunei'), ('Asia/Chita', 'Asia/Chita'), ('Asia/Choibalsan', 'Asia/Choibalsan'), ('Asia/Chongqing', 'Asia/Chongqing'), ('Asia/Colombo', 'Asia/Colombo'), ('Asia/Damascus', 'Asia/Damascus'), ('Asia/Dhaka', 'Asia/Dhaka'), ('Asia/Dili', 'Asia/Dili'), ('Asia/Dubai', 'Asia/Dubai'), ('Asia/Dushanbe', 'Asia/Dushanbe'), ('Asia/Famagusta', 'Asia/Famagusta'), ('Asia/Gaza', 'Asia/Gaza'), ('Asia/Hanoi', 'Asia/Hanoi'), ('Asia/Harbin', 'Asia/Harbin'), ('Asia/Hebron', 'Asia/Hebron'), ('Asia/Ho_Chi_Minh', 'Asia/Ho_Chi_Minh'), ('Asia/Hong_Kong', 'Asia/Hong_Kong'), ('Asia/Hovd', 'Asia/Hovd'), ('Asia/Irkutsk', 'Asia/Irkutsk'), ('Asia/Jakarta', 'Asia/Jakarta'), ('Asia/Jayapura', 'Asia/Jayapura'), ('Asia/Jerusalem', 'Asia/Jerusalem'), ('Asia/Kabul', 'Asia/Kabul'), ('Asia/Kamchatka', 'Asia/Kamchatka'), ('Asia/Karachi', 'Asia/Karachi'), ('Asia/Kashgar', 'Asia/Kashgar'), ('Asia/Kathmandu', 'Asia/Kathmandu'), ('Asia/Khandyga', 'Asia/Khandyga'), ('Asia/Kolkata', 'Asia/Kolkata'), ('Asia/Krasnoyarsk', 'Asia/Krasnoyarsk'), ('Asia/Kuala_Lumpur', 'Asia/Kuala_Lumpur'), ('Asia/Kuching', 'Asia/Kuching'), ('Asia/Kuwait', 'Asia/Kuwait'), ('Asia/Macau', 'Asia/Macau'), ('Asia/Magadan', 'Asia/Magadan'), ('Asia/Makassar', 'Asia/Makassar'), ('Asia/Manila', 'Asia/Manila'), ('Asia/Muscat', 'Asia/Muscat'), ('Asia/Nicosia', 'Asia/Nicosia'), ('Asia/Novokuznetsk', 'Asia/Novokuznetsk'), ('Asia/Novosibirsk', 'Asia/Novosibirsk'), ('Asia/Omsk', 'Asia/Omsk'), ('Asia/Oral', 'Asia/Oral'), ('Asia/Phnom_Penh', 'Asia/Phnom_Penh'), ('Asia/Pontianak', 'Asia/Pontianak'), ('Asia/Pyongyang', 'Asia/Pyongyang'), ('Asia/Qatar', 'Asia/Qatar'), ('Asia/Qostanay', 'Asia/Qostanay'), ('Asia/Qyzylorda', 'Asia/Qyzylorda'), ('Asia/Riyadh', 'Asia/Riyadh'), ('Asia/Sakhalin', 'Asia/Sakhalin'), ('Asia/Samarkand', 'Asia/Samarkand'), ('Asia/Seoul', 'Asia/Seoul'), ('Asia/Shanghai', 'Asia/Shanghai'), ('Asia/Singapore', 'Asia/Singapore'), ('Asia/Srednekolymsk', 'Asia/Srednekolymsk'), ('Asia/Taipei', 'Asia/Taipei'), ('Asia/Tashkent', 'Asia/Tashkent'), ('Asia/Tbilisi', 'Asia/Tbilisi'), ('Asia/Tehran', 'Asia/Tehran'), ('Asia/Thimphu', 'Asia/Thimphu'), ('Asia/Tokyo', 'Asia/Tokyo'), ('Asia/Tomsk', 'Asia/Tomsk'), ('Asia/Ulaanbaatar', 'Asia/Ulaanbaatar'), ('Asia/Urumqi', 'Asia/Urumqi'), ('Asia/Ust-Nera', 'Asia/Ust-Nera'), ('Asia/Vientiane', 'Asia/Vientiane'), ('Asia/Vladivostok', 'Asia/Vladivostok'), ('Asia/Yakutsk', 'Asia/Yakutsk'), ('Asia/Yangon', 'Asia/Yangon'), ('Asia/Yekaterinburg', 'Asia/Yekaterinburg'), ('Asia/Yerevan', 'Asia/Yerevan'), ('Atlantic/Azores', 'Atlantic/Azores'), ('Atlantic/Bermuda', 'Atlantic/Bermuda'), ('Atlantic/Canary', 'Atlantic/Canary'), ('Atlantic/Cape_Verde', 'Atlantic/Cape_Verde'), ('Atlantic/Faroe', 'Atlantic/Faroe'), ('Atlantic/Madeira', 'Atlantic/Madeira'), ('Atlantic/Reykjavik', 'Atlantic/Reykjavik'), ('Atlantic/South_Georgia', 'Atlantic/South_Georgia'), ('Atlantic/St_Helena', 'Atlantic/St_Helena'), ('Atlantic/Stanley', 'Atlantic/Stanley'), ('Australia/Adelaide', 'Australia/Adelaide'), ('Australia/Brisbane', 'Australia/Brisbane'), ('Australia/Broken_Hill', 'Australia/Broken_Hill'), ('Australia/Currie', 'Australia/Currie'), ('Australia/Darwin', 'Australia/Darwin'), ('Australia/Eucla', 'Australia/Eucla'), ('Australia/Hobart', 'Australia/Hobart'), ('Australia/Lindeman', 'Australia/Lindeman'), ('Australia/Lord_Howe', 'Australia/Lord_Howe'), ('Australia/Melbourne', 'Australia/Melbourne'), ('Australia/Perth', 'Australia/Perth'), ('Australia/Sydney', 'Australia/Sydney'), ('Europe/Amsterdam', 'Europe/Amsterdam'), ('Europe/Andorra', 'Europe/Andorra'), ('Europe/Astrakhan', 'Europe/Astrakhan'), ('Europe/Athens', 'Europe/Athens'), ('Europe/Belgrade', 'Europe/Belgrade'), ('Europe/Berlin', 'Europe/Berlin'), ('Europe/Brussels', 'Europe/Brussels'), ('Europe/Bucharest', 'Europe/Bucharest'), ('Europe/Budapest', 'Europe/Budapest'), ('Europe/Chisinau', 'Europe/Chisinau'), ('Europe/Copenhagen', 'Europe/Copenhagen'), ('Europe/Dublin', 'Europe/Dublin'), ('Europe/Gibraltar', 'Europe/Gibraltar'), ('Europe/Helsinki', 'Europe/Helsinki'), ('Europe/Istanbul', 'Europe/Istanbul'), ('Europe/Kaliningrad', 'Europe/Kaliningrad'), ('Europe/Kirov', 'Europe/Kirov'), ('Europe/Kyiv', 'Europe/Kyiv'), ('Europe/Lisbon', 'Europe/Lisbon'), ('Europe/London', 'Europe/London'), ('Europe/Luxembourg', 'Europe/Luxembourg'), ('Europe/Madrid', 'Europe/Madrid'), ('Europe/Malta', 'Europe/Malta'), ('Europe/Minsk', 'Europe/Minsk'), ('Europe/Monaco', 'Europe/Monaco'), ('Europe/Moscow', 'Europe/Moscow'), ('Europe/Oslo', 'Europe/Oslo'), ('Europe/Paris', 'Europe/Paris'), ('Europe/Prague', 'Europe/Prague'), ('Europe/Riga', 'Europe/Riga'), ('Europe/Rome', 'Europe/Rome'), ('Europe/Samara', 'Europe/Samara'), ('Europe/Saratov', 'Europe/Saratov'), ('Europe/Simferopol', 'Europe/Simferopol'), ('Europe/Sofia', 'Europe/Sofia'), ('Europe/Stockholm', 'Europe/Stockholm'), ('Europe/Tallinn', 'Europe/Tallinn'), ('Europe/Tirane', 'Europe/Tirane'), ('Europe/Ulyanovsk', 'Europe/Ulyanovsk'), ('Europe/Uzhgorod', 'Europe/Uzhgorod'), ('Europe/Vaduz', 'Europe/Vaduz'), ('Europe/Vienna', 'Europe/Vienna'), ('Europe/Vilnius', 'Europe/Vilnius'), ('Europe/Volgograd', 'Europe/Volgograd'), ('Europe/Warsaw', 'Europe/Warsaw'), ('Europe/Zaporozhye', 'Europe/Zaporozhye'), ('Europe/Zurich', 'Europe/Zurich'), ('GMT', 'GMT'), ('Indian/Antananarivo', 'Indian/Antananarivo'), ('Indian/Chagos', 'Indian/Chagos'), ('Indian/Christmas', 'Indian/Christmas'), ('Indian/Cocos', 'Indian/Cocos'), ('Indian/Comoro', 'Indian/Comoro'), ('Indian/Kerguelen', 'Indian/Kerguelen'), ('Indian/Mahe', 'Indian/Mahe'), ('Indian/Maldives', 'Indian/Maldives'), ('Indian/Mauritius', 'Indian/Mauritius'), ('Indian/Mayotte', 'Indian/Mayotte'), ('Indian/Reunion', 'Indian/Reunion'), ('Pacific/Apia', 'Pacific/Apia'), ('Pacific/Auckland', 'Pacific/Auckland'), ('Pacific/Bougainville', 'Pacific/Bougainville'), ('Pacific/Chatham', 'Pacific/Chatham'), ('Pacific/Chuuk', 'Pacific/Chuuk'), ('Pacific/Easter', 'Pacific/Easter'), ('Pacific/Efate', 'Pacific/Efate'), ('Pacific/Enderbury', 'Pacific/Enderbury'), ('Pacific/Fakaofo', 'Pacific/Fakaofo'), ('Pacific/Fiji', 'Pacific/Fiji'), ('Pacific/Funafuti', 'Pacific/Funafuti'), ('Pacific/Galapagos', 'Pacific/Galapagos'), ('Pacific/Gambier', 'Pacific/Gambier'), ('Pacific/Guadalcanal', 'Pacific/Guadalcanal'), ('Pacific/Guam', 'Pacific/Guam'), ('Pacific/Honolulu', 'Pacific/Honolulu'), ('Pacific/Johnston', 'Pacific/Johnston'), ('Pacific/Kanton', 'Pacific/Kanton'), ('Pacific/Kiritimati', 'Pacific/Kiritimati'), ('Pacific/Kosrae', 'Pacific/Kosrae'), ('Pacific/Kwajalein', 'Pacific/Kwajalein'), ('Pacific/Majuro', 'Pacific/Majuro'), ('Pacific/Marquesas', 'Pacific/Marquesas'), ('Pacific/Midway', 'Pacific/Midway'), ('Pacific/Nauru', 'Pacific/Nauru'), ('Pacific/Niue', 'Pacific/Niue'), ('Pacific/Norfolk', 'Pacific/Norfolk'), ('Pacific/Noumea', 'Pacific/Noumea'), ('Pacific/Pago_Pago', 'Pacific/Pago_Pago'), ('Pacific/Palau', 'Pacific/Palau'), ('Pacific/Pitcairn', 'Pacific/Pitcairn'), ('Pacific/Pohnpei', 'Pacific/Pohnpei'), ('Pacific/Port_Moresby', 'Pacific/Port_Moresby'), ('Pacific/Rarotonga', 'Pacific/Rarotonga'), ('Pacific/Saipan', 'Pacific/Saipan'), ('Pacific/Tahiti', 'Pacific/Tahiti'), ('Pacific/Tarawa', 'Pacific/Tarawa'), ('Pacific/Tongatapu', 'Pacific/Tongatapu'), ('Pacific/Wake', 'Pacific/Wake'), ('Pacific/Wallis', 'Pacific/Wallis'), ('UTC', 'UTC')], max_length=255), - ), - ] diff --git a/ietf/meeting/migrations/0055_pytz_2022_2_1.py b/ietf/meeting/migrations/0055_pytz_2022_2_1.py deleted file mode 100644 index 8ad9b4b827..0000000000 --- a/ietf/meeting/migrations/0055_pytz_2022_2_1.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.2.28 on 2022-08-18 08:31 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0054_pytz_2002_2'), - ] - - operations = [ - migrations.AlterField( - model_name='meeting', - name='time_zone', - field=models.CharField(blank=True, choices=[('', '---------'), ('Africa/Abidjan', 'Africa/Abidjan'), ('Africa/Accra', 'Africa/Accra'), ('Africa/Addis_Ababa', 'Africa/Addis_Ababa'), ('Africa/Algiers', 'Africa/Algiers'), ('Africa/Asmara', 'Africa/Asmara'), ('Africa/Bamako', 'Africa/Bamako'), ('Africa/Bangui', 'Africa/Bangui'), ('Africa/Banjul', 'Africa/Banjul'), ('Africa/Bissau', 'Africa/Bissau'), ('Africa/Blantyre', 'Africa/Blantyre'), ('Africa/Brazzaville', 'Africa/Brazzaville'), ('Africa/Bujumbura', 'Africa/Bujumbura'), ('Africa/Cairo', 'Africa/Cairo'), ('Africa/Casablanca', 'Africa/Casablanca'), ('Africa/Ceuta', 'Africa/Ceuta'), ('Africa/Conakry', 'Africa/Conakry'), ('Africa/Dakar', 'Africa/Dakar'), ('Africa/Dar_es_Salaam', 'Africa/Dar_es_Salaam'), ('Africa/Djibouti', 'Africa/Djibouti'), ('Africa/Douala', 'Africa/Douala'), ('Africa/El_Aaiun', 'Africa/El_Aaiun'), ('Africa/Freetown', 'Africa/Freetown'), ('Africa/Gaborone', 'Africa/Gaborone'), ('Africa/Harare', 'Africa/Harare'), ('Africa/Johannesburg', 'Africa/Johannesburg'), ('Africa/Juba', 'Africa/Juba'), ('Africa/Kampala', 'Africa/Kampala'), ('Africa/Khartoum', 'Africa/Khartoum'), ('Africa/Kigali', 'Africa/Kigali'), ('Africa/Kinshasa', 'Africa/Kinshasa'), ('Africa/Lagos', 'Africa/Lagos'), ('Africa/Libreville', 'Africa/Libreville'), ('Africa/Lome', 'Africa/Lome'), ('Africa/Luanda', 'Africa/Luanda'), ('Africa/Lubumbashi', 'Africa/Lubumbashi'), ('Africa/Lusaka', 'Africa/Lusaka'), ('Africa/Malabo', 'Africa/Malabo'), ('Africa/Maputo', 'Africa/Maputo'), ('Africa/Maseru', 'Africa/Maseru'), ('Africa/Mbabane', 'Africa/Mbabane'), ('Africa/Mogadishu', 'Africa/Mogadishu'), ('Africa/Monrovia', 'Africa/Monrovia'), ('Africa/Nairobi', 'Africa/Nairobi'), ('Africa/Ndjamena', 'Africa/Ndjamena'), ('Africa/Niamey', 'Africa/Niamey'), ('Africa/Nouakchott', 'Africa/Nouakchott'), ('Africa/Ouagadougou', 'Africa/Ouagadougou'), ('Africa/Porto-Novo', 'Africa/Porto-Novo'), ('Africa/Sao_Tome', 'Africa/Sao_Tome'), ('Africa/Tripoli', 'Africa/Tripoli'), ('Africa/Tunis', 'Africa/Tunis'), ('Africa/Windhoek', 'Africa/Windhoek'), ('America/Adak', 'America/Adak'), ('America/Anchorage', 'America/Anchorage'), ('America/Anguilla', 'America/Anguilla'), ('America/Antigua', 'America/Antigua'), ('America/Araguaina', 'America/Araguaina'), ('America/Argentina/Buenos_Aires', 'America/Argentina/Buenos_Aires'), ('America/Argentina/Catamarca', 'America/Argentina/Catamarca'), ('America/Argentina/Cordoba', 'America/Argentina/Cordoba'), ('America/Argentina/Jujuy', 'America/Argentina/Jujuy'), ('America/Argentina/La_Rioja', 'America/Argentina/La_Rioja'), ('America/Argentina/Mendoza', 'America/Argentina/Mendoza'), ('America/Argentina/Rio_Gallegos', 'America/Argentina/Rio_Gallegos'), ('America/Argentina/Salta', 'America/Argentina/Salta'), ('America/Argentina/San_Juan', 'America/Argentina/San_Juan'), ('America/Argentina/San_Luis', 'America/Argentina/San_Luis'), ('America/Argentina/Tucuman', 'America/Argentina/Tucuman'), ('America/Argentina/Ushuaia', 'America/Argentina/Ushuaia'), ('America/Aruba', 'America/Aruba'), ('America/Asuncion', 'America/Asuncion'), ('America/Atikokan', 'America/Atikokan'), ('America/Bahia', 'America/Bahia'), ('America/Bahia_Banderas', 'America/Bahia_Banderas'), ('America/Barbados', 'America/Barbados'), ('America/Belem', 'America/Belem'), ('America/Belize', 'America/Belize'), ('America/Blanc-Sablon', 'America/Blanc-Sablon'), ('America/Boa_Vista', 'America/Boa_Vista'), ('America/Bogota', 'America/Bogota'), ('America/Boise', 'America/Boise'), ('America/Cambridge_Bay', 'America/Cambridge_Bay'), ('America/Campo_Grande', 'America/Campo_Grande'), ('America/Cancun', 'America/Cancun'), ('America/Caracas', 'America/Caracas'), ('America/Cayenne', 'America/Cayenne'), ('America/Cayman', 'America/Cayman'), ('America/Chicago', 'America/Chicago'), ('America/Chihuahua', 'America/Chihuahua'), ('America/Costa_Rica', 'America/Costa_Rica'), ('America/Creston', 'America/Creston'), ('America/Cuiaba', 'America/Cuiaba'), ('America/Curacao', 'America/Curacao'), ('America/Danmarkshavn', 'America/Danmarkshavn'), ('America/Dawson', 'America/Dawson'), ('America/Dawson_Creek', 'America/Dawson_Creek'), ('America/Denver', 'America/Denver'), ('America/Detroit', 'America/Detroit'), ('America/Dominica', 'America/Dominica'), ('America/Edmonton', 'America/Edmonton'), ('America/Eirunepe', 'America/Eirunepe'), ('America/El_Salvador', 'America/El_Salvador'), ('America/Fort_Nelson', 'America/Fort_Nelson'), ('America/Fortaleza', 'America/Fortaleza'), ('America/Glace_Bay', 'America/Glace_Bay'), ('America/Goose_Bay', 'America/Goose_Bay'), ('America/Grand_Turk', 'America/Grand_Turk'), ('America/Grenada', 'America/Grenada'), ('America/Guadeloupe', 'America/Guadeloupe'), ('America/Guatemala', 'America/Guatemala'), ('America/Guayaquil', 'America/Guayaquil'), ('America/Guyana', 'America/Guyana'), ('America/Halifax', 'America/Halifax'), ('America/Havana', 'America/Havana'), ('America/Hermosillo', 'America/Hermosillo'), ('America/Indiana/Indianapolis', 'America/Indiana/Indianapolis'), ('America/Indiana/Knox', 'America/Indiana/Knox'), ('America/Indiana/Marengo', 'America/Indiana/Marengo'), ('America/Indiana/Petersburg', 'America/Indiana/Petersburg'), ('America/Indiana/Tell_City', 'America/Indiana/Tell_City'), ('America/Indiana/Vevay', 'America/Indiana/Vevay'), ('America/Indiana/Vincennes', 'America/Indiana/Vincennes'), ('America/Indiana/Winamac', 'America/Indiana/Winamac'), ('America/Inuvik', 'America/Inuvik'), ('America/Iqaluit', 'America/Iqaluit'), ('America/Jamaica', 'America/Jamaica'), ('America/Juneau', 'America/Juneau'), ('America/Kentucky/Louisville', 'America/Kentucky/Louisville'), ('America/Kentucky/Monticello', 'America/Kentucky/Monticello'), ('America/La_Paz', 'America/La_Paz'), ('America/Lima', 'America/Lima'), ('America/Los_Angeles', 'America/Los_Angeles'), ('America/Maceio', 'America/Maceio'), ('America/Managua', 'America/Managua'), ('America/Manaus', 'America/Manaus'), ('America/Martinique', 'America/Martinique'), ('America/Matamoros', 'America/Matamoros'), ('America/Mazatlan', 'America/Mazatlan'), ('America/Menominee', 'America/Menominee'), ('America/Merida', 'America/Merida'), ('America/Metlakatla', 'America/Metlakatla'), ('America/Mexico_City', 'America/Mexico_City'), ('America/Miquelon', 'America/Miquelon'), ('America/Moncton', 'America/Moncton'), ('America/Monterrey', 'America/Monterrey'), ('America/Montevideo', 'America/Montevideo'), ('America/Montserrat', 'America/Montserrat'), ('America/Nassau', 'America/Nassau'), ('America/New_York', 'America/New_York'), ('America/Nipigon', 'America/Nipigon'), ('America/Nome', 'America/Nome'), ('America/Noronha', 'America/Noronha'), ('America/North_Dakota/Beulah', 'America/North_Dakota/Beulah'), ('America/North_Dakota/Center', 'America/North_Dakota/Center'), ('America/North_Dakota/New_Salem', 'America/North_Dakota/New_Salem'), ('America/Nuuk', 'America/Nuuk'), ('America/Ojinaga', 'America/Ojinaga'), ('America/Panama', 'America/Panama'), ('America/Pangnirtung', 'America/Pangnirtung'), ('America/Paramaribo', 'America/Paramaribo'), ('America/Phoenix', 'America/Phoenix'), ('America/Port-au-Prince', 'America/Port-au-Prince'), ('America/Port_of_Spain', 'America/Port_of_Spain'), ('America/Porto_Velho', 'America/Porto_Velho'), ('America/Puerto_Rico', 'America/Puerto_Rico'), ('America/Punta_Arenas', 'America/Punta_Arenas'), ('America/Rainy_River', 'America/Rainy_River'), ('America/Rankin_Inlet', 'America/Rankin_Inlet'), ('America/Recife', 'America/Recife'), ('America/Regina', 'America/Regina'), ('America/Resolute', 'America/Resolute'), ('America/Rio_Branco', 'America/Rio_Branco'), ('America/Santarem', 'America/Santarem'), ('America/Santiago', 'America/Santiago'), ('America/Santo_Domingo', 'America/Santo_Domingo'), ('America/Sao_Paulo', 'America/Sao_Paulo'), ('America/Scoresbysund', 'America/Scoresbysund'), ('America/Sitka', 'America/Sitka'), ('America/St_Johns', 'America/St_Johns'), ('America/St_Kitts', 'America/St_Kitts'), ('America/St_Lucia', 'America/St_Lucia'), ('America/St_Thomas', 'America/St_Thomas'), ('America/St_Vincent', 'America/St_Vincent'), ('America/Swift_Current', 'America/Swift_Current'), ('America/Tegucigalpa', 'America/Tegucigalpa'), ('America/Thule', 'America/Thule'), ('America/Thunder_Bay', 'America/Thunder_Bay'), ('America/Tijuana', 'America/Tijuana'), ('America/Toronto', 'America/Toronto'), ('America/Tortola', 'America/Tortola'), ('America/Vancouver', 'America/Vancouver'), ('America/Whitehorse', 'America/Whitehorse'), ('America/Winnipeg', 'America/Winnipeg'), ('America/Yakutat', 'America/Yakutat'), ('America/Yellowknife', 'America/Yellowknife'), ('Antarctica/Casey', 'Antarctica/Casey'), ('Antarctica/Davis', 'Antarctica/Davis'), ('Antarctica/DumontDUrville', 'Antarctica/DumontDUrville'), ('Antarctica/Macquarie', 'Antarctica/Macquarie'), ('Antarctica/Mawson', 'Antarctica/Mawson'), ('Antarctica/McMurdo', 'Antarctica/McMurdo'), ('Antarctica/Palmer', 'Antarctica/Palmer'), ('Antarctica/Rothera', 'Antarctica/Rothera'), ('Antarctica/Syowa', 'Antarctica/Syowa'), ('Antarctica/Troll', 'Antarctica/Troll'), ('Antarctica/Vostok', 'Antarctica/Vostok'), ('Asia/Aden', 'Asia/Aden'), ('Asia/Almaty', 'Asia/Almaty'), ('Asia/Amman', 'Asia/Amman'), ('Asia/Anadyr', 'Asia/Anadyr'), ('Asia/Aqtau', 'Asia/Aqtau'), ('Asia/Aqtobe', 'Asia/Aqtobe'), ('Asia/Ashgabat', 'Asia/Ashgabat'), ('Asia/Atyrau', 'Asia/Atyrau'), ('Asia/Baghdad', 'Asia/Baghdad'), ('Asia/Bahrain', 'Asia/Bahrain'), ('Asia/Baku', 'Asia/Baku'), ('Asia/Bangkok', 'Asia/Bangkok'), ('Asia/Barnaul', 'Asia/Barnaul'), ('Asia/Beirut', 'Asia/Beirut'), ('Asia/Bishkek', 'Asia/Bishkek'), ('Asia/Brunei', 'Asia/Brunei'), ('Asia/Chita', 'Asia/Chita'), ('Asia/Choibalsan', 'Asia/Choibalsan'), ('Asia/Colombo', 'Asia/Colombo'), ('Asia/Damascus', 'Asia/Damascus'), ('Asia/Dhaka', 'Asia/Dhaka'), ('Asia/Dili', 'Asia/Dili'), ('Asia/Dubai', 'Asia/Dubai'), ('Asia/Dushanbe', 'Asia/Dushanbe'), ('Asia/Famagusta', 'Asia/Famagusta'), ('Asia/Gaza', 'Asia/Gaza'), ('Asia/Hebron', 'Asia/Hebron'), ('Asia/Ho_Chi_Minh', 'Asia/Ho_Chi_Minh'), ('Asia/Hong_Kong', 'Asia/Hong_Kong'), ('Asia/Hovd', 'Asia/Hovd'), ('Asia/Irkutsk', 'Asia/Irkutsk'), ('Asia/Jakarta', 'Asia/Jakarta'), ('Asia/Jayapura', 'Asia/Jayapura'), ('Asia/Jerusalem', 'Asia/Jerusalem'), ('Asia/Kabul', 'Asia/Kabul'), ('Asia/Kamchatka', 'Asia/Kamchatka'), ('Asia/Karachi', 'Asia/Karachi'), ('Asia/Kathmandu', 'Asia/Kathmandu'), ('Asia/Khandyga', 'Asia/Khandyga'), ('Asia/Kolkata', 'Asia/Kolkata'), ('Asia/Krasnoyarsk', 'Asia/Krasnoyarsk'), ('Asia/Kuala_Lumpur', 'Asia/Kuala_Lumpur'), ('Asia/Kuching', 'Asia/Kuching'), ('Asia/Kuwait', 'Asia/Kuwait'), ('Asia/Macau', 'Asia/Macau'), ('Asia/Magadan', 'Asia/Magadan'), ('Asia/Makassar', 'Asia/Makassar'), ('Asia/Manila', 'Asia/Manila'), ('Asia/Muscat', 'Asia/Muscat'), ('Asia/Nicosia', 'Asia/Nicosia'), ('Asia/Novokuznetsk', 'Asia/Novokuznetsk'), ('Asia/Novosibirsk', 'Asia/Novosibirsk'), ('Asia/Omsk', 'Asia/Omsk'), ('Asia/Oral', 'Asia/Oral'), ('Asia/Phnom_Penh', 'Asia/Phnom_Penh'), ('Asia/Pontianak', 'Asia/Pontianak'), ('Asia/Pyongyang', 'Asia/Pyongyang'), ('Asia/Qatar', 'Asia/Qatar'), ('Asia/Qostanay', 'Asia/Qostanay'), ('Asia/Qyzylorda', 'Asia/Qyzylorda'), ('Asia/Riyadh', 'Asia/Riyadh'), ('Asia/Sakhalin', 'Asia/Sakhalin'), ('Asia/Samarkand', 'Asia/Samarkand'), ('Asia/Seoul', 'Asia/Seoul'), ('Asia/Shanghai', 'Asia/Shanghai'), ('Asia/Singapore', 'Asia/Singapore'), ('Asia/Srednekolymsk', 'Asia/Srednekolymsk'), ('Asia/Taipei', 'Asia/Taipei'), ('Asia/Tashkent', 'Asia/Tashkent'), ('Asia/Tbilisi', 'Asia/Tbilisi'), ('Asia/Tehran', 'Asia/Tehran'), ('Asia/Thimphu', 'Asia/Thimphu'), ('Asia/Tokyo', 'Asia/Tokyo'), ('Asia/Tomsk', 'Asia/Tomsk'), ('Asia/Ulaanbaatar', 'Asia/Ulaanbaatar'), ('Asia/Urumqi', 'Asia/Urumqi'), ('Asia/Ust-Nera', 'Asia/Ust-Nera'), ('Asia/Vientiane', 'Asia/Vientiane'), ('Asia/Vladivostok', 'Asia/Vladivostok'), ('Asia/Yakutsk', 'Asia/Yakutsk'), ('Asia/Yangon', 'Asia/Yangon'), ('Asia/Yekaterinburg', 'Asia/Yekaterinburg'), ('Asia/Yerevan', 'Asia/Yerevan'), ('Atlantic/Azores', 'Atlantic/Azores'), ('Atlantic/Bermuda', 'Atlantic/Bermuda'), ('Atlantic/Canary', 'Atlantic/Canary'), ('Atlantic/Cape_Verde', 'Atlantic/Cape_Verde'), ('Atlantic/Faroe', 'Atlantic/Faroe'), ('Atlantic/Madeira', 'Atlantic/Madeira'), ('Atlantic/Reykjavik', 'Atlantic/Reykjavik'), ('Atlantic/South_Georgia', 'Atlantic/South_Georgia'), ('Atlantic/St_Helena', 'Atlantic/St_Helena'), ('Atlantic/Stanley', 'Atlantic/Stanley'), ('Australia/Adelaide', 'Australia/Adelaide'), ('Australia/Brisbane', 'Australia/Brisbane'), ('Australia/Broken_Hill', 'Australia/Broken_Hill'), ('Australia/Darwin', 'Australia/Darwin'), ('Australia/Eucla', 'Australia/Eucla'), ('Australia/Hobart', 'Australia/Hobart'), ('Australia/Lindeman', 'Australia/Lindeman'), ('Australia/Lord_Howe', 'Australia/Lord_Howe'), ('Australia/Melbourne', 'Australia/Melbourne'), ('Australia/Perth', 'Australia/Perth'), ('Australia/Sydney', 'Australia/Sydney'), ('Europe/Amsterdam', 'Europe/Amsterdam'), ('Europe/Andorra', 'Europe/Andorra'), ('Europe/Astrakhan', 'Europe/Astrakhan'), ('Europe/Athens', 'Europe/Athens'), ('Europe/Belgrade', 'Europe/Belgrade'), ('Europe/Berlin', 'Europe/Berlin'), ('Europe/Brussels', 'Europe/Brussels'), ('Europe/Bucharest', 'Europe/Bucharest'), ('Europe/Budapest', 'Europe/Budapest'), ('Europe/Chisinau', 'Europe/Chisinau'), ('Europe/Copenhagen', 'Europe/Copenhagen'), ('Europe/Dublin', 'Europe/Dublin'), ('Europe/Gibraltar', 'Europe/Gibraltar'), ('Europe/Helsinki', 'Europe/Helsinki'), ('Europe/Istanbul', 'Europe/Istanbul'), ('Europe/Kaliningrad', 'Europe/Kaliningrad'), ('Europe/Kirov', 'Europe/Kirov'), ('Europe/Kyiv', 'Europe/Kyiv'), ('Europe/Lisbon', 'Europe/Lisbon'), ('Europe/London', 'Europe/London'), ('Europe/Luxembourg', 'Europe/Luxembourg'), ('Europe/Madrid', 'Europe/Madrid'), ('Europe/Malta', 'Europe/Malta'), ('Europe/Minsk', 'Europe/Minsk'), ('Europe/Monaco', 'Europe/Monaco'), ('Europe/Moscow', 'Europe/Moscow'), ('Europe/Oslo', 'Europe/Oslo'), ('Europe/Paris', 'Europe/Paris'), ('Europe/Prague', 'Europe/Prague'), ('Europe/Riga', 'Europe/Riga'), ('Europe/Rome', 'Europe/Rome'), ('Europe/Samara', 'Europe/Samara'), ('Europe/Saratov', 'Europe/Saratov'), ('Europe/Simferopol', 'Europe/Simferopol'), ('Europe/Sofia', 'Europe/Sofia'), ('Europe/Stockholm', 'Europe/Stockholm'), ('Europe/Tallinn', 'Europe/Tallinn'), ('Europe/Tirane', 'Europe/Tirane'), ('Europe/Ulyanovsk', 'Europe/Ulyanovsk'), ('Europe/Uzhgorod', 'Europe/Uzhgorod'), ('Europe/Vaduz', 'Europe/Vaduz'), ('Europe/Vienna', 'Europe/Vienna'), ('Europe/Vilnius', 'Europe/Vilnius'), ('Europe/Volgograd', 'Europe/Volgograd'), ('Europe/Warsaw', 'Europe/Warsaw'), ('Europe/Zaporozhye', 'Europe/Zaporozhye'), ('Europe/Zurich', 'Europe/Zurich'), ('GMT', 'GMT'), ('Indian/Antananarivo', 'Indian/Antananarivo'), ('Indian/Chagos', 'Indian/Chagos'), ('Indian/Christmas', 'Indian/Christmas'), ('Indian/Cocos', 'Indian/Cocos'), ('Indian/Comoro', 'Indian/Comoro'), ('Indian/Kerguelen', 'Indian/Kerguelen'), ('Indian/Mahe', 'Indian/Mahe'), ('Indian/Maldives', 'Indian/Maldives'), ('Indian/Mauritius', 'Indian/Mauritius'), ('Indian/Mayotte', 'Indian/Mayotte'), ('Indian/Reunion', 'Indian/Reunion'), ('Pacific/Apia', 'Pacific/Apia'), ('Pacific/Auckland', 'Pacific/Auckland'), ('Pacific/Bougainville', 'Pacific/Bougainville'), ('Pacific/Chatham', 'Pacific/Chatham'), ('Pacific/Chuuk', 'Pacific/Chuuk'), ('Pacific/Easter', 'Pacific/Easter'), ('Pacific/Efate', 'Pacific/Efate'), ('Pacific/Fakaofo', 'Pacific/Fakaofo'), ('Pacific/Fiji', 'Pacific/Fiji'), ('Pacific/Funafuti', 'Pacific/Funafuti'), ('Pacific/Galapagos', 'Pacific/Galapagos'), ('Pacific/Gambier', 'Pacific/Gambier'), ('Pacific/Guadalcanal', 'Pacific/Guadalcanal'), ('Pacific/Guam', 'Pacific/Guam'), ('Pacific/Honolulu', 'Pacific/Honolulu'), ('Pacific/Kanton', 'Pacific/Kanton'), ('Pacific/Kiritimati', 'Pacific/Kiritimati'), ('Pacific/Kosrae', 'Pacific/Kosrae'), ('Pacific/Kwajalein', 'Pacific/Kwajalein'), ('Pacific/Majuro', 'Pacific/Majuro'), ('Pacific/Marquesas', 'Pacific/Marquesas'), ('Pacific/Midway', 'Pacific/Midway'), ('Pacific/Nauru', 'Pacific/Nauru'), ('Pacific/Niue', 'Pacific/Niue'), ('Pacific/Norfolk', 'Pacific/Norfolk'), ('Pacific/Noumea', 'Pacific/Noumea'), ('Pacific/Pago_Pago', 'Pacific/Pago_Pago'), ('Pacific/Palau', 'Pacific/Palau'), ('Pacific/Pitcairn', 'Pacific/Pitcairn'), ('Pacific/Pohnpei', 'Pacific/Pohnpei'), ('Pacific/Port_Moresby', 'Pacific/Port_Moresby'), ('Pacific/Rarotonga', 'Pacific/Rarotonga'), ('Pacific/Saipan', 'Pacific/Saipan'), ('Pacific/Tahiti', 'Pacific/Tahiti'), ('Pacific/Tarawa', 'Pacific/Tarawa'), ('Pacific/Tongatapu', 'Pacific/Tongatapu'), ('Pacific/Wake', 'Pacific/Wake'), ('Pacific/Wallis', 'Pacific/Wallis'), ('UTC', 'UTC')], max_length=255), - ), - ] diff --git a/ietf/meeting/migrations/0056_use_timezone_now_for_meeting_models.py b/ietf/meeting/migrations/0056_use_timezone_now_for_meeting_models.py deleted file mode 100644 index 83d20fd57a..0000000000 --- a/ietf/meeting/migrations/0056_use_timezone_now_for_meeting_models.py +++ /dev/null @@ -1,19 +0,0 @@ -# Generated by Django 2.2.28 on 2022-07-12 11:24 - -from django.db import migrations, models -import django.utils.timezone - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0055_pytz_2022_2_1'), - ] - - operations = [ - migrations.AlterField( - model_name='schedulingevent', - name='time', - field=models.DateTimeField(default=django.utils.timezone.now, help_text='When the event happened'), - ), - ] diff --git a/ietf/meeting/migrations/0057_fill_in_empty_meeting_time_zone.py b/ietf/meeting/migrations/0057_fill_in_empty_meeting_time_zone.py deleted file mode 100644 index f009b08f3b..0000000000 --- a/ietf/meeting/migrations/0057_fill_in_empty_meeting_time_zone.py +++ /dev/null @@ -1,47 +0,0 @@ -# Generated by Django 2.2.28 on 2022-08-08 11:37 - -import datetime - -from django.db import migrations - - -# date of last meeting with an empty time_zone before this migration -LAST_EMPTY_TZ = datetime.date(2022, 7, 1) - - -def forward(apps, schema_editor): - Meeting = apps.get_model('meeting', 'Meeting') - - # Check that we will be able to identify the migrated meetings later - old_meetings_in_pst8pdt = Meeting.objects.filter(type_id='interim', time_zone='PST8PDT', date__lte=LAST_EMPTY_TZ) - assert old_meetings_in_pst8pdt.count() == 0, 'not expecting interim meetings in PST8PDT time_zone' - - meetings_with_empty_tz = Meeting.objects.filter(time_zone='') - # check our expected conditions - for mtg in meetings_with_empty_tz: - assert mtg.type_id == 'interim', 'was not expecting non-interim meetings to be affected' - assert mtg.date <= LAST_EMPTY_TZ, 'affected meeting outside expected date range' - mtg.time_zone = 'PST8PDT' - - # commit the changes - Meeting.objects.bulk_update(meetings_with_empty_tz, ['time_zone']) - - -def reverse(apps, schema_editor): - Meeting = apps.get_model('meeting', 'Meeting') - meetings_to_restore = Meeting.objects.filter(time_zone='PST8PDT', date__lte=LAST_EMPTY_TZ) - for mtg in meetings_to_restore: - mtg.time_zone = '' - # commit the changes - Meeting.objects.bulk_update(meetings_to_restore, ['time_zone']) - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0056_use_timezone_now_for_meeting_models'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/meeting/migrations/0058_meeting_time_zone_not_blank.py b/ietf/meeting/migrations/0058_meeting_time_zone_not_blank.py deleted file mode 100644 index 0adec77a28..0000000000 --- a/ietf/meeting/migrations/0058_meeting_time_zone_not_blank.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.2.28 on 2022-08-25 12:09 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('meeting', '0057_fill_in_empty_meeting_time_zone'), - ] - - operations = [ - migrations.AlterField( - model_name='meeting', - name='time_zone', - field=models.CharField(choices=[('', '---------'), ('Africa/Abidjan', 'Africa/Abidjan'), ('Africa/Accra', 'Africa/Accra'), ('Africa/Addis_Ababa', 'Africa/Addis_Ababa'), ('Africa/Algiers', 'Africa/Algiers'), ('Africa/Asmara', 'Africa/Asmara'), ('Africa/Bamako', 'Africa/Bamako'), ('Africa/Bangui', 'Africa/Bangui'), ('Africa/Banjul', 'Africa/Banjul'), ('Africa/Bissau', 'Africa/Bissau'), ('Africa/Blantyre', 'Africa/Blantyre'), ('Africa/Brazzaville', 'Africa/Brazzaville'), ('Africa/Bujumbura', 'Africa/Bujumbura'), ('Africa/Cairo', 'Africa/Cairo'), ('Africa/Casablanca', 'Africa/Casablanca'), ('Africa/Ceuta', 'Africa/Ceuta'), ('Africa/Conakry', 'Africa/Conakry'), ('Africa/Dakar', 'Africa/Dakar'), ('Africa/Dar_es_Salaam', 'Africa/Dar_es_Salaam'), ('Africa/Djibouti', 'Africa/Djibouti'), ('Africa/Douala', 'Africa/Douala'), ('Africa/El_Aaiun', 'Africa/El_Aaiun'), ('Africa/Freetown', 'Africa/Freetown'), ('Africa/Gaborone', 'Africa/Gaborone'), ('Africa/Harare', 'Africa/Harare'), ('Africa/Johannesburg', 'Africa/Johannesburg'), ('Africa/Juba', 'Africa/Juba'), ('Africa/Kampala', 'Africa/Kampala'), ('Africa/Khartoum', 'Africa/Khartoum'), ('Africa/Kigali', 'Africa/Kigali'), ('Africa/Kinshasa', 'Africa/Kinshasa'), ('Africa/Lagos', 'Africa/Lagos'), ('Africa/Libreville', 'Africa/Libreville'), ('Africa/Lome', 'Africa/Lome'), ('Africa/Luanda', 'Africa/Luanda'), ('Africa/Lubumbashi', 'Africa/Lubumbashi'), ('Africa/Lusaka', 'Africa/Lusaka'), ('Africa/Malabo', 'Africa/Malabo'), ('Africa/Maputo', 'Africa/Maputo'), ('Africa/Maseru', 'Africa/Maseru'), ('Africa/Mbabane', 'Africa/Mbabane'), ('Africa/Mogadishu', 'Africa/Mogadishu'), ('Africa/Monrovia', 'Africa/Monrovia'), ('Africa/Nairobi', 'Africa/Nairobi'), ('Africa/Ndjamena', 'Africa/Ndjamena'), ('Africa/Niamey', 'Africa/Niamey'), ('Africa/Nouakchott', 'Africa/Nouakchott'), ('Africa/Ouagadougou', 'Africa/Ouagadougou'), ('Africa/Porto-Novo', 'Africa/Porto-Novo'), ('Africa/Sao_Tome', 'Africa/Sao_Tome'), ('Africa/Tripoli', 'Africa/Tripoli'), ('Africa/Tunis', 'Africa/Tunis'), ('Africa/Windhoek', 'Africa/Windhoek'), ('America/Adak', 'America/Adak'), ('America/Anchorage', 'America/Anchorage'), ('America/Anguilla', 'America/Anguilla'), ('America/Antigua', 'America/Antigua'), ('America/Araguaina', 'America/Araguaina'), ('America/Argentina/Buenos_Aires', 'America/Argentina/Buenos_Aires'), ('America/Argentina/Catamarca', 'America/Argentina/Catamarca'), ('America/Argentina/Cordoba', 'America/Argentina/Cordoba'), ('America/Argentina/Jujuy', 'America/Argentina/Jujuy'), ('America/Argentina/La_Rioja', 'America/Argentina/La_Rioja'), ('America/Argentina/Mendoza', 'America/Argentina/Mendoza'), ('America/Argentina/Rio_Gallegos', 'America/Argentina/Rio_Gallegos'), ('America/Argentina/Salta', 'America/Argentina/Salta'), ('America/Argentina/San_Juan', 'America/Argentina/San_Juan'), ('America/Argentina/San_Luis', 'America/Argentina/San_Luis'), ('America/Argentina/Tucuman', 'America/Argentina/Tucuman'), ('America/Argentina/Ushuaia', 'America/Argentina/Ushuaia'), ('America/Aruba', 'America/Aruba'), ('America/Asuncion', 'America/Asuncion'), ('America/Atikokan', 'America/Atikokan'), ('America/Bahia', 'America/Bahia'), ('America/Bahia_Banderas', 'America/Bahia_Banderas'), ('America/Barbados', 'America/Barbados'), ('America/Belem', 'America/Belem'), ('America/Belize', 'America/Belize'), ('America/Blanc-Sablon', 'America/Blanc-Sablon'), ('America/Boa_Vista', 'America/Boa_Vista'), ('America/Bogota', 'America/Bogota'), ('America/Boise', 'America/Boise'), ('America/Cambridge_Bay', 'America/Cambridge_Bay'), ('America/Campo_Grande', 'America/Campo_Grande'), ('America/Cancun', 'America/Cancun'), ('America/Caracas', 'America/Caracas'), ('America/Cayenne', 'America/Cayenne'), ('America/Cayman', 'America/Cayman'), ('America/Chicago', 'America/Chicago'), ('America/Chihuahua', 'America/Chihuahua'), ('America/Costa_Rica', 'America/Costa_Rica'), ('America/Creston', 'America/Creston'), ('America/Cuiaba', 'America/Cuiaba'), ('America/Curacao', 'America/Curacao'), ('America/Danmarkshavn', 'America/Danmarkshavn'), ('America/Dawson', 'America/Dawson'), ('America/Dawson_Creek', 'America/Dawson_Creek'), ('America/Denver', 'America/Denver'), ('America/Detroit', 'America/Detroit'), ('America/Dominica', 'America/Dominica'), ('America/Edmonton', 'America/Edmonton'), ('America/Eirunepe', 'America/Eirunepe'), ('America/El_Salvador', 'America/El_Salvador'), ('America/Fort_Nelson', 'America/Fort_Nelson'), ('America/Fortaleza', 'America/Fortaleza'), ('America/Glace_Bay', 'America/Glace_Bay'), ('America/Goose_Bay', 'America/Goose_Bay'), ('America/Grand_Turk', 'America/Grand_Turk'), ('America/Grenada', 'America/Grenada'), ('America/Guadeloupe', 'America/Guadeloupe'), ('America/Guatemala', 'America/Guatemala'), ('America/Guayaquil', 'America/Guayaquil'), ('America/Guyana', 'America/Guyana'), ('America/Halifax', 'America/Halifax'), ('America/Havana', 'America/Havana'), ('America/Hermosillo', 'America/Hermosillo'), ('America/Indiana/Indianapolis', 'America/Indiana/Indianapolis'), ('America/Indiana/Knox', 'America/Indiana/Knox'), ('America/Indiana/Marengo', 'America/Indiana/Marengo'), ('America/Indiana/Petersburg', 'America/Indiana/Petersburg'), ('America/Indiana/Tell_City', 'America/Indiana/Tell_City'), ('America/Indiana/Vevay', 'America/Indiana/Vevay'), ('America/Indiana/Vincennes', 'America/Indiana/Vincennes'), ('America/Indiana/Winamac', 'America/Indiana/Winamac'), ('America/Inuvik', 'America/Inuvik'), ('America/Iqaluit', 'America/Iqaluit'), ('America/Jamaica', 'America/Jamaica'), ('America/Juneau', 'America/Juneau'), ('America/Kentucky/Louisville', 'America/Kentucky/Louisville'), ('America/Kentucky/Monticello', 'America/Kentucky/Monticello'), ('America/La_Paz', 'America/La_Paz'), ('America/Lima', 'America/Lima'), ('America/Los_Angeles', 'America/Los_Angeles'), ('America/Maceio', 'America/Maceio'), ('America/Managua', 'America/Managua'), ('America/Manaus', 'America/Manaus'), ('America/Martinique', 'America/Martinique'), ('America/Matamoros', 'America/Matamoros'), ('America/Mazatlan', 'America/Mazatlan'), ('America/Menominee', 'America/Menominee'), ('America/Merida', 'America/Merida'), ('America/Metlakatla', 'America/Metlakatla'), ('America/Mexico_City', 'America/Mexico_City'), ('America/Miquelon', 'America/Miquelon'), ('America/Moncton', 'America/Moncton'), ('America/Monterrey', 'America/Monterrey'), ('America/Montevideo', 'America/Montevideo'), ('America/Montserrat', 'America/Montserrat'), ('America/Nassau', 'America/Nassau'), ('America/New_York', 'America/New_York'), ('America/Nipigon', 'America/Nipigon'), ('America/Nome', 'America/Nome'), ('America/Noronha', 'America/Noronha'), ('America/North_Dakota/Beulah', 'America/North_Dakota/Beulah'), ('America/North_Dakota/Center', 'America/North_Dakota/Center'), ('America/North_Dakota/New_Salem', 'America/North_Dakota/New_Salem'), ('America/Nuuk', 'America/Nuuk'), ('America/Ojinaga', 'America/Ojinaga'), ('America/Panama', 'America/Panama'), ('America/Pangnirtung', 'America/Pangnirtung'), ('America/Paramaribo', 'America/Paramaribo'), ('America/Phoenix', 'America/Phoenix'), ('America/Port-au-Prince', 'America/Port-au-Prince'), ('America/Port_of_Spain', 'America/Port_of_Spain'), ('America/Porto_Velho', 'America/Porto_Velho'), ('America/Puerto_Rico', 'America/Puerto_Rico'), ('America/Punta_Arenas', 'America/Punta_Arenas'), ('America/Rainy_River', 'America/Rainy_River'), ('America/Rankin_Inlet', 'America/Rankin_Inlet'), ('America/Recife', 'America/Recife'), ('America/Regina', 'America/Regina'), ('America/Resolute', 'America/Resolute'), ('America/Rio_Branco', 'America/Rio_Branco'), ('America/Santarem', 'America/Santarem'), ('America/Santiago', 'America/Santiago'), ('America/Santo_Domingo', 'America/Santo_Domingo'), ('America/Sao_Paulo', 'America/Sao_Paulo'), ('America/Scoresbysund', 'America/Scoresbysund'), ('America/Sitka', 'America/Sitka'), ('America/St_Johns', 'America/St_Johns'), ('America/St_Kitts', 'America/St_Kitts'), ('America/St_Lucia', 'America/St_Lucia'), ('America/St_Thomas', 'America/St_Thomas'), ('America/St_Vincent', 'America/St_Vincent'), ('America/Swift_Current', 'America/Swift_Current'), ('America/Tegucigalpa', 'America/Tegucigalpa'), ('America/Thule', 'America/Thule'), ('America/Thunder_Bay', 'America/Thunder_Bay'), ('America/Tijuana', 'America/Tijuana'), ('America/Toronto', 'America/Toronto'), ('America/Tortola', 'America/Tortola'), ('America/Vancouver', 'America/Vancouver'), ('America/Whitehorse', 'America/Whitehorse'), ('America/Winnipeg', 'America/Winnipeg'), ('America/Yakutat', 'America/Yakutat'), ('America/Yellowknife', 'America/Yellowknife'), ('Antarctica/Casey', 'Antarctica/Casey'), ('Antarctica/Davis', 'Antarctica/Davis'), ('Antarctica/DumontDUrville', 'Antarctica/DumontDUrville'), ('Antarctica/Macquarie', 'Antarctica/Macquarie'), ('Antarctica/Mawson', 'Antarctica/Mawson'), ('Antarctica/McMurdo', 'Antarctica/McMurdo'), ('Antarctica/Palmer', 'Antarctica/Palmer'), ('Antarctica/Rothera', 'Antarctica/Rothera'), ('Antarctica/Syowa', 'Antarctica/Syowa'), ('Antarctica/Troll', 'Antarctica/Troll'), ('Antarctica/Vostok', 'Antarctica/Vostok'), ('Asia/Aden', 'Asia/Aden'), ('Asia/Almaty', 'Asia/Almaty'), ('Asia/Amman', 'Asia/Amman'), ('Asia/Anadyr', 'Asia/Anadyr'), ('Asia/Aqtau', 'Asia/Aqtau'), ('Asia/Aqtobe', 'Asia/Aqtobe'), ('Asia/Ashgabat', 'Asia/Ashgabat'), ('Asia/Atyrau', 'Asia/Atyrau'), ('Asia/Baghdad', 'Asia/Baghdad'), ('Asia/Bahrain', 'Asia/Bahrain'), ('Asia/Baku', 'Asia/Baku'), ('Asia/Bangkok', 'Asia/Bangkok'), ('Asia/Barnaul', 'Asia/Barnaul'), ('Asia/Beirut', 'Asia/Beirut'), ('Asia/Bishkek', 'Asia/Bishkek'), ('Asia/Brunei', 'Asia/Brunei'), ('Asia/Chita', 'Asia/Chita'), ('Asia/Choibalsan', 'Asia/Choibalsan'), ('Asia/Colombo', 'Asia/Colombo'), ('Asia/Damascus', 'Asia/Damascus'), ('Asia/Dhaka', 'Asia/Dhaka'), ('Asia/Dili', 'Asia/Dili'), ('Asia/Dubai', 'Asia/Dubai'), ('Asia/Dushanbe', 'Asia/Dushanbe'), ('Asia/Famagusta', 'Asia/Famagusta'), ('Asia/Gaza', 'Asia/Gaza'), ('Asia/Hebron', 'Asia/Hebron'), ('Asia/Ho_Chi_Minh', 'Asia/Ho_Chi_Minh'), ('Asia/Hong_Kong', 'Asia/Hong_Kong'), ('Asia/Hovd', 'Asia/Hovd'), ('Asia/Irkutsk', 'Asia/Irkutsk'), ('Asia/Jakarta', 'Asia/Jakarta'), ('Asia/Jayapura', 'Asia/Jayapura'), ('Asia/Jerusalem', 'Asia/Jerusalem'), ('Asia/Kabul', 'Asia/Kabul'), ('Asia/Kamchatka', 'Asia/Kamchatka'), ('Asia/Karachi', 'Asia/Karachi'), ('Asia/Kathmandu', 'Asia/Kathmandu'), ('Asia/Khandyga', 'Asia/Khandyga'), ('Asia/Kolkata', 'Asia/Kolkata'), ('Asia/Krasnoyarsk', 'Asia/Krasnoyarsk'), ('Asia/Kuala_Lumpur', 'Asia/Kuala_Lumpur'), ('Asia/Kuching', 'Asia/Kuching'), ('Asia/Kuwait', 'Asia/Kuwait'), ('Asia/Macau', 'Asia/Macau'), ('Asia/Magadan', 'Asia/Magadan'), ('Asia/Makassar', 'Asia/Makassar'), ('Asia/Manila', 'Asia/Manila'), ('Asia/Muscat', 'Asia/Muscat'), ('Asia/Nicosia', 'Asia/Nicosia'), ('Asia/Novokuznetsk', 'Asia/Novokuznetsk'), ('Asia/Novosibirsk', 'Asia/Novosibirsk'), ('Asia/Omsk', 'Asia/Omsk'), ('Asia/Oral', 'Asia/Oral'), ('Asia/Phnom_Penh', 'Asia/Phnom_Penh'), ('Asia/Pontianak', 'Asia/Pontianak'), ('Asia/Pyongyang', 'Asia/Pyongyang'), ('Asia/Qatar', 'Asia/Qatar'), ('Asia/Qostanay', 'Asia/Qostanay'), ('Asia/Qyzylorda', 'Asia/Qyzylorda'), ('Asia/Riyadh', 'Asia/Riyadh'), ('Asia/Sakhalin', 'Asia/Sakhalin'), ('Asia/Samarkand', 'Asia/Samarkand'), ('Asia/Seoul', 'Asia/Seoul'), ('Asia/Shanghai', 'Asia/Shanghai'), ('Asia/Singapore', 'Asia/Singapore'), ('Asia/Srednekolymsk', 'Asia/Srednekolymsk'), ('Asia/Taipei', 'Asia/Taipei'), ('Asia/Tashkent', 'Asia/Tashkent'), ('Asia/Tbilisi', 'Asia/Tbilisi'), ('Asia/Tehran', 'Asia/Tehran'), ('Asia/Thimphu', 'Asia/Thimphu'), ('Asia/Tokyo', 'Asia/Tokyo'), ('Asia/Tomsk', 'Asia/Tomsk'), ('Asia/Ulaanbaatar', 'Asia/Ulaanbaatar'), ('Asia/Urumqi', 'Asia/Urumqi'), ('Asia/Ust-Nera', 'Asia/Ust-Nera'), ('Asia/Vientiane', 'Asia/Vientiane'), ('Asia/Vladivostok', 'Asia/Vladivostok'), ('Asia/Yakutsk', 'Asia/Yakutsk'), ('Asia/Yangon', 'Asia/Yangon'), ('Asia/Yekaterinburg', 'Asia/Yekaterinburg'), ('Asia/Yerevan', 'Asia/Yerevan'), ('Atlantic/Azores', 'Atlantic/Azores'), ('Atlantic/Bermuda', 'Atlantic/Bermuda'), ('Atlantic/Canary', 'Atlantic/Canary'), ('Atlantic/Cape_Verde', 'Atlantic/Cape_Verde'), ('Atlantic/Faroe', 'Atlantic/Faroe'), ('Atlantic/Madeira', 'Atlantic/Madeira'), ('Atlantic/Reykjavik', 'Atlantic/Reykjavik'), ('Atlantic/South_Georgia', 'Atlantic/South_Georgia'), ('Atlantic/St_Helena', 'Atlantic/St_Helena'), ('Atlantic/Stanley', 'Atlantic/Stanley'), ('Australia/Adelaide', 'Australia/Adelaide'), ('Australia/Brisbane', 'Australia/Brisbane'), ('Australia/Broken_Hill', 'Australia/Broken_Hill'), ('Australia/Darwin', 'Australia/Darwin'), ('Australia/Eucla', 'Australia/Eucla'), ('Australia/Hobart', 'Australia/Hobart'), ('Australia/Lindeman', 'Australia/Lindeman'), ('Australia/Lord_Howe', 'Australia/Lord_Howe'), ('Australia/Melbourne', 'Australia/Melbourne'), ('Australia/Perth', 'Australia/Perth'), ('Australia/Sydney', 'Australia/Sydney'), ('Europe/Amsterdam', 'Europe/Amsterdam'), ('Europe/Andorra', 'Europe/Andorra'), ('Europe/Astrakhan', 'Europe/Astrakhan'), ('Europe/Athens', 'Europe/Athens'), ('Europe/Belgrade', 'Europe/Belgrade'), ('Europe/Berlin', 'Europe/Berlin'), ('Europe/Brussels', 'Europe/Brussels'), ('Europe/Bucharest', 'Europe/Bucharest'), ('Europe/Budapest', 'Europe/Budapest'), ('Europe/Chisinau', 'Europe/Chisinau'), ('Europe/Copenhagen', 'Europe/Copenhagen'), ('Europe/Dublin', 'Europe/Dublin'), ('Europe/Gibraltar', 'Europe/Gibraltar'), ('Europe/Helsinki', 'Europe/Helsinki'), ('Europe/Istanbul', 'Europe/Istanbul'), ('Europe/Kaliningrad', 'Europe/Kaliningrad'), ('Europe/Kirov', 'Europe/Kirov'), ('Europe/Kyiv', 'Europe/Kyiv'), ('Europe/Lisbon', 'Europe/Lisbon'), ('Europe/London', 'Europe/London'), ('Europe/Luxembourg', 'Europe/Luxembourg'), ('Europe/Madrid', 'Europe/Madrid'), ('Europe/Malta', 'Europe/Malta'), ('Europe/Minsk', 'Europe/Minsk'), ('Europe/Monaco', 'Europe/Monaco'), ('Europe/Moscow', 'Europe/Moscow'), ('Europe/Oslo', 'Europe/Oslo'), ('Europe/Paris', 'Europe/Paris'), ('Europe/Prague', 'Europe/Prague'), ('Europe/Riga', 'Europe/Riga'), ('Europe/Rome', 'Europe/Rome'), ('Europe/Samara', 'Europe/Samara'), ('Europe/Saratov', 'Europe/Saratov'), ('Europe/Simferopol', 'Europe/Simferopol'), ('Europe/Sofia', 'Europe/Sofia'), ('Europe/Stockholm', 'Europe/Stockholm'), ('Europe/Tallinn', 'Europe/Tallinn'), ('Europe/Tirane', 'Europe/Tirane'), ('Europe/Ulyanovsk', 'Europe/Ulyanovsk'), ('Europe/Uzhgorod', 'Europe/Uzhgorod'), ('Europe/Vaduz', 'Europe/Vaduz'), ('Europe/Vienna', 'Europe/Vienna'), ('Europe/Vilnius', 'Europe/Vilnius'), ('Europe/Volgograd', 'Europe/Volgograd'), ('Europe/Warsaw', 'Europe/Warsaw'), ('Europe/Zaporozhye', 'Europe/Zaporozhye'), ('Europe/Zurich', 'Europe/Zurich'), ('GMT', 'GMT'), ('Indian/Antananarivo', 'Indian/Antananarivo'), ('Indian/Chagos', 'Indian/Chagos'), ('Indian/Christmas', 'Indian/Christmas'), ('Indian/Cocos', 'Indian/Cocos'), ('Indian/Comoro', 'Indian/Comoro'), ('Indian/Kerguelen', 'Indian/Kerguelen'), ('Indian/Mahe', 'Indian/Mahe'), ('Indian/Maldives', 'Indian/Maldives'), ('Indian/Mauritius', 'Indian/Mauritius'), ('Indian/Mayotte', 'Indian/Mayotte'), ('Indian/Reunion', 'Indian/Reunion'), ('Pacific/Apia', 'Pacific/Apia'), ('Pacific/Auckland', 'Pacific/Auckland'), ('Pacific/Bougainville', 'Pacific/Bougainville'), ('Pacific/Chatham', 'Pacific/Chatham'), ('Pacific/Chuuk', 'Pacific/Chuuk'), ('Pacific/Easter', 'Pacific/Easter'), ('Pacific/Efate', 'Pacific/Efate'), ('Pacific/Fakaofo', 'Pacific/Fakaofo'), ('Pacific/Fiji', 'Pacific/Fiji'), ('Pacific/Funafuti', 'Pacific/Funafuti'), ('Pacific/Galapagos', 'Pacific/Galapagos'), ('Pacific/Gambier', 'Pacific/Gambier'), ('Pacific/Guadalcanal', 'Pacific/Guadalcanal'), ('Pacific/Guam', 'Pacific/Guam'), ('Pacific/Honolulu', 'Pacific/Honolulu'), ('Pacific/Kanton', 'Pacific/Kanton'), ('Pacific/Kiritimati', 'Pacific/Kiritimati'), ('Pacific/Kosrae', 'Pacific/Kosrae'), ('Pacific/Kwajalein', 'Pacific/Kwajalein'), ('Pacific/Majuro', 'Pacific/Majuro'), ('Pacific/Marquesas', 'Pacific/Marquesas'), ('Pacific/Midway', 'Pacific/Midway'), ('Pacific/Nauru', 'Pacific/Nauru'), ('Pacific/Niue', 'Pacific/Niue'), ('Pacific/Norfolk', 'Pacific/Norfolk'), ('Pacific/Noumea', 'Pacific/Noumea'), ('Pacific/Pago_Pago', 'Pacific/Pago_Pago'), ('Pacific/Palau', 'Pacific/Palau'), ('Pacific/Pitcairn', 'Pacific/Pitcairn'), ('Pacific/Pohnpei', 'Pacific/Pohnpei'), ('Pacific/Port_Moresby', 'Pacific/Port_Moresby'), ('Pacific/Rarotonga', 'Pacific/Rarotonga'), ('Pacific/Saipan', 'Pacific/Saipan'), ('Pacific/Tahiti', 'Pacific/Tahiti'), ('Pacific/Tarawa', 'Pacific/Tarawa'), ('Pacific/Tongatapu', 'Pacific/Tongatapu'), ('Pacific/Wake', 'Pacific/Wake'), ('Pacific/Wallis', 'Pacific/Wallis'), ('UTC', 'UTC')], default='UTC', max_length=255), - ), - ] diff --git a/ietf/meeting/models.py b/ietf/meeting/models.py index 068358790b..7d9e318aab 100644 --- a/ietf/meeting/models.py +++ b/ietf/meeting/models.py @@ -1,5 +1,4 @@ -# Copyright The IETF Trust 2007-2022, All Rights Reserved -# -*- coding: utf-8 -*- +# Copyright The IETF Trust 2007-2025, All Rights Reserved # old meeting models can be found in ../proceedings/models.py @@ -26,7 +25,6 @@ from django.urls import reverse as urlreverse from django.utils import timezone from django.utils.text import slugify -from django.utils.safestring import mark_safe from ietf.dbtemplate.models import DBTemplate from ietf.doc.models import Document @@ -35,12 +33,12 @@ from ietf.name.models import ( MeetingTypeName, TimeSlotTypeName, SessionStatusName, ConstraintName, RoomResourceName, ImportantDateName, TimerangeName, SlideSubmissionStatusName, ProceedingsMaterialTypeName, - SessionPurposeName, + SessionPurposeName, AttendanceTypeName, RegistrationTicketTypeName ) from ietf.person.models import Person from ietf.utils.decorators import memoize from ietf.utils.history import find_history_replacements_active_at, find_history_active_at -from ietf.utils.storage import NoLocationMigrationFileSystemStorage +from ietf.utils.storage import BlobShadowFileSystemStorage from ietf.utils.text import xslugify from ietf.utils.timezone import datetime_from_date, date_today from ietf.utils.models import ForeignKey @@ -50,15 +48,20 @@ ) from ietf.utils.fields import MissingOkImageField -countries = list(pytz.country_names.items()) -countries.sort(key=lambda x: x[1]) +# Set up countries / timezones, including an empty choice for fields +EMPTY_CHOICE = ("", "-" * 9) +COUNTRIES = (EMPTY_CHOICE,) + tuple( + sorted(pytz.country_names.items(), key=lambda x: x[1]) +) -timezones = [] -for name in pytz.common_timezones: - tzfn = os.path.join(settings.TZDATA_ICS_PATH, name + ".ics") - if not os.path.islink(tzfn): - timezones.append((name, name)) -timezones.sort() +_tzdata_ics_path = Path(settings.TZDATA_ICS_PATH) +TIMEZONES = (EMPTY_CHOICE,) + tuple( + sorted( + (name, name) + for name in pytz.common_timezones + if name != "GMT" and not (_tzdata_ics_path / f"{name}.ics").is_symlink() + ) +) class Meeting(models.Model): @@ -73,11 +76,11 @@ class Meeting(models.Model): days = models.IntegerField(default=7, null=False, validators=[MinValueValidator(1)], help_text="The number of days the meeting lasts") city = models.CharField(blank=True, max_length=255) - country = models.CharField(blank=True, max_length=2, choices=countries) + country = models.CharField(blank=True, max_length=2, choices=COUNTRIES) # We can't derive time-zone from country, as there are some that have # more than one timezone, and the pytz module doesn't provide timezone # lookup information for all relevant city/country combinations. - time_zone = models.CharField(max_length=255, choices=timezones, default='UTC') + time_zone = models.CharField(max_length=255, choices=TIMEZONES, default='UTC') idsubmit_cutoff_day_offset_00 = models.IntegerField(blank=True, default=settings.IDSUBMIT_DEFAULT_CUTOFF_DAY_OFFSET_00, help_text = "The number of days before the meeting start date when the submission of -00 drafts will be closed.") @@ -146,7 +149,7 @@ def get_00_cutoff(self): cutoff_date = importantdate.date else: cutoff_date = self.date + datetime.timedelta(days=ImportantDateName.objects.get(slug='idcutoff').default_offset_days) - cutoff_time = datetime_from_date(cutoff_date, datetime.timezone.utc) + self.idsubmit_cutoff_time_utc + cutoff_time = datetime_from_date(cutoff_date, datetime.UTC) + self.idsubmit_cutoff_time_utc return cutoff_time def get_01_cutoff(self): @@ -158,7 +161,7 @@ def get_01_cutoff(self): cutoff_date = importantdate.date else: cutoff_date = self.date + datetime.timedelta(days=ImportantDateName.objects.get(slug='idcutoff').default_offset_days) - cutoff_time = datetime_from_date(cutoff_date, datetime.timezone.utc) + self.idsubmit_cutoff_time_utc + cutoff_time = datetime_from_date(cutoff_date, datetime.UTC) + self.idsubmit_cutoff_time_utc return cutoff_time def get_reopen_time(self): @@ -234,9 +237,9 @@ def get_proceedings_materials(self): ).order_by('type__order') def get_attendance(self): - """Get the meeting attendance from the MeetingRegistrations + """Get the meeting attendance from the Registrations - Returns a NamedTuple with onsite and online attributes. Returns None if the record is unavailable + Returns a NamedTuple with onsite and remote attributes. Returns None if the record is unavailable for this meeting. """ number = self.get_number() @@ -247,25 +250,39 @@ def get_attendance(self): # MeetingRegistration.attended started conflating badge-pickup and session attendance before IETF 114. # We've separated session attendance off to ietf.meeting.Attended, but need to report attendance at older # meetings correctly. - - attended_per_meetingregistration = ( - Q(meetingregistration__meeting=self) & ( - Q(meetingregistration__attended=True) | - Q(meetingregistration__checkedin=True) + # + # Looking up by registration and attendance records separately and joining in + # python is far faster than combining the Q objects in the query (~100x). + # Further optimization may be possible, but the queries are tricky... + attended_per_meeting_registration = ( + Q(registration__meeting=self) & ( + Q(registration__attended=True) | + Q(registration__checkedin=True) + ) + ) + attendees_by_reg = set( + Person.objects.filter(attended_per_meeting_registration).values_list( + "pk", flat=True ) ) + attended_per_meeting_attended = ( Q(attended__session__meeting=self) # Note that we are not filtering to plenary, wg, or rg sessions # as we do for nomcom eligibility - if picking up a badge (see above) # is good enough, just attending e.g. a training session is also good enough ) - attended = Person.objects.filter( - attended_per_meetingregistration | attended_per_meeting_attended - ).distinct() - - onsite=set(attended.filter(meetingregistration__meeting=self, meetingregistration__reg_type='onsite')) - remote=set(attended.filter(meetingregistration__meeting=self, meetingregistration__reg_type='remote')) + attendees_by_att = set( + Person.objects.filter(attended_per_meeting_attended).values_list( + "pk", flat=True + ) + ) + + attendees = Person.objects.filter( + pk__in=attendees_by_att | attendees_by_reg + ) + onsite = set(attendees.filter(registration__meeting=self, registration__tickets__attendance_type__slug='onsite')) + remote = set(attendees.filter(registration__meeting=self, registration__tickets__attendance_type__slug='remote')) remote.difference_update(onsite) return Attendance( @@ -298,26 +315,6 @@ def proceedings_format_version(self): self._proceedings_format_version = version # save this for later return self._proceedings_format_version - @property - def session_constraintnames(self): - """Gets a list of the constraint names that should be used for this meeting - - Anticipated that this will soon become a many-to-many relationship with ConstraintName - (see issue #2770). Making this a @property allows use of the .all(), .filter(), etc, - so that other code should not need changes when this is replaced. - """ - try: - mtg_num = int(self.number) - except ValueError: - mtg_num = None # should not come up, but this method should not fail - if mtg_num is None or mtg_num >= 106: - # These meetings used the old 'conflic?' constraint types labeled as though - # they were the new types. - slugs = ('chair_conflict', 'tech_overlap', 'key_participant') - else: - slugs = ('conflict', 'conflic2', 'conflic3') - return ConstraintName.objects.filter(slug__in=slugs) - def base_url(self): return "/meeting/%s" % (self.number, ) @@ -387,27 +384,39 @@ def vtimezone(self): pass return None - def set_official_schedule(self, schedule): - if self.schedule != schedule: - self.schedule = schedule - self.save() def updated(self): # should be Meeting.modified, but we don't have that - min_time = pytz.utc.localize(datetime.datetime(1970, 1, 1, 0, 0, 0)) - timeslots_updated = self.timeslot_set.aggregate(Max('modified'))["modified__max"] or min_time - sessions_updated = self.session_set.aggregate(Max('modified'))["modified__max"] or min_time - assignments_updated = min_time + timeslots_updated = self.timeslot_set.aggregate(Max('modified'))["modified__max"] + sessions_updated = self.session_set.aggregate(Max('modified'))["modified__max"] + assignments_updated = None if self.schedule: - assignments_updated = SchedTimeSessAssignment.objects.filter(schedule__in=[self.schedule, self.schedule.base if self.schedule else None]).aggregate(Max('modified'))["modified__max"] or min_time - return max(timeslots_updated, sessions_updated, assignments_updated) + assignments_updated = SchedTimeSessAssignment.objects.filter(schedule__in=[self.schedule, self.schedule.base if self.schedule else None]).aggregate(Max('modified'))["modified__max"] + dts = [timeslots_updated, sessions_updated, assignments_updated] + valid_only = [dt for dt in dts if dt is not None] + return max(valid_only) if valid_only else None @memoize def previous_meeting(self): return Meeting.objects.filter(type_id=self.type_id,date__lt=self.date).order_by('-date').first() def uses_notes(self): - return self.date>=datetime.date(2020,7,6) + if self.type_id != 'ietf': + return True + num = self.get_number() + return num is not None and num >= 108 + + def has_recordings(self): + if self.type_id != 'ietf': + return True + num = self.get_number() + return num is not None and num >= 80 + + def has_chat_logs(self): + if self.type_id != 'ietf': + return True; + num = self.get_number() + return num is not None and num >= 60 def meeting_start(self): """Meeting-local midnight at the start of the meeting date""" @@ -476,24 +485,9 @@ class Room(models.Model): # end floorplan-related stuff def __str__(self): - return u"%s size: %s" % (self.name, self.capacity) - - def delete_timeslots(self): - for ts in self.timeslot_set.all(): - ts.sessionassignments.all().delete() - ts.delete() - - def create_timeslots(self): - days, time_slices, slots = self.meeting.build_timeslices() - for day in days: - for ts in slots[day]: - TimeSlot.objects.create(type_id=ts.type_id, - meeting=self.meeting, - name=ts.name, - time=ts.time, - location=self, - duration=ts.duration) - #self.meeting.create_all_timeslots() + if len(self.functional_name) > 0 and self.functional_name != self.name: + return f"{self.name} [{self.functional_name}] (size: {self.capacity})" + return f"{self.name} (size: {self.capacity})" def dom_id(self): return "room%u" % (self.pk) @@ -517,14 +511,6 @@ def right(self): return max(self.x1, self.x2) if (self.x1 and self.x2) else 0 def bottom(self): return max(self.y1, self.y2) if (self.y1 and self.y2) else 0 - def functional_display_name(self): - if not self.functional_name: - return "" - if 'breakout' in self.functional_name.lower(): - return "" - if self.functional_name[0].isdigit(): - return "" - return self.functional_name # audio stream support def audio_stream_url(self): urlresources = [ur for ur in self.urlresource_set.all() if ur.name_id == 'audiostream'] @@ -559,7 +545,12 @@ class FloorPlan(models.Model): modified= models.DateTimeField(auto_now=True) meeting = ForeignKey(Meeting) order = models.SmallIntegerField() - image = models.ImageField(storage=NoLocationMigrationFileSystemStorage(), upload_to=floorplan_path, blank=True, default=None) + image = models.ImageField( + storage=BlobShadowFileSystemStorage(kind="floorplan"), + upload_to=floorplan_path, + blank=False, + default=None, + ) # class Meta: ordering = ['-id',] @@ -590,7 +581,7 @@ class TimeSlot(models.Model): duration = models.DurationField(default=datetime.timedelta(0)) location = ForeignKey(Room, blank=True, null=True) show_location = models.BooleanField(default=True, help_text="Show location in agenda.") - sessions = models.ManyToManyField('Session', related_name='slots', through='SchedTimeSessAssignment', blank=True, help_text="Scheduled session, if any.") + sessions = models.ManyToManyField('meeting.Session', related_name='slots', through='meeting.SchedTimeSessAssignment', blank=True, help_text="Scheduled session, if any.") modified = models.DateTimeField(auto_now=True) # @@ -602,23 +593,23 @@ def session(self): self._session_cache = self.sessions.filter(timeslotassignments__schedule__in=[self.meeting.schedule, self.meeting.schedule.base if self.meeting else None]).first() return self._session_cache - @property - def time_desc(self): - return "%s-%s" % (self.time.strftime("%H%M"), (self.time + self.duration).strftime("%H%M")) - - def meeting_date(self): - return self.time.date() - - def registration(self): - # below implements a object local cache - # it tries to find a timeslot of type registration which starts at the same time as this slot - # so that it can be shown at the top of the agenda. - if not hasattr(self, '_reg_info'): - try: - self._reg_info = TimeSlot.objects.get(meeting=self.meeting, time__month=self.time.month, time__day=self.time.day, type="reg") - except TimeSlot.DoesNotExist: - self._reg_info = None - return self._reg_info + # Unused + # + # def meeting_date(self): + # return self.time.date() + + # Unused + # + # def registration(self): + # # below implements a object local cache + # # it tries to find a timeslot of type registration which starts at the same time as this slot + # # so that it can be shown at the top of the agenda. + # if not hasattr(self, '_reg_info'): + # try: + # self._reg_info = TimeSlot.objects.get(meeting=self.meeting, time__month=self.time.month, time__day=self.time.day, type="reg") + # except TimeSlot.DoesNotExist: + # self._reg_info = None + # return self._reg_info def __str__(self): location = self.get_location() @@ -645,30 +636,33 @@ def get_hidden_location(self): def get_location(self): return self.get_hidden_location() if self.show_location else "" - def get_functional_location(self): - name_parts = [] - room = self.location - if room and room.functional_name: - name_parts.append(room.functional_name) - location = self.get_hidden_location() - if location: - name_parts.append(location) - return ' - '.join(name_parts) - - def get_html_location(self): - if not hasattr(self, '_cached_html_location'): - self._cached_html_location = self.get_location() - if len(self._cached_html_location) > 8: - self._cached_html_location = mark_safe(self._cached_html_location.replace('/', '/')) - else: - self._cached_html_location = mark_safe(self._cached_html_location.replace(' ', ' ')) - return self._cached_html_location + # Unused + # + # def get_functional_location(self): + # name_parts = [] + # room = self.location + # if room and room.functional_name: + # name_parts.append(room.functional_name) + # location = self.get_hidden_location() + # if location: + # name_parts.append(location) + # return ' - '.join(name_parts) + + # def get_html_location(self): + # if not hasattr(self, '_cached_html_location'): + # self._cached_html_location = self.get_location() + # if len(self._cached_html_location) > 8: + # self._cached_html_location = mark_safe(self._cached_html_location.replace('/', '/')) + # else: + # self._cached_html_location = mark_safe(self._cached_html_location.replace(' ', ' ')) + # return self._cached_html_location def tz(self): return self.meeting.tz() - def tzname(self): - return self.tz().tzname(self.time) + # Unused + # def tzname(self): + # return self.tz().tzname(self.time) def utc_start_time(self): return self.time.astimezone(pytz.utc) # USE_TZ is True, so time is aware @@ -682,30 +676,32 @@ def local_start_time(self): def local_end_time(self): return (self.time.astimezone(pytz.utc) + self.duration).astimezone(self.tz()) - @property - def js_identifier(self): - # this returns a unique identifier that is js happy. - # {{s.timeslot.time|date:'Y-m-d'}}_{{ s.timeslot.time|date:'Hi' }}" - # also must match: - # {{r|slugify}}_{{day}}_{{slot.0|date:'Hi'}} - dom_id="ts%u" % (self.pk) - if self.location is not None: - dom_id = self.location.dom_id() - return "%s_%s_%s" % (dom_id, self.time.strftime('%Y-%m-%d'), self.time.strftime('%H%M')) - - def delete_concurrent_timeslots(self): - """Delete all timeslots which are in the same time as this slot""" - # can not include duration in filter, because there is no support - # for having it a WHERE clause. - # below will delete self as well. - for ts in self.meeting.timeslot_set.filter(time=self.time).all(): - if ts.duration!=self.duration: - continue - - # now remove any schedule that might have been made to this - # timeslot. - ts.sessionassignments.all().delete() - ts.delete() + # Unused + # + # @property + # def js_identifier(self): + # # this returns a unique identifier that is js happy. + # # {{s.timeslot.time|date:'Y-m-d'}}_{{ s.timeslot.time|date:'Hi' }}" + # # also must match: + # # {{r|slugify}}_{{day}}_{{slot.0|date:'Hi'}} + # dom_id="ts%u" % (self.pk) + # if self.location is not None: + # dom_id = self.location.dom_id() + # return "%s_%s_%s" % (dom_id, self.time.strftime('%Y-%m-%d'), self.time.strftime('%H%M')) + + # def delete_concurrent_timeslots(self): + # """Delete all timeslots which are in the same time as this slot""" + # # can not include duration in filter, because there is no support + # # for having it a WHERE clause. + # # below will delete self as well. + # for ts in self.meeting.timeslot_set.filter(time=self.time).all(): + # if ts.duration!=self.duration: + # continue + + # # now remove any schedule that might have been made to this + # # timeslot. + # ts.sessionassignments.all().delete() + # ts.delete() """ Find a timeslot that comes next, in the same room. It must be on the same day, @@ -791,9 +787,6 @@ def official_token(self): else: return "unofficial" - def delete_assignments(self): - self.assignments.all().delete() - @property def qs_assignments_with_sessions(self): return self.assignments.filter(session__isnull=False) @@ -806,10 +799,6 @@ def qs_sessions_scheduled(self): """Get QuerySet containing sessions assigned to timeslots by this schedule""" return Session.objects.filter(timeslotassignments__schedule=self) - def delete_schedule(self): - self.assignments.all().delete() - self.delete() - # to be renamed SchedTimeSessAssignments (stsa) class SchedTimeSessAssignment(models.Model): """ @@ -822,7 +811,6 @@ class SchedTimeSessAssignment(models.Model): schedule = ForeignKey('Schedule', null=False, blank=False, related_name='assignments') extendedfrom = ForeignKey('self', null=True, default=None, help_text="Timeslot this session is an extension of.") modified = models.DateTimeField(auto_now=True) - notes = models.TextField(blank=True) badness = models.IntegerField(default=0, blank=True, null=True) pinned = models.BooleanField(default=False, help_text="Do not move session during automatic placement.") @@ -955,8 +943,8 @@ def brief_display(self): class SessionPresentation(models.Model): - session = ForeignKey('Session') - document = ForeignKey(Document) + session = ForeignKey('Session', related_name="presentations") + document = ForeignKey(Document, related_name="presentations") rev = models.CharField(verbose_name="revision", max_length=16, null=True, blank=True) order = models.PositiveSmallIntegerField(default=0) @@ -968,8 +956,6 @@ class Meta: def __str__(self): return u"%s -> %s-%s" % (self.session, self.document.name, self.rev) -constraint_cache_uses = 0 -constraint_cache_initials = 0 class SessionQuerySet(models.QuerySet): def with_current_status(self): @@ -1068,13 +1054,16 @@ class Session(models.Model): group = ForeignKey(Group) # The group type historically determined the session type. BOFs also need to be added as a group. Note that not all meeting requests have a natural group to associate with. joint_with_groups = models.ManyToManyField(Group, related_name='sessions_joint_in',blank=True) attendees = models.IntegerField(null=True, blank=True) - agenda_note = models.CharField(blank=True, max_length=255) + agenda_note = models.CharField(blank=True, max_length=512) requested_duration = models.DurationField(default=datetime.timedelta(0)) comments = models.TextField(blank=True) scheduled = models.DateTimeField(null=True, blank=True) modified = models.DateTimeField(auto_now=True) remote_instructions = models.CharField(blank=True,max_length=1024) on_agenda = models.BooleanField(default=True, help_text='Is this session visible on the meeting agenda?') + has_onsite_tool = models.BooleanField(default=False, help_text="Does this session use the officially supported onsite and remote tooling?") + chat_room = models.CharField(blank=True, max_length=32, help_text='Name of Zulip stream, if different from group acronym') + meetecho_recording_name = models.CharField(blank=True, max_length=64, help_text="Name of the meetecho recording") tombstone_for = models.ForeignKey('Session', blank=True, null=True, help_text="This session is the tombstone for a session that was rescheduled", on_delete=models.CASCADE) @@ -1093,7 +1082,7 @@ def get_material(self, material_type, only_one): for d in l: d.meeting_related = lambda: True else: - l = self.materials.filter(type=material_type).exclude(states__type=material_type, states__slug='deleted').order_by('sessionpresentation__order') + l = self.materials.filter(type=material_type).exclude(states__type=material_type, states__slug='deleted').order_by('presentations__order') if only_one: if l: @@ -1113,16 +1102,25 @@ def minutes(self): self._cached_minutes = self.get_material("minutes", only_one=True) return self._cached_minutes + def narrative_minutes(self): + if not hasattr(self, '_cached_narrative_minutes'): + self._cached_minutes = self.get_material("narrativeminutes", only_one=True) + return self._cached_minutes + def recordings(self): return list(self.get_material("recording", only_one=False)) def bluesheets(self): return list(self.get_material("bluesheets", only_one=False)) + def chatlogs(self): + return list(self.get_material("chatlog", only_one=False)) + def slides(self): if not hasattr(self, "_slides_cache"): self._slides_cache = list(self.get_material("slides", only_one=False)) return self._slides_cache + def drafts(self): return list(self.materials.filter(type='draft')) @@ -1157,30 +1155,6 @@ def order_in_meeting(self): self._order_in_meeting = session_list.index(self) + 1 if self in session_list else 0 return self._order_in_meeting - def all_meeting_sessions_cancelled(self): - return set(s.current_status for s in self.all_meeting_sessions_for_group()) == {'canceled'} - - def all_meeting_recordings(self): - recordings = [] # These are not sets because we need to preserve relative ordering or redo the ordering work later - sessions = self.all_meeting_sessions_for_group() - for session in sessions: - recordings.extend([r for r in session.recordings() if r not in recordings]) - return recordings - - def all_meeting_bluesheets(self): - bluesheets = [] - sessions = self.all_meeting_sessions_for_group() - for session in sessions: - bluesheets.extend([b for b in session.bluesheets() if b not in bluesheets]) - return bluesheets - - def all_meeting_drafts(self): - drafts = [] - sessions = self.all_meeting_sessions_for_group() - for session in sessions: - drafts.extend([d for d in session.drafts() if d not in drafts]) - return drafts - def all_meeting_agendas(self): agendas = [] sessions = self.all_meeting_sessions_for_group() @@ -1210,7 +1184,7 @@ def can_manage_materials(self, user): return can_manage_materials(user,self.group) def is_material_submission_cutoff(self): - return date_today(datetime.timezone.utc) > self.meeting.get_submission_correction_date() + return date_today(datetime.UTC) > self.meeting.get_submission_correction_date() def joint_with_groups_acronyms(self): return [group.acronym for group in self.joint_with_groups.all()] @@ -1253,19 +1227,30 @@ def special_request_token(self): else: return "" + @staticmethod + def _alpha_str(n: int): + """Convert integer to string of a-z characters (a, b, c, ..., aa, ab, ...)""" + chars = [] + while True: + chars.append(string.ascii_lowercase[n % 26]) + n //= 26 + # for 2nd letter and beyond, 0 means end the string + if n == 0: + break + # beyond the first letter, no need to represent a 0, so decrement + n -= 1 + return "".join(chars[::-1]) + def docname_token(self): sess_mtg = Session.objects.filter(meeting=self.meeting, group=self.group).order_by('pk') index = list(sess_mtg).index(self) - return 'sess%s' % (string.ascii_lowercase[index]) + return f"sess{self._alpha_str(index)}" def docname_token_only_for_multiple(self): sess_mtg = Session.objects.filter(meeting=self.meeting, group=self.group).order_by('pk') if len(list(sess_mtg)) > 1: index = list(sess_mtg).index(self) - if index < 26: - token = 'sess%s' % (string.ascii_lowercase[index]) - else: - token = 'sess%s%s' % (string.ascii_lowercase[index//26],string.ascii_lowercase[index%26]) + token = f"sess{self._alpha_str(index)}" return token return None @@ -1276,7 +1261,10 @@ def reverse_constraints(self): return Constraint.objects.filter(target=self.group, meeting=self.meeting).order_by('name__name') def official_timeslotassignment(self): - return self.timeslotassignments.filter(schedule__in=[self.meeting.schedule, self.meeting.schedule.base if self.meeting.schedule else None]).first() + # cache only non-None values + if getattr(self, "_cache_official_timeslotassignment", None) is None: + self._cache_official_timeslotassignment = self.timeslotassignments.filter(schedule__in=[self.meeting.schedule, self.meeting.schedule.base if self.meeting.schedule else None]).first() + return self._cache_official_timeslotassignment @property def people_constraints(self): @@ -1294,21 +1282,11 @@ def agenda_text(self): else: return "The agenda has not been uploaded yet." - def agenda_file(self): - if not hasattr(self, '_agenda_file'): - self._agenda_file = "" - - agenda = self.agenda() - if not agenda: - return "" - - # FIXME: uploaded_filename should be replaced with a function that computes filenames when they are of a fixed schema and not uploaded names - self._agenda_file = "%s/agenda/%s" % (self.meeting.number, agenda.uploaded_filename) - - return self._agenda_file - def chat_room_name(self): - if self.type_id=='plenary': + if self.chat_room: + return self.chat_room + # At some point, add a migration to add "plenary" chat room name to existing sessions in the database. + elif self.type_id=='plenary': return 'plenary' else: return self.group_at_the_time().acronym @@ -1317,8 +1295,22 @@ def chat_room_url(self): return settings.CHAT_URL_PATTERN.format(chat_room_name=self.chat_room_name()) def chat_archive_url(self): - # datatracker 8.8.0 released on 2022 July 15; before that, fall back to old log URL + + if hasattr(self,"prefetched_active_materials"): + chatlog_doc = None + for doc in self.prefetched_active_materials: + if doc.type_id=="chatlog": + chatlog_doc = doc + break + if chatlog_doc is not None: + return chatlog_doc.get_href() + else: + chatlog = self.presentations.filter(document__type__slug='chatlog').first() + if chatlog is not None: + return chatlog.document.get_href() + if self.meeting.date <= datetime.date(2022, 7, 15): + # datatracker 8.8.0 released on 2022 July 15; before that, fall back to old log URL return f'https://www.ietf.org/jabber/logs/{ self.chat_room_name() }?C=M;O=D' elif hasattr(settings,'CHAT_ARCHIVE_URL_PATTERN'): return settings.CHAT_ARCHIVE_URL_PATTERN.format(chat_room_name=self.chat_room_name()) @@ -1333,13 +1325,54 @@ def notes_id(self): def notes_url(self): return urljoin(settings.IETF_NOTES_URL, self.notes_id()) + def group_at_the_time(self): - return self.meeting.group_at_the_time(self.group) + if not hasattr(self,"_cached_group_at_the_time"): + self._cached_group_at_the_time = self.meeting.group_at_the_time(self.group) + return self._cached_group_at_the_time def group_parent_at_the_time(self): if self.group_at_the_time().parent: return self.meeting.group_at_the_time(self.group_at_the_time().parent) + def audio_stream_url(self): + url = getattr(settings, "MEETECHO_AUDIO_STREAM_URL", "") + if self.meeting.type.slug == "ietf" and self.has_onsite_tool and url: + return url.format(session=self) + return None + + def video_stream_url(self): + url = getattr(settings, "MEETECHO_VIDEO_STREAM_URL", "") + if self.meeting.type.slug == "ietf" and self.has_onsite_tool and url: + return url.format(session=self) + return None + + def onsite_tool_url(self): + url = getattr(settings, "MEETECHO_ONSITE_TOOL_URL", "") + if self.meeting.type.slug == "ietf" and self.has_onsite_tool and url: + return url.format(session=self) + return None + + def _session_recording_url_label(self): + otsa = self.official_timeslotassignment() + if otsa is None: + return None + if self.meeting.type.slug == "ietf" and self.has_onsite_tool: + session_label = f"IETF{self.meeting.number}-{self.group.acronym.upper()}-{otsa.timeslot.time.strftime('%Y%m%d-%H%M')}" + else: + session_label = f"IETF-{self.group.acronym.upper()}-{otsa.timeslot.time.strftime('%Y%m%d-%H%M')}" + return session_label + + def session_recording_url(self): + url_formatter = getattr(settings, "MEETECHO_SESSION_RECORDING_URL", "") + url = None + name = self.meetecho_recording_name + if name is None or name.strip() == "": + name = self._session_recording_url_label() + if url_formatter.strip() != "" and name is not None: + url = url_formatter.format(session_label=name) + return url + class SchedulingEvent(models.Model): session = ForeignKey(Session) @@ -1368,7 +1401,7 @@ class SlideSubmission(models.Model): apply_to_all = models.BooleanField(default=False) submitter = ForeignKey(Person) status = ForeignKey(SlideSubmissionStatusName, null=True, default='pending', on_delete=models.SET_NULL) - doc = ForeignKey(Document, null=True, on_delete=models.SET_NULL) + doc = ForeignKey(Document, blank=True, null=True, on_delete=models.SET_NULL) def staged_filepath(self): return os.path.join(settings.SLIDE_STAGING_PATH , self.filename) @@ -1419,8 +1452,12 @@ class MeetingHost(models.Model): """Meeting sponsor""" meeting = ForeignKey(Meeting, related_name='meetinghosts') name = models.CharField(max_length=255, blank=False) + # TODO-BLOBSTORE - capture these logos and look for other ImageField like model fields. logo = MissingOkImageField( - storage=NoLocationMigrationFileSystemStorage(location=settings.MEETINGHOST_LOGO_PATH), + storage=BlobShadowFileSystemStorage( + kind="meetinghostlogo", + location=settings.MEETINGHOST_LOGO_PATH, + ), upload_to=_host_upload_path, width_field='logo_width', height_field='logo_height', @@ -1435,7 +1472,7 @@ class MeetingHost(models.Model): validate_file_extension, settings.MEETING_VALID_UPLOAD_EXTENSIONS['meetinghostlogo'], ), - WrappedValidator( + WrappedValidator( validate_mime_type, settings.MEETING_VALID_UPLOAD_MIME_TYPES['meetinghostlogo'], True, @@ -1454,9 +1491,56 @@ class Meta: class Attended(models.Model): person = ForeignKey(Person) session = ForeignKey(Session) + time = models.DateTimeField(default=timezone.now, null=True, blank=True) + origin = models.CharField(max_length=32, default='datatracker') class Meta: unique_together = (('person', 'session'),) def __str__(self): return f'{self.person} at {self.session}' + + +class RegistrationManager(models.Manager): + def onsite(self): + return self.get_queryset().filter(tickets__attendance_type__slug='onsite') + + def remote(self): + return self.get_queryset().filter(tickets__attendance_type__slug='remote').exclude(tickets__attendance_type__slug='onsite') + +class Registration(models.Model): + """Registration attendee records from the IETF registration system""" + meeting = ForeignKey(Meeting) + first_name = models.CharField(max_length=255) + last_name = models.CharField(max_length=255) + affiliation = models.CharField(blank=True, max_length=255) + country_code = models.CharField(max_length=2) # ISO 3166 + person = ForeignKey(Person, blank=True, null=True, on_delete=models.PROTECT) + email = models.EmailField(blank=True, null=True) + # attended was used prior to the introduction of the ietf.meeting.Attended model and is still used by + # Meeting.get_attendance() for older meetings. It should not be used except for dealing with legacy data. + attended = models.BooleanField(default=False) + # checkedin indicates that the badge was picked up + checkedin = models.BooleanField(default=False) + + # custom manager + objects = RegistrationManager() + + def __str__(self): + return "{} {}".format(self.first_name, self.last_name) + + @property + def attendance_type(self): + if self.tickets.filter(attendance_type__slug='onsite').exists(): + return 'onsite' + elif self.tickets.filter(attendance_type__slug='remote').exists(): + return 'remote' + return None + +class RegistrationTicket(models.Model): + registration = ForeignKey(Registration, related_name='tickets') + attendance_type = ForeignKey(AttendanceTypeName, on_delete=models.PROTECT) + ticket_type = ForeignKey(RegistrationTicketTypeName, on_delete=models.PROTECT) + + def __str__(self): + return "{}:{}".format(self.attendance_type, self.ticket_type) diff --git a/ietf/meeting/resources.py b/ietf/meeting/resources.py index dc273c04cf..490b75f925 100644 --- a/ietf/meeting/resources.py +++ b/ietf/meeting/resources.py @@ -11,12 +11,23 @@ from ietf import api -from ietf.meeting.models import ( Meeting, ResourceAssociation, Constraint, Room, Schedule, Session, - TimeSlot, SchedTimeSessAssignment, SessionPresentation, FloorPlan, - UrlResource, ImportantDate, SlideSubmission, SchedulingEvent, - BusinessConstraint, ProceedingsMaterial, MeetingHost, Attended) +from ietf.meeting.models import (Meeting, ResourceAssociation, Constraint, Room, + Schedule, Session, + TimeSlot, SchedTimeSessAssignment, SessionPresentation, + FloorPlan, + UrlResource, ImportantDate, SlideSubmission, + SchedulingEvent, + BusinessConstraint, ProceedingsMaterial, MeetingHost, + Attended, + Registration, RegistrationTicket) + +from ietf.name.resources import ( + AttendanceTypeNameResource, + MeetingTypeNameResource, + RegistrationTicketTypeNameResource, +) + -from ietf.name.resources import MeetingTypeNameResource class MeetingResource(ModelResource): type = ToOneField(MeetingTypeNameResource, 'type') schedule = ToOneField('ietf.meeting.resources.ScheduleResource', 'schedule', null=True) @@ -269,7 +280,6 @@ class Meta: filtering = { "id": ALL, "modified": ALL, - "notes": ALL, "badness": ALL, "pinned": ALL, "timeslot": ALL_WITH_RELATIONS, @@ -432,3 +442,52 @@ class Meta: "session": ALL_WITH_RELATIONS, } api.meeting.register(AttendedResource()) + +from ietf.person.resources import PersonResource +class RegistrationResource(ModelResource): + meeting = ToOneField(MeetingResource, 'meeting') + person = ToOneField(PersonResource, 'person', null=True) + tickets = ToManyField( + 'ietf.meeting.resources.RegistrationTicketResource', + 'tickets', + full=True, + ) + + class Meta: + queryset = Registration.objects.all() + serializer = api.Serializer() + cache = SimpleCache() + #resource_name = 'registration' + ordering = ['id', ] + filtering = { + "id": ALL, + "first_name": ALL, + "last_name": ALL, + "affiliation": ALL, + "country_code": ALL, + "email": ALL, + "attended": ALL, + "checkedin": ALL, + "meeting": ALL_WITH_RELATIONS, + "person": ALL_WITH_RELATIONS, + "tickets": ALL_WITH_RELATIONS, + } +api.meeting.register(RegistrationResource()) + +class RegistrationTicketResource(ModelResource): + registration = ToOneField(RegistrationResource, 'registration') + attendance_type = ToOneField(AttendanceTypeNameResource, 'attendance_type') + ticket_type = ToOneField(RegistrationTicketTypeNameResource, 'ticket_type') + class Meta: + queryset = RegistrationTicket.objects.all() + serializer = api.Serializer() + cache = SimpleCache() + #resource_name = 'registrationticket' + ordering = ['id', ] + filtering = { + "id": ALL, + "ticket_type": ALL_WITH_RELATIONS, + "attendance_type": ALL_WITH_RELATIONS, + "registration": ALL_WITH_RELATIONS, + } +api.meeting.register(RegistrationTicketResource()) diff --git a/ietf/meeting/tasks.py b/ietf/meeting/tasks.py new file mode 100644 index 0000000000..a73763560b --- /dev/null +++ b/ietf/meeting/tasks.py @@ -0,0 +1,248 @@ +# Copyright The IETF Trust 2024-2026, All Rights Reserved +# +# Celery task definitions +# +import datetime + +from itertools import batched + +from celery import shared_task, chain +from django.db.models import IntegerField +from django.db.models.functions import Cast +from django.utils import timezone + +from ietf.utils import log +from .models import Meeting +from .utils import ( + generate_proceedings_content, + resolve_materials_for_one_meeting, + store_blobs_for_one_meeting, +) +from .views import generate_agenda_data +from .utils import fetch_attendance_from_meetings + + +@shared_task +def agenda_data_refresh_task(num=None): + """Refresh agenda data for one plenary meeting + + If `num` is `None`, refreshes data for the current meeting. + """ + log.log( + f"Refreshing agenda data for {f"IETF-{num}" if num else "current IETF meeting"}" + ) + try: + generate_agenda_data(num, force_refresh=True) + except Exception as err: + # Log and swallow exceptions so failure on one meeting won't break a chain of + # tasks. This is used by agenda_data_refresh_all_task(). + log.log(f"ERROR: Refreshing agenda data failed for num={num}: {err}") + + +@shared_task +def agenda_data_refresh(): + """Deprecated. Use agenda_data_refresh_task() instead. + + TODO remove this after switching the periodic task to the new name + """ + log.log("Deprecated agenda_data_refresh task called!") + agenda_data_refresh_task() + + +@shared_task +def agenda_data_refresh_all_task(*, batch_size=10): + """Refresh agenda data for all plenary meetings + + Executes as a chain of tasks, each computing up to `batch_size` meetings + in a single task. + """ + meeting_numbers = sorted( + Meeting.objects.annotate( + number_as_int=Cast("number", output_field=IntegerField()) + ) + .filter(type_id="ietf", number_as_int__gt=64) + .values_list("number_as_int", flat=True) + ) + # Batch using chained maps rather than celery.chunk so we only use one worker + # at a time. + batched_task_chain = chain( + *( + agenda_data_refresh_task.map(nums) + for nums in batched(meeting_numbers, batch_size) + ) + ) + batched_task_chain.delay() + + +@shared_task +def proceedings_content_refresh_task(*, all=False): + """Refresh meeting proceedings cache + + If `all` is `False`, then refreshes the cache for meetings whose numbers modulo + 24 equal the current hour number (0-23). Scheduling the task once per hour will + then result in all proceedings being recomputed daily, with no more than two per + hour (now) or a few per hour in the next decade. That keeps the computation time + to under a couple minutes on our current production system. + + If `all` is True, refreshes all meetings + """ + now = timezone.now() + + for meeting in Meeting.objects.filter(type_id="ietf").order_by("number"): + if meeting.proceedings_format_version == 1: + continue # skip v1 proceedings, they're stored externally + num = meeting.get_number() # convert str -> int + if num is None: + log.log( + f"Not refreshing proceedings for meeting {meeting.number}: " + f"type is 'ietf' but get_number() returned None" + ) + elif all or (num % 24 == now.hour): + log.log(f"Refreshing proceedings for meeting {meeting.number}...") + generate_proceedings_content(meeting, force_refresh=True) + + +@shared_task +def fetch_meeting_attendance_task(): + # fetch most recent two meetings + meetings = Meeting.objects.filter(type="ietf", date__lte=timezone.now()).order_by( + "-date" + )[:2] + try: + stats = fetch_attendance_from_meetings(meetings) + except RuntimeError as err: + log.log(f"Error in fetch_meeting_attendance_task: {err}") + else: + for meeting, meeting_stats in zip(meetings, stats): + log.log( + "Fetched data for meeting {:>3}: {:4d} created, {:4d} updated, {:4d} deleted, {:4d} processed".format( + meeting.number, + meeting_stats["created"], + meeting_stats["updated"], + meeting_stats["deleted"], + meeting_stats["processed"], + ) + ) + + +def _select_meetings( + meetings: list[str] | None = None, + meetings_since: str | None = None, + meetings_until: str | None = None, +): # nyah + """Select meetings by number or date range""" + # IETF-1 = 1986-01-16 + EARLIEST_MEETING_DATE = datetime.datetime(1986, 1, 1) + meetings_since_dt: datetime.datetime | None = None + meetings_until_dt: datetime.datetime | None = None + + if meetings_since == "zero": + meetings_since_dt = EARLIEST_MEETING_DATE + elif meetings_since is not None: + try: + meetings_since_dt = datetime.datetime.fromisoformat(meetings_since) + except ValueError: + log.log( + "Failed to parse meetings_since='{meetings_since}' with fromisoformat" + ) + raise + + if meetings_until is not None: + try: + meetings_until_dt = datetime.datetime.fromisoformat(meetings_until) + except ValueError: + log.log( + "Failed to parse meetings_until='{meetings_until}' with fromisoformat" + ) + raise + if meetings_since_dt is None: + # if we only got meetings_until, start from the first meeting + meetings_since_dt = EARLIEST_MEETING_DATE + + if meetings is None: + if meetings_since_dt is None: + log.log("No meetings requested, doing nothing.") + return Meeting.objects.none() + meetings_qs = Meeting.objects.filter(date__gte=meetings_since_dt) + if meetings_until_dt is not None: + meetings_qs = meetings_qs.filter(date__lte=meetings_until_dt) + log.log( + "Selecting meetings between " + f"{meetings_since_dt} and {meetings_until_dt}" + ) + else: + log.log(f"Selecting meetings since {meetings_since_dt}") + else: + if meetings_since_dt is not None: + log.log( + "Ignoring meetings_since and meetings_until " + "because specific meetings were requested." + ) + meetings_qs = Meeting.objects.filter(number__in=meetings) + return meetings_qs + + +@shared_task +def resolve_meeting_materials_task( + *, # only allow kw arguments + meetings: list[str] | None = None, + meetings_since: str | None = None, + meetings_until: str | None = None, +): + """Run materials resolver on meetings + + Can request a set of meetings by number by passing a list in the meetings arg, or + by range by passing an iso-format timestamps in meetings_since / meetings_until. + To select all meetings, set meetings_since="zero" and omit other parameters. + """ + meetings_qs = _select_meetings(meetings, meetings_since, meetings_until) + for meeting in meetings_qs.order_by("date"): + log.log( + f"Resolving materials for {meeting.type_id} " + f"meeting {meeting.number} ({meeting.date})..." + ) + mark = timezone.now() + try: + resolve_materials_for_one_meeting(meeting) + except Exception as err: + log.log( + "Exception raised while resolving materials for " + f"meeting {meeting.number}: {err}" + ) + else: + log.log( + f"Resolved in {(timezone.now() - mark).total_seconds():0.3f} seconds." + ) + + +@shared_task +def store_meeting_materials_as_blobs_task( + *, # only allow kw arguments + meetings: list[str] | None = None, + meetings_since: str | None = None, + meetings_until: str | None = None, +): + """Push meeting materials into the blob store + + Can request a set of meetings by number by passing a list in the meetings arg, or + by range by passing an iso-format timestamps in meetings_since / meetings_until. + To select all meetings, set meetings_since="zero" and omit other parameters. + """ + meetings_qs = _select_meetings(meetings, meetings_since, meetings_until) + for meeting in meetings_qs.order_by("date"): + log.log( + f"Creating blobs for materials for {meeting.type_id} " + f"meeting {meeting.number} ({meeting.date})..." + ) + mark = timezone.now() + try: + store_blobs_for_one_meeting(meeting) + except Exception as err: + log.log( + "Exception raised while creating blobs for " + f"meeting {meeting.number}: {err}" + ) + else: + log.log( + f"Blobs created in {(timezone.now() - mark).total_seconds():0.3f} seconds." + ) diff --git a/ietf/meeting/templatetags/ams_filters.py b/ietf/meeting/templatetags/ams_filters.py new file mode 100644 index 0000000000..a8175a81d6 --- /dev/null +++ b/ietf/meeting/templatetags/ams_filters.py @@ -0,0 +1,67 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +from django import template +from ietf.person.models import Person + + +register = template.Library() + +@register.filter +def abbr_status(value): + """ + Converts RFC Status to a short abbreviation + """ + d = {'Proposed Standard':'PS', + 'Draft Standard':'DS', + 'Standard':'S', + 'Historic':'H', + 'Informational':'I', + 'Experimental':'E', + 'Best Current Practice':'BCP', + 'Internet Standard':'IS'} + + return d.get(value,value) + +@register.filter(name='display_duration') +def display_duration(value): + """ + Maps a session requested duration from select index to + label.""" + if value in (None, ''): + return 'unspecified' + value = int(value) + map = {0: 'None', + 1800: '30 Minutes', + 3600: '1 Hour', + 5400: '1.5 Hours', + 7200: '2 Hours', + 9000: '2.5 Hours', + 10800: '3 Hours', + 12600: '3.5 Hours', + 14400: '4 Hours'} + if value in map: + return map[value] + else: + return "%d Hours %d Minutes %d Seconds"%(value//3600,(value%3600)//60,value%60) + +@register.filter +def is_ppt(value): + ''' + Checks if the value ends in ppt or pptx + ''' + if value.endswith('ppt') or value.endswith('pptx'): + return True + else: + return False + +@register.filter +def smart_login(user): + ''' + Expects a Person object. If person is a Secretariat returns "on behalf of the" + ''' + if not isinstance (user, Person): + return user + if user.role_set.filter(name='secr',group__acronym='secretariat'): + return '%s, on behalf of the' % user + else: + return '%s, a chair of the' % user diff --git a/ietf/meeting/templatetags/meetings_filters.py b/ietf/meeting/templatetags/meetings_filters.py new file mode 100644 index 0000000000..3fa209daa2 --- /dev/null +++ b/ietf/meeting/templatetags/meetings_filters.py @@ -0,0 +1,21 @@ +# Copyright The IETF Trust 2023, All Rights Reserved +# -*- coding: utf-8 -*- + +from django import template +from ietf.meeting.helpers import can_request_interim_meeting + +import debug # pyflakes:ignore + +register = template.Library() + +@register.filter +def can_request_interim(user): + """Determine whether the user can request an interim meeting + + Usage: can_request_interim + Returns Boolean. True means user can request an interim meeting. + """ + + if not user: + return False + return can_request_interim_meeting(user) diff --git a/ietf/meeting/templatetags/proceedings_filters.py b/ietf/meeting/templatetags/proceedings_filters.py index f5fe0e1f14..a2a4932e7c 100644 --- a/ietf/meeting/templatetags/proceedings_filters.py +++ b/ietf/meeting/templatetags/proceedings_filters.py @@ -11,7 +11,7 @@ def hack_recording_title(recording,add_timestamp=False): if recording.title.startswith('Audio recording for') or recording.title.startswith('Video recording for'): hacked_title = recording.title[:15] if add_timestamp: - hacked_title += ' '+recording.sessionpresentation_set.first().session.official_timeslotassignment().timeslot.time.strftime("%a %H:%M") + hacked_title += ' '+recording.presentations.first().session.official_timeslotassignment().timeslot.time.strftime("%a %H:%M") return hacked_title else: return recording.title diff --git a/ietf/meeting/templatetags/session_filters.py b/ietf/meeting/templatetags/session_filters.py index 2a3da70283..3846dab49e 100644 --- a/ietf/meeting/templatetags/session_filters.py +++ b/ietf/meeting/templatetags/session_filters.py @@ -1,17 +1,56 @@ +# Copyright The IETF Trust 2023, All Rights Reserved from django import template +from ietf.name.models import SessionStatusName + register = template.Library() + @register.filter -def presented_versions(session,doc): - sp = session.sessionpresentation_set.filter(document=doc) - if not sp: - return "Document not in session" - else: - rev = sp.first().rev - return rev if rev else "(current)" +def presented_versions(session, doc): + sp = session.presentations.filter(document=doc) + if not sp: + return "Document not in session" + else: + rev = sp.first().rev + return rev if rev else "(current)" + @register.filter -def can_manage_materials(session,user): +def can_manage_materials(session, user): return session.can_manage_materials(user) + +@register.filter +def describe_with_tz(session): + # Very similar to session.__str__, but doesn't treat interims differently from sessions at an IETF meeting + # and displays the timeslot in the meeting's timezone. + + if session is None: + return "" + + status_id = None + if hasattr(session, "current_status"): + status_id = session.current_status + elif session.pk is not None: + latest_event = session.schedulingevent_set.order_by("-time", "-id").first() + if latest_event: + status_id = latest_event.status_id + + if status_id in ("canceled", "disappr", "notmeet", "deleted"): + ss0name = "(%s)" % SessionStatusName.objects.get(slug=status_id).name + else: + ss0name = "(unscheduled)" + ss = session.timeslotassignments.filter( + schedule__in=[ + session.meeting.schedule, + session.meeting.schedule.base if session.meeting.schedule else None, + ] + ).order_by("timeslot__time") + if ss: + ss0name = ",".join( + x.timeslot.time.astimezone(session.meeting.tz()).strftime("%a-%H%M") + for x in ss + ) + ss0name += f" {session.meeting.tz()}" + return f"{session.meeting}: {session.group.acronym} {session.name} {ss0name}" diff --git a/ietf/meeting/test_data.py b/ietf/meeting/test_data.py index 5ecb494df2..8be55b47a2 100644 --- a/ietf/meeting/test_data.py +++ b/ietf/meeting/test_data.py @@ -51,7 +51,7 @@ def make_interim_meeting(group,date,status='sched',tz='UTC'): doc = DocumentFactory.create(name=name, type_id='agenda', title="Agenda", uploaded_filename=file, group=group, rev=rev, states=[('draft','active')]) pres = SessionPresentation.objects.create(session=session, document=doc, rev=doc.rev) - session.sessionpresentation_set.add(pres) + session.presentations.add(pres) # minutes name = "minutes-%s-%s" % (meeting.number, time.strftime("%Y%m%d%H%M")) rev = '00' @@ -59,7 +59,7 @@ def make_interim_meeting(group,date,status='sched',tz='UTC'): doc = DocumentFactory.create(name=name, type_id='minutes', title="Minutes", uploaded_filename=file, group=group, rev=rev, states=[('draft','active')]) pres = SessionPresentation.objects.create(session=session, document=doc, rev=doc.rev) - session.sessionpresentation_set.add(pres) + session.presentations.add(pres) # slides title = "Slideshow" @@ -70,7 +70,7 @@ def make_interim_meeting(group,date,status='sched',tz='UTC'): uploaded_filename=file, group=group, rev=rev, states=[('slides','active'), ('reuse_policy', 'single')]) pres = SessionPresentation.objects.create(session=session, document=doc, rev=doc.rev) - session.sessionpresentation_set.add(pres) + session.presentations.add(pres) # return meeting @@ -198,24 +198,24 @@ def make_meeting_test_data(meeting=None, create_interims=False): doc = DocumentFactory.create(name='agenda-72-mars', type_id='agenda', title="Agenda", uploaded_filename="agenda-72-mars.txt", group=mars, rev='00', states=[('agenda','active')]) pres = SessionPresentation.objects.create(session=mars_session,document=doc,rev=doc.rev) - mars_session.sessionpresentation_set.add(pres) # + mars_session.presentations.add(pres) # doc = DocumentFactory.create(name='minutes-72-mars', type_id='minutes', title="Minutes", uploaded_filename="minutes-72-mars.md", group=mars, rev='00', states=[('minutes','active')]) pres = SessionPresentation.objects.create(session=mars_session,document=doc,rev=doc.rev) - mars_session.sessionpresentation_set.add(pres) + mars_session.presentations.add(pres) doc = DocumentFactory.create(name='slides-72-mars-1-active', type_id='slides', title="Slideshow", uploaded_filename="slides-72-mars.txt", group=mars, rev='00', states=[('slides','active'), ('reuse_policy', 'single')]) pres = SessionPresentation.objects.create(session=mars_session,document=doc,rev=doc.rev) - mars_session.sessionpresentation_set.add(pres) + mars_session.presentations.add(pres) doc = DocumentFactory.create(name='slides-72-mars-2-deleted', type_id='slides', title="Bad Slideshow", uploaded_filename="slides-72-mars-2-deleted.txt", group=mars, rev='00', states=[('slides','deleted'), ('reuse_policy', 'single')]) pres = SessionPresentation.objects.create(session=mars_session,document=doc,rev=doc.rev) - mars_session.sessionpresentation_set.add(pres) + mars_session.presentations.add(pres) # Future Interim Meetings date = date_today() + datetime.timedelta(days=365) diff --git a/ietf/meeting/tests_helpers.py b/ietf/meeting/tests_helpers.py index 9ce3c21cbc..b118b9f041 100644 --- a/ietf/meeting/tests_helpers.py +++ b/ietf/meeting/tests_helpers.py @@ -487,7 +487,7 @@ def test_create_interim_session_conferences(self, mock): mock.reset_mock() mock_conf_mgr.create.return_value = [ Conference( - manager=mock_conf_mgr, id=1, public_id='some-uuid', description='desc', + manager=mock_conf_mgr, id=int(sessions[0].pk), public_id='some-uuid', description='desc', start_time=timeslots[0].utc_start_time(), duration=timeslots[0].duration, url='fake-meetecho-url', deletion_token='please-delete-me', ), @@ -498,6 +498,7 @@ def test_create_interim_session_conferences(self, mock): mock_conf_mgr.create.call_args[1], { 'group': sessions[0].group, + 'session_id': sessions[0].id, 'description': str(sessions[0]), 'start_time': timeslots[0].utc_start_time(), 'duration': timeslots[0].duration, @@ -512,12 +513,12 @@ def test_create_interim_session_conferences(self, mock): mock.reset_mock() mock_conf_mgr.create.side_effect = [ [Conference( - manager=mock_conf_mgr, id=1, public_id='some-uuid', description='desc', + manager=mock_conf_mgr, id=int(sessions[0].pk), public_id='some-uuid', description='desc', start_time=timeslots[0].utc_start_time(), duration=timeslots[0].duration, url='different-fake-meetecho-url', deletion_token='please-delete-me', )], [Conference( - manager=mock_conf_mgr, id=2, public_id='another-uuid', description='desc', + manager=mock_conf_mgr, id=int(sessions[1].pk), public_id='another-uuid', description='desc', start_time=timeslots[1].utc_start_time(), duration=timeslots[1].duration, url='another-fake-meetecho-url', deletion_token='please-delete-me-too', )], @@ -528,16 +529,18 @@ def test_create_interim_session_conferences(self, mock): mock_conf_mgr.create.call_args_list, [ ({ - 'group': sessions[0].group, - 'description': str(sessions[0]), - 'start_time': timeslots[0].utc_start_time(), - 'duration': timeslots[0].duration, + 'group': sessions[0].group, + 'session_id': sessions[0].id, + 'description': str(sessions[0]), + 'start_time': timeslots[0].utc_start_time(), + 'duration': timeslots[0].duration, },), ({ - 'group': sessions[1].group, - 'description': str(sessions[1]), - 'start_time': timeslots[1].utc_start_time(), - 'duration': timeslots[1].duration, + 'group': sessions[1].group, + 'session_id': sessions[1].id, + 'description': str(sessions[1]), + 'start_time': timeslots[1].utc_start_time(), + 'duration': timeslots[1].duration, },), ] ) diff --git a/ietf/meeting/tests_js.py b/ietf/meeting/tests_js.py index 0e6a423437..3269342924 100644 --- a/ietf/meeting/tests_js.py +++ b/ietf/meeting/tests_js.py @@ -5,11 +5,9 @@ import time import datetime import shutil -import os +import tempfile import re -from unittest import skipIf -import django from django.utils import timezone from django.utils.text import slugify from django.db.models import F @@ -134,7 +132,7 @@ def test_edit_meeting_schedule(self): self.assertEqual(session_info_container.find_element(By.CSS_SELECTOR, ".other-session .time").text, "not yet scheduled") # deselect - self.driver.find_element(By.CSS_SELECTOR, '.drop-target').click() + self.driver.find_element(By.CSS_SELECTOR, '.timeslot[data-type="regular"] .drop-target').click() self.assertEqual(session_info_container.find_elements(By.CSS_SELECTOR, ".title"), []) self.assertNotIn('other-session-selected', s2b_element.get_attribute('class')) @@ -193,9 +191,9 @@ def test_edit_meeting_schedule(self): # violated due to constraints - both the timeslot and its timeslot label self.assertTrue(self.driver.find_elements(By.CSS_SELECTOR, '#timeslot{}.would-violate-hint'.format(slot1.pk))) - # Find the timeslot label for slot1 - it's the first timeslot in the first room group + # Find the timeslot label for slot1 - it's the first timeslot in the room group containing room 1 slot1_roomgroup_elt = self.driver.find_element(By.CSS_SELECTOR, - '.day-flow .day:first-child .room-group:nth-child(2)' # count from 2 - first-child is the day label + f'.day-flow .day:first-child .room-group[data-rooms="{room1.pk}"]' ) self.assertTrue( slot1_roomgroup_elt.find_elements(By.CSS_SELECTOR, @@ -251,7 +249,9 @@ def test_edit_meeting_schedule(self): self.assertTrue(s1_element.is_displayed()) # should still be displayed self.assertIn('hidden-parent', s1_element.get_attribute('class'), 'Session should be hidden when parent disabled') - s1_element.click() # try to select + + self.scroll_and_click((By.CSS_SELECTOR, '#session{}'.format(s1.pk))) + self.assertNotIn('selected', s1_element.get_attribute('class'), 'Session should not be selectable when parent disabled') @@ -301,9 +301,9 @@ def test_edit_meeting_schedule(self): 'Session s1 should have moved to second meeting day') # swap timeslot column - put session in a differently-timed timeslot - self.driver.find_element(By.CSS_SELECTOR, + self.scroll_and_click((By.CSS_SELECTOR, '.day .swap-timeslot-col[data-timeslot-pk="{}"]'.format(slot1b.pk) - ).click() # open modal on the second timeslot for room1 + )) # open modal on the second timeslot for room1 self.assertTrue(self.driver.find_element(By.CSS_SELECTOR, "#swap-timeslot-col-modal").is_displayed()) self.driver.find_element(By.CSS_SELECTOR, '#swap-timeslot-col-modal input[name="target_timeslot"][value="{}"]'.format(slot4.pk) @@ -501,7 +501,7 @@ def test_past_swap_days_buttons(self): clicked_index = 1 # scroll so the button we want to click is just below the navbar, otherwise it may # fall beneath the sessions panel - navbar = self.driver.find_element_by_class_name('navbar') + navbar = self.driver.find_element(By.CSS_SELECTOR, '.navbar') self.driver.execute_script( 'window.scrollBy({top: %s, behavior: "instant"})' % ( future_swap_days_buttons[1].location['y'] - navbar.size['height'] @@ -833,7 +833,7 @@ def test_unassigned_sessions_drop_target_visible_when_empty(self): def test_session_constraint_hints(self): """Selecting a session should mark conflicting sessions - To test for recurrence of https://trac.ietf.org/trac/ietfdb/ticket/3327 need to have some constraints that + To test for recurrence of https://github.com/ietf-tools/datatracker/issues/3327 need to have some constraints that do not conflict. Testing with only violated constraints does not exercise the code adequately. """ meeting = MeetingFactory(type_id='ietf', date=date_today(), populate_schedule=False) @@ -880,51 +880,15 @@ def test_session_constraint_hints(self): self.assertNotIn('would-violate-hint', session_elements[4].get_attribute('class'), 'Constraint violation should not be indicated on non-conflicting session') -@ifSeleniumEnabled -@skipIf(django.VERSION[0]==2, "Skipping test with race conditions under Django 2") -class ScheduleEditTests(IetfSeleniumTestCase): - def testUnschedule(self): - - meeting = make_meeting_test_data() - - self.assertEqual(SchedTimeSessAssignment.objects.filter(session__meeting=meeting, session__group__acronym='mars', schedule__name='test-schedule').count(),1) - - - ss = list(SchedTimeSessAssignment.objects.filter(session__meeting__number=72,session__group__acronym='mars',schedule__name='test-schedule')) # pyflakes:ignore - - self.login() - url = self.absreverse('ietf.meeting.views.edit_meeting_schedule',kwargs=dict(num='72',name='test-schedule',owner='plain@example.com')) - self.driver.get(url) - - # driver.get() will wait for scripts to finish, but not ajax - # requests. Wait for completion of the permissions check: - read_only_note = self.driver.find_element(By.ID, 'read_only') - WebDriverWait(self.driver, 10).until(expected_conditions.invisibility_of_element(read_only_note), "Read-only schedule") - - s1 = Session.objects.filter(group__acronym='mars', meeting=meeting).first() - selector = "#session_{}".format(s1.pk) - WebDriverWait(self.driver, 30).until(expected_conditions.presence_of_element_located((By.CSS_SELECTOR, selector)), "Did not find %s"%selector) - - self.assertEqual(self.driver.find_elements(By.CSS_SELECTOR, "#sortable-list #session_{}".format(s1.pk)), []) - - element = self.driver.find_element(By.ID, 'session_{}'.format(s1.pk)) - target = self.driver.find_element(By.ID, 'sortable-list') - ActionChains(self.driver).drag_and_drop(element,target).perform() - - self.assertTrue(self.driver.find_elements(By.CSS_SELECTOR, "#sortable-list #session_{}".format(s1.pk))) - - time.sleep(0.1) # The API that modifies the database runs async - - self.assertEqual(SchedTimeSessAssignment.objects.filter(session__meeting__number=72,session__group__acronym='mars',schedule__name='test-schedule').count(),0) @ifSeleniumEnabled class SlideReorderTests(IetfSeleniumTestCase): def setUp(self): super(SlideReorderTests, self).setUp() self.session = SessionFactory(meeting__type_id='ietf', status_id='sched') - self.session.sessionpresentation_set.create(document=DocumentFactory(type_id='slides',name='one'),order=1) - self.session.sessionpresentation_set.create(document=DocumentFactory(type_id='slides',name='two'),order=2) - self.session.sessionpresentation_set.create(document=DocumentFactory(type_id='slides',name='three'),order=3) + self.session.presentations.create(document=DocumentFactory(type_id='slides',name='one'),order=1) + self.session.presentations.create(document=DocumentFactory(type_id='slides',name='two'),order=2) + self.session.presentations.create(document=DocumentFactory(type_id='slides',name='three'),order=3) def secr_login(self): self.login('secretary') @@ -944,7 +908,7 @@ def testReorderSlides(self): ActionChains(self.driver).drag_and_drop(second,third).perform() time.sleep(0.1) # The API that modifies the database runs async - names=self.session.sessionpresentation_set.values_list('document__name',flat=True) + names=self.session.presentations.values_list('document__name',flat=True) self.assertEqual(list(names),['one','three','two']) @ifSeleniumEnabled @@ -977,13 +941,8 @@ def tearDown(self): def tempdir(self, label): # Borrowed from test_utils.TestCase slug = slugify(self.__class__.__name__.replace('.','-')) - dirname = "tmp-{label}-{slug}-dir".format(**locals()) - if 'VIRTUAL_ENV' in os.environ: - dirname = os.path.join(os.environ['VIRTUAL_ENV'], dirname) - path = os.path.abspath(dirname) - if not os.path.exists(path): - os.mkdir(path) - return path + suffix = "-{label}-{slug}-dir".format(**locals()) + return tempfile.mkdtemp(suffix=suffix) def displayed_interims(self, groups=None): sessions = add_event_info_to_session_qs( @@ -1083,6 +1042,7 @@ def advance_month(): def do_upcoming_view_filter_test(self, querystring, visible_meetings=()): self.login() self.driver.get(self.absreverse('ietf.meeting.views.upcoming') + querystring) + time.sleep(0.2) # gross, but give the filter JS time to do its thing self.assert_upcoming_meeting_visibility(visible_meetings) self.assert_upcoming_meeting_calendar(visible_meetings) self.assert_upcoming_view_filter_matches_ics_filter(querystring) @@ -1270,10 +1230,13 @@ def _assert_ietf_tz_correct(meetings, tz): self.driver.get(self.absreverse('ietf.meeting.views.upcoming')) tz_select_input = self.driver.find_element(By.ID, 'timezone-select') tz_select_bottom_input = self.driver.find_element(By.ID, 'timezone-select-bottom') - local_tz_link = self.driver.find_element(By.ID, 'local-timezone') - utc_tz_link = self.driver.find_element(By.ID, 'utc-timezone') - local_tz_bottom_link = self.driver.find_element(By.ID, 'local-timezone-bottom') - utc_tz_bottom_link = self.driver.find_element(By.ID, 'utc-timezone-bottom') + + # For things we click, need to click the labels / actually visible items. The actual inputs are hidden + # and managed by the JS. + local_tz_link = self.driver.find_element(By.CSS_SELECTOR, 'label[for="local-timezone"]') + utc_tz_link = self.driver.find_element(By.CSS_SELECTOR, 'label[for="utc-timezone"]') + local_tz_bottom_link = self.driver.find_element(By.CSS_SELECTOR, 'label[for="local-timezone-bottom"]') + utc_tz_bottom_link = self.driver.find_element(By.CSS_SELECTOR, 'label[for="utc-timezone-bottom"]') # wait for the select box to be updated - look for an arbitrary time zone to be in # its options list to detect this @@ -1283,7 +1246,10 @@ def _assert_ietf_tz_correct(meetings, tz): (By.CSS_SELECTOR, '#timezone-select > option[value="%s"]' % arbitrary_tz) ) ) - + tz_selector_clickables = self.driver.find_elements(By.CSS_SELECTOR, ".tz-display .select2") + self.assertEqual(len(tz_selector_clickables), 2) + (tz_selector_top, tz_selector_bottom) = tz_selector_clickables + arbitrary_tz_bottom_opt = tz_select_bottom_input.find_element(By.CSS_SELECTOR, '#timezone-select-bottom > option[value="%s"]' % arbitrary_tz) @@ -1294,7 +1260,7 @@ def _assert_ietf_tz_correct(meetings, tz): # to inherit Django's settings.TIME_ZONE but I don't know whether that's guaranteed to be consistent. # To avoid test fragility, ask Moment what it considers local and expect that. local_tz = self.driver.execute_script('return moment.tz.guess();') - local_tz_opt = tz_select_input.find_element(By.CSS_SELECTOR, 'option[value=%s]' % local_tz) + local_tz_opt = tz_select_input.find_element(By.CSS_SELECTOR, 'option[value="%s"]' % local_tz) local_tz_bottom_opt = tz_select_bottom_input.find_element(By.CSS_SELECTOR, 'option[value="%s"]' % local_tz) # Should start off in local time zone @@ -1304,8 +1270,7 @@ def _assert_ietf_tz_correct(meetings, tz): _assert_ietf_tz_correct(ietf_meetings, local_tz) # click 'utc' button - self.driver.execute_script("arguments[0].click();", utc_tz_link) # FIXME-LARS: not working: - # utc_tz_link.click() + utc_tz_link.click() self.wait.until(expected_conditions.element_to_be_selected(utc_tz_opt)) self.assertFalse(local_tz_opt.is_selected()) self.assertFalse(local_tz_bottom_opt.is_selected()) @@ -1317,8 +1282,7 @@ def _assert_ietf_tz_correct(meetings, tz): _assert_ietf_tz_correct(ietf_meetings, 'UTC') # click back to 'local' - self.driver.execute_script("arguments[0].click();", local_tz_link) # FIXME-LARS: not working: - # local_tz_link.click() + local_tz_link.click() self.wait.until(expected_conditions.element_to_be_selected(local_tz_opt)) self.assertTrue(local_tz_opt.is_selected()) self.assertTrue(local_tz_bottom_opt.is_selected()) @@ -1330,7 +1294,12 @@ def _assert_ietf_tz_correct(meetings, tz): _assert_ietf_tz_correct(ietf_meetings, local_tz) # Now select a different item from the select input - arbitrary_tz_opt.click() + tz_selector_top.click() + self.wait.until( + expected_conditions.presence_of_element_located( + (By.CSS_SELECTOR, 'span.select2-container .select2-results li[id$="America/Halifax"]') + ) + ).click() self.wait.until(expected_conditions.element_to_be_selected(arbitrary_tz_opt)) self.assertFalse(local_tz_opt.is_selected()) self.assertFalse(local_tz_bottom_opt.is_selected()) @@ -1343,8 +1312,8 @@ def _assert_ietf_tz_correct(meetings, tz): # Now repeat those tests using the widgets at the bottom of the page # click 'utc' button - self.driver.execute_script("arguments[0].click();", utc_tz_bottom_link) # FIXME-LARS: not working: - # utc_tz_bottom_link.click() + self.scroll_to_element(utc_tz_bottom_link) + utc_tz_bottom_link.click() self.wait.until(expected_conditions.element_to_be_selected(utc_tz_opt)) self.assertFalse(local_tz_opt.is_selected()) self.assertFalse(local_tz_bottom_opt.is_selected()) @@ -1356,8 +1325,8 @@ def _assert_ietf_tz_correct(meetings, tz): _assert_ietf_tz_correct(ietf_meetings, 'UTC') # click back to 'local' - self.driver.execute_script("arguments[0].click();", local_tz_bottom_link) # FIXME-LARS: not working: - # local_tz_bottom_link.click() + self.scroll_to_element(local_tz_bottom_link) + local_tz_bottom_link.click() self.wait.until(expected_conditions.element_to_be_selected(local_tz_opt)) self.assertTrue(local_tz_opt.is_selected()) self.assertTrue(local_tz_bottom_opt.is_selected()) @@ -1369,7 +1338,13 @@ def _assert_ietf_tz_correct(meetings, tz): _assert_ietf_tz_correct(ietf_meetings, local_tz) # Now select a different item from the select input - arbitrary_tz_bottom_opt.click() + self.scroll_to_element(tz_selector_bottom) + tz_selector_bottom.click() + self.wait.until( + expected_conditions.presence_of_element_located( + (By.CSS_SELECTOR, 'span.select2-container .select2-results li[id$="America/Halifax"]') + ) + ).click() self.wait.until(expected_conditions.element_to_be_selected(arbitrary_tz_opt)) self.assertFalse(local_tz_opt.is_selected()) self.assertFalse(local_tz_bottom_opt.is_selected()) @@ -1400,13 +1375,8 @@ def test_upcoming_materials_modal(self): self.assertFalse(modal_div.is_displayed()) # Click the 'materials' button - open_modal_button = self.wait.until( - expected_conditions.element_to_be_clickable( - (By.CSS_SELECTOR, '[data-bs-target="#modal-%s"]' % slug) - ), - 'Modal open button not found or not clickable', - ) - open_modal_button.click() + open_modal_button_locator = (By.CSS_SELECTOR, '[data-bs-target="#modal-%s"]' % slug) + self.scroll_and_click(open_modal_button_locator) self.wait.until( expected_conditions.visibility_of(modal_div), 'Modal did not become visible after clicking open button', @@ -1420,6 +1390,7 @@ def test_upcoming_materials_modal(self): ), 'Modal close button not found or not clickable', ) + time.sleep(0.3) # gross, but the button is clickable while still fading in close_modal_button.click() self.wait.until( expected_conditions.invisibility_of_element(modal_div), @@ -1531,7 +1502,7 @@ class EditTimeslotsTests(IetfSeleniumTestCase): """Test the timeslot editor""" def setUp(self): super().setUp() - self.meeting: Meeting = MeetingFactory( + self.meeting: Meeting = MeetingFactory( # type: ignore[annotation-unchecked] type_id='ietf', number=120, date=date_today() + datetime.timedelta(days=10), @@ -1605,16 +1576,16 @@ def test_delete_timeslot_cancel(self): def do_delete_time_interval_test(self, cancel=False): delete_time_local = datetime_from_date(self.meeting.date, self.meeting.tz()).replace(hour=10) - delete_time = delete_time_local.astimezone(datetime.timezone.utc) + delete_time = delete_time_local.astimezone(datetime.UTC) duration = datetime.timedelta(minutes=60) - delete: [TimeSlot] = TimeSlotFactory.create_batch( + delete: [TimeSlot] = TimeSlotFactory.create_batch( # type: ignore[annotation-unchecked] 2, meeting=self.meeting, time=delete_time_local, duration=duration, ) - keep: [TimeSlot] = [ + keep: [TimeSlot] = [ # type: ignore[annotation-unchecked] TimeSlotFactory( meeting=self.meeting, time=keep_time, @@ -1651,14 +1622,14 @@ def do_delete_day_test(self, cancel=False): hours = [10, 12] other_days = [self.meeting.get_meeting_date(d) for d in range(1, 3)] - delete: [TimeSlot] = [ + delete: [TimeSlot] = [ # type: ignore[annotation-unchecked] TimeSlotFactory( meeting=self.meeting, time=datetime_from_date(delete_day, self.meeting.tz()).replace(hour=hour), ) for hour in hours ] - keep: [TimeSlot] = [ + keep: [TimeSlot] = [ # type: ignore[annotation-unchecked] TimeSlotFactory( meeting=self.meeting, time=datetime_from_date(day, self.meeting.tz()).replace(hour=hour), diff --git a/ietf/meeting/tests_models.py b/ietf/meeting/tests_models.py index 8ea0d0c5b3..869d9ec814 100644 --- a/ietf/meeting/tests_models.py +++ b/ietf/meeting/tests_models.py @@ -1,13 +1,18 @@ -# Copyright The IETF Trust 2021, All Rights Reserved +# Copyright The IETF Trust 2021-2024, All Rights Reserved # -*- coding: utf-8 -*- """Tests of models in the Meeting application""" import datetime -from mock import patch +from unittest.mock import patch +from django.conf import settings +from django.test import override_settings + +import ietf.meeting.models from ietf.group.factories import GroupFactory, GroupHistoryFactory -from ietf.meeting.factories import MeetingFactory, SessionFactory, AttendedFactory -from ietf.stats.factories import MeetingRegistrationFactory +from ietf.meeting.factories import MeetingFactory, SessionFactory, AttendedFactory, SessionPresentationFactory +from ietf.meeting.factories import RegistrationFactory +from ietf.meeting.models import Session from ietf.utils.test_utils import TestCase from ietf.utils.timezone import date_today, datetime_today @@ -16,9 +21,9 @@ class MeetingTests(TestCase): def test_get_attendance_pre110(self): """Pre-110 meetings do not calculate attendance""" meeting = MeetingFactory(type_id='ietf', number='109') - MeetingRegistrationFactory.create_batch(3, meeting=meeting, reg_type='') - MeetingRegistrationFactory.create_batch(4, meeting=meeting, reg_type='remote') - MeetingRegistrationFactory.create_batch(5, meeting=meeting, reg_type='in_person') + RegistrationFactory.create_batch(3, meeting=meeting, with_ticket={'attendance_type_id': 'unknown'}) + RegistrationFactory.create_batch(4, meeting=meeting, with_ticket={'attendance_type_id': 'remote'}) + RegistrationFactory.create_batch(5, meeting=meeting, with_ticket={'attendance_type_id': 'onsite'}) self.assertIsNone(meeting.get_attendance()) def test_get_attendance_110(self): @@ -26,31 +31,31 @@ def test_get_attendance_110(self): meeting = MeetingFactory(type_id='ietf', number='110') # start with attendees that should be ignored - MeetingRegistrationFactory.create_batch(3, meeting=meeting, reg_type='', attended=True) - MeetingRegistrationFactory(meeting=meeting, reg_type='', attended=False) + RegistrationFactory.create_batch(3, meeting=meeting, with_ticket={'attendance_type_id': 'unknown'}, attended=True) + RegistrationFactory(meeting=meeting, with_ticket={'attendance_type_id': 'unknown'}, attended=False) attendance = meeting.get_attendance() self.assertIsNotNone(attendance) self.assertEqual(attendance.remote, 0) self.assertEqual(attendance.onsite, 0) # add online attendees with at least one who registered but did not attend - MeetingRegistrationFactory.create_batch(4, meeting=meeting, reg_type='remote', attended=True) - MeetingRegistrationFactory(meeting=meeting, reg_type='remote', attended=False) + RegistrationFactory.create_batch(4, meeting=meeting, with_ticket={'attendance_type_id': 'remote'}, attended=True) + RegistrationFactory(meeting=meeting, with_ticket={'attendance_type_id': 'remote'}, attended=False) attendance = meeting.get_attendance() self.assertIsNotNone(attendance) self.assertEqual(attendance.remote, 4) self.assertEqual(attendance.onsite, 0) # and the same for onsite attendees - MeetingRegistrationFactory.create_batch(5, meeting=meeting, reg_type='onsite', attended=True) - MeetingRegistrationFactory(meeting=meeting, reg_type='in_person', attended=False) + RegistrationFactory.create_batch(5, meeting=meeting, with_ticket={'attendance_type_id': 'onsite'}, attended=True) + RegistrationFactory(meeting=meeting, with_ticket={'attendance_type_id': 'onsite'}, attended=False) attendance = meeting.get_attendance() self.assertIsNotNone(attendance) self.assertEqual(attendance.remote, 4) self.assertEqual(attendance.onsite, 5) # and once more after removing all the online attendees - meeting.meetingregistration_set.filter(reg_type='remote').delete() + meeting.registration_set.remote().delete() attendance = meeting.get_attendance() self.assertIsNotNone(attendance) self.assertEqual(attendance.remote, 0) @@ -59,11 +64,11 @@ def test_get_attendance_110(self): def test_get_attendance_113(self): """Simulate IETF 113 attendance gathering data""" meeting = MeetingFactory(type_id='ietf', number='113') - MeetingRegistrationFactory(meeting=meeting, reg_type='onsite', attended=True, checkedin=False) - MeetingRegistrationFactory(meeting=meeting, reg_type='onsite', attended=False, checkedin=True) - p1 = MeetingRegistrationFactory(meeting=meeting, reg_type='onsite', attended=False, checkedin=False).person + RegistrationFactory(meeting=meeting, with_ticket={'attendance_type_id': 'onsite'}, attended=True, checkedin=False) + RegistrationFactory(meeting=meeting, with_ticket={'attendance_type_id': 'onsite'}, attended=False, checkedin=True) + p1 = RegistrationFactory(meeting=meeting, with_ticket={'attendance_type_id': 'onsite'}, attended=False, checkedin=False).person AttendedFactory(session__meeting=meeting, person=p1) - p2 = MeetingRegistrationFactory(meeting=meeting, reg_type='remote', attended=False, checkedin=False).person + p2 = RegistrationFactory(meeting=meeting, with_ticket={'attendance_type_id': 'remote'}, attended=False, checkedin=False).person AttendedFactory(session__meeting=meeting, person=p2) attendance = meeting.get_attendance() self.assertEqual(attendance.onsite, 3) @@ -77,9 +82,9 @@ def test_get_attendance_keeps_meetings_distinct(self): # Create a person who attended a remote session for first_mtg and onsite for second_mtg without # checking in for either. - p = MeetingRegistrationFactory(meeting=second_mtg, reg_type='onsite', attended=False, checkedin=False).person + p = RegistrationFactory(meeting=second_mtg, with_ticket={'attendance_type_id': 'onsite'}, attended=False, checkedin=False).person AttendedFactory(session__meeting=first_mtg, person=p) - MeetingRegistrationFactory(meeting=first_mtg, person=p, reg_type='remote', attended=False, checkedin=False) + RegistrationFactory(meeting=first_mtg, person=p, with_ticket={'attendance_type_id': 'remote'}, attended=False, checkedin=False) AttendedFactory(session__meeting=second_mtg, person=p) att = first_mtg.get_attendance() @@ -116,7 +121,80 @@ def test_group_at_the_time(self): class SessionTests(TestCase): - def test_chat_archive_url_with_jabber(self): + def test_chat_archive_url(self): + session = SessionFactory( + meeting__date=datetime.date.today(), + meeting__number=120, # needs to use proceedings_format_version > 1 + ) + with override_settings(): + if hasattr(settings, 'CHAT_ARCHIVE_URL_PATTERN'): + del settings.CHAT_ARCHIVE_URL_PATTERN + self.assertEqual(session.chat_archive_url(), session.chat_room_url()) + settings.CHAT_ARCHIVE_URL_PATTERN = 'http://chat.example.com' + self.assertEqual(session.chat_archive_url(), 'http://chat.example.com') + chatlog = SessionPresentationFactory(session=session, document__type_id='chatlog').document + self.assertEqual(session.chat_archive_url(), chatlog.get_href()) + # datatracker 8.8.0 rolled out on 2022-07-15. Before that, chat logs were jabber logs hosted at www.ietf.org. session_with_jabber = SessionFactory(group__acronym='fakeacronym', meeting__date=datetime.date(2022,7,14)) self.assertEqual(session_with_jabber.chat_archive_url(), 'https://www.ietf.org/jabber/logs/fakeacronym?C=M;O=D') + chatlog = SessionPresentationFactory(session=session_with_jabber, document__type_id='chatlog').document + self.assertEqual(session_with_jabber.chat_archive_url(), chatlog.get_href()) + + def test_chat_room_name(self): + session = SessionFactory(group__acronym='xyzzy') + self.assertEqual(session.chat_room_name(), 'xyzzy') + session.type_id = 'plenary' + self.assertEqual(session.chat_room_name(), 'plenary') + session.chat_room = 'fnord' + self.assertEqual(session.chat_room_name(), 'fnord') + + def test_alpha_str(self): + self.assertEqual(Session._alpha_str(0), "a") + self.assertEqual(Session._alpha_str(1), "b") + self.assertEqual(Session._alpha_str(25), "z") + self.assertEqual(Session._alpha_str(26), "aa") + self.assertEqual(Session._alpha_str(27 * 26 - 1), "zz") + self.assertEqual(Session._alpha_str(27 * 26), "aaa") + + @patch.object(ietf.meeting.models.Session, "_session_recording_url_label", return_value="LABEL") + def test_session_recording_url(self, mock): + for session_type in ["ietf", "interim"]: + session = SessionFactory(meeting__type_id=session_type) + with override_settings(): + if hasattr(settings, "MEETECHO_SESSION_RECORDING_URL"): + del settings.MEETECHO_SESSION_RECORDING_URL + self.assertIsNone(session.session_recording_url()) + + settings.MEETECHO_SESSION_RECORDING_URL = "http://player.example.com" + self.assertEqual(session.session_recording_url(), "http://player.example.com") + + settings.MEETECHO_SESSION_RECORDING_URL = "http://player.example.com?{session_label}" + self.assertEqual(session.session_recording_url(), "http://player.example.com?LABEL") + + session.meetecho_recording_name="actualname" + session.save() + self.assertEqual(session.session_recording_url(), "http://player.example.com?actualname") + + def test_session_recording_url_label_ietf(self): + session = SessionFactory( + meeting__type_id='ietf', + meeting__date=date_today(), + meeting__number="123", + group__acronym="acro", + ) + session_time = session.official_timeslotassignment().timeslot.time + self.assertEqual( + f"IETF123-ACRO-{session_time:%Y%m%d-%H%M}", # n.b., time in label is UTC + session._session_recording_url_label()) + + def test_session_recording_url_label_interim(self): + session = SessionFactory( + meeting__type_id='interim', + meeting__date=date_today(), + group__acronym="acro", + ) + session_time = session.official_timeslotassignment().timeslot.time + self.assertEqual( + f"IETF-ACRO-{session_time:%Y%m%d-%H%M}", # n.b., time in label is UTC + session._session_recording_url_label()) diff --git a/ietf/meeting/tests_schedule_forms.py b/ietf/meeting/tests_schedule_forms.py index 58c1332bd5..426d26dc2d 100644 --- a/ietf/meeting/tests_schedule_forms.py +++ b/ietf/meeting/tests_schedule_forms.py @@ -140,13 +140,13 @@ def test_location_options(self): rendered = str(TimeSlotEditForm(instance=ts)['location']) # noinspection PyTypeChecker self.assertInHTML( - f'', + f'', rendered, ) for room in rooms: # noinspection PyTypeChecker self.assertInHTML( - f'', + f'', rendered, ) diff --git a/ietf/meeting/tests_schedule_generator.py b/ietf/meeting/tests_schedule_generator.py index 0cc430e18a..8c5f71a74e 100644 --- a/ietf/meeting/tests_schedule_generator.py +++ b/ietf/meeting/tests_schedule_generator.py @@ -78,12 +78,13 @@ def test_normal_schedule(self): self.stdout.seek(0) output = self.stdout.read() - self.assertIn('WARNING: session wg2 (pk 13) has no attendees set', output) + wg2_no_attendees_session_pk = [s.session_pk for s in generator.schedule.sessions if s.group == "wg2" and not s.attendees][0] + self.assertIn(f'WARNING: session wg2 (pk {wg2_no_attendees_session_pk}) has no attendees set', output) self.assertIn('scheduling 13 sessions in 20 timeslots', output) self.assertIn('Optimiser starting run 1', output) self.assertIn('Optimiser found an optimal schedule', output) - schedule = self.meeting.schedule_set.get(name__startswith='Auto-') + schedule = self.meeting.schedule_set.get(name__startswith='auto-') self.assertEqual(schedule.assignments.count(), 13) def test_unresolvable_schedule(self): @@ -160,7 +161,8 @@ def test_base_schedule(self): self.assertIn('Applying schedule {} as base schedule'.format( generate_schedule.ScheduleId.from_schedule(base_schedule) ), output) - self.assertIn('WARNING: session wg2 (pk 13) has no attendees set', output) + wg2_no_attendees_session_pk = [s.session_pk for s in generator.schedule.sessions if s.group == "wg2" and not s.attendees][0] + self.assertIn(f'WARNING: session wg2 (pk {wg2_no_attendees_session_pk}) has no attendees set', output) self.assertIn('scheduling 13 sessions in 19 timeslots', output) # 19 because base is using one self.assertIn('Optimiser starting run 1', output) self.assertIn('Optimiser found an optimal schedule', output) diff --git a/ietf/meeting/tests_session_requests.py b/ietf/meeting/tests_session_requests.py new file mode 100644 index 0000000000..42dbee5f23 --- /dev/null +++ b/ietf/meeting/tests_session_requests.py @@ -0,0 +1,1133 @@ +# Copyright The IETF Trust 2013-2025, All Rights Reserved +# -*- coding: utf-8 -*- + + +import datetime + +from django.urls import reverse + +import debug # pyflakes:ignore + +from ietf.utils.test_utils import TestCase +from ietf.group.factories import GroupFactory, RoleFactory +from ietf.meeting.models import Session, ResourceAssociation, SchedulingEvent, Constraint +from ietf.meeting.factories import MeetingFactory, SessionFactory +from ietf.name.models import ConstraintName, TimerangeName +from ietf.person.factories import PersonFactory +from ietf.person.models import Person +from ietf.meeting.forms import SessionRequestForm +from ietf.utils.mail import outbox, empty_outbox, get_payload_text, send_mail +from ietf.utils.timezone import date_today + + +from pyquery import PyQuery + +SECR_USER = 'secretary' + + +class SessionRequestTestCase(TestCase): + def test_main(self): + meeting = MeetingFactory(type_id='ietf', date=date_today()) + SessionFactory.create_batch(2, meeting=meeting, status_id='sched') + SessionFactory.create_batch(2, meeting=meeting, status_id='disappr') + # Several unscheduled groups come from make_immutable_base_data + url = reverse('ietf.meeting.views_session_request.list_view') + self.client.login(username="secretary", password="secretary+password") + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + sched = r.context['scheduled_groups'] + self.assertEqual(len(sched), 2) + unsched = r.context['unscheduled_groups'] + self.assertEqual(len(unsched), 12) + + def test_approve(self): + meeting = MeetingFactory(type_id='ietf', date=date_today()) + ad = Person.objects.get(user__username='ad') + area = RoleFactory(name_id='ad', person=ad, group__type_id='area').group + mars = GroupFactory(parent=area, acronym='mars') + # create session waiting for approval + session = SessionFactory(meeting=meeting, group=mars, status_id='apprw') + url = reverse('ietf.meeting.views_session_request.approve_request', kwargs={'acronym': 'mars'}) + self.client.login(username="ad", password="ad+password") + r = self.client.get(url) + self.assertRedirects(r, reverse('ietf.meeting.views_session_request.view_request', kwargs={'acronym': 'mars'})) + self.assertEqual(SchedulingEvent.objects.filter(session=session).order_by('-id')[0].status_id, 'appr') + + def test_cancel(self): + meeting = MeetingFactory(type_id='ietf', date=date_today()) + ad = Person.objects.get(user__username='ad') + area = RoleFactory(name_id='ad', person=ad, group__type_id='area').group + session = SessionFactory(meeting=meeting, group__parent=area, group__acronym='mars', status_id='sched') + url = reverse('ietf.meeting.views_session_request.cancel_request', kwargs={'acronym': 'mars'}) + self.client.login(username="ad", password="ad+password") + r = self.client.get(url) + self.assertRedirects(r, reverse('ietf.meeting.views_session_request.list_view')) + self.assertEqual(SchedulingEvent.objects.filter(session=session).order_by('-id')[0].status_id, 'deleted') + + def test_cancel_notification_msg(self): + to = "" + subject = "Dummy subject" + template = "meeting/session_cancel_notification.txt" + meeting = MeetingFactory(type_id="ietf", date=date_today()) + requester = PersonFactory(name="James O'Rourke", user__username="jimorourke") + context = {"meeting": meeting, "requester": requester} + cc = "cc.a@example.com, cc.b@example.com" + bcc = "bcc@example.com" + + msg = send_mail( + None, + to, + None, + subject, + template, + context, + cc=cc, + bcc=bcc, + ) + self.assertEqual(requester.name, "James O'Rourke") # note ' (single quote) in the name + self.assertIn( + f"A request to cancel a meeting session has just been submitted by {requester.name}.", + get_payload_text(msg), + ) + + def test_edit(self): + meeting = MeetingFactory(type_id='ietf', date=date_today()) + mars = RoleFactory(name_id='chair', person__user__username='marschairman', group__acronym='mars').group + group2 = GroupFactory() + group3 = GroupFactory() + group4 = GroupFactory() + iabprog = GroupFactory(type_id='program') + + SessionFactory(meeting=meeting, group=mars, status_id='sched') + + url = reverse('ietf.meeting.views_session_request.edit_request', kwargs={'acronym': 'mars'}) + self.client.login(username="marschairman", password="marschairman+password") + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + attendees = 10 + comments = 'need lights' + mars_sessions = meeting.session_set.filter(group__acronym='mars') + empty_outbox() + post_data = {'num_session': '2', + 'attendees': attendees, + 'constraint_chair_conflict': iabprog.acronym, + 'session_time_relation': 'subsequent-days', + 'adjacent_with_wg': group2.acronym, + 'joint_with_groups': group3.acronym + ' ' + group4.acronym, + 'joint_for_session': '2', + 'timeranges': ['thursday-afternoon-early', 'thursday-afternoon-late'], + 'session_set-TOTAL_FORMS': '3', # matches what view actually sends, even with only 2 filled in + 'session_set-INITIAL_FORMS': '1', + 'session_set-MIN_NUM_FORMS': '1', + 'session_set-MAX_NUM_FORMS': '3', + 'session_set-0-id': mars_sessions[0].pk, + 'session_set-0-name': mars_sessions[0].name, + 'session_set-0-short': mars_sessions[0].short, + 'session_set-0-purpose': mars_sessions[0].purpose_id, + 'session_set-0-type': mars_sessions[0].type_id, + 'session_set-0-requested_duration': '3600', + 'session_set-0-on_agenda': mars_sessions[0].on_agenda, + 'session_set-0-remote_instructions': mars_sessions[0].remote_instructions, + 'session_set-0-attendees': attendees, + 'session_set-0-comments': comments, + 'session_set-0-DELETE': '', + # no session_set-1-id because it's a new request + 'session_set-1-name': '', + 'session_set-1-short': '', + 'session_set-1-purpose': 'regular', + 'session_set-1-type': 'regular', + 'session_set-1-requested_duration': '3600', + 'session_set-1-on_agenda': True, + 'session_set-1-remote_instructions': mars_sessions[0].remote_instructions, + 'session_set-1-attendees': attendees, + 'session_set-1-comments': comments, + 'session_set-1-DELETE': '', + 'session_set-2-id': '', + 'session_set-2-name': '', + 'session_set-2-short': '', + 'session_set-2-purpose': 'regular', + 'session_set-2-type': 'regular', + 'session_set-2-requested_duration': '', + 'session_set-2-on_agenda': 'True', + 'session_set-2-attendees': attendees, + 'session_set-2-comments': '', + 'session_set-2-DELETE': 'on', + 'submit': 'Continue'} + r = self.client.post(url, post_data, HTTP_HOST='example.com') + redirect_url = reverse('ietf.meeting.views_session_request.view_request', kwargs={'acronym': 'mars'}) + self.assertRedirects(r, redirect_url) + + # Check whether updates were stored in the database + sessions = Session.objects.filter(meeting=meeting, group=mars).order_by("id") # order to match edit() view + self.assertEqual(len(sessions), 2) + session = sessions[0] + + self.assertEqual(session.constraints().get(name='chair_conflict').target.acronym, iabprog.acronym) + self.assertEqual(session.constraints().get(name='time_relation').time_relation, 'subsequent-days') + self.assertEqual(session.constraints().get(name='wg_adjacent').target.acronym, group2.acronym) + self.assertEqual( + list(session.constraints().get(name='timerange').timeranges.all().values('name')), + list(TimerangeName.objects.filter(name__in=['thursday-afternoon-early', 'thursday-afternoon-late']).values('name')) + ) + self.assertFalse(sessions[0].joint_with_groups.count()) + self.assertEqual(set(sessions[1].joint_with_groups.all()), {group3, group4}) + + # Check whether the updated data is visible on the view page + r = self.client.get(redirect_url) + self.assertContains(r, 'Schedule the sessions on subsequent days') + self.assertContains(r, 'Thursday early afternoon, Thursday late afternoon') + self.assertContains(r, group2.acronym) + # The sessions can be in any order in the HTML, deal with that + self.assertRegex(r.content.decode(), r'Second session with: ({} {}|{} {})'.format(group3.acronym, group4.acronym, group4.acronym, group3.acronym)) + + # check that a notification was sent + self.assertEqual(len(outbox), 1) + notification_payload = get_payload_text(outbox[0]) + self.assertIn('1 Hour, 1 Hour', notification_payload) + self.assertNotIn('1 Hour, 1 Hour, 1 Hour', notification_payload) + + # Edit again, changing the joint sessions and clearing some fields. The behaviour of + # edit is different depending on whether previous joint sessions were recorded. + empty_outbox() + post_data = {'num_session': '2', + 'attendees': attendees, + 'constraint_chair_conflict': '', + 'comments': 'need lights', + 'joint_with_groups': group2.acronym, + 'joint_for_session': '1', + 'session_set-TOTAL_FORMS': '3', # matches what view actually sends, even with only 2 filled in + 'session_set-INITIAL_FORMS': '2', + 'session_set-MIN_NUM_FORMS': '1', + 'session_set-MAX_NUM_FORMS': '3', + 'session_set-0-id': sessions[0].pk, + 'session_set-0-name': sessions[0].name, + 'session_set-0-short': sessions[0].short, + 'session_set-0-purpose': sessions[0].purpose_id, + 'session_set-0-type': sessions[0].type_id, + 'session_set-0-requested_duration': '3600', + 'session_set-0-on_agenda': sessions[0].on_agenda, + 'session_set-0-remote_instructions': sessions[0].remote_instructions, + 'session_set-0-attendees': sessions[0].attendees, + 'session_set-0-comments': sessions[1].comments, + 'session_set-0-DELETE': '', + 'session_set-1-id': sessions[1].pk, + 'session_set-1-name': sessions[1].name, + 'session_set-1-short': sessions[1].short, + 'session_set-1-purpose': sessions[1].purpose_id, + 'session_set-1-type': sessions[1].type_id, + 'session_set-1-requested_duration': '3600', + 'session_set-1-on_agenda': sessions[1].on_agenda, + 'session_set-1-remote_instructions': sessions[1].remote_instructions, + 'session_set-1-attendees': sessions[1].attendees, + 'session_set-1-comments': sessions[1].comments, + 'session_set-1-DELETE': '', + 'session_set-2-id': '', + 'session_set-2-name': '', + 'session_set-2-short': '', + 'session_set-2-purpose': 'regular', + 'session_set-2-type': 'regular', + 'session_set-2-requested_duration': '', + 'session_set-2-on_agenda': 'True', + 'session_set-2-attendees': attendees, + 'session_set-2-comments': '', + 'session_set-2-DELETE': 'on', + 'submit': 'Continue'} + r = self.client.post(url, post_data, HTTP_HOST='example.com') + self.assertRedirects(r, redirect_url) + + # Check whether updates were stored in the database + sessions = Session.objects.filter(meeting=meeting, group=mars).order_by("id") + self.assertEqual(len(sessions), 2) + session = sessions[0] + self.assertFalse(session.constraints().filter(name='time_relation')) + self.assertFalse(session.constraints().filter(name='wg_adjacent')) + self.assertFalse(session.constraints().filter(name='timerange')) + self.assertEqual(list(sessions[0].joint_with_groups.all()), [group2]) + self.assertFalse(sessions[1].joint_with_groups.count()) + + # check that a notification was sent + self.assertEqual(len(outbox), 1) + notification_payload = get_payload_text(outbox[0]) + self.assertIn('1 Hour, 1 Hour', notification_payload) + self.assertNotIn('1 Hour, 1 Hour, 1 Hour', notification_payload) + + # Check whether the updated data is visible on the view page + r = self.client.get(redirect_url) + self.assertContains(r, 'First session with: {}'.format(group2.acronym)) + + def test_edit_constraint_bethere(self): + meeting = MeetingFactory(type_id='ietf', date=date_today()) + mars = RoleFactory(name_id='chair', person__user__username='marschairman', group__acronym='mars').group + session = SessionFactory(meeting=meeting, group=mars, status_id='sched') + Constraint.objects.create( + meeting=meeting, + source=mars, + person=Person.objects.get(user__username='marschairman'), + name_id='bethere', + ) + self.assertEqual(session.people_constraints.count(), 1) + url = reverse('ietf.meeting.views_session_request.edit_request', kwargs=dict(acronym='mars')) + self.client.login(username='marschairman', password='marschairman+password') + attendees = '10' + ad = Person.objects.get(user__username='ad') + post_data = { + 'num_session': '1', + 'attendees': attendees, + 'bethere': str(ad.pk), + 'constraint_chair_conflict': '', + 'comments': '', + 'joint_with_groups': '', + 'joint_for_session': '', + 'delete_conflict': 'on', + 'session_set-TOTAL_FORMS': '3', # matches what view actually sends, even with only 2 filled in + 'session_set-INITIAL_FORMS': '1', + 'session_set-MIN_NUM_FORMS': '1', + 'session_set-MAX_NUM_FORMS': '3', + 'session_set-0-id': session.pk, + 'session_set-0-name': session.name, + 'session_set-0-short': session.short, + 'session_set-0-purpose': session.purpose_id, + 'session_set-0-type': session.type_id, + 'session_set-0-requested_duration': '3600', + 'session_set-0-on_agenda': session.on_agenda, + 'session_set-0-remote_instructions': session.remote_instructions, + 'session_set-0-attendees': attendees, + 'session_set-0-comments': '', + 'session_set-0-DELETE': '', + 'session_set-1-id': '', + 'session_set-1-name': '', + 'session_set-1-short': '', + 'session_set-1-purpose': 'regular', + 'session_set-1-type': 'regular', + 'session_set-1-requested_duration': '', + 'session_set-1-on_agenda': 'True', + 'session_set-1-attendees': attendees, + 'session_set-1-comments': '', + 'session_set-1-DELETE': 'on', + 'session_set-2-id': '', + 'session_set-2-name': '', + 'session_set-2-short': '', + 'session_set-2-purpose': 'regular', + 'session_set-2-type': 'regular', + 'session_set-2-requested_duration': '', + 'session_set-2-on_agenda': 'True', + 'session_set-2-attendees': attendees, + 'session_set-2-comments': '', + 'session_set-2-DELETE': 'on', + 'submit': 'Save', + } + r = self.client.post(url, post_data, HTTP_HOST='example.com') + redirect_url = reverse('ietf.meeting.views_session_request.view_request', kwargs={'acronym': 'mars'}) + self.assertRedirects(r, redirect_url) + self.assertEqual([pc.person for pc in session.people_constraints.all()], [ad]) + + def test_edit_inactive_conflicts(self): + """Inactive conflicts should be displayed and removable""" + meeting = MeetingFactory(type_id='ietf', date=date_today(), group_conflicts=['chair_conflict']) + mars = RoleFactory(name_id='chair', person__user__username='marschairman', group__acronym='mars').group + session = SessionFactory(meeting=meeting, group=mars, status_id='sched') + other_group = GroupFactory() + Constraint.objects.create( + meeting=meeting, + name_id='conflict', # not in group_conflicts for the meeting + source=mars, + target=other_group, + ) + + url = reverse('ietf.meeting.views_session_request.edit_request', kwargs=dict(acronym='mars')) + self.client.login(username='marschairman', password='marschairman+password') + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + + # check that the inactive session is displayed + found = q('input#id_delete_conflict[type="checkbox"]') + self.assertEqual(len(found), 1) + delete_checkbox = found[0] + self.assertIn('Delete this conflict', delete_checkbox.label.text) + # check that the target is displayed correctly in the UI + row = found.parent().parent() + self.assertIn(other_group.acronym, row.find('input[@type="text"]').val()) + + attendees = '10' + post_data = { + 'num_session': '1', + 'attendees': attendees, + 'constraint_chair_conflict': '', + 'comments': '', + 'joint_with_groups': '', + 'joint_for_session': '', + 'delete_conflict': 'on', + 'session_set-TOTAL_FORMS': '1', + 'session_set-INITIAL_FORMS': '1', + 'session_set-MIN_NUM_FORMS': '1', + 'session_set-MAX_NUM_FORMS': '3', + 'session_set-0-id': session.pk, + 'session_set-0-name': session.name, + 'session_set-0-short': session.short, + 'session_set-0-purpose': session.purpose_id, + 'session_set-0-type': session.type_id, + 'session_set-0-requested_duration': '3600', + 'session_set-0-on_agenda': session.on_agenda, + 'session_set-0-remote_instructions': session.remote_instructions, + 'session_set-0-attendees': attendees, + 'session_set-0-comments': '', + 'session_set-0-DELETE': '', + 'submit': 'Save', + } + r = self.client.post(url, post_data, HTTP_HOST='example.com') + redirect_url = reverse('ietf.meeting.views_session_request.view_request', kwargs={'acronym': 'mars'}) + self.assertRedirects(r, redirect_url) + self.assertEqual(len(mars.constraint_source_set.filter(name_id='conflict')), 0) + + def test_tool_status(self): + MeetingFactory(type_id='ietf', date=date_today()) + url = reverse('ietf.meeting.views_session_request.status') + self.client.login(username="secretary", password="secretary+password") + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + r = self.client.post(url, {'message': 'locked', 'submit': 'Lock'}) + self.assertRedirects(r, reverse('ietf.meeting.views_session_request.list_view')) + + def test_new_req_constraint_types(self): + """Configurable constraint types should be handled correctly in a new request + + Relies on SessionRequestForm representing constraint values with element IDs + like id_constraint_ + """ + meeting = MeetingFactory(type_id='ietf', date=date_today()) + RoleFactory(name_id='chair', person__user__username='marschairman', group__acronym='mars') + url = reverse('ietf.meeting.views_session_request.new_request', kwargs=dict(acronym='mars')) + self.client.login(username="marschairman", password="marschairman+password") + + for expected in [ + ['conflict', 'conflic2', 'conflic3'], + ['chair_conflict', 'tech_overlap', 'key_participant'], + ]: + meeting.group_conflict_types.clear() + for slug in expected: + meeting.group_conflict_types.add(ConstraintName.objects.get(slug=slug)) + + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertCountEqual( + [elt.attr('id') for elt in q.items('*[id^=id_constraint_]')], + ['id_constraint_{}'.format(conf_name) for conf_name in expected], + ) + + def test_edit_req_constraint_types(self): + """Editing a request constraint should show the expected constraints""" + meeting = MeetingFactory(type_id='ietf', date=date_today()) + SessionFactory(group__acronym='mars', + status_id='schedw', + meeting=meeting, + add_to_schedule=False) + RoleFactory(name_id='chair', person__user__username='marschairman', group__acronym='mars') + + url = reverse('ietf.meeting.views_session_request.edit_request', kwargs=dict(acronym='mars')) + self.client.login(username='marschairman', password='marschairman+password') + + for expected in [ + ['conflict', 'conflic2', 'conflic3'], + ['chair_conflict', 'tech_overlap', 'key_participant'], + ]: + meeting.group_conflict_types.clear() + for slug in expected: + meeting.group_conflict_types.add(ConstraintName.objects.get(slug=slug)) + + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertCountEqual( + [elt.attr('id') for elt in q.items('*[id^=id_constraint_]')], + ['id_constraint_{}'.format(conf_name) for conf_name in expected], + ) + + +class SubmitRequestCase(TestCase): + def setUp(self): + super(SubmitRequestCase, self).setUp() + # Ensure meeting numbers are predictable. Temporarily needed while basing + # constraint types on meeting number, expected to go away when #2770 is resolved. + MeetingFactory.reset_sequence(0) + + def test_submit_request(self): + meeting = MeetingFactory(type_id='ietf', date=date_today()) + ad = Person.objects.get(user__username='ad') + area = RoleFactory(name_id='ad', person=ad, group__type_id='area').group + group = GroupFactory(parent=area) + group2 = GroupFactory(parent=area) + group3 = GroupFactory(parent=area) + group4 = GroupFactory(parent=area) + session_count_before = Session.objects.filter(meeting=meeting, group=group).count() + url = reverse('ietf.meeting.views_session_request.new_request', kwargs={'acronym': group.acronym}) + confirm_url = reverse('ietf.meeting.views_session_request.confirm', kwargs={'acronym': group.acronym}) + main_url = reverse('ietf.meeting.views_session_request.list_view') + attendees = '10' + comments = 'need projector' + post_data = {'num_session': '1', + 'attendees': attendees, + 'constraint_chair_conflict': '', + 'comments': comments, + 'adjacent_with_wg': group2.acronym, + 'timeranges': ['thursday-afternoon-early', 'thursday-afternoon-late'], + 'joint_with_groups': group3.acronym + ' ' + group4.acronym, + 'joint_for_session': '1', + 'session_set-TOTAL_FORMS': '1', + 'session_set-INITIAL_FORMS': '0', + 'session_set-MIN_NUM_FORMS': '1', + 'session_set-MAX_NUM_FORMS': '3', + # no 'session_set-0-id' to create a new session + 'session_set-0-name': '', + 'session_set-0-short': '', + 'session_set-0-purpose': 'regular', + 'session_set-0-type': 'regular', + 'session_set-0-requested_duration': '3600', + 'session_set-0-on_agenda': True, + 'session_set-0-remote_instructions': '', + 'session_set-0-attendees': attendees, + 'session_set-0-comments': comments, + 'session_set-0-DELETE': '', + 'submit': 'Continue'} + self.client.login(username="secretary", password="secretary+password") + r = self.client.post(url, post_data) + self.assertEqual(r.status_code, 200) + + # Verify the contents of the confirm view + self.assertContains(r, 'Thursday early afternoon, Thursday late afternoon') + self.assertContains(r, group2.acronym) + self.assertContains(r, 'First session with: {} {}'.format(group3.acronym, group4.acronym)) + + post_data['submit'] = 'Submit' + r = self.client.post(confirm_url, post_data) + self.assertRedirects(r, main_url) + session_count_after = Session.objects.filter(meeting=meeting, group=group, type='regular').count() + self.assertEqual(session_count_after, session_count_before + 1) + + # test that second confirm does not add sessions + r = self.client.post(confirm_url, post_data) + self.assertRedirects(r, main_url) + session_count_after = Session.objects.filter(meeting=meeting, group=group, type='regular').count() + self.assertEqual(session_count_after, session_count_before + 1) + + # Verify database content + session = Session.objects.get(meeting=meeting, group=group) + self.assertEqual(session.constraints().get(name='wg_adjacent').target.acronym, group2.acronym) + self.assertEqual( + list(session.constraints().get(name='timerange').timeranges.all().values('name')), + list(TimerangeName.objects.filter(name__in=['thursday-afternoon-early', 'thursday-afternoon-late']).values('name')) + ) + self.assertEqual(set(list(session.joint_with_groups.all())), set([group3, group4])) + + def test_submit_request_check_constraints(self): + m1 = MeetingFactory(type_id='ietf', date=date_today() - datetime.timedelta(days=100)) + MeetingFactory(type_id='ietf', date=date_today(), + group_conflicts=['chair_conflict', 'conflic2', 'conflic3']) + ad = Person.objects.get(user__username='ad') + area = RoleFactory(name_id='ad', person=ad, group__type_id='area').group + group = GroupFactory(parent=area) + still_active_group = GroupFactory(parent=area) + Constraint.objects.create( + meeting=m1, + source=group, + target=still_active_group, + name_id='chair_conflict', + ) + inactive_group = GroupFactory(parent=area, state_id='conclude') + inactive_group.save() + Constraint.objects.create( + meeting=m1, + source=group, + target=inactive_group, + name_id='chair_conflict', + ) + session = SessionFactory(group=group, meeting=m1) + + self.client.login(username="secretary", password="secretary+password") + + url = reverse('ietf.meeting.views_session_request.new_request', kwargs={'acronym': group.acronym}) + r = self.client.get(url + '?previous') + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + conflict1 = q('[name="constraint_chair_conflict"]').val() + self.assertIn(still_active_group.acronym, conflict1) + self.assertNotIn(inactive_group.acronym, conflict1) + + attendees = '10' + comments = 'need projector' + post_data = {'num_session': '1', + 'attendees': attendees, + 'constraint_chair_conflict': group.acronym, + 'comments': comments, + 'session_set-TOTAL_FORMS': '3', + 'session_set-INITIAL_FORMS': '1', + 'session_set-MIN_NUM_FORMS': '1', + 'session_set-MAX_NUM_FORMS': '3', + # no 'session_set-0-id' to create a new session + 'session_set-0-name': '', + 'session_set-0-short': '', + 'session_set-0-purpose': session.purpose_id, + 'session_set-0-type': session.type_id, + 'session_set-0-requested_duration': '3600', + 'session_set-0-on_agenda': session.on_agenda, + 'session_set-0-remote_instructions': session.remote_instructions, + 'session_set-0-attendees': attendees, + 'session_set-0-comments': comments, + 'session_set-0-DELETE': '', + 'session_set-1-name': '', + 'session_set-1-short': '', + 'session_set-1-purpose': session.purpose_id, + 'session_set-1-type': session.type_id, + 'session_set-1-requested_duration': '', + 'session_set-1-on_agenda': session.on_agenda, + 'session_set-1-remote_instructions': '', + 'session_set-1-attendees': attendees, + 'session_set-1-comments': '', + 'session_set-1-DELETE': 'on', + 'session_set-2-name': '', + 'session_set-2-short': '', + 'session_set-2-purpose': session.purpose_id, + 'session_set-2-type': session.type_id, + 'session_set-2-requested_duration': '', + 'session_set-2-on_agenda': session.on_agenda, + 'session_set-2-remote_instructions': '', + 'session_set-2-attendees': attendees, + 'session_set-2-comments': '', + 'session_set-2-DELETE': 'on', + 'submit': 'Continue'} + r = self.client.post(url, post_data) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q('#session-request-form')), 1) + self.assertContains(r, "Cannot declare a conflict with the same group") + + def test_request_notification(self): + meeting = MeetingFactory(type_id='ietf', date=date_today()) + ad = Person.objects.get(user__username='ad') + area = GroupFactory(type_id='area') + RoleFactory(name_id='ad', person=ad, group=area) + group = GroupFactory(acronym='ames', parent=area) + group2 = GroupFactory(acronym='ames2', parent=area) + group3 = GroupFactory(acronym='ames2', parent=area) + group4 = GroupFactory(acronym='ames3', parent=area) + RoleFactory(name_id='chair', group=group, person__user__username='ameschairman') + resource = ResourceAssociation.objects.create(name_id='project') + # Bit of a test data hack - the fixture now has no used resources to pick from + resource.name.used = True + resource.name.save() + + url = reverse('ietf.meeting.views_session_request.new_request', kwargs={'acronym': group.acronym}) + confirm_url = reverse('ietf.meeting.views_session_request.confirm', kwargs={'acronym': group.acronym}) + len_before = len(outbox) + attendees = '10' + post_data = {'num_session': '2', + 'attendees': attendees, + 'bethere': str(ad.pk), + 'constraint_chair_conflict': group4.acronym, + 'comments': '', + 'resources': resource.pk, + 'session_time_relation': 'subsequent-days', + 'adjacent_with_wg': group2.acronym, + 'joint_with_groups': group3.acronym, + 'joint_for_session': '2', + 'timeranges': ['thursday-afternoon-early', 'thursday-afternoon-late'], + 'session_set-TOTAL_FORMS': '2', + 'session_set-INITIAL_FORMS': '0', + 'session_set-MIN_NUM_FORMS': '1', + 'session_set-MAX_NUM_FORMS': '3', + # no 'session_set-0-id' for new session + 'session_set-0-name': '', + 'session_set-0-short': '', + 'session_set-0-purpose': 'regular', + 'session_set-0-type': 'regular', + 'session_set-0-requested_duration': '3600', + 'session_set-0-on_agenda': True, + 'session_set-0-remote_instructions': '', + 'session_set-0-attendees': attendees, + 'session_set-0-comments': '', + 'session_set-0-DELETE': '', + # no 'session_set-1-id' for new session + 'session_set-1-name': '', + 'session_set-1-short': '', + 'session_set-1-purpose': 'regular', + 'session_set-1-type': 'regular', + 'session_set-1-requested_duration': '3600', + 'session_set-1-on_agenda': True, + 'session_set-1-remote_instructions': '', + 'session_set-1-attendees': attendees, + 'session_set-1-comments': '', + 'session_set-1-DELETE': '', + 'submit': 'Continue'} + self.client.login(username="ameschairman", password="ameschairman+password") + # submit + r = self.client.post(url, post_data) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertTrue('Confirm' in str(q("title")), r.context['form'].errors) + # confirm + post_data['submit'] = 'Submit' + r = self.client.post(confirm_url, post_data) + self.assertRedirects(r, reverse('ietf.meeting.views_session_request.list_view')) + self.assertEqual(len(outbox), len_before + 1) + notification = outbox[-1] + notification_payload = get_payload_text(notification) + sessions = Session.objects.filter(meeting=meeting, group=group) + self.assertEqual(len(sessions), 2) + session = sessions[0] + + self.assertEqual(session.resources.count(), 1) + self.assertEqual(session.people_constraints.count(), 1) + self.assertEqual(session.constraints().get(name='time_relation').time_relation, 'subsequent-days') + self.assertEqual(session.constraints().get(name='wg_adjacent').target.acronym, group2.acronym) + self.assertEqual( + list(session.constraints().get(name='timerange').timeranges.all().values('name')), + list(TimerangeName.objects.filter(name__in=['thursday-afternoon-early', 'thursday-afternoon-late']).values('name')) + ) + resource = session.resources.first() + self.assertTrue(resource.desc in notification_payload) + self.assertTrue('Schedule the sessions on subsequent days' in notification_payload) + self.assertTrue(group2.acronym in notification_payload) + self.assertTrue("Can't meet: Thursday early afternoon, Thursday late" in notification_payload) + self.assertTrue('Second session joint with: {}'.format(group3.acronym) in notification_payload) + self.assertTrue(ad.ascii_name() in notification_payload) + self.assertIn(ConstraintName.objects.get(slug='chair_conflict').name, notification_payload) + self.assertIn(group.acronym, notification_payload) + self.assertIn('1 Hour, 1 Hour', notification_payload) + self.assertNotIn('1 Hour, 1 Hour, 1 Hour', notification_payload) + self.assertNotIn('The third session requires your approval', notification_payload) + + def test_request_notification_msg(self): + to = "" + subject = "Dummy subject" + template = "meeting/session_request_notification.txt" + header = "A new" + meeting = MeetingFactory(type_id="ietf", date=date_today()) + requester = PersonFactory(name="James O'Rourke", user__username="jimorourke") + context = {"header": header, "meeting": meeting, "requester": requester} + cc = "cc.a@example.com, cc.b@example.com" + bcc = "bcc@example.com" + + msg = send_mail( + None, + to, + None, + subject, + template, + context, + cc=cc, + bcc=bcc, + ) + self.assertEqual(requester.name, "James O'Rourke") # note ' (single quote) in the name + self.assertIn( + f"{header} meeting session request has just been submitted by {requester.name}.", + get_payload_text(msg), + ) + + def test_request_notification_third_session(self): + meeting = MeetingFactory(type_id='ietf', date=date_today()) + ad = Person.objects.get(user__username='ad') + area = GroupFactory(type_id='area') + RoleFactory(name_id='ad', person=ad, group=area) + group = GroupFactory(acronym='ames', parent=area) + group2 = GroupFactory(acronym='ames2', parent=area) + group3 = GroupFactory(acronym='ames2', parent=area) + group4 = GroupFactory(acronym='ames3', parent=area) + RoleFactory(name_id='chair', group=group, person__user__username='ameschairman') + resource = ResourceAssociation.objects.create(name_id='project') + # Bit of a test data hack - the fixture now has no used resources to pick from + resource.name.used = True + resource.name.save() + + url = reverse('ietf.meeting.views_session_request.new_request', kwargs={'acronym': group.acronym}) + confirm_url = reverse('ietf.meeting.views_session_request.confirm', kwargs={'acronym': group.acronym}) + len_before = len(outbox) + attendees = '10' + post_data = {'num_session': '2', + 'third_session': 'true', + 'attendees': attendees, + 'bethere': str(ad.pk), + 'constraint_chair_conflict': group4.acronym, + 'comments': '', + 'resources': resource.pk, + 'session_time_relation': 'subsequent-days', + 'adjacent_with_wg': group2.acronym, + 'joint_with_groups': group3.acronym, + 'joint_for_session': '2', + 'timeranges': ['thursday-afternoon-early', 'thursday-afternoon-late'], + 'session_set-TOTAL_FORMS': '3', + 'session_set-INITIAL_FORMS': '0', + 'session_set-MIN_NUM_FORMS': '1', + 'session_set-MAX_NUM_FORMS': '3', + # no 'session_set-0-id' for new session + 'session_set-0-name': '', + 'session_set-0-short': '', + 'session_set-0-purpose': 'regular', + 'session_set-0-type': 'regular', + 'session_set-0-requested_duration': '3600', + 'session_set-0-on_agenda': True, + 'session_set-0-remote_instructions': '', + 'session_set-0-attendees': attendees, + 'session_set-0-comments': '', + 'session_set-0-DELETE': '', + # no 'session_set-1-id' for new session + 'session_set-1-name': '', + 'session_set-1-short': '', + 'session_set-1-purpose': 'regular', + 'session_set-1-type': 'regular', + 'session_set-1-requested_duration': '3600', + 'session_set-1-on_agenda': True, + 'session_set-1-remote_instructions': '', + 'session_set-1-attendees': attendees, + 'session_set-1-comments': '', + 'session_set-1-DELETE': '', + # no 'session_set-2-id' for new session + 'session_set-2-name': '', + 'session_set-2-short': '', + 'session_set-2-purpose': 'regular', + 'session_set-2-type': 'regular', + 'session_set-2-requested_duration': '3600', + 'session_set-2-on_agenda': True, + 'session_set-2-remote_instructions': '', + 'session_set-2-attendees': attendees, + 'session_set-2-comments': '', + 'session_set-2-DELETE': '', + 'submit': 'Continue'} + self.client.login(username="ameschairman", password="ameschairman+password") + # submit + r = self.client.post(url, post_data) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertTrue('Confirm' in str(q("title")), r.context['form'].errors) + # confirm + post_data['submit'] = 'Submit' + r = self.client.post(confirm_url, post_data) + self.assertRedirects(r, reverse('ietf.meeting.views_session_request.list_view')) + self.assertEqual(len(outbox), len_before + 1) + notification = outbox[-1] + notification_payload = get_payload_text(notification) + sessions = Session.objects.filter(meeting=meeting, group=group) + self.assertEqual(len(sessions), 3) + session = sessions[0] + + self.assertEqual(session.resources.count(), 1) + self.assertEqual(session.people_constraints.count(), 1) + self.assertEqual(session.constraints().get(name='time_relation').time_relation, 'subsequent-days') + self.assertEqual(session.constraints().get(name='wg_adjacent').target.acronym, group2.acronym) + self.assertEqual( + list(session.constraints().get(name='timerange').timeranges.all().values('name')), + list(TimerangeName.objects.filter(name__in=['thursday-afternoon-early', 'thursday-afternoon-late']).values('name')) + ) + resource = session.resources.first() + self.assertTrue(resource.desc in notification_payload) + self.assertTrue('Schedule the sessions on subsequent days' in notification_payload) + self.assertTrue(group2.acronym in notification_payload) + self.assertTrue("Can't meet: Thursday early afternoon, Thursday late" in notification_payload) + self.assertTrue('Second session joint with: {}'.format(group3.acronym) in notification_payload) + self.assertTrue(ad.ascii_name() in notification_payload) + self.assertIn(ConstraintName.objects.get(slug='chair_conflict').name, notification_payload) + self.assertIn(group.acronym, notification_payload) + self.assertIn('1 Hour, 1 Hour, 1 Hour', notification_payload) + self.assertIn('The third session requires your approval', notification_payload) + + +class LockAppTestCase(TestCase): + def setUp(self): + super().setUp() + self.meeting = MeetingFactory(type_id='ietf', date=date_today(), session_request_lock_message='locked') + self.group = GroupFactory(acronym='mars') + RoleFactory(name_id='chair', group=self.group, person__user__username='marschairman') + SessionFactory(group=self.group, meeting=self.meeting) + + def test_edit_request(self): + url = reverse('ietf.meeting.views_session_request.edit_request', kwargs={'acronym': self.group.acronym}) + self.client.login(username="secretary", password="secretary+password") + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q(':disabled[name="submit"]')), 0) + chair = self.group.role_set.filter(name_id='chair').first().person.user.username + self.client.login(username=chair, password=f'{chair}+password') + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q(':disabled[name="submit"]')), 1) + + def test_view_request(self): + url = reverse('ietf.meeting.views_session_request.view_request', kwargs={'acronym': self.group.acronym}) + self.client.login(username="secretary", password="secretary+password") + r = self.client.get(url, follow=True) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q(':enabled[name="edit"]')), 1) # secretary can edit + chair = self.group.role_set.filter(name_id='chair').first().person.user.username + self.client.login(username=chair, password=f'{chair}+password') + r = self.client.get(url, follow=True) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q(':disabled[name="edit"]')), 1) # chair cannot edit + + def test_new_request(self): + url = reverse('ietf.meeting.views_session_request.new_request', kwargs={'acronym': self.group.acronym}) + + # try as WG Chair + self.client.login(username="marschairman", password="marschairman+password") + r = self.client.get(url, follow=True) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q('#session-request-form')), 0) + + # try as Secretariat + self.client.login(username="secretary", password="secretary+password") + r = self.client.get(url, follow=True) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q('#session-request-form')), 1) + + +class NotMeetingCase(TestCase): + def test_not_meeting(self): + MeetingFactory(type_id='ietf', date=date_today()) + group = GroupFactory(acronym='mars') + url = reverse('ietf.meeting.views_session_request.no_session', kwargs={'acronym': group.acronym}) + self.client.login(username="secretary", password="secretary+password") + + empty_outbox() + + r = self.client.get(url, follow=True) + # If the view invoked by that get throws an exception (such as an integrity error), + # the traceback from this test will talk about a TransactionManagementError and + # yell about executing queries before the end of an 'atomic' block + + # This is a sign of a problem - a get shouldn't have a side-effect like this one does + self.assertEqual(r.status_code, 200) + self.assertContains(r, 'A message was sent to notify not having a session') + + r = self.client.get(url, follow=True) + self.assertEqual(r.status_code, 200) + self.assertContains(r, 'is already marked as not meeting') + + self.assertEqual(len(outbox), 1) + self.assertTrue('Not having a session' in outbox[0]['Subject']) + self.assertTrue('session-request@' in outbox[0]['To']) + + +class RetrievePreviousCase(TestCase): + pass + + # test error if already scheduled + # test get previous exists/doesn't exist + # test that groups scheduled and unscheduled add up to total groups + # test access by unauthorized + + +class SessionRequestFormTest(TestCase): + def setUp(self): + super().setUp() + self.meeting = MeetingFactory(type_id='ietf') + self.group1 = GroupFactory() + self.group2 = GroupFactory() + self.group3 = GroupFactory() + self.group4 = GroupFactory() + self.group5 = GroupFactory() + self.group6 = GroupFactory() + + attendees = '10' + comments = 'need lights' + self.valid_form_data = { + 'num_session': '2', + 'third_session': 'true', + 'attendees': attendees, + 'constraint_chair_conflict': self.group2.acronym, + 'constraint_tech_overlap': self.group3.acronym, + 'constraint_key_participant': self.group4.acronym, + 'comments': comments, + 'session_time_relation': 'subsequent-days', + 'adjacent_with_wg': self.group5.acronym, + 'joint_with_groups': self.group6.acronym, + 'joint_for_session': '3', + 'timeranges': ['thursday-afternoon-early', 'thursday-afternoon-late'], + 'submit': 'Continue', + 'session_set-TOTAL_FORMS': '3', + 'session_set-INITIAL_FORMS': '0', + 'session_set-MIN_NUM_FORMS': '1', + 'session_set-MAX_NUM_FORMS': '3', + # no 'session_set-0-id' for new session + 'session_set-0-name': '', + 'session_set-0-short': '', + 'session_set-0-purpose': 'regular', + 'session_set-0-type': 'regular', + 'session_set-0-requested_duration': '3600', + 'session_set-0-on_agenda': True, + 'session_set-0-remote_instructions': '', + 'session_set-0-attendees': attendees, + 'session_set-0-comments': '', + 'session_set-0-DELETE': '', + # no 'session_set-1-id' for new session + 'session_set-1-name': '', + 'session_set-1-short': '', + 'session_set-1-purpose': 'regular', + 'session_set-1-type': 'regular', + 'session_set-1-requested_duration': '3600', + 'session_set-1-on_agenda': True, + 'session_set-1-remote_instructions': '', + 'session_set-1-attendees': attendees, + 'session_set-1-comments': '', + 'session_set-1-DELETE': '', + # no 'session_set-2-id' for new session + 'session_set-2-name': '', + 'session_set-2-short': '', + 'session_set-2-purpose': 'regular', + 'session_set-2-type': 'regular', + 'session_set-2-requested_duration': '3600', + 'session_set-2-on_agenda': True, + 'session_set-2-remote_instructions': '', + 'session_set-2-attendees': attendees, + 'session_set-2-comments': '', + 'session_set-2-DELETE': '', + } + + def test_valid(self): + # Test with three sessions + form = SessionRequestForm(data=self.valid_form_data, group=self.group1, meeting=self.meeting) + self.assertTrue(form.is_valid()) + + # Test with two sessions + self.valid_form_data.update({ + 'third_session': '', + 'session_set-TOTAL_FORMS': '2', + 'joint_for_session': '2' + }) + form = SessionRequestForm(data=self.valid_form_data, group=self.group1, meeting=self.meeting) + self.assertTrue(form.is_valid()) + + # Test with one session + self.valid_form_data.update({ + 'num_session': 1, + 'session_set-TOTAL_FORMS': '1', + 'joint_for_session': '1', + 'session_time_relation': '', + }) + form = SessionRequestForm(data=self.valid_form_data, group=self.group1, meeting=self.meeting) + self.assertTrue(form.is_valid()) + + def test_invalid_groups(self): + new_form_data = { + 'constraint_chair_conflict': 'doesnotexist', + 'constraint_tech_overlap': 'doesnotexist', + 'constraint_key_participant': 'doesnotexist', + 'adjacent_with_wg': 'doesnotexist', + 'joint_with_groups': 'doesnotexist', + } + form = self._invalid_test_helper(new_form_data) + self.assertEqual(set(form.errors.keys()), set(new_form_data.keys())) + + def test_valid_group_appears_in_multiple_conflicts(self): + """Some conflict types allow overlapping groups""" + new_form_data = { + 'constraint_chair_conflict': self.group2.acronym, + 'constraint_tech_overlap': self.group2.acronym, + } + self.valid_form_data.update(new_form_data) + form = SessionRequestForm(data=self.valid_form_data, group=self.group1, meeting=self.meeting) + self.assertTrue(form.is_valid()) + + def test_invalid_group_appears_in_multiple_conflicts(self): + """Some conflict types do not allow overlapping groups""" + self.meeting.group_conflict_types.clear() + self.meeting.group_conflict_types.add(ConstraintName.objects.get(slug='conflict')) + self.meeting.group_conflict_types.add(ConstraintName.objects.get(slug='conflic2')) + new_form_data = { + 'constraint_conflict': self.group2.acronym, + 'constraint_conflic2': self.group2.acronym, + } + form = self._invalid_test_helper(new_form_data) + self.assertEqual(form.non_field_errors(), ['%s appears in conflicts more than once' % self.group2.acronym]) + + def test_invalid_conflict_with_self(self): + new_form_data = { + 'constraint_chair_conflict': self.group1.acronym, + } + self._invalid_test_helper(new_form_data) + + def test_invalid_session_time_relation(self): + form = self._invalid_test_helper({ + 'third_session': '', + 'session_set-TOTAL_FORMS': 1, + 'num_session': 1, + 'joint_for_session': '1', + }) + self.assertEqual(form.errors, + { + 'session_time_relation': ['Time between sessions can only be used when two ' + 'sessions are requested.'] + }) + + def test_invalid_joint_for_session(self): + form = self._invalid_test_helper({ + 'third_session': '', + 'session_set-TOTAL_FORMS': '2', + 'num_session': 2, + 'joint_for_session': '3', + }) + self.assertEqual(form.errors, + { + 'joint_for_session': [ + 'Session 3 can not be the joint session, the session has not been requested.'] + }) + + form = self._invalid_test_helper({ + 'third_session': '', + 'session_set-TOTAL_FORMS': '1', + 'num_session': 1, + 'joint_for_session': '2', + 'session_time_relation': '', + }) + self.assertEqual(form.errors, + { + 'joint_for_session': [ + 'Session 2 can not be the joint session, the session has not been requested.'] + }) + + def test_invalid_missing_session_length(self): + form = self._invalid_test_helper({ + 'session_set-TOTAL_FORMS': '2', + 'session_set-1-requested_duration': '', + 'third_session': 'false', + 'joint_for_session': None, + }) + self.assertEqual(form.session_forms.errors, + [ + {}, + {'requested_duration': ['This field is required.']}, + ]) + + form = self._invalid_test_helper({ + 'session_set-1-requested_duration': '', + 'session_set-2-requested_duration': '', + 'joint_for_session': None, + }) + self.assertEqual( + form.session_forms.errors, + [ + {}, + {'requested_duration': ['This field is required.']}, + {'requested_duration': ['This field is required.']}, + ]) + + form = self._invalid_test_helper({ + 'session_set-2-requested_duration': '', + 'joint_for_session': None, + }) + self.assertEqual(form.session_forms.errors, + [ + {}, + {}, + {'requested_duration': ['This field is required.']}, + ]) + + def _invalid_test_helper(self, new_form_data): + form_data = dict(self.valid_form_data, **new_form_data) + form = SessionRequestForm(data=form_data, group=self.group1, meeting=self.meeting) + self.assertFalse(form.is_valid()) + return form diff --git a/ietf/meeting/tests_tasks.py b/ietf/meeting/tests_tasks.py new file mode 100644 index 0000000000..2c5120a39d --- /dev/null +++ b/ietf/meeting/tests_tasks.py @@ -0,0 +1,121 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +import datetime +from unittest.mock import patch, call +from ietf.utils.test_utils import TestCase +from ietf.utils.timezone import date_today +from .factories import MeetingFactory +from .tasks import ( + proceedings_content_refresh_task, + agenda_data_refresh_task, + agenda_data_refresh_all_task, +) +from .tasks import fetch_meeting_attendance_task + + +class TaskTests(TestCase): + @patch("ietf.meeting.tasks.generate_agenda_data") + def test_agenda_data_refresh_task(self, mock_generate): + agenda_data_refresh_task() + self.assertTrue(mock_generate.called) + self.assertEqual(mock_generate.call_args, call(None, force_refresh=True)) + + mock_generate.reset_mock() + mock_generate.side_effect = RuntimeError + try: + agenda_data_refresh_task() + except Exception as err: + self.fail( + f"agenda_data_refresh_task should not raise exceptions (got {repr(err)})" + ) + + @patch("ietf.meeting.tasks.agenda_data_refresh_task") + @patch("ietf.meeting.tasks.chain") + def test_agenda_data_refresh_all_task(self, mock_chain, mock_agenda_data_refresh): + # Patch the agenda_data_refresh_task task with a mock whose `.map` attribute + # converts its argument, which is expected to be an iterator, to a list + # and returns it. We'll use this to check that the expected task chain + # was set up, but we don't actually run any celery tasks. + mock_agenda_data_refresh.map.side_effect = lambda x: list(x) + + meetings = MeetingFactory.create_batch(5, type_id="ietf") + numbers = sorted(int(m.number) for m in meetings) + agenda_data_refresh_all_task(batch_size=2) + self.assertTrue(mock_chain.called) + # The lists in the call() below are the output of the lambda we patched in + # via mock_agenda_data_refresh.map.side_effect above. I.e., this tests that + # map() was called with the correct batched data. + self.assertEqual( + mock_chain.call_args, + call( + [numbers[0], numbers[1]], + [numbers[2], numbers[3]], + [numbers[4]], + ), + ) + self.assertEqual(mock_agenda_data_refresh.call_count, 0) + self.assertEqual(mock_agenda_data_refresh.map.call_count, 3) + + @patch("ietf.meeting.tasks.generate_proceedings_content") + def test_proceedings_content_refresh_task(self, mock_generate): + # Generate a couple of meetings + meeting120 = MeetingFactory(type_id="ietf", number="120") # 24 * 5 + meeting127 = MeetingFactory(type_id="ietf", number="127") # 24 * 5 + 7 + + # Times to be returned + now_utc = datetime.datetime.now(tz=datetime.UTC) + hour_00_utc = now_utc.replace(hour=0) + hour_01_utc = now_utc.replace(hour=1) + hour_07_utc = now_utc.replace(hour=7) + + # hour 00 - should call meeting with number % 24 == 0 + with patch("ietf.meeting.tasks.timezone.now", return_value=hour_00_utc): + proceedings_content_refresh_task() + self.assertEqual(mock_generate.call_count, 1) + self.assertEqual(mock_generate.call_args, call(meeting120, force_refresh=True)) + mock_generate.reset_mock() + + # hour 01 - should call no meetings + with patch("ietf.meeting.tasks.timezone.now", return_value=hour_01_utc): + proceedings_content_refresh_task() + self.assertEqual(mock_generate.call_count, 0) + + # hour 07 - should call meeting with number % 24 == 0 + with patch("ietf.meeting.tasks.timezone.now", return_value=hour_07_utc): + proceedings_content_refresh_task() + self.assertEqual(mock_generate.call_count, 1) + self.assertEqual(mock_generate.call_args, call(meeting127, force_refresh=True)) + mock_generate.reset_mock() + + # With all=True, all should be called regardless of time. Reuse hour_01_utc which called none before + with patch("ietf.meeting.tasks.timezone.now", return_value=hour_01_utc): + proceedings_content_refresh_task(all=True) + self.assertEqual(mock_generate.call_count, 2) + + @patch("ietf.meeting.tasks.fetch_attendance_from_meetings") + def test_fetch_meeting_attendance_task(self, mock_fetch_attendance): + today = date_today() + meetings = [ + MeetingFactory(type_id="ietf", date=today - datetime.timedelta(days=1)), + MeetingFactory(type_id="ietf", date=today - datetime.timedelta(days=2)), + MeetingFactory(type_id="ietf", date=today - datetime.timedelta(days=3)), + ] + data = { + "created": 1, + "updated": 2, + "deleted": 0, + "processed": 3, + } + + mock_fetch_attendance.return_value = [data, data] + + fetch_meeting_attendance_task() + self.assertEqual(mock_fetch_attendance.call_count, 1) + self.assertCountEqual(mock_fetch_attendance.call_args[0][0], meetings[0:2]) + + # test handling of RuntimeError + mock_fetch_attendance.reset_mock() + mock_fetch_attendance.side_effect = RuntimeError + fetch_meeting_attendance_task() + self.assertTrue(mock_fetch_attendance.called) + # Good enough that we got here without raising an exception diff --git a/ietf/meeting/tests_utils.py b/ietf/meeting/tests_utils.py new file mode 100644 index 0000000000..7dd8f435e1 --- /dev/null +++ b/ietf/meeting/tests_utils.py @@ -0,0 +1,309 @@ +# Copyright The IETF Trust 2025, All Rights Reserved +# -*- coding: utf-8 -*- + +import copy +import datetime +import debug # pyflakes: ignore +import json +import jsonschema +from json import JSONDecodeError +from unittest.mock import patch, Mock + +from django.http import HttpResponse, JsonResponse +from ietf.meeting.factories import MeetingFactory, RegistrationFactory, RegistrationTicketFactory +from ietf.meeting.models import Registration +from ietf.meeting.utils import ( + process_single_registration, + get_registration_data, + sync_registration_data, + fetch_attendance_from_meetings, + get_activity_stats +) +from ietf.nomcom.models import Volunteer +from ietf.nomcom.factories import NomComFactory, nomcom_kwargs_for_year +from ietf.person.factories import PersonFactory +from ietf.utils.test_utils import TestCase +from ietf.meeting.test_data import make_meeting_test_data +from ietf.doc.factories import NewRevisionDocEventFactory, DocEventFactory + + +class JsonResponseWithJson(JsonResponse): + def json(self): + return json.loads(self.content) + + +class ActivityStatsTests(TestCase): + + def test_activity_stats(self): + utc = datetime.timezone.utc + make_meeting_test_data() + sdate = datetime.date(2016,4,3) + edate = datetime.date(2016,7,14) + MeetingFactory(type_id='ietf', date=sdate, number="96") + MeetingFactory(type_id='ietf', date=edate, number="97") + + NewRevisionDocEventFactory(time=datetime.datetime(2016,4,5,12,0,0,0,tzinfo=utc)) + NewRevisionDocEventFactory(time=datetime.datetime(2016,4,6,12,0,0,0,tzinfo=utc)) + NewRevisionDocEventFactory(time=datetime.datetime(2016,4,7,12,0,0,0,tzinfo=utc)) + + NewRevisionDocEventFactory(time=datetime.datetime(2016,6,30,12,0,0,0,tzinfo=utc)) + NewRevisionDocEventFactory(time=datetime.datetime(2016,6,30,13,0,0,0,tzinfo=utc)) + + DocEventFactory(doc__std_level_id="ps", doc__type_id="rfc", type="published_rfc", time=datetime.datetime(2016,4,5,12,0,0,0,tzinfo=utc)) + DocEventFactory(doc__std_level_id="bcp", doc__type_id="rfc", type="published_rfc", time=datetime.datetime(2016,4,6,12,0,0,0,tzinfo=utc)) + DocEventFactory(doc__std_level_id="inf", doc__type_id="rfc", type="published_rfc", time=datetime.datetime(2016,4,7,12,0,0,0,tzinfo=utc)) + DocEventFactory(doc__std_level_id="exp", doc__type_id="rfc", type="published_rfc", time=datetime.datetime(2016,4,8,12,0,0,0,tzinfo=utc)) + + data = get_activity_stats(sdate, edate) + self.assertEqual(data['new_drafts_count'], len(data['new_docs'])) + self.assertEqual(data['ffw_new_count'], 2) + self.assertEqual(data['ffw_new_percent'], '40%') + rfc_count = 0 + for c in data['counts']: + rfc_count += data['counts'].get(c) + self.assertEqual(rfc_count, len(data['rfcs'])) + + +class GetRegistrationsTests(TestCase): + + @patch('ietf.meeting.utils.requests.get') + def test_get_registation_data(self, mock_get): + meeting = MeetingFactory(type_id='ietf', number='122') + person = PersonFactory() + reg_details = dict( + first_name=person.first_name(), + last_name=person.last_name(), + email=person.email().address, + affiliation='Microsoft', + country_code='US', + meeting=meeting.number, + checkedin=True, + is_nomcom_volunteer=False, + cancelled=False, + tickets=[{'attendance_type': 'onsite', 'ticket_type': 'week_pass'}], + ) + reg_data = {'objects': {person.email().address: reg_details}} + reg_data_bad = copy.deepcopy(reg_data) + del reg_data_bad['objects'][person.email().address]['email'] + response1 = HttpResponse('Invalid apikey', status=403) + response2 = JsonResponseWithJson(reg_data) + response3 = Mock() + response3.status_code = 200 + response3.json.side_effect = JSONDecodeError("Expecting value", doc="", pos=0) + response4 = JsonResponseWithJson(reg_data_bad) + mock_get.side_effect = [response1, response2, response3, response4] + # test status 403 + with self.assertRaises(Exception): + get_registration_data(meeting) + # test status 200 good + returned_data = get_registration_data(meeting) + self.assertEqual(returned_data, reg_data) + # test decode error + with self.assertRaises(ValueError): + get_registration_data(meeting) + # test validation error + with self.assertRaises(jsonschema.exceptions.ValidationError): + get_registration_data(meeting) + + @patch('ietf.meeting.utils.get_registration_data') + def test_sync_registation_data(self, mock_get): + meeting = MeetingFactory(type_id='ietf', number='122') + person1 = PersonFactory() + person2 = PersonFactory() + items = [] + for person in [person1, person2]: + items.append(dict( + first_name=person.first_name(), + last_name=person.last_name(), + email=person.email().address, + affiliation='Microsoft', + country_code='US', + meeting=meeting.number, + checkedin=True, + is_nomcom_volunteer=False, + cancelled=False, + tickets=[{'attendance_type': 'onsite', 'ticket_type': 'week_pass'}], + )) + reg_data = {'objects': {items[0]['email']: items[0], items[1]['email']: items[1]}} + mock_get.return_value = reg_data + self.assertEqual(Registration.objects.filter(meeting=meeting).count(), 0) + stats = sync_registration_data(meeting) + self.assertEqual(Registration.objects.filter(meeting=meeting).count(), 2) + self.assertEqual(stats['created'], 2) + # test idempotent + stats = sync_registration_data(meeting) + self.assertEqual(Registration.objects.filter(meeting=meeting).count(), 2) + self.assertEqual(stats['created'], 0) + # test delete cancelled registration + del reg_data['objects'][items[1]['email']] + stats = sync_registration_data(meeting) + self.assertEqual(Registration.objects.filter(meeting=meeting).count(), 1) + self.assertEqual(stats['deleted'], 1) + + def test_process_single_registration(self): + # test new registration + meeting = MeetingFactory(type_id='ietf', number='122') + person = PersonFactory() + reg_data = dict( + first_name=person.first_name(), + last_name=person.last_name(), + email=person.email().address, + affiliation='Microsoft', + country_code='US', + meeting=meeting.number, + checkedin=True, + is_nomcom_volunteer=False, + cancelled=False, + tickets=[{'attendance_type': 'onsite', 'ticket_type': 'week_pass'}], + ) + self.assertEqual(meeting.registration_set.count(), 0) + new_reg, action = process_single_registration(reg_data, meeting) + self.assertEqual(meeting.registration_set.count(), 1) + reg = meeting.registration_set.first() + self.assertEqual(new_reg, reg) + self.assertEqual(action, 'created') + self.assertEqual(reg.first_name, person.first_name()) + self.assertEqual(reg.last_name, person.last_name()) + self.assertEqual(reg.email, person.email().address) + self.assertEqual(reg.affiliation, 'Microsoft') + self.assertEqual(reg.meeting, meeting) + self.assertEqual(reg.checkedin, True) + self.assertEqual(reg.tickets.count(), 1) + ticket = reg.tickets.first() + self.assertEqual(ticket.attendance_type.slug, 'onsite') + self.assertEqual(ticket.ticket_type.slug, 'week_pass') + + # test no change + new_reg, action = process_single_registration(reg_data, meeting) + self.assertEqual(meeting.registration_set.count(), 1) + reg = meeting.registration_set.first() + self.assertEqual(new_reg, reg) + self.assertEqual(action, None) + + # test update fields + reg_data['affiliation'] = 'Cisco' + new_reg, action = process_single_registration(reg_data, meeting) + self.assertEqual(meeting.registration_set.count(), 1) + reg = meeting.registration_set.first() + self.assertEqual(new_reg, reg) + self.assertEqual(action, 'updated') + self.assertEqual(reg.affiliation, 'Cisco') + + # test update tickets + reg_data['tickets'] = [{'attendance_type': 'remote', 'ticket_type': 'week_pass'}] + new_reg, action = process_single_registration(reg_data, meeting) + self.assertEqual(meeting.registration_set.count(), 1) + reg = meeting.registration_set.first() + self.assertEqual(new_reg, reg) + self.assertEqual(action, 'updated') + self.assertEqual(reg.tickets.count(), 1) + ticket = reg.tickets.first() + self.assertEqual(ticket.attendance_type.slug, 'remote') + + # test tickets, two of same + reg_data['tickets'] = [ + {'attendance_type': 'onsite', 'ticket_type': 'one_day'}, + {'attendance_type': 'onsite', 'ticket_type': 'one_day'}, + {'attendance_type': 'remote', 'ticket_type': 'week_pass'}, + ] + new_reg, action = process_single_registration(reg_data, meeting) + self.assertEqual(meeting.registration_set.count(), 1) + reg = meeting.registration_set.first() + self.assertEqual(new_reg, reg) + self.assertEqual(action, 'updated') + self.assertEqual(reg.tickets.count(), 3) + self.assertEqual(reg.tickets.filter(attendance_type__slug='onsite', ticket_type__slug='one_day').count(), 2) + self.assertEqual(reg.tickets.filter(attendance_type__slug='remote', ticket_type__slug='week_pass').count(), 1) + + # test tickets, two of same, delete one + reg_data['tickets'] = [ + {'attendance_type': 'onsite', 'ticket_type': 'one_day'}, + {'attendance_type': 'remote', 'ticket_type': 'week_pass'}, + ] + new_reg, action = process_single_registration(reg_data, meeting) + self.assertEqual(meeting.registration_set.count(), 1) + reg = meeting.registration_set.first() + self.assertEqual(new_reg, reg) + self.assertEqual(action, 'updated') + self.assertEqual(reg.tickets.count(), 2) + self.assertEqual(reg.tickets.filter(attendance_type__slug='onsite', ticket_type__slug='one_day').count(), 1) + self.assertEqual(reg.tickets.filter(attendance_type__slug='remote', ticket_type__slug='week_pass').count(), 1) + + def test_process_single_registration_nomcom(self): + '''Test that Volunteer is created if is_nomcom_volunteer=True''' + meeting = MeetingFactory(type_id='ietf', number='122') + person = PersonFactory() + reg_data = dict( + first_name=person.first_name(), + last_name=person.last_name(), + email=person.email().address, + affiliation='Microsoft', + country_code='US', + meeting=meeting.number, + checkedin=True, + is_nomcom_volunteer=True, + cancelled=False, + tickets=[{'attendance_type': 'onsite', 'ticket_type': 'week_pass'}], + ) + now = datetime.datetime.now() + if now.month > 10: + year = now.year + 1 + else: + year = now.year + # create appropriate group and nomcom objects + nomcom = NomComFactory.create(is_accepting_volunteers=True, **nomcom_kwargs_for_year(year)) + # assert no Volunteers exists + self.assertEqual(Volunteer.objects.count(), 0) + new_reg, action = process_single_registration(reg_data, meeting) + self.assertEqual(action, 'created') + # assert Volunteer exists + self.assertEqual(Volunteer.objects.count(), 1) + volunteer = Volunteer.objects.last() + self.assertEqual(volunteer.person, person) + self.assertEqual(volunteer.nomcom, nomcom) + self.assertEqual(volunteer.origin, 'registration') + + def test_process_single_registration_cancelled(self): + # test cancelled registration, one of two tickets + meeting = MeetingFactory(type_id='ietf', number='122') + person = PersonFactory() + reg = RegistrationFactory(meeting=meeting, person=person, checkedin=False, with_ticket={'attendance_type_id': 'onsite'}) + RegistrationTicketFactory(registration=reg, attendance_type_id='remote', ticket_type_id='week_pass') + reg_data = dict( + first_name=person.first_name(), + last_name=person.last_name(), + email=person.email().address, + affiliation='Microsoft', + country_code='US', + meeting=meeting.number, + checkedin=False, + is_nomcom_volunteer=False, + cancelled=True, + tickets=[{'attendance_type': 'onsite', 'ticket_type': 'week_pass'}], + ) + self.assertEqual(meeting.registration_set.count(), 1) + self.assertEqual(reg.tickets.count(), 2) + new_reg, action = process_single_registration(reg_data, meeting) + self.assertEqual((new_reg, action), (None, 'deleted')) + self.assertEqual(meeting.registration_set.count(), 1) + self.assertEqual(reg.tickets.count(), 1) + self.assertTrue(reg.tickets.filter(attendance_type__slug='remote').exists()) + # test cancelled registration, last ticket + reg_data['tickets'][0]['attendance_type'] = 'remote' + new_reg, action = process_single_registration(reg_data, meeting) + self.assertEqual((new_reg, action), (None, 'deleted')) + self.assertEqual(meeting.registration_set.count(), 0) + + @patch("ietf.meeting.utils.sync_registration_data") + def test_fetch_attendance_from_meetings(self, mock_sync_reg_data): + mock_meetings = [object(), object(), object()] + d1 = dict(created=1, updated=2, deleted=0, processed=3) + d2 = dict(created=2, updated=2, deleted=0, processed=4) + d3 = dict(created=1, updated=4, deleted=1, processed=5) + mock_sync_reg_data.side_effect = (d1, d2, d3) + stats = fetch_attendance_from_meetings(mock_meetings) + self.assertEqual( + [mock_sync_reg_data.call_args_list[n][0][0] for n in range(3)], + mock_meetings, + ) + self.assertEqual(stats, [d1, d2, d3]) diff --git a/ietf/meeting/tests_views.py b/ietf/meeting/tests_views.py index 9d7fc8f287..17988e50be 100644 --- a/ietf/meeting/tests_views.py +++ b/ietf/meeting/tests_views.py @@ -1,4 +1,4 @@ -# Copyright The IETF Trust 2009-2020, All Rights Reserved +# Copyright The IETF Trust 2009-2025, All Rights Reserved # -*- coding: utf-8 -*- import datetime import io @@ -12,9 +12,10 @@ import requests_mock from unittest import skipIf -from mock import patch, PropertyMock +from unittest.mock import call, patch, PropertyMock from pyquery import PyQuery from lxml.etree import tostring +from icalendar import Calendar from io import StringIO, BytesIO from bs4 import BeautifulSoup from urllib.parse import urlparse, urlsplit @@ -26,40 +27,51 @@ from django.urls import reverse as urlreverse from django.conf import settings from django.contrib.auth.models import User +from django.core.serializers.json import DjangoJSONEncoder from django.test import Client, override_settings from django.db.models import F, Max from django.http import QueryDict, FileResponse from django.template import Context, Template from django.utils import timezone +from django.utils.html import escape +from django.utils.safestring import mark_safe from django.utils.text import slugify import debug # pyflakes:ignore -from ietf.doc.models import Document +from ietf.doc.models import Document, NewRevisionDocEvent +from ietf.doc.storage_utils import exists_in_storage, remove_from_storage, retrieve_bytes, retrieve_str from ietf.group.models import Group, Role, GroupFeatures from ietf.group.utils import can_manage_group from ietf.person.models import Person -from ietf.meeting.helpers import can_approve_interim_request, can_view_interim_request, preprocess_assignments_for_agenda +from ietf.meeting.helpers import can_approve_interim_request, can_request_interim_meeting, can_view_interim_request, preprocess_assignments_for_agenda from ietf.meeting.helpers import send_interim_approval_request, AgendaKeywordTagger from ietf.meeting.helpers import send_interim_meeting_cancellation_notice, send_interim_session_cancellation_notice from ietf.meeting.helpers import send_interim_minutes_reminder, populate_important_dates, update_important_dates from ietf.meeting.models import Session, TimeSlot, Meeting, SchedTimeSessAssignment, Schedule, SessionPresentation, SlideSubmission, SchedulingEvent, Room, Constraint, ConstraintName from ietf.meeting.test_data import make_meeting_test_data, make_interim_meeting, make_interim_test_data -from ietf.meeting.utils import finalize, condition_slide_order -from ietf.meeting.utils import add_event_info_to_session_qs +from ietf.meeting.utils import ( + condition_slide_order, + generate_proceedings_content, + diff_meeting_schedules, +) +from ietf.meeting.utils import add_event_info_to_session_qs, participants_for_meeting +from ietf.meeting.utils import create_recording, delete_recording, get_next_sequence, bluesheet_data from ietf.meeting.views import session_draft_list, parse_agenda_filter_params, sessions_post_save, agenda_extract_schedule +from ietf.meeting.views import get_summary_by_area, get_summary_by_type, get_summary_by_purpose, generate_agenda_data from ietf.name.models import SessionStatusName, ImportantDateName, RoleName, ProceedingsMaterialTypeName -from ietf.utils.decorators import skip_coverage from ietf.utils.mail import outbox, empty_outbox, get_payload_text +from ietf.utils.test_runner import TestBlobstoreManager, disable_coverage from ietf.utils.test_utils import TestCase, login_testing_unauthorized, unicontent from ietf.utils.timezone import date_today, time_now -from ietf.person.factories import PersonFactory +from ietf.person.factories import PersonFactory, PersonalApiKeyFactory from ietf.group.factories import GroupFactory, GroupEventFactory, RoleFactory -from ietf.meeting.factories import ( SessionFactory, ScheduleFactory, +from ietf.meeting.factories import (SessionFactory, ScheduleFactory, SessionPresentationFactory, MeetingFactory, FloorPlanFactory, TimeSlotFactory, SlideSubmissionFactory, RoomFactory, - ConstraintFactory, MeetingHostFactory, ProceedingsMaterialFactory ) + ConstraintFactory, MeetingHostFactory, ProceedingsMaterialFactory, + AttendedFactory, RegistrationFactory) from ietf.doc.factories import DocumentFactory, WgDraftFactory from ietf.submit.tests import submission_file from ietf.utils.test_utils import assert_ical_response_is_valid @@ -106,7 +118,7 @@ def setUp(self): # files will upload to the locations specified in settings.py. # Note that this will affect any use of the storage class in # meeting.models - i.e., FloorPlan.image and MeetingHost.logo - self.patcher = patch('ietf.meeting.models.NoLocationMigrationFileSystemStorage.base_location', + self.patcher = patch('ietf.meeting.models.BlobShadowFileSystemStorage.base_location', new_callable=PropertyMock) mocked = self.patcher.start() mocked.return_value = self.storage_dir @@ -121,8 +133,12 @@ def tearDown(self): settings.MEETINGHOST_LOGO_PATH = self.saved_meetinghost_logo_path super().tearDown() - def write_materials_file(self, meeting, doc, content, charset="utf-8"): - path = os.path.join(self.materials_dir, "%s/%s/%s" % (meeting.number, doc.type_id, doc.uploaded_filename)) + def write_materials_file(self, meeting, doc, content, charset="utf-8", with_ext=None): + if with_ext is None: + filename = doc.uploaded_filename + else: + filename = Path(doc.uploaded_filename).with_suffix(with_ext) + path = os.path.join(self.materials_dir, "%s/%s/%s" % (meeting.number, doc.type_id, filename)) dirname = os.path.dirname(path) if not os.path.exists(dirname): @@ -162,7 +178,7 @@ def test_agenda_extract_schedule_location(self): meeting ) AgendaKeywordTagger(assignments=processed).apply() - extracted = {item.pk: agenda_extract_schedule(item) for item in processed} + extracted = {item.session.pk: agenda_extract_schedule(item) for item in processed} hidden = extracted[hidden_sess.pk] self.assertIsNone(hidden['room']) @@ -172,12 +188,56 @@ def test_agenda_extract_schedule_location(self): self.assertEqual(shown['room'], room.name) self.assertEqual(shown['location'], {'name': room.floorplan.name, 'short': room.floorplan.short}) + def test_agenda_extract_schedule_names(self): + meeting = MeetingFactory(type_id='ietf') + named_timeslots = TimeSlotFactory.create_batch(2, meeting=meeting, name='Timeslot Name') + unnamed_timeslots = TimeSlotFactory.create_batch(2, meeting=meeting, name='') + named_sessions = SessionFactory.create_batch(2, meeting=meeting, name='Session Name') + unnamed_sessions = SessionFactory.create_batch(2, meeting=meeting, name='') + pk_with = { + 'both named': named_sessions[0].timeslotassignments.create( + schedule=meeting.schedule, + timeslot=named_timeslots[0], + ).pk, + 'session named': named_sessions[1].timeslotassignments.create( + schedule=meeting.schedule, + timeslot=unnamed_timeslots[0], + ).pk, + 'timeslot named': unnamed_sessions[0].timeslotassignments.create( + schedule=meeting.schedule, + timeslot=named_timeslots[1], + ).pk, + 'neither named': unnamed_sessions[1].timeslotassignments.create( + schedule=meeting.schedule, + timeslot=unnamed_timeslots[1], + ).pk, + } + processed = preprocess_assignments_for_agenda(meeting.schedule.assignments.all(), meeting) + AgendaKeywordTagger(assignments=processed).apply() + extracted = {item.pk: agenda_extract_schedule(item) for item in processed} + self.assertEqual(extracted[pk_with['both named']]['name'], 'Session Name') + self.assertEqual(extracted[pk_with['both named']]['slotName'], 'Timeslot Name') + self.assertEqual(extracted[pk_with['session named']]['name'], 'Session Name') + self.assertEqual(extracted[pk_with['session named']]['slotName'], '') + self.assertEqual(extracted[pk_with['timeslot named']]['name'], '') + self.assertEqual(extracted[pk_with['timeslot named']]['slotName'], 'Timeslot Name') + self.assertEqual(extracted[pk_with['neither named']]['name'], '') + self.assertEqual(extracted[pk_with['neither named']]['slotName'], '') + class MeetingTests(BaseMeetingTestCase): + @override_settings( + MEETECHO_ONSITE_TOOL_URL="https://onsite.example.com", + MEETECHO_VIDEO_STREAM_URL="https://meetecho.example.com", + ) def test_meeting_agenda(self): meeting = make_meeting_test_data() session = Session.objects.filter(meeting=meeting, group__acronym="mars").first() + session.remote_instructions='https://remote.example.com' + session.save() slot = TimeSlot.objects.get(sessionassignments__session=session,sessionassignments__schedule=meeting.schedule) + meeting.timeslot_set.filter(type_id="break").update(show_location=False) + meeting.importantdate_set.create(name_id='prelimagenda',date=date_today() + datetime.timedelta(days=20)) # self.write_materials_files(meeting, session) # @@ -188,39 +248,41 @@ def test_meeting_agenda(self): registration_text = "Registration" - # utc - time_interval = r"%s-%s" % (slot.utc_start_time().strftime("%H:%M").lstrip("0"), (slot.utc_start_time() + slot.duration).strftime("%H:%M").lstrip("0")) - # Extremely rudementary test of agenda-neue - to be replaced with back-end tests as the front-end tests are developed. r = self.client.get(urlreverse("agenda", kwargs=dict(num=meeting.number,utc='-utc'))) self.assertEqual(r.status_code, 200) # Agenda API tests # -> Meeting data - r = self.client.get(urlreverse("ietf.meeting.views.api_get_agenda_data", kwargs=dict(num=meeting.number))) - self.assertEqual(r.status_code, 200) - rjson = json.loads(r.content.decode("utf8")) - self.assertJSONEqual( - r.content.decode("utf8"), + # First, check that the generation function does the right thing + generated_data = generate_agenda_data(meeting.number) + self.assertEqual( + generated_data, { "meeting": { "number": meeting.number, "city": meeting.city, "startDate": meeting.date.isoformat(), "endDate": meeting.end_date().isoformat(), - "updated": rjson.get("meeting").get("updated"), # Just expect the value to exist + "updated": generated_data.get("meeting").get("updated"), # Just expect the value to exist "timezone": meeting.time_zone, "infoNote": meeting.agenda_info_note, - "warningNote": meeting.agenda_warning_note + "warningNote": meeting.agenda_warning_note, + "prelimAgendaDate": (date_today() + datetime.timedelta(days=20)).isoformat() }, - "categories": rjson.get("categories"), # Just expect the value to exist + "categories": generated_data.get("categories"), # Just expect the value to exist "isCurrentMeeting": True, - "useNotes": True, - "schedule": rjson.get("schedule"), # Just expect the value to exist + "usesNotes": False, # make_meeting_test_data sets number=72 + "schedule": generated_data.get("schedule"), # Just expect the value to exist "floors": [] } ) - # -> Session Materials + with patch("ietf.meeting.views.generate_agenda_data", return_value=generated_data): + r = self.client.get(urlreverse("ietf.meeting.views.api_get_agenda_data", kwargs=dict(num=meeting.number))) + self.assertEqual(r.status_code, 200) + # json.dumps using the DjangoJSONEncoder to handle timestamps consistently + self.assertJSONEqual(r.content.decode("utf8"), json.dumps(generated_data, cls=DjangoJSONEncoder)) + # -> Session MaterialM r = self.client.get(urlreverse("ietf.meeting.views.api_get_session_materials", kwargs=dict(session_id=session.id))) self.assertEqual(r.status_code, 200) rjson = json.loads(r.content.decode("utf8")) @@ -239,23 +301,44 @@ def test_meeting_agenda(self): } ) - # plain - time_interval = r"{}-{}".format( - slot.time.astimezone(meeting.tz()).strftime("%H:%M").lstrip("0"), - slot.end_time().astimezone(meeting.tz()).strftime("%H:%M").lstrip("0"), - ) - # text - # the rest of the results don't have as nicely formatted times - time_interval = "%s-%s" % (slot.time.strftime("%H%M").lstrip("0"), (slot.time + slot.duration).strftime("%H%M").lstrip("0")) - r = self.client.get(urlreverse("ietf.meeting.views.agenda_plain", kwargs=dict(num=meeting.number, ext=".txt"))) self.assertContains(r, session.group.acronym) self.assertContains(r, session.group.name) self.assertContains(r, session.group.parent.acronym.upper()) self.assertContains(r, slot.location.name) - - self.assertContains(r, time_interval) + self.assertContains(r, "{}-{}".format( + slot.time.astimezone(meeting.tz()).strftime("%H%M"), + (slot.time + slot.duration).astimezone(meeting.tz()).strftime("%H%M"), + )) + self.assertContains(r, f"shown in the {meeting.tz()} time zone") + updated = meeting.updated().astimezone(meeting.tz()).strftime("%Y-%m-%d %H:%M:%S %Z") + self.assertContains(r, f"Updated {updated}") + + # text, UTC + r = self.client.get(urlreverse( + "ietf.meeting.views.agenda_plain", + kwargs=dict(num=meeting.number, ext=".txt", utc="-utc"), + )) + self.assertContains(r, session.group.acronym) + self.assertContains(r, session.group.name) + self.assertContains(r, session.group.parent.acronym.upper()) + self.assertContains(r, slot.location.name) + self.assertContains(r, "{}-{}".format( + slot.time.astimezone(datetime.UTC).strftime("%H%M"), + (slot.time + slot.duration).astimezone(datetime.UTC).strftime("%H%M"), + )) + self.assertContains(r, "shown in UTC") + updated = meeting.updated().astimezone(datetime.UTC).strftime("%Y-%m-%d %H:%M:%S %Z") + self.assertContains(r, f"Updated {updated}") + + # text, invalid updated (none) + with patch("ietf.meeting.models.Meeting.updated", return_value=None): + r = self.client.get(urlreverse( + "ietf.meeting.views.agenda_plain", + kwargs=dict(num=meeting.number, ext=".txt", utc="-utc"), + )) + self.assertNotContains(r, "Updated ") # future meeting, no agenda r = self.client.get(urlreverse("ietf.meeting.views.agenda_plain", kwargs=dict(num=future_meeting.number, ext=".txt"))) @@ -269,33 +352,204 @@ def test_meeting_agenda(self): self.assertContains(r, session.group.parent.acronym.upper()) self.assertContains(r, slot.location.name) self.assertContains(r, registration_text) + start_time = slot.time.astimezone(meeting.tz()) + end_time = slot.end_time().astimezone(meeting.tz()) + self.assertContains(r, '"{}","{}","{}"'.format( + start_time.strftime("%Y-%m-%d"), + start_time.strftime("%H%M"), + end_time.strftime("%H%M"), + )) + self.assertContains(r, session.materials.get(type='agenda').uploaded_filename) + self.assertContains(r, session.materials.filter(type='slides').exclude(states__type__slug='slides',states__slug='deleted').first().uploaded_filename) + self.assertNotContains(r, session.materials.filter(type='slides',states__type__slug='slides',states__slug='deleted').first().uploaded_filename) + # CSV, utc + r = self.client.get(urlreverse( + "ietf.meeting.views.agenda_plain", + kwargs=dict(num=meeting.number, ext=".csv", utc="-utc"), + )) + self.assertContains(r, session.group.acronym) + self.assertContains(r, session.group.name) + self.assertContains(r, session.group.parent.acronym.upper()) + self.assertContains(r, slot.location.name) + self.assertContains(r, registration_text) + start_time = slot.time.astimezone(datetime.UTC) + end_time = slot.end_time().astimezone(datetime.UTC) + self.assertContains(r, '"{}","{}","{}"'.format( + start_time.strftime("%Y-%m-%d"), + start_time.strftime("%H%M"), + end_time.strftime("%H%M"), + )) self.assertContains(r, session.materials.get(type='agenda').uploaded_filename) self.assertContains(r, session.materials.filter(type='slides').exclude(states__type__slug='slides',states__slug='deleted').first().uploaded_filename) self.assertNotContains(r, session.materials.filter(type='slides',states__type__slug='slides',states__slug='deleted').first().uploaded_filename) - # iCal - r = self.client.get(urlreverse("ietf.meeting.views.agenda_ical", kwargs=dict(num=meeting.number)) - + "?show=" + session.group.parent.acronym.upper()) + # iCal, no session filtering + ical_url = urlreverse("ietf.meeting.views.agenda_ical", kwargs=dict(num=meeting.number)) + r = self.client.get(ical_url) + + assert_ical_response_is_valid(self, r) + self.assertNotEqual( + meeting.time_zone, + meeting.time_zone.lower(), + "meeting needs a mixed-case tz for this test", + ) + self.assertNotContains(r, meeting.time_zone.lower(), msg_prefix="time_zone should not be lower-cased") + self.assertNotEqual( + meeting.time_zone, + meeting.time_zone.upper(), + "meeting needs a mixed-case tz for this test", + ) + self.assertNotContains(r, meeting.time_zone.upper(), msg_prefix="time_zone should not be upper-cased") + + # iCal, single group + r = self.client.get(ical_url + "?show=" + session.group.parent.acronym.upper()) assert_ical_response_is_valid(self, r) self.assertContains(r, session.group.acronym) self.assertContains(r, session.group.name) - self.assertContains(r, slot.location.name) - self.assertContains(r, "BEGIN:VTIMEZONE") - self.assertContains(r, "END:VTIMEZONE") - self.assertContains(r, session.agenda().get_href()) - self.assertContains( - r, + cal = Calendar.from_ical(r.content) + events = [component for component in cal.walk() if component.name == "VEVENT"] + + self.assertEqual(len(events), 2) + self.assertIn(session.remote_instructions, events[0].get('description')) + self.assertIn("Onsite tool: https://onsite.example.com", events[0].get('description')) + self.assertIn("Meetecho: https://meetecho.example.com", events[0].get('description')) + self.assertIn(f"Agenda {session.agenda().get_href()}", events[0].get('description')) + session_materials_url = settings.IDTRACKER_BASE_URL + urlreverse( + 'ietf.meeting.views.session_details', + kwargs=dict(num=meeting.number, acronym=session.group.acronym) + ) + self.assertIn(f"Session materials: {session_materials_url}", events[0].get('description')) + self.assertIn( urlreverse( 'ietf.meeting.views.session_details', kwargs=dict(num=meeting.number, acronym=session.group.acronym)), - msg_prefix='ical should contain link to meeting materials page for session') + events[0].get('description')) + self.assertEqual( + session_materials_url, + events[0].get('url') + ) + self.assertContains(r, f"LOCATION:{slot.location.name}") + # Floor Plan r = self.client.get(urlreverse('floor-plan', kwargs=dict(num=meeting.number))) self.assertEqual(r.status_code, 200) + def test_session_recordings_via_factories(self): + session = SessionFactory(meeting__type_id="ietf", meeting__date=date_today()-datetime.timedelta(days=180), meeting__number=str(random.randint(108,150))) + self.assertEqual(session.meetecho_recording_name, "") + self.assertEqual(len(session.recordings()), 0) + url = urlreverse("ietf.meeting.views.session_details", kwargs=dict(num=session.meeting.number, acronym=session.group.acronym)) + r = self.client.get(url) + q = PyQuery(r.content) + # debug.show("q(f'#notes_and_recordings_{session.pk}')") + self.assertEqual(len(q(f"#notes_and_recordings_{session.pk} tr")), 2) + links = q(f"#notes_and_recordings_{session.pk} tr a") + self.assertEqual(len(links), 2) + self.assertEqual(links[0].attrib['href'], str(session.notes_url())) + self.assertEqual(links[1].attrib['href'], str(session.session_recording_url())) + + session.meetecho_recording_name = 'my_test_session_name' + session.save() + r = self.client.get(url) + q = PyQuery(r.content) + self.assertEqual(len(q(f"#notes_and_recordings_{session.pk} tr")), 2) + links = q(f"#notes_and_recordings_{session.pk} tr a") + self.assertEqual(len(links), 2) + self.assertEqual(links[0].attrib['href'], str(session.notes_url())) + self.assertEqual(links[1].attrib['href'], str(session.session_recording_url())) + + new_recording_url = "https://www.youtube.com/watch?v=jNQXAC9IVRw" + new_recording_title = "Me at the zoo" + create_recording(session, new_recording_url, new_recording_title) + r = self.client.get(url) + q = PyQuery(r.content) + self.assertEqual(len(q(f"#notes_and_recordings_{session.pk} tr")), 3) + links = q(f"#notes_and_recordings_{session.pk} tr a") + self.assertEqual(len(links), 3) + self.assertEqual(links[0].attrib['href'], str(session.notes_url())) + self.assertEqual(links[1].attrib['href'], new_recording_url) + self.assertIn(new_recording_title, links[1].text_content()) + self.assertEqual(links[2].attrib['href'], str(session.session_recording_url())) + #debug.show("q(f'#notes_and_recordings_{session_pk}')") + + def test_delete_recordings(self): + # No user specified, active recording state + sp = SessionPresentationFactory( + document__type_id="recording", + document__external_url="https://example.com/some-recording", + document__states=[("recording", "active")], + ) + doc = sp.document + doc.docevent_set.all().delete() # clear this out + delete_recording(sp) + self.assertFalse(SessionPresentation.objects.filter(pk=sp.pk).exists()) + self.assertEqual(doc.get_state("recording").slug, "deleted", "recording state updated") + self.assertEqual(doc.docevent_set.count(), 1, "one event added") + event = doc.docevent_set.first() + self.assertEqual(event.type, "changed_state", "event is a changed_state event") + self.assertEqual(event.by.name, "(System)", "system user is responsible") + + # Specified user, no recording state + sp = SessionPresentationFactory( + document__type_id="recording", + document__external_url="https://example.com/some-recording", + document__states=[], + ) + doc = sp.document + doc.docevent_set.all().delete() # clear this out + user = PersonFactory() # naming matches the methods - user is a Person, not a User + delete_recording(sp, user=user) + self.assertFalse(SessionPresentation.objects.filter(pk=sp.pk).exists()) + self.assertEqual(doc.get_state("recording").slug, "deleted", "recording state updated") + self.assertEqual(doc.docevent_set.count(), 1, "one event added") + event = doc.docevent_set.first() + self.assertEqual(event.type, "changed_state", "event is a changed_state event") + self.assertEqual(event.by, user, "user is responsible") + + # Document is not a recording + sp = SessionPresentationFactory( + document__type_id="draft", + document__external_url="https://example.com/some-recording", + ) + with self.assertRaises(ValueError): + delete_recording(sp) + + def test_agenda_ical_next_meeting_type(self): + # start with no upcoming IETF meetings, just an interim + MeetingFactory( + type_id="interim", date=date_today() + datetime.timedelta(days=15) + ) + r = self.client.get(urlreverse("ietf.meeting.views.agenda_ical", kwargs={})) + self.assertEqual( + r.status_code, 404, "Should not return an interim meeting as next meeting" + ) + # create an IETF meeting after the interim - it should be found as "next" + ietf_meeting = MeetingFactory( + type_id="ietf", date=date_today() + datetime.timedelta(days=30) + ) + SessionFactory(meeting=ietf_meeting, name="Session at IETF meeting") + r = self.client.get(urlreverse("ietf.meeting.views.agenda_ical", kwargs={})) + self.assertContains(r, "Session at IETF meeting", status_code=200) + + def test_agenda_json_next_meeting_type(self): + # start with no upcoming IETF meetings, just an interim + MeetingFactory( + type_id="interim", date=date_today() + datetime.timedelta(days=15) + ) + r = self.client.get(urlreverse("ietf.meeting.views.agenda_json", kwargs={})) + self.assertEqual( + r.status_code, 404, "Should not return an interim meeting as next meeting" + ) + # create an IETF meeting after the interim - it should be found as "next" + ietf_meeting = MeetingFactory( + type_id="ietf", date=date_today() + datetime.timedelta(days=30) + ) + SessionFactory(meeting=ietf_meeting, name="Session at IETF meeting") + r = self.client.get(urlreverse("ietf.meeting.views.agenda_json", kwargs={})) + self.assertContains(r, "Session at IETF meeting", status_code=200) @override_settings(PROCEEDINGS_V1_BASE_URL='https://example.com/{meeting.number}') def test_agenda_redirects_for_old_meetings(self): @@ -345,16 +599,16 @@ def test_materials_through_cdn(self): doc = DocumentFactory.create(name='agenda-172-mars', type_id='agenda', title="Agenda", uploaded_filename="agenda-172-mars.txt", group=session107.group, rev='00', states=[('agenda','active')]) pres = SessionPresentation.objects.create(session=session107,document=doc,rev=doc.rev) - session107.sessionpresentation_set.add(pres) # + session107.presentations.add(pres) # doc = DocumentFactory.create(name='minutes-172-mars', type_id='minutes', title="Minutes", uploaded_filename="minutes-172-mars.md", group=session107.group, rev='00', states=[('minutes','active')]) pres = SessionPresentation.objects.create(session=session107,document=doc,rev=doc.rev) - session107.sessionpresentation_set.add(pres) + session107.presentations.add(pres) doc = DocumentFactory.create(name='slides-172-mars-1-active', type_id='slides', title="Slideshow", uploaded_filename="slides-172-mars.txt", group=session107.group, rev='00', states=[('slides','active'), ('reuse_policy', 'single')]) pres = SessionPresentation.objects.create(session=session107,document=doc,rev=doc.rev) - session107.sessionpresentation_set.add(pres) + session107.presentations.add(pres) for session in ( Session.objects.filter(meeting=meeting, group__acronym="mars").first(), @@ -387,6 +641,66 @@ def test_interim_materials(self): self.do_test_materials(meeting, session) + def test_named_session(self): + """Session with a name should appear separately in the materials""" + meeting = MeetingFactory(type_id='ietf', number='100') + meeting.importantdate_set.create(name_id='revsub',date=date_today() + datetime.timedelta(days=20)) + group = GroupFactory() + plain_session = SessionFactory(meeting=meeting, group=group) + named_session = SessionFactory(meeting=meeting, group=group, name='I Got a Name') + for doc_type_id in ('agenda', 'minutes', 'slides', 'draft'): + # Set up sessions materials that will have distinct URLs for each session. + # This depends on settings.MEETING_DOC_HREFS and may need updating if that changes. + SessionPresentationFactory( + session=plain_session, + document__type_id=doc_type_id, + document__uploaded_filename=f'upload-{doc_type_id}-plain', + document__external_url=f'external_url-{doc_type_id}-plain', + ) + SessionPresentationFactory( + session=named_session, + document__type_id=doc_type_id, + document__uploaded_filename=f'upload-{doc_type_id}-named', + document__external_url=f'external_url-{doc_type_id}-named', + ) + + url = urlreverse('ietf.meeting.views.materials', kwargs={'num': meeting.number}) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + + plain_label = q(f'div#{group.acronym}') + self.assertEqual(plain_label.text(), group.acronym) + plain_row = plain_label.closest('tr') + self.assertTrue(plain_row) + + named_label = q(f'div#{slugify(named_session.name)}') + self.assertEqual(named_label.text(), named_session.name) + named_row = named_label.closest('tr') + self.assertTrue(named_row) + + for material in (sp.document for sp in plain_session.presentations.all()): + if material.type_id == 'draft': + expected_url = urlreverse( + 'ietf.doc.views_doc.document_main', + kwargs={'name': material.name}, + ) + else: + expected_url = material.get_href(meeting) + self.assertTrue(plain_row.find(f'a[href="{expected_url}"]')) + self.assertFalse(named_row.find(f'a[href="{expected_url}"]')) + + for material in (sp.document for sp in named_session.presentations.all()): + if material.type_id == 'draft': + expected_url = urlreverse( + 'ietf.doc.views_doc.document_main', + kwargs={'name': material.name}, + ) + else: + expected_url = material.get_href(meeting) + self.assertFalse(plain_row.find(f'a[href="{expected_url}"]')) + self.assertTrue(named_row.find(f'a[href="{expected_url}"]')) + @override_settings(MEETING_MATERIALS_SERVE_LOCALLY=True) def test_meeting_materials_non_utf8(self): meeting = make_meeting_test_data() @@ -435,7 +749,7 @@ def do_test_materials(self, meeting, session): self.assertContains(r, "1. More work items underway") - cont_disp = r._headers.get('content-disposition', ('Content-Disposition', ''))[1] + cont_disp = r.headers.get('content-disposition', ('Content-Disposition', ''))[1] cont_disp = re.split('; ?', cont_disp) cont_disp_settings = dict( e.split('=', 1) for e in cont_disp if '=' in e ) filename = cont_disp_settings.get('filename', '').strip('"') @@ -467,6 +781,20 @@ def do_test_materials(self, meeting, session): self.assertFalse(row.find("a:contains(\"Bad Slideshow\")")) # test with no meeting number in url + # Add various group sessions + groups = [] + parent_groups = [ + GroupFactory.create(type_id="area", acronym="gen"), + GroupFactory.create(acronym="iab"), + GroupFactory.create(acronym="irtf"), + ] + for parent in parent_groups: + groups.append(GroupFactory.create(parent=parent)) + for acronym in ["rsab", "edu"]: + groups.append(GroupFactory.create(acronym=acronym)) + for group in groups: + SessionFactory(meeting=meeting, group=group) + self.write_materials_files(meeting, session) url = urlreverse("ietf.meeting.views.materials", kwargs=dict()) r = self.client.get(url) self.assertEqual(r.status_code, 200) @@ -476,6 +804,10 @@ def do_test_materials(self, meeting, session): self.assertTrue(row.find('a:contains("Minutes")')) self.assertTrue(row.find('a:contains("Slideshow")')) self.assertFalse(row.find("a:contains(\"Bad Slideshow\")")) + # test for different sections + sections = ["plenaries", "gen", "iab", "editorial", "irtf", "training"] + for section in sections: + self.assertEqual(len(q(f"#{section}")), 1, f"{section} section should exists in proceedings") # test with a loggged-in wg chair self.client.login(username="marschairman", password="marschairman+password") @@ -540,7 +872,56 @@ def test_materials_has_edit_links(self): ) self.assertEqual(len(q(f'a[href^="{edit_url}#session"]')), 1, f'Link to session_details page for {acro}') + def test_materials_document_extension_choice(self): + def _url(**kwargs): + return urlreverse("ietf.meeting.views.materials_document", kwargs=kwargs) + + presentation = SessionPresentationFactory( + document__rev="00", + document__name="slides-whatever", + document__uploaded_filename="slides-whatever-00.txt", + document__type_id="slides", + document__states=(("reuse_policy", "single"),) + ) + session = presentation.session + meeting = session.meeting + # This is not a realistic set of files to exist, but is useful for testing. Normally, + # we'd have _either_ txt, pdf, or pptx + pdf. + self.write_materials_file(meeting, presentation.document, "Hi I'm a txt", with_ext=".txt") + self.write_materials_file(meeting, presentation.document, "Hi I'm a pptx", with_ext=".pptx") + + # with no rev, prefers the uploaded_filename + r = self.client.get(_url(document="slides-whatever", num=meeting.number)) # no rev + self.assertEqual(r.status_code, 200) + self.assertEqual(r.content.decode(), "Hi I'm a txt") + + # with a rev, prefers pptx because it comes first alphabetically + r = self.client.get(_url(document="slides-whatever-00", num=meeting.number)) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.content.decode(), "Hi I'm a pptx") + + # now create a pdf + self.write_materials_file(meeting, presentation.document, "Hi I'm a pdf", with_ext=".pdf") + + # with no rev, still prefers uploaded_filename + r = self.client.get(_url(document="slides-whatever", num=meeting.number)) # no rev + self.assertEqual(r.status_code, 200) + self.assertEqual(r.content.decode(), "Hi I'm a txt") + # pdf should be preferred with a rev + r = self.client.get(_url(document="slides-whatever-00", num=meeting.number)) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.content.decode(), "Hi I'm a pdf") + + # and explicit extensions should, of course, be respected + for ext in ["pdf", "pptx", "txt"]: + r = self.client.get(_url(document="slides-whatever-00", num=meeting.number, ext=f".{ext}")) + self.assertEqual(r.status_code, 200) + self.assertEqual(r.content.decode(), f"Hi I'm a {ext}") + + # and 404 should come up if the ext is not found + r = self.client.get(_url(document="slides-whatever-00", num=meeting.number, ext=".docx")) + self.assertEqual(r.status_code, 404) def test_materials_editable_groups(self): meeting = make_meeting_test_data() @@ -564,13 +945,73 @@ def test_materials_editable_groups(self): @override_settings(MEETING_MATERIALS_SERVE_LOCALLY=True) def test_materials_name_endswith_hyphen_number_number(self): - sp = SessionPresentationFactory(document__name='slides-junk-15',document__type_id='slides',document__states=[('reuse_policy','single')]) - sp.document.uploaded_filename = '%s-%s.pdf'%(sp.document.name,sp.document.rev) + # be sure a shadowed filename without the hyphen does not interfere + shadow = SessionPresentationFactory( + document__name="slides-115-junk", + document__type_id="slides", + document__states=[("reuse_policy", "single")], + ) + shadow.document.uploaded_filename = ( + f"{shadow.document.name}-{shadow.document.rev}.pdf" + ) + shadow.document.save() + # create the material we want to find for the test + sp = SessionPresentationFactory( + document__name="slides-115-junk-15", + document__type_id="slides", + document__states=[("reuse_policy", "single")], + ) + sp.document.uploaded_filename = f"{sp.document.name}-{sp.document.rev}.pdf" sp.document.save() - self.write_materials_file(sp.session.meeting, sp.document, 'Fake slide contents') - url = urlreverse("ietf.meeting.views.materials_document", kwargs=dict(document=sp.document.name,num=sp.session.meeting.number)) + self.write_materials_file( + sp.session.meeting, sp.document, "Fake slide contents rev 00" + ) + + # create rev 01 + sp.document.rev = "01" + sp.document.uploaded_filename = f"{sp.document.name}-{sp.document.rev}.pdf" + sp.document.save_with_history( + [ + NewRevisionDocEvent.objects.create( + type="new_revision", + doc=sp.document, + rev=sp.document.rev, + by=Person.objects.get(name="(System)"), + desc=f"New version available: {sp.document.name}-{sp.document.rev}.txt", + ) + ] + ) + self.write_materials_file( + sp.session.meeting, sp.document, "Fake slide contents rev 01" + ) + url = urlreverse( + "ietf.meeting.views.materials_document", + kwargs=dict(document=sp.document.name, num=sp.session.meeting.number), + ) r = self.client.get(url) - self.assertEqual(r.status_code, 200) + self.assertContains( + r, + "Fake slide contents rev 01", + status_code=200, + msg_prefix="Should return latest rev by default", + ) + url = urlreverse( + "ietf.meeting.views.materials_document", + kwargs=dict(document=sp.document.name + "-00", num=sp.session.meeting.number), + ) + r = self.client.get(url) + self.assertContains( + r, + "Fake slide contents rev 00", + status_code=200, + msg_prefix="Should return existing version on request", + ) + url = urlreverse( + "ietf.meeting.views.materials_document", + kwargs=dict(document=sp.document.name + "-02", num=sp.session.meeting.number), + ) + r = self.client.get(url) + self.assertEqual(r.status_code, 404, "Should not find nonexistent version") def test_important_dates(self): meeting=MeetingFactory(type_id='ietf') @@ -598,37 +1039,59 @@ def test_important_dates_ical(self): for d in meeting.importantdate_set.all(): self.assertContains(r, d.date.isoformat()) + updated = meeting.updated() + self.assertIsNotNone(updated) + expected_updated = updated.astimezone(datetime.UTC).strftime("%Y%m%dT%H%M%SZ") + self.assertContains(r, f"DTSTAMP:{expected_updated}") + dtstamps_count = r.content.decode("utf-8").count(f"DTSTAMP:{expected_updated}") + self.assertEqual(dtstamps_count, meeting.importantdate_set.count()) + + # With default cached_updated, 1970-01-01 + with patch("ietf.meeting.models.Meeting.updated", return_value=None): + r = self.client.get(url) + for d in meeting.importantdate_set.all(): + self.assertContains(r, d.date.isoformat()) + + expected_updated = "19700101T000000Z" + self.assertContains(r, f"DTSTAMP:{expected_updated}") + dtstamps_count = r.content.decode("utf-8").count(f"DTSTAMP:{expected_updated}") + self.assertEqual(dtstamps_count, meeting.importantdate_set.count()) + def test_group_ical(self): meeting = make_meeting_test_data() s1 = Session.objects.filter(meeting=meeting, group__acronym="mars").first() a1 = s1.official_timeslotassignment() t1 = a1.timeslot + # Create an extra session t2 = TimeSlotFactory.create( meeting=meeting, - time=meeting.tz().localize( + time=pytz.utc.localize( datetime.datetime.combine(meeting.date, datetime.time(11, 30)) ) ) + s2 = SessionFactory.create(meeting=meeting, group=s1.group, add_to_schedule=False) SchedTimeSessAssignment.objects.create(timeslot=t2, session=s2, schedule=meeting.schedule) - # + url = urlreverse('ietf.meeting.views.agenda_ical', kwargs={'num':meeting.number, 'acronym':s1.group.acronym, }) r = self.client.get(url) assert_ical_response_is_valid(self, r, expected_event_summaries=['mars - Martian Special Interest Group'], expected_event_count=2) - self.assertContains(r, t1.local_start_time().strftime('%Y%m%dT%H%M%S')) - self.assertContains(r, t2.local_start_time().strftime('%Y%m%dT%H%M%S')) - # + self.assertContains(r, f"DTSTART:{t1.time.strftime('%Y%m%dT%H%M%SZ')}") + self.assertContains(r, f"DTEND:{(t1.time + t1.duration).strftime('%Y%m%dT%H%M%SZ')}") + self.assertContains(r, f"DTSTART:{t2.time.strftime('%Y%m%dT%H%M%SZ')}") + self.assertContains(r, f"DTEND:{(t2.time + t2.duration).strftime('%Y%m%dT%H%M%SZ')}") + url = urlreverse('ietf.meeting.views.agenda_ical', kwargs={'num':meeting.number, 'session_id':s1.id, }) r = self.client.get(url) assert_ical_response_is_valid(self, r, expected_event_summaries=['mars - Martian Special Interest Group'], expected_event_count=1) - self.assertContains(r, t1.local_start_time().strftime('%Y%m%dT%H%M%S')) - self.assertNotContains(r, t2.local_start_time().strftime('%Y%m%dT%H%M%S')) + self.assertContains(r, f"DTSTART:{t1.time.strftime('%Y%m%dT%H%M%SZ')}") + self.assertNotContains(r, f"DTSTART:{t2.time.strftime('%Y%m%dT%H%M%SZ')}") def test_parse_agenda_filter_params(self): def _r(show=(), hide=(), showtypes=(), hidetypes=()): @@ -694,10 +1157,10 @@ def build_session_setup(self): # but lists a different on in its agenda. The expectation is that the pdf and tgz views will return both. session = SessionFactory(group__type_id='wg',meeting__type_id='ietf') draft1 = WgDraftFactory(group=session.group) - session.sessionpresentation_set.create(document=draft1) + session.presentations.create(document=draft1) draft2 = WgDraftFactory(group=session.group) agenda = DocumentFactory(type_id='agenda',group=session.group, uploaded_filename='agenda-%s-%s' % (session.meeting.number,session.group.acronym), states=[('agenda','active')]) - session.sessionpresentation_set.create(document=agenda) + session.presentations.create(document=agenda) self.write_materials_file(session.meeting, session.materials.get(type="agenda"), "1. WG status (15 minutes)\n\n2. Status of %s\n\n" % draft2.name) filenames = [] @@ -712,23 +1175,27 @@ def build_session_setup(self): def test_session_draft_tarfile(self): session, filenames = self.build_session_setup() - url = urlreverse('ietf.meeting.views.session_draft_tarfile', kwargs={'num':session.meeting.number,'acronym':session.group.acronym}) - response = self.client.get(url) - self.assertEqual(response.status_code, 200) - self.assertEqual(response.get('Content-Type'), 'application/octet-stream') - for filename in filenames: - os.unlink(filename) + try: + url = urlreverse('ietf.meeting.views.session_draft_tarfile', kwargs={'num':session.meeting.number,'acronym':session.group.acronym}) + response = self.client.get(url) + self.assertEqual(response.status_code, 200) + self.assertEqual(response.get('Content-Type'), 'application/octet-stream') + finally: + for filename in filenames: + os.unlink(filename) @skipIf(skip_pdf_tests, skip_message) - @skip_coverage - def test_session_draft_pdf(self): + @disable_coverage() + def test_session_draft_pdf(self): # pragma: no cover session, filenames = self.build_session_setup() - url = urlreverse('ietf.meeting.views.session_draft_pdf', kwargs={'num':session.meeting.number,'acronym':session.group.acronym}) - response = self.client.get(url) - self.assertEqual(response.status_code, 200) - self.assertEqual(response.get('Content-Type'), 'application/pdf') - for filename in filenames: - os.unlink(filename) + try: + url = urlreverse('ietf.meeting.views.session_draft_pdf', kwargs={'num':session.meeting.number,'acronym':session.group.acronym}) + response = self.client.get(url) + self.assertEqual(response.status_code, 200) + self.assertEqual(response.get('Content-Type'), 'application/pdf') + finally: + for filename in filenames: + os.unlink(filename) def test_current_materials(self): url = urlreverse('ietf.meeting.views.current_materials') @@ -1654,8 +2121,8 @@ def test_editor_time_zone(self): # strftime() does not seem to support hours without leading 0, so do this manually time_label_string = f'{ts_start.hour:d}:{ts_start.minute:02d} - {ts_end.hour:d}:{ts_end.minute:02d}' self.assertIn(time_label_string, time_label.text()) - self.assertEqual(time_label.attr('data-start'), ts_start.astimezone(datetime.timezone.utc).isoformat()) - self.assertEqual(time_label.attr('data-end'), ts_end.astimezone(datetime.timezone.utc).isoformat()) + self.assertEqual(time_label.attr('data-start'), ts_start.astimezone(datetime.UTC).isoformat()) + self.assertEqual(time_label.attr('data-end'), ts_end.astimezone(datetime.UTC).isoformat()) ts_swap = time_label.find('.swap-timeslot-col') origin_label = ts_swap.attr('data-origin-label') @@ -1666,8 +2133,8 @@ def test_editor_time_zone(self): timeslot_elt = pq(f'#timeslot{timeslot.pk}') self.assertEqual(len(timeslot_elt), 1) - self.assertEqual(timeslot_elt.attr('data-start'), ts_start.astimezone(datetime.timezone.utc).isoformat()) - self.assertEqual(timeslot_elt.attr('data-end'), ts_end.astimezone(datetime.timezone.utc).isoformat()) + self.assertEqual(timeslot_elt.attr('data-start'), ts_start.astimezone(datetime.UTC).isoformat()) + self.assertEqual(timeslot_elt.attr('data-end'), ts_end.astimezone(datetime.UTC).isoformat()) timeslot_label = pq(f'#timeslot{timeslot.pk} .time-label') self.assertEqual(len(timeslot_label), 1) @@ -1695,7 +2162,8 @@ def create_timeslots_url(meeting): @staticmethod def create_bare_meeting(number=120) -> Meeting: """Create a basic IETF meeting""" - return MeetingFactory( + # Call create() explicitly so mypy sees the correct type + return MeetingFactory.create( type_id='ietf', number=number, date=date_today() + datetime.timedelta(days=10), @@ -2120,24 +2588,30 @@ def test_edit_timeslot(self): ) self.login() + url = self.edit_timeslot_url(ts) + + # check that sched parameter is preserved + r = self.client.get(url) + self.assertNotContains(r, '?sched=', status_code=200) + r = self.client.get(url + '?sched=1234') + self.assertContains(r, '?sched=1234', status_code=200) # could check in more detail + name_after = 'New Name (tm)' type_after = 'plenary' time_after = (time_utc + datetime.timedelta(days=1, hours=2)).astimezone(meeting.tz()) duration_after = duration_before * 2 show_location_after = False location_after = meeting.room_set.last() - r = self.client.post( - self.edit_timeslot_url(ts), - data=dict( - name=name_after, - type=type_after, - time_0=time_after.strftime('%Y-%m-%d'), # date for SplitDateTimeField - time_1=time_after.strftime('%H:%M'), # time for SplitDateTimeField - duration=str(duration_after), - # show_location=show_location_after, # False values are omitted from form - location=location_after.pk, - ) - ) + post_data = dict( + name=name_after, + type=type_after, + time_0=time_after.strftime('%Y-%m-%d'), # date for SplitDateTimeField + time_1=time_after.strftime('%H:%M'), # time for SplitDateTimeField + duration=str(duration_after), + # show_location=show_location_after, # False values are omitted from form + location=location_after.pk, + ) + r = self.client.post(url, data=post_data) self.assertEqual(r.status_code, 302) # expect redirect to timeslot edit url self.assertEqual(r['Location'], self.edit_timeslots_url(meeting), 'Expected to be redirected to meeting timeslots edit page') @@ -2158,9 +2632,15 @@ def test_edit_timeslot(self): self.assertEqual(ts.show_location, show_location_after) self.assertEqual(ts.location, location_after) + # and check with sched param set + r = self.client.post(url + '?sched=1234', data=post_data) + self.assertEqual(r.status_code, 302) # expect redirect to timeslot edit url + self.assertEqual(r['Location'], self.edit_timeslots_url(meeting) + '?sched=1234', + 'Expected to be redirected to meeting timeslots edit page with sched param set') + def test_invalid_edit_timeslot(self): meeting = self.create_bare_meeting() - ts: TimeSlot = TimeSlotFactory(meeting=meeting, name='slot') # n.b., colon indicates type hinting + ts: TimeSlot = TimeSlotFactory(meeting=meeting, name='slot') # type: ignore[annotation-unchecked] self.login() r = self.client.post( self.edit_timeslot_url(ts), @@ -2280,6 +2760,7 @@ def test_create_single_timeslot(self): meeting = self.create_meeting() timeslots_before = set(ts.pk for ts in meeting.timeslot_set.all()) + url = self.create_timeslots_url(meeting) post_data = dict( name='some name', type='regular', @@ -2290,10 +2771,14 @@ def test_create_single_timeslot(self): locations=str(meeting.room_set.first().pk), ) self.login() - r = self.client.post( - self.create_timeslots_url(meeting), - data=post_data, - ) + + # check that sched parameter is preserved + r = self.client.get(url) + self.assertNotContains(r, '?sched=', status_code=200) + r = self.client.get(url + '?sched=1234') + self.assertContains(r, '?sched=1234', status_code=200) # could check in more detail + + r = self.client.post(url, data=post_data) self.assertEqual(r.status_code, 302) self.assertEqual(r['Location'], self.edit_timeslots_url(meeting), 'Expected to be redirected to meeting timeslots edit page') @@ -2308,6 +2793,12 @@ def test_create_single_timeslot(self): self.assertEqual(ts.show_location, post_data['show_location']) self.assertEqual(str(ts.location.pk), post_data['locations']) + # check again with sched parameter + r = self.client.post(url + '?sched=1234', data=post_data) + self.assertEqual(r.status_code, 302) + self.assertEqual(r['Location'], self.edit_timeslots_url(meeting) + '?sched=1234', + 'Expected to be redirected to meeting timeslots edit page with sched parameter set') + def test_create_single_timeslot_outside_meeting_days(self): """Creating a single timeslot outside the official meeting days should work""" meeting = self.create_meeting() @@ -2591,6 +3082,17 @@ def test_create_bulk_timeslots(self): day_locs.discard((ts.time.date(), ts.location)) self.assertEqual(day_locs, set(), 'Not all day/location combinations created') + def test_sched_param_preserved(self): + meeting = MeetingFactory(type_id='ietf') + url = urlreverse('ietf.meeting.views.edit_timeslots', kwargs={'num': meeting.number}) + self.client.login(username='secretary', password='secretary+password') + r = self.client.get(url) + self.assertNotContains(r, '?sched=', status_code=200) + self.assertNotContains(r, "Back to agenda") + r = self.client.get(url + '?sched=1234') + self.assertContains(r, '?sched=1234', status_code=200) # could check in more detail + self.assertContains(r, "Back to agenda") + def test_ajax_delete_timeslot(self): """AJAX call to delete timeslot should work""" meeting = self.create_bare_meeting() @@ -2721,7 +3223,9 @@ def test_ajax_delete_timeslots_invalid(self): class ReorderSlidesTests(TestCase): - def test_add_slides_to_session(self): + @override_settings(MEETECHO_API_CONFIG="fake settings") # enough to trigger API calls + @patch("ietf.meeting.views.SlidesManager") + def test_add_slides_to_session(self, mock_slides_manager_cls): for type_id in ('ietf','interim'): chair_role = RoleFactory(name_id='chair') session = SessionFactory(group=chair_role.group, meeting__date=date_today() - datetime.timedelta(days=90), meeting__type_id=type_id) @@ -2732,6 +3236,7 @@ def test_add_slides_to_session(self): r = self.client.post(url, {'order':1, 'name':slides.name }) self.assertEqual(r.status_code, 403) self.assertIn('have permission', unicontent(r)) + self.assertFalse(mock_slides_manager_cls.called) self.client.login(username=chair_role.person.user.username, password=chair_role.person.user.username+"+password") @@ -2739,6 +3244,7 @@ def test_add_slides_to_session(self): r = self.client.post(url, {'order':0, 'name':slides.name }) self.assertEqual(r.status_code, 403) self.assertIn('materials cutoff', unicontent(r)) + self.assertFalse(mock_slides_manager_cls.called) session.meeting.date = date_today() session.meeting.save() @@ -2748,54 +3254,67 @@ def test_add_slides_to_session(self): self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('No data',r.json()['error']) + self.assertFalse(mock_slides_manager_cls.called) r = self.client.post(url, {'garbage':'garbage'}) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('order is not valid',r.json()['error']) + self.assertFalse(mock_slides_manager_cls.called) r = self.client.post(url, {'order':0, 'name':slides.name }) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('order is not valid',r.json()['error']) + self.assertFalse(mock_slides_manager_cls.called) r = self.client.post(url, {'order':2, 'name':slides.name }) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('order is not valid',r.json()['error']) + self.assertFalse(mock_slides_manager_cls.called) r = self.client.post(url, {'order':'garbage', 'name':slides.name }) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('order is not valid',r.json()['error']) + self.assertFalse(mock_slides_manager_cls.called) # Invalid name r = self.client.post(url, {'order':1 }) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('name is not valid',r.json()['error']) + self.assertFalse(mock_slides_manager_cls.called) r = self.client.post(url, {'order':1, 'name':'garbage' }) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('name is not valid',r.json()['error']) + self.assertFalse(mock_slides_manager_cls.called) # Valid post r = self.client.post(url, {'order':1, 'name':slides.name }) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],True) - self.assertEqual(session.sessionpresentation_set.count(),1) + self.assertEqual(session.presentations.count(),1) + self.assertTrue(mock_slides_manager_cls.called) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertTrue(mock_slides_manager_cls.return_value.add.called) + self.assertEqual(mock_slides_manager_cls.return_value.add.call_args, call(session=session, slides=slides, order=1)) + mock_slides_manager_cls.reset_mock() # Ignore a request to add slides that are already in a session r = self.client.post(url, {'order':1, 'name':slides.name }) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],True) - self.assertEqual(session.sessionpresentation_set.count(),1) + self.assertEqual(session.presentations.count(),1) + self.assertFalse(mock_slides_manager_cls.called) session2 = SessionFactory(group=session.group, meeting=session.meeting) SessionPresentationFactory.create_batch(3, document__type_id='slides', session=session2) - for num, sp in enumerate(session2.sessionpresentation_set.filter(document__type_id='slides'),start=1): + for num, sp in enumerate(session2.presentations.filter(document__type_id='slides'),start=1): sp.order = num sp.save() @@ -2807,24 +3326,41 @@ def test_add_slides_to_session(self): r = self.client.post(url, {'order':1, 'name':more_slides[0].name}) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],True) - self.assertEqual(session2.sessionpresentation_set.get(document=more_slides[0]).order,1) - self.assertEqual(list(session2.sessionpresentation_set.order_by('order').values_list('order',flat=True)), list(range(1,5))) + self.assertEqual(session2.presentations.get(document=more_slides[0]).order,1) + self.assertEqual(list(session2.presentations.order_by('order').values_list('order',flat=True)), list(range(1,5))) + self.assertTrue(mock_slides_manager_cls.called) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertTrue(mock_slides_manager_cls.return_value.add.called) + self.assertEqual(mock_slides_manager_cls.return_value.add.call_args, call(session=session2, slides=more_slides[0], order=1)) + mock_slides_manager_cls.reset_mock() # Insert at end r = self.client.post(url, {'order':5, 'name':more_slides[1].name}) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],True) - self.assertEqual(session2.sessionpresentation_set.get(document=more_slides[1]).order,5) - self.assertEqual(list(session2.sessionpresentation_set.order_by('order').values_list('order',flat=True)), list(range(1,6))) + self.assertEqual(session2.presentations.get(document=more_slides[1]).order,5) + self.assertEqual(list(session2.presentations.order_by('order').values_list('order',flat=True)), list(range(1,6))) + self.assertTrue(mock_slides_manager_cls.called) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertTrue(mock_slides_manager_cls.return_value.add.called) + self.assertEqual(mock_slides_manager_cls.return_value.add.call_args, call(session=session2, slides=more_slides[1], order=5)) + mock_slides_manager_cls.reset_mock() # Insert in middle r = self.client.post(url, {'order':3, 'name':more_slides[2].name}) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],True) - self.assertEqual(session2.sessionpresentation_set.get(document=more_slides[2]).order,3) - self.assertEqual(list(session2.sessionpresentation_set.order_by('order').values_list('order',flat=True)), list(range(1,7))) - - def test_remove_slides_from_session(self): + self.assertEqual(session2.presentations.get(document=more_slides[2]).order,3) + self.assertEqual(list(session2.presentations.order_by('order').values_list('order',flat=True)), list(range(1,7))) + self.assertTrue(mock_slides_manager_cls.called) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertTrue(mock_slides_manager_cls.return_value.add.called) + self.assertEqual(mock_slides_manager_cls.return_value.add.call_args, call(session=session2, slides=more_slides[2], order=3)) + mock_slides_manager_cls.reset_mock() + + @override_settings(MEETECHO_API_CONFIG="fake settings") # enough to trigger API calls + @patch("ietf.meeting.views.SlidesManager") + def test_remove_slides_from_session(self, mock_slides_manager_cls): for type_id in ['ietf','interim']: chair_role = RoleFactory(name_id='chair') session = SessionFactory(group=chair_role.group, meeting__date=date_today()-datetime.timedelta(days=90), meeting__type_id=type_id) @@ -2835,6 +3371,7 @@ def test_remove_slides_from_session(self): r = self.client.post(url, {'oldIndex':1, 'name':slides.name }) self.assertEqual(r.status_code, 403) self.assertIn('have permission', unicontent(r)) + self.assertFalse(mock_slides_manager_cls.called) self.client.login(username=chair_role.person.user.username, password=chair_role.person.user.username+"+password") @@ -2842,6 +3379,7 @@ def test_remove_slides_from_session(self): r = self.client.post(url, {'oldIndex':0, 'name':slides.name }) self.assertEqual(r.status_code, 403) self.assertIn('materials cutoff', unicontent(r)) + self.assertFalse(mock_slides_manager_cls.called) session.meeting.date = date_today() session.meeting.save() @@ -2851,40 +3389,47 @@ def test_remove_slides_from_session(self): self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('No data',r.json()['error']) + self.assertFalse(mock_slides_manager_cls.called) r = self.client.post(url, {'garbage':'garbage'}) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('index is not valid',r.json()['error']) + self.assertFalse(mock_slides_manager_cls.called) r = self.client.post(url, {'oldIndex':0, 'name':slides.name }) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('index is not valid',r.json()['error']) + self.assertFalse(mock_slides_manager_cls.called) r = self.client.post(url, {'oldIndex':'garbage', 'name':slides.name }) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('index is not valid',r.json()['error']) - + self.assertFalse(mock_slides_manager_cls.called) + # No matching thing to delete r = self.client.post(url, {'oldIndex':1, 'name':slides.name }) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('index is not valid',r.json()['error']) + self.assertFalse(mock_slides_manager_cls.called) - session.sessionpresentation_set.create(document=slides, rev=slides.rev, order=1) + session.presentations.create(document=slides, rev=slides.rev, order=1) # Bad names r = self.client.post(url, {'oldIndex':1}) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('name is not valid',r.json()['error']) + self.assertFalse(mock_slides_manager_cls.called) r = self.client.post(url, {'oldIndex':1, 'name':'garbage' }) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('name is not valid',r.json()['error']) + self.assertFalse(mock_slides_manager_cls.called) slides2 = DocumentFactory(type_id='slides') @@ -2893,22 +3438,29 @@ def test_remove_slides_from_session(self): self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('SessionPresentation not found',r.json()['error']) + self.assertFalse(mock_slides_manager_cls.called) - session.sessionpresentation_set.create(document=slides2, rev=slides2.rev, order=2) + session.presentations.create(document=slides2, rev=slides2.rev, order=2) r = self.client.post(url, {'oldIndex':1, 'name':slides2.name }) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('Name does not match index',r.json()['error']) + self.assertFalse(mock_slides_manager_cls.called) # valid removal r = self.client.post(url, {'oldIndex':1, 'name':slides.name }) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],True) - self.assertEqual(session.sessionpresentation_set.count(),1) + self.assertEqual(session.presentations.count(),1) + self.assertTrue(mock_slides_manager_cls.called) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertTrue(mock_slides_manager_cls.return_value.delete.called) + self.assertEqual(mock_slides_manager_cls.return_value.delete.call_args, call(session=session, slides=slides)) + mock_slides_manager_cls.reset_mock() session2 = SessionFactory(group=session.group, meeting=session.meeting) sp_list = SessionPresentationFactory.create_batch(5, document__type_id='slides', session=session2) - for num, sp in enumerate(session2.sessionpresentation_set.filter(document__type_id='slides'),start=1): + for num, sp in enumerate(session2.presentations.filter(document__type_id='slides'),start=1): sp.order = num sp.save() @@ -2918,28 +3470,47 @@ def test_remove_slides_from_session(self): r = self.client.post(url, {'oldIndex':1, 'name':sp_list[0].document.name }) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],True) - self.assertFalse(session2.sessionpresentation_set.filter(pk=sp_list[0].pk).exists()) - self.assertEqual(list(session2.sessionpresentation_set.order_by('order').values_list('order',flat=True)), list(range(1,5))) + self.assertFalse(session2.presentations.filter(pk=sp_list[0].pk).exists()) + self.assertEqual(list(session2.presentations.order_by('order').values_list('order',flat=True)), list(range(1,5))) + self.assertTrue(mock_slides_manager_cls.called) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertTrue(mock_slides_manager_cls.return_value.delete.called) + self.assertEqual(mock_slides_manager_cls.return_value.delete.call_args, call(session=session2, slides=sp_list[0].document)) + mock_slides_manager_cls.reset_mock() # delete in middle of list r = self.client.post(url, {'oldIndex':4, 'name':sp_list[4].document.name }) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],True) - self.assertFalse(session2.sessionpresentation_set.filter(pk=sp_list[4].pk).exists()) - self.assertEqual(list(session2.sessionpresentation_set.order_by('order').values_list('order',flat=True)), list(range(1,4))) + self.assertFalse(session2.presentations.filter(pk=sp_list[4].pk).exists()) + self.assertEqual(list(session2.presentations.order_by('order').values_list('order',flat=True)), list(range(1,4))) + self.assertTrue(mock_slides_manager_cls.called) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertTrue(mock_slides_manager_cls.return_value.delete.called) + self.assertEqual(mock_slides_manager_cls.return_value.delete.call_args, call(session=session2, slides=sp_list[4].document)) + mock_slides_manager_cls.reset_mock() # delete at end of list r = self.client.post(url, {'oldIndex':2, 'name':sp_list[2].document.name }) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],True) - self.assertFalse(session2.sessionpresentation_set.filter(pk=sp_list[2].pk).exists()) - self.assertEqual(list(session2.sessionpresentation_set.order_by('order').values_list('order',flat=True)), list(range(1,3))) - - - def test_reorder_slides_in_session(self): + self.assertFalse(session2.presentations.filter(pk=sp_list[2].pk).exists()) + self.assertEqual(list(session2.presentations.order_by('order').values_list('order',flat=True)), list(range(1,3))) + self.assertTrue(mock_slides_manager_cls.called) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertTrue(mock_slides_manager_cls.return_value.delete.called) + self.assertEqual(mock_slides_manager_cls.return_value.delete.call_args, call(session=session2, slides=sp_list[2].document)) + mock_slides_manager_cls.reset_mock() + + @override_settings(MEETECHO_API_CONFIG="fake settings") # enough to trigger API calls + @patch("ietf.meeting.views.SlidesManager") + def test_reorder_slides_in_session(self, mock_slides_manager_cls): + def _sppk_at(sppk, positions): + return [sppk[p-1] for p in positions] chair_role = RoleFactory(name_id='chair') session = SessionFactory(group=chair_role.group, meeting__date=date_today() - datetime.timedelta(days=90)) sp_list = SessionPresentationFactory.create_batch(5, document__type_id='slides', session=session) + sppk = [o.pk for o in sp_list] for num, sp in enumerate(sp_list, start=1): sp.order = num sp.save() @@ -2955,6 +3526,7 @@ def test_reorder_slides_in_session(self): r = self.client.post(url, {'oldIndex':1, 'newIndex':2 }) self.assertEqual(r.status_code, 403) self.assertIn('have permission', unicontent(r)) + self.assertFalse(mock_slides_manager_cls.called) self.client.login(username=chair_role.person.user.username, password=chair_role.person.user.username+"+password") @@ -2962,6 +3534,7 @@ def test_reorder_slides_in_session(self): r = self.client.post(url, {'oldIndex':1, 'newIndex':2 }) self.assertEqual(r.status_code, 403) self.assertIn('materials cutoff', unicontent(r)) + self.assertFalse(mock_slides_manager_cls.called) session.meeting.date = date_today() session.meeting.save() @@ -2971,60 +3544,98 @@ def test_reorder_slides_in_session(self): self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('index is not valid',r.json()['error']) + self.assertFalse(mock_slides_manager_cls.called) r = self.client.post(url, {'oldIndex':2, 'newIndex':6 }) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('index is not valid',r.json()['error']) + self.assertFalse(mock_slides_manager_cls.called) r = self.client.post(url, {'oldIndex':2, 'newIndex':2 }) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],False) self.assertIn('index is not valid',r.json()['error']) + self.assertFalse(mock_slides_manager_cls.called) # Move from beginning r = self.client.post(url, {'oldIndex':1, 'newIndex':3}) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],True) - self.assertEqual(list(session.sessionpresentation_set.order_by('order').values_list('pk',flat=True)),[2,3,1,4,5]) + self.assertEqual(list(session.presentations.order_by('order').values_list('pk',flat=True)),_sppk_at(sppk,[2,3,1,4,5])) + self.assertTrue(mock_slides_manager_cls.called) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertTrue(mock_slides_manager_cls.return_value.send_update.called) + self.assertEqual(mock_slides_manager_cls.return_value.send_update.call_args, call(session)) + mock_slides_manager_cls.reset_mock() # Move to beginning r = self.client.post(url, {'oldIndex':3, 'newIndex':1}) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],True) - self.assertEqual(list(session.sessionpresentation_set.order_by('order').values_list('pk',flat=True)),[1,2,3,4,5]) - + self.assertEqual(list(session.presentations.order_by('order').values_list('pk',flat=True)),_sppk_at(sppk,[1,2,3,4,5])) + self.assertTrue(mock_slides_manager_cls.called) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertTrue(mock_slides_manager_cls.return_value.send_update.called) + self.assertEqual(mock_slides_manager_cls.return_value.send_update.call_args, call(session)) + mock_slides_manager_cls.reset_mock() + # Move from end r = self.client.post(url, {'oldIndex':5, 'newIndex':3}) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],True) - self.assertEqual(list(session.sessionpresentation_set.order_by('order').values_list('pk',flat=True)),[1,2,5,3,4]) + self.assertEqual(list(session.presentations.order_by('order').values_list('pk',flat=True)),_sppk_at(sppk,[1,2,5,3,4])) + self.assertTrue(mock_slides_manager_cls.called) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertTrue(mock_slides_manager_cls.return_value.send_update.called) + self.assertEqual(mock_slides_manager_cls.return_value.send_update.call_args, call(session)) + mock_slides_manager_cls.reset_mock() # Move to end r = self.client.post(url, {'oldIndex':3, 'newIndex':5}) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],True) - self.assertEqual(list(session.sessionpresentation_set.order_by('order').values_list('pk',flat=True)),[1,2,3,4,5]) + self.assertEqual(list(session.presentations.order_by('order').values_list('pk',flat=True)),_sppk_at(sppk,[1,2,3,4,5])) + self.assertTrue(mock_slides_manager_cls.called) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertTrue(mock_slides_manager_cls.return_value.send_update.called) + self.assertEqual(mock_slides_manager_cls.return_value.send_update.call_args, call(session)) + mock_slides_manager_cls.reset_mock() # Move beginning to end r = self.client.post(url, {'oldIndex':1, 'newIndex':5}) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],True) - self.assertEqual(list(session.sessionpresentation_set.order_by('order').values_list('pk',flat=True)),[2,3,4,5,1]) + self.assertEqual(list(session.presentations.order_by('order').values_list('pk',flat=True)),_sppk_at(sppk,[2,3,4,5,1])) + self.assertTrue(mock_slides_manager_cls.called) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertTrue(mock_slides_manager_cls.return_value.send_update.called) + self.assertEqual(mock_slides_manager_cls.return_value.send_update.call_args, call(session)) + mock_slides_manager_cls.reset_mock() # Move middle to middle r = self.client.post(url, {'oldIndex':3, 'newIndex':4}) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],True) - self.assertEqual(list(session.sessionpresentation_set.order_by('order').values_list('pk',flat=True)),[2,3,5,4,1]) + self.assertEqual(list(session.presentations.order_by('order').values_list('pk',flat=True)),_sppk_at(sppk,[2,3,5,4,1])) + self.assertTrue(mock_slides_manager_cls.called) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertTrue(mock_slides_manager_cls.return_value.send_update.called) + self.assertEqual(mock_slides_manager_cls.return_value.send_update.call_args, call(session)) + mock_slides_manager_cls.reset_mock() r = self.client.post(url, {'oldIndex':3, 'newIndex':2}) self.assertEqual(r.status_code, 200) self.assertEqual(r.json()['success'],True) - self.assertEqual(list(session.sessionpresentation_set.order_by('order').values_list('pk',flat=True)),[2,5,3,4,1]) + self.assertEqual(list(session.presentations.order_by('order').values_list('pk',flat=True)),_sppk_at(sppk,[2,5,3,4,1])) + self.assertTrue(mock_slides_manager_cls.called) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertTrue(mock_slides_manager_cls.return_value.send_update.called) + self.assertEqual(mock_slides_manager_cls.return_value.send_update.call_args, call(session)) + mock_slides_manager_cls.reset_mock() # Reset for next iteration in the loop - session.sessionpresentation_set.update(order=F('pk')) + session.presentations.update(order=F('pk')) self.client.logout() @@ -3041,7 +3652,7 @@ def test_slide_order_reconditioning(self): except AssertionError: pass - self.assertEqual(list(session.sessionpresentation_set.order_by('order').values_list('order',flat=True)),list(range(1,6))) + self.assertEqual(list(session.presentations.order_by('order').values_list('order',flat=True)),list(range(1,6))) class EditTests(TestCase): @@ -3161,10 +3772,11 @@ def test_edit_meeting_schedule(self): e = q("#session{}".format(s.pk)) # should be link to edit/cancel session + edit_session_url = urlreverse( + 'ietf.meeting.views.edit_session', kwargs={'session_id': s.pk} + ) + f'?sched={meeting.schedule.pk}' self.assertTrue( - e.find('a[href="{}"]'.format( - urlreverse('ietf.meeting.views.edit_session', kwargs={'session_id': s.pk}), - )) + e.find(f'a[href="{edit_session_url}"]') ) self.assertTrue( e.find('a[href="{}?sched={}"]'.format( @@ -3706,7 +4318,7 @@ def test_new_meeting_schedule_rejects_invalid_names(self): 'base': meeting.schedule.base_id, }) self.assertEqual(r.status_code, 200) - self.assertFormError(r, 'form', 'name', 'Enter a valid value.') + self.assertFormError(r.context["form"], 'name', 'Enter a valid value.') self.assertEqual(meeting.schedule_set.count(), orig_schedule_count, 'Schedule should not be created') r = self.client.post(url, { @@ -3716,7 +4328,7 @@ def test_new_meeting_schedule_rejects_invalid_names(self): 'base': meeting.schedule.base_id, }) self.assertEqual(r.status_code, 200) - self.assertFormError(r, 'form', 'name', 'Enter a valid value.') + self.assertFormError(r.context["form"], 'name', 'Enter a valid value.') self.assertEqual(meeting.schedule_set.count(), orig_schedule_count, 'Schedule should not be created') # Non-ASCII alphanumeric characters @@ -3727,16 +4339,20 @@ def test_new_meeting_schedule_rejects_invalid_names(self): 'base': meeting.schedule.base_id, }) self.assertEqual(r.status_code, 200) - self.assertFormError(r, 'form', 'name', 'Enter a valid value.') + self.assertFormError(r.context["form"], 'name', 'Enter a valid value.') self.assertEqual(meeting.schedule_set.count(), orig_schedule_count, 'Schedule should not be created') def test_edit_session(self): session = SessionFactory(meeting__type_id='ietf', group__type_id='team') # type determines allowed session purposes + edit_meeting_url = urlreverse('ietf.meeting.views.edit_meeting_schedule', kwargs={'num': session.meeting.number}) self.client.login(username='secretary', password='secretary+password') url = urlreverse('ietf.meeting.views.edit_session', kwargs={'session_id': session.pk}) r = self.client.get(url) self.assertContains(r, 'Edit session', status_code=200) - r = self.client.post(url, { + pq = PyQuery(r.content) + back_button = pq(f'a[href="{edit_meeting_url}"]') + self.assertEqual(len(back_button), 1) + post_data = { 'name': 'this is a name', 'short': 'tian', 'purpose': 'coding', @@ -3746,10 +4362,11 @@ def test_edit_session(self): 'remote_instructions': 'Do this do that', 'attendees': '103', 'comments': 'So much to say', - }) + 'chat_room': 'xyzzy', + } + r = self.client.post(url, post_data) self.assertNoFormPostErrors(r) - self.assertRedirects(r, urlreverse('ietf.meeting.views.edit_meeting_schedule', - kwargs={'num': session.meeting.number})) + self.assertRedirects(r, edit_meeting_url) session = Session.objects.get(pk=session.pk) # refresh objects from DB self.assertEqual(session.name, 'this is a name') self.assertEqual(session.short, 'tian') @@ -3760,6 +4377,24 @@ def test_edit_session(self): self.assertEqual(session.remote_instructions, 'Do this do that') self.assertEqual(session.attendees, 103) self.assertEqual(session.comments, 'So much to say') + self.assertEqual(session.chat_room, 'xyzzy') + + # Verify return to correct schedule when sched query parameter is present + other_schedule = ScheduleFactory(meeting=session.meeting) + r = self.client.get(url + f'?sched={other_schedule.pk}') + edit_meeting_url = urlreverse( + 'ietf.meeting.views.edit_meeting_schedule', + kwargs={ + 'num': session.meeting.number, + 'owner': other_schedule.owner.email(), + 'name': other_schedule.name, + }, + ) + pq = PyQuery(r.content) + back_button = pq(f'a[href="{edit_meeting_url}"]') + self.assertEqual(len(back_button), 1) + r = self.client.post(url + f'?sched={other_schedule.pk}', post_data) + self.assertRedirects(r, edit_meeting_url) def test_cancel_session(self): # session for testing with official schedule @@ -3791,9 +4426,9 @@ def test_cancel_session(self): self.assertIn(return_url_unofficial, r.content.decode()) r = self.client.post(url, {}) - self.assertFormError(r, 'form', 'confirmed', 'This field is required.') + self.assertFormError(r.context["form"], 'confirmed', 'This field is required.') r = self.client.post(url_unofficial, {}) - self.assertFormError(r, 'form', 'confirmed', 'This field is required.') + self.assertFormError(r.context["form"], 'confirmed', 'This field is required.') r = self.client.post(url, {'confirmed': 'on'}) self.assertRedirects(r, return_url) @@ -3910,6 +4545,7 @@ def test_persistent_enabled_timeslot_types(self): class SessionDetailsTests(TestCase): + settings_temp_path_overrides = TestCase.settings_temp_path_overrides + ['SLIDE_STAGING_PATH'] def test_session_details(self): @@ -4006,7 +4642,7 @@ def test_add_session_drafts(self): group.role_set.create(name_id='chair',person = group_chair, email = group_chair.email()) session = SessionFactory.create(meeting__type_id='ietf',group=group, meeting__date=date_today() + datetime.timedelta(days=90)) SessionPresentationFactory.create(session=session,document__type_id='draft',rev=None) - old_draft = session.sessionpresentation_set.filter(document__type='draft').first().document + old_draft = session.presentations.filter(document__type='draft').first().document new_draft = DocumentFactory(type_id='draft') url = urlreverse('ietf.meeting.views.add_session_drafts', kwargs=dict(num=session.meeting.number, session_id=session.pk)) @@ -4027,10 +4663,10 @@ def test_add_session_drafts(self): q = PyQuery(r.content) self.assertIn("Already linked:", q('form .text-danger').text()) - self.assertEqual(1,session.sessionpresentation_set.count()) + self.assertEqual(1,session.presentations.count()) r = self.client.post(url,dict(drafts=[new_draft.pk,])) self.assertTrue(r.status_code, 302) - self.assertEqual(2,session.sessionpresentation_set.count()) + self.assertEqual(2,session.presentations.count()) session.meeting.date -= datetime.timedelta(days=180) session.meeting.save() @@ -4042,6 +4678,85 @@ def test_add_session_drafts(self): q = PyQuery(r.content) self.assertEqual(1,len(q(".alert-warning:contains('may affect published proceedings')"))) + def test_proposed_slides_for_approval(self): + # This test overlaps somewhat with MaterialsTests of proposed slides handling. The focus + # here is on the display of slides, not the approval action. + group = GroupFactory() + meeting = MeetingFactory( + type_id="ietf", date=date_today() + datetime.timedelta(days=10) + ) + sessions = SessionFactory.create_batch( + 2, + group=group, + meeting=meeting, + ) + + # slides submission _not_ in the `pending` state + do_not_show = [ + SlideSubmissionFactory( + session=sessions[0], + title="already approved", + status_id="approved", + ), + SlideSubmissionFactory( + session=sessions[1], + title="already rejected", + status_id="rejected", + ), + ] + + # pending submissions + first_session_pending = SlideSubmissionFactory( + session=sessions[0], title="first session title" + ) + second_session_pending = SlideSubmissionFactory( + session=sessions[1], title="second session title" + ) + + # and their approval URLs + def _approval_url(slidesub): + return urlreverse( + "ietf.meeting.views.approve_proposed_slides", + kwargs={"slidesubmission_id": slidesub.pk, "num": meeting.number}, + ) + + first_approval_url = _approval_url(first_session_pending) + second_approval_url = _approval_url(second_session_pending) + do_not_show_urls = [_approval_url(ss) for ss in do_not_show] + + # Retrieve the URL as a group chair + url = urlreverse( + "ietf.meeting.views.session_details", + kwargs={ + "num": meeting.number, + "acronym": group.acronym, + }, + ) + chair = RoleFactory(group=group, name_id="chair").person + self.client.login( + username=chair.user.username, password=f"{chair.user.username}+password" + ) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + pq = PyQuery(r.content) + self.assertEqual( + len(pq(f'a[href="{first_approval_url}"]')), + 1, + "first session proposed slides should be linked for approval", + ) + self.assertEqual( + len(pq(f'a[href="{second_approval_url}"]')), + 1, + "second session proposed slides should be linked for approval", + ) + for no_show_url in do_not_show_urls: + self.assertEqual( + len(pq(f'a[href="{no_show_url}"]')), + 0, + "second session proposed slides should be linked for approval", + ) + + class EditScheduleListTests(TestCase): def setUp(self): super().setUp() @@ -4055,73 +4770,151 @@ def test_list_schedules(self): self.assertTrue(r.status_code, 200) def test_diff_schedules(self): - meeting = make_meeting_test_data() - - url = urlreverse('ietf.meeting.views.diff_schedules',kwargs={'num':meeting.number}) - login_testing_unauthorized(self,"secretary", url) - r = self.client.get(url) - self.assertTrue(r.status_code, 200) - - from_schedule = Schedule.objects.get(meeting=meeting, name="test-unofficial-schedule") + # Create meeting and some time slots + meeting = MeetingFactory(type_id="ietf", populate_schedule=False) + rooms = RoomFactory.create_batch(2, meeting=meeting) + # first index is room, second is time + timeslots = [ + [ + TimeSlotFactory( + location=room, + meeting=meeting, + time=datetime.datetime.combine( + meeting.date, datetime.time(9, 0, tzinfo=datetime.UTC) + ) + ), + TimeSlotFactory( + location=room, + meeting=meeting, + time=datetime.datetime.combine( + meeting.date, datetime.time(10, 0, tzinfo=datetime.UTC) + ) + ), + TimeSlotFactory( + location=room, + meeting=meeting, + time=datetime.datetime.combine( + meeting.date, datetime.time(11, 0, tzinfo=datetime.UTC) + ) + ), + ] + for room in rooms + ] + sessions = SessionFactory.create_batch( + 5, meeting=meeting, add_to_schedule=False + ) - session1 = Session.objects.filter(meeting=meeting, group__acronym='mars').first() - session2 = Session.objects.filter(meeting=meeting, group__acronym='ames').first() - session3 = SessionFactory(meeting=meeting, group=Group.objects.get(acronym='mars'), - attendees=10, requested_duration=datetime.timedelta(minutes=70), - add_to_schedule=False) - SchedulingEvent.objects.create(session=session3, status_id='schedw', by=Person.objects.first()) + from_schedule = ScheduleFactory(meeting=meeting) + to_schedule = ScheduleFactory(meeting=meeting) - slot2 = TimeSlot.objects.filter(meeting=meeting, type='regular').order_by('-time').first() - slot3 = TimeSlot.objects.create( - meeting=meeting, type_id='regular', location=slot2.location, - duration=datetime.timedelta(minutes=60), - time=slot2.time + datetime.timedelta(minutes=60), + # sessions[0]: not scheduled in from_schedule, scheduled in to_schedule + SchedTimeSessAssignment.objects.create( + schedule=to_schedule, + session=sessions[0], + timeslot=timeslots[0][0], + ) + # sessions[1]: scheduled in from_schedule, not scheduled in to_schedule + SchedTimeSessAssignment.objects.create( + schedule=from_schedule, + session=sessions[1], + timeslot=timeslots[0][0], + ) + # sessions[2]: moves rooms, not time + SchedTimeSessAssignment.objects.create( + schedule=from_schedule, + session=sessions[2], + timeslot=timeslots[0][1], + ) + SchedTimeSessAssignment.objects.create( + schedule=to_schedule, + session=sessions[2], + timeslot=timeslots[1][1], + ) + # sessions[3]: moves time, not room + SchedTimeSessAssignment.objects.create( + schedule=from_schedule, + session=sessions[3], + timeslot=timeslots[1][1], + ) + SchedTimeSessAssignment.objects.create( + schedule=to_schedule, + session=sessions[3], + timeslot=timeslots[1][2], + ) + # sessions[4]: moves room and time + SchedTimeSessAssignment.objects.create( + schedule=from_schedule, + session=sessions[4], + timeslot=timeslots[1][0], + ) + SchedTimeSessAssignment.objects.create( + schedule=to_schedule, + session=sessions[4], + timeslot=timeslots[0][2], ) - # copy - new_url = urlreverse("ietf.meeting.views.new_meeting_schedule", kwargs=dict(num=meeting.number, owner=from_schedule.owner_email(), name=from_schedule.name)) - r = self.client.post(new_url, { - 'name': "newtest", - 'public': "on", - }) - self.assertNoFormPostErrors(r) - - to_schedule = Schedule.objects.get(meeting=meeting, name='newtest') + # Check the raw diffs + raw_diffs = diff_meeting_schedules(from_schedule, to_schedule) + self.assertCountEqual( + raw_diffs, + [ + { + "change": "schedule", + "session": sessions[0].pk, + "to": timeslots[0][0].pk, + }, + { + "change": "unschedule", + "session": sessions[1].pk, + "from": timeslots[0][0].pk, + }, + { + "change": "move", + "session": sessions[2].pk, + "from": timeslots[0][1].pk, + "to": timeslots[1][1].pk, + }, + { + "change": "move", + "session": sessions[3].pk, + "from": timeslots[1][1].pk, + "to": timeslots[1][2].pk, + }, + { + "change": "move", + "session": sessions[4].pk, + "from": timeslots[1][0].pk, + "to": timeslots[0][2].pk, + }, + ] + ) - # make some changes + # Check the view + url = urlreverse("ietf.meeting.views.diff_schedules", + kwargs={"num": meeting.number}) + login_testing_unauthorized(self, "secretary", url) + r = self.client.get(url) + self.assertTrue(r.status_code, 200) - edit_url = urlreverse("ietf.meeting.views.edit_meeting_schedule", kwargs=dict(num=meeting.number, owner=to_schedule.owner_email(), name=to_schedule.name)) - - # schedule session - r = self.client.post(edit_url, { - 'action': 'assign', - 'timeslot': slot3.pk, - 'session': session3.pk, - }) - self.assertEqual(json.loads(r.content)['success'], True) - # unschedule session - r = self.client.post(edit_url, { - 'action': 'unassign', - 'session': session1.pk, - }) - self.assertEqual(json.loads(r.content)['success'], True) - # move session - r = self.client.post(edit_url, { - 'action': 'assign', - 'timeslot': slot2.pk, - 'session': session2.pk, + # with show room changes disabled - does not show sessions[2] because it did + # not change time + r = self.client.get(url, { + "from_schedule": from_schedule.name, + "to_schedule": to_schedule.name, }) - self.assertEqual(json.loads(r.content)['success'], True) + self.assertTrue(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q(".schedule-diffs tr")), 4 + 1) - # now get differences + # with show room changes enabled - shows all changes r = self.client.get(url, { - 'from_schedule': from_schedule.name, - 'to_schedule': to_schedule.name, + "from_schedule": from_schedule.name, + "to_schedule": to_schedule.name, + "show_room_changes": "on", }) self.assertTrue(r.status_code, 200) - q = PyQuery(r.content) - self.assertEqual(len(q(".schedule-diffs tr")), 3+1) + self.assertEqual(len(q(".schedule-diffs tr")), 5 + 1) def test_delete_schedule(self): url = urlreverse('ietf.meeting.views.delete_schedule', @@ -4280,10 +5073,11 @@ def do_interim_send_announcement_test(self, base_session=False, extra_session=Fa if sess: timeslot = sess.official_timeslotassignment().timeslot self.assertIn(timeslot.time.strftime('%Y-%m-%d'), announcement_text) - self.assertIn( - '(%s to %s UTC)' % ( + self.assertRegex( + announcement_text, + r'(%s\s+to\s+%s\s+UTC)' % ( timeslot.utc_start_time().strftime('%H:%M'),timeslot.utc_end_time().strftime('%H:%M') - ), announcement_text) + )) # Count number of sessions listed if base_session and extra_session: expected_session_matches = 3 @@ -4490,20 +5284,55 @@ def test_upcoming_ical(self): meeting = make_meeting_test_data(create_interims=True) populate_important_dates(meeting) url = urlreverse("ietf.meeting.views.upcoming_ical") - + + # Expect events 3 sessions - one for each WG and one for the IETF meeting + expected_event_summaries = [ + 'ames - Asteroid Mining Equipment Standardization Group', + 'mars - Martian Special Interest Group', + 'IETF 72', + ] + + Session.objects.filter( + meeting__type_id='interim', + group__acronym="mars", + ).update( + remote_instructions='https://someurl.example.com', + ) r = self.client.get(url) + self.assertEqual(r.status_code, 200) + assert_ical_response_is_valid(self, r, + expected_event_summaries=expected_event_summaries, + expected_event_count=len(expected_event_summaries)) + # Unfold long lines that might have been folded by iCal + content_unfolded = r.content.decode('utf-8').replace('\r\n ', '') + self.assertIn('Remote instructions: https://someurl.example.com', content_unfolded) + Session.objects.filter(meeting__type_id='interim').update(remote_instructions='') + r = self.client.get(url) self.assertEqual(r.status_code, 200) - # Expect events 3 sessions - one for each WG and one for the IETF meeting assert_ical_response_is_valid(self, r, - expected_event_summaries=[ - 'ames - Asteroid Mining Equipment Standardization Group', - 'mars - Martian Special Interest Group', - 'IETF 72', - ], - expected_event_count=3) + expected_event_summaries=expected_event_summaries, + expected_event_count=len(expected_event_summaries)) + content_unfolded = r.content.decode('utf-8').replace('\r\n ', '') + self.assertNotIn('Remote instructions:', content_unfolded) + + updated = meeting.updated() + self.assertIsNotNone(updated) + expected_updated = updated.astimezone(datetime.UTC).strftime("%Y%m%dT%H%M%SZ") + self.assertContains(r, f"DTSTAMP:{expected_updated}") + + # With default cached_updated, 1970-01-01 + with patch("ietf.meeting.models.Meeting.updated", return_value=None): + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + + self.assertEqual(meeting.type_id, "ietf") - def test_upcoming_ical_filter(self): + expected_updated = "19700101T000000Z" + self.assertEqual(1, r.content.decode("utf-8").count(f"DTSTAMP:{expected_updated}")) + + @patch("ietf.meeting.utils.preprocess_meeting_important_dates") + def test_upcoming_ical_filter(self, mock_preprocess_meeting_important_dates): # Just a quick check of functionality - details tested by test_js.InterimTests make_meeting_test_data(create_interims=True) url = urlreverse("ietf.meeting.views.upcoming_ical") @@ -4525,6 +5354,8 @@ def test_upcoming_ical_filter(self): ], expected_event_count=2) + # Verify preprocess_meeting_important_dates isn't being called + mock_preprocess_meeting_important_dates.assert_not_called() def test_upcoming_json(self): make_meeting_test_data(create_interims=True) @@ -4595,6 +5426,7 @@ def test_interim_request_options(self): def do_interim_request_single_virtual(self, emails_expected): make_meeting_test_data() + TestBlobstoreManager().emptyTestBlobstores() group = Group.objects.get(acronym='mars') date = date_today() + datetime.timedelta(days=30) time = time_now().replace(microsecond=0,second=0) @@ -4645,6 +5477,12 @@ def do_interim_request_single_virtual(self, emails_expected): doc = session.materials.first() path = os.path.join(doc.get_file_path(),doc.filename_with_rev()) self.assertTrue(os.path.exists(path)) + with Path(path).open() as f: + self.assertEqual(f.read(), agenda) + self.assertEqual( + retrieve_str("agenda",doc.uploaded_filename), + agenda + ) # check notices to secretariat and chairs self.assertEqual(len(outbox), length_before + emails_expected) return meeting @@ -4666,6 +5504,7 @@ def test_interim_request_single_virtual_settings_approval_not_required(self): def test_interim_request_single_in_person(self): make_meeting_test_data() + TestBlobstoreManager().emptyTestBlobstores() group = Group.objects.get(acronym='mars') date = date_today() + datetime.timedelta(days=30) time = time_now().replace(microsecond=0,second=0) @@ -4712,6 +5551,10 @@ def test_interim_request_single_in_person(self): timeslot = session.official_timeslotassignment().timeslot self.assertEqual(timeslot.time,dt) self.assertEqual(timeslot.duration,duration) + self.assertEqual( + retrieve_str("agenda",session.agenda().uploaded_filename), + agenda + ) def test_interim_request_multi_day(self): make_meeting_test_data() @@ -4779,6 +5622,11 @@ def test_interim_request_multi_day(self): self.assertEqual(timeslot.time,dt2) self.assertEqual(timeslot.duration,duration) self.assertEqual(session.agenda_note,agenda_note) + for session in meeting.session_set.all(): + self.assertEqual( + retrieve_str("agenda",session.agenda().uploaded_filename), + agenda + ) def test_interim_request_multi_day_non_consecutive(self): make_meeting_test_data() @@ -4841,6 +5689,7 @@ def test_interim_request_multi_day_cancel(self): def test_interim_request_series(self): make_meeting_test_data() + TestBlobstoreManager().emptyTestBlobstores() meeting_count_before = Meeting.objects.filter(type='interim').count() date = date_today() + datetime.timedelta(days=30) if (date.month, date.day) == (12, 31): @@ -4928,6 +5777,11 @@ def test_interim_request_series(self): self.assertEqual(timeslot.time,dt2) self.assertEqual(timeslot.duration,duration) self.assertEqual(session.agenda_note,agenda_note) + for session in meeting.session_set.all(): + self.assertEqual( + retrieve_str("agenda",session.agenda().uploaded_filename), + agenda + ) # test_interim_pending subsumed by test_appears_on_pending @@ -5250,8 +6104,17 @@ def test_interim_request_cancel(self, mock): self.assertEqual(r.status_code, 403) self.assertFalse(mock.called, 'Should not cancel sessions if request rejected') - # test cancelling before announcement + # test with overly-long comments + comments += '0123456789abcdef'*32 self.client.login(username="marschairman", password="marschairman+password") + r = self.client.post(url, {'comments': comments}) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertTrue(q('form .is-invalid')) + # truncate to max_length + comments = comments[:512] + + # test cancelling before announcement length_before = len(outbox) r = self.client.post(url, {'comments': comments}) self.assertRedirects(r, urlreverse('ietf.meeting.views.upcoming')) @@ -5273,6 +6136,7 @@ def test_interim_request_cancel(self, mock): self.assertEqual(session.agenda_note, comments) self.assertEqual(len(outbox), length_before + 1) self.assertIn('Interim Meeting Cancelled', outbox[-1]['Subject']) + self.assertIn(comments, get_payload_text(outbox[-1])) self.assertTrue(mock.called, 'Should cancel sessions if request handled') self.assertCountEqual(mock.call_args[0][1], meeting.session_set.all()) @@ -5456,6 +6320,7 @@ def strfdelta(self, tdelta, fmt): def test_interim_request_edit_agenda_updates_doc(self): """Updating the agenda through the request edit form should update the doc correctly""" make_interim_test_data() + TestBlobstoreManager().emptyTestBlobstores() meeting = add_event_info_to_session_qs(Session.objects.filter(meeting__type='interim', group__acronym='mars')).filter(current_status='sched').first().meeting group = meeting.session_set.first().group url = urlreverse('ietf.meeting.views.interim_request_edit', kwargs={'number': meeting.number}) @@ -5491,6 +6356,10 @@ def test_interim_request_edit_agenda_updates_doc(self): self.assertNotEqual(agenda_doc.uploaded_filename, uploaded_filename_before, 'Uploaded filename should be updated') with (Path(agenda_doc.get_file_path()) / agenda_doc.uploaded_filename).open() as f: self.assertEqual(f.read(), 'modified agenda contents', 'New agenda contents should be saved') + self.assertEqual( + retrieve_str(agenda_doc.type_id, agenda_doc.uploaded_filename), + "modified agenda contents" + ) def test_interim_request_details_permissions(self): make_interim_test_data() @@ -5551,6 +6420,7 @@ def test_group_ical(self): make_interim_test_data() meeting = Meeting.objects.filter(type='interim', session__group__acronym='mars').first() s1 = Session.objects.filter(meeting=meeting, group__acronym="mars").first() + self.assertGreater(len(s1.remote_instructions), 0, 'Expected remote_instructions to be set') a1 = s1.official_timeslotassignment() t1 = a1.timeslot # Create an extra session @@ -5569,6 +6439,7 @@ def test_group_ical(self): self.assertEqual(r.content.count(b'UID'), 2) self.assertContains(r, 'SUMMARY:mars - Martian Special Interest Group') self.assertContains(r, t1.local_start_time().strftime('%Y%m%dT%H%M%S')) + self.assertContains(r, s1.remote_instructions) self.assertContains(r, t2.local_start_time().strftime('%Y%m%dT%H%M%S')) self.assertContains(r, 'END:VEVENT') # @@ -5579,39 +6450,11 @@ def test_group_ical(self): self.assertEqual(r.content.count(b'UID'), 1) self.assertContains(r, 'SUMMARY:mars - Martian Special Interest Group') self.assertContains(r, t1.time.strftime('%Y%m%dT%H%M%S')) + self.assertContains(r, s1.remote_instructions) self.assertNotContains(r, t2.time.strftime('%Y%m%dT%H%M%S')) self.assertContains(r, 'END:VEVENT') -class AjaxTests(TestCase): - def test_ajax_get_utc(self): - # test bad queries - url = urlreverse('ietf.meeting.views.ajax_get_utc') + "?date=2016-1-1&time=badtime&timezone=UTC" - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - data = r.json() - self.assertEqual(data["error"], True) - url = urlreverse('ietf.meeting.views.ajax_get_utc') + "?date=2016-1-1&time=25:99&timezone=UTC" - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - data = r.json() - self.assertEqual(data["error"], True) - url = urlreverse('ietf.meeting.views.ajax_get_utc') + "?date=2016-1-1&time=10:00am&timezone=UTC" - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - data = r.json() - self.assertEqual(data["error"], True) - # test good query - url = urlreverse('ietf.meeting.views.ajax_get_utc') + "?date=2016-1-1&time=12:00&timezone=America/Los_Angeles" - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - data = r.json() - self.assertIn('timezone', data) - self.assertIn('time', data) - self.assertIn('utc', data) - self.assertNotIn('error', data) - self.assertEqual(data['utc'], '20:00') - class IphoneAppJsonTests(TestCase): def test_iphone_app_json_interim(self): make_interim_test_data() @@ -5650,16 +6493,10 @@ def test_iphone_app_json(self): self.assertTrue(msessions.filter(group__acronym=s['group']['acronym']).exists()) class FinalizeProceedingsTests(TestCase): - @override_settings(STATS_REGISTRATION_ATTENDEES_JSON_URL='https://ietf.example.com/{number}') - @requests_mock.Mocker() - def test_finalize_proceedings(self, mock): + def test_finalize_proceedings(self): make_meeting_test_data() meeting = Meeting.objects.filter(type_id='ietf').order_by('id').last() - meeting.session_set.filter(group__acronym='mars').first().sessionpresentation_set.create(document=Document.objects.filter(type='draft').first(),rev=None) - mock.get( - settings.STATS_REGISTRATION_ATTENDEES_JSON_URL.format(number=meeting.number), - text=json.dumps([{"LastName": "Smith", "FirstName": "John", "Company": "ABC", "Country": "US"}]), - ) + meeting.session_set.filter(group__acronym='mars').first().presentations.create(document=Document.objects.filter(type='draft').first(),rev=None) url = urlreverse('ietf.meeting.views.finalize_proceedings',kwargs={'num':meeting.number}) login_testing_unauthorized(self,"secretary",url) @@ -5667,13 +6504,41 @@ def test_finalize_proceedings(self, mock): self.assertEqual(r.status_code, 200) self.assertEqual(meeting.proceedings_final,False) - self.assertEqual(meeting.session_set.filter(group__acronym="mars").first().sessionpresentation_set.filter(document__type="draft").first().rev,None) + self.assertEqual(meeting.session_set.filter(group__acronym="mars").first().presentations.filter(document__type="draft").first().rev,None) r = self.client.post(url,{'finalize':1}) self.assertEqual(r.status_code, 302) meeting = Meeting.objects.get(pk=meeting.pk) self.assertEqual(meeting.proceedings_final,True) - self.assertEqual(meeting.session_set.filter(group__acronym="mars").first().sessionpresentation_set.filter(document__type="draft").first().rev,'00') + self.assertEqual(meeting.session_set.filter(group__acronym="mars").first().presentations.filter(document__type="draft").first().rev,'00') + @patch("ietf.meeting.utils.generate_bluesheet") + def test_bluesheet_generation(self, mock): + meeting = MeetingFactory(type_id="ietf", number="107") # number where generate_bluesheets should not be called + SessionFactory.create_batch(5, meeting=meeting) + url = urlreverse("ietf.meeting.views.finalize_proceedings", kwargs={"num": meeting.number}) + self.client.login(username="secretary", password="secretary+password") + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + self.assertFalse(mock.called) + r = self.client.post(url,{'finalize': 1}) + self.assertEqual(r.status_code, 302) + self.assertFalse(mock.called) + + meeting = MeetingFactory(type_id="ietf", number="108") # number where generate_bluesheets should be called + SessionFactory.create_batch(5, meeting=meeting) + url = urlreverse("ietf.meeting.views.finalize_proceedings", kwargs={"num": meeting.number}) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + self.assertFalse(mock.called) + r = self.client.post(url,{'finalize': 1}) + self.assertEqual(r.status_code, 302) + self.assertTrue(mock.called) + self.assertCountEqual( + [call_args[0][1] for call_args in mock.call_args_list], + [sess for sess in meeting.session_set.all()], + ) + + class MaterialsTests(TestCase): settings_temp_path_overrides = TestCase.settings_temp_path_overrides + [ 'AGENDA_PATH', @@ -5714,13 +6579,15 @@ def test_upload_bluesheets(self): self.assertEqual(r.status_code, 200) q = PyQuery(r.content) self.assertIn('Upload', str(q("title"))) - self.assertFalse(session.sessionpresentation_set.exists()) - test_file = StringIO('%PDF-1.4\n%âãÏÓ\nthis is some text for a test') + self.assertFalse(session.presentations.exists()) + test_content = '%PDF-1.4\n%âãÏÓ\nthis is some text for a test' + test_file = StringIO(test_content) test_file.name = "not_really.pdf" r = self.client.post(url,dict(file=test_file)) self.assertEqual(r.status_code, 302) - bs_doc = session.sessionpresentation_set.filter(document__type_id='bluesheets').first().document + bs_doc = session.presentations.filter(document__type_id='bluesheets').first().document self.assertEqual(bs_doc.rev,'00') + self.assertEqual(retrieve_str("bluesheets", f"{bs_doc.name}-{bs_doc.rev}.pdf"), test_content) r = self.client.get(url) self.assertEqual(r.status_code, 200) q = PyQuery(r.content) @@ -5749,13 +6616,15 @@ def test_upload_bluesheets_interim(self): self.assertEqual(r.status_code, 200) q = PyQuery(r.content) self.assertIn('Upload', str(q("title"))) - self.assertFalse(session.sessionpresentation_set.exists()) - test_file = StringIO('%PDF-1.4\n%âãÏÓ\nthis is some text for a test') + self.assertFalse(session.presentations.exists()) + test_content = '%PDF-1.4\n%âãÏÓ\nthis is some text for a test' + test_file = StringIO(test_content) test_file.name = "not_really.pdf" r = self.client.post(url,dict(file=test_file)) self.assertEqual(r.status_code, 302) - bs_doc = session.sessionpresentation_set.filter(document__type_id='bluesheets').first().document + bs_doc = session.presentations.filter(document__type_id='bluesheets').first().document self.assertEqual(bs_doc.rev,'00') + self.assertEqual(retrieve_str("bluesheets", f"{bs_doc.name}-{bs_doc.rev}.pdf"), test_content) def test_upload_bluesheets_interim_chair_access(self): make_meeting_test_data() @@ -5769,96 +6638,130 @@ def test_upload_bluesheets_interim_chair_access(self): self.assertIn('Upload', str(q("title"))) - def test_upload_minutes_agenda(self): - for doctype in ('minutes','agenda'): - session = SessionFactory(meeting__type_id='ietf') - if doctype == 'minutes': - url = urlreverse('ietf.meeting.views.upload_session_minutes',kwargs={'num':session.meeting.number,'session_id':session.id}) - else: - url = urlreverse('ietf.meeting.views.upload_session_agenda',kwargs={'num':session.meeting.number,'session_id':session.id}) - self.client.logout() - login_testing_unauthorized(self,"secretary",url) - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - q = PyQuery(r.content) - self.assertIn('Upload', str(q("Title"))) - self.assertFalse(session.sessionpresentation_set.exists()) - self.assertFalse(q('form input[type="checkbox"]')) - - session2 = SessionFactory(meeting=session.meeting,group=session.group) - r = self.client.get(url) - self.assertEqual(r.status_code, 200) - q = PyQuery(r.content) - self.assertTrue(q('form input[type="checkbox"]')) - - test_file = BytesIO(b'this is some text for a test') - test_file.name = "not_really.json" - r = self.client.post(url,dict(file=test_file)) - self.assertEqual(r.status_code, 200) - q = PyQuery(r.content) - self.assertTrue(q('form .is-invalid')) - - test_file = BytesIO(b'this is some text for a test'*1510000) - test_file.name = "not_really.pdf" - r = self.client.post(url,dict(file=test_file)) - self.assertEqual(r.status_code, 200) - q = PyQuery(r.content) - self.assertTrue(q('form .is-invalid')) - - test_file = BytesIO(b'') - test_file.name = "not_really.html" - r = self.client.post(url,dict(file=test_file)) - self.assertEqual(r.status_code, 200) - q = PyQuery(r.content) - self.assertTrue(q('form .is-invalid')) - - # Test html sanitization - test_file = BytesIO(b'Title

Title

Some text
') - test_file.name = "some.html" - r = self.client.post(url,dict(file=test_file)) - self.assertEqual(r.status_code, 302) - doc = session.sessionpresentation_set.filter(document__type_id=doctype).first().document - self.assertEqual(doc.rev,'00') - text = doc.text() - self.assertIn('Some text', text) - self.assertNotIn('
', text) - self.assertIn('charset="utf-8"', text) - - # txt upload - test_file = BytesIO(b'This is some text for a test, with the word\nvirtual at the beginning of a line.') - test_file.name = "some.txt" - r = self.client.post(url,dict(file=test_file,apply_to_all=False)) - self.assertEqual(r.status_code, 302) - doc = session.sessionpresentation_set.filter(document__type_id=doctype).first().document - self.assertEqual(doc.rev,'01') - self.assertFalse(session2.sessionpresentation_set.filter(document__type_id=doctype)) - + def test_label_future_sessions(self): + self.client.login(username='secretary', password='secretary+password') + for future in (True, False): + mtg_date = date_today()+datetime.timedelta(days=180 if future else -180) + session = SessionFactory(meeting__type_id='ietf', meeting__date=mtg_date) + # Verify future warning shows on the session details panel + url = urlreverse('ietf.meeting.views.session_details', kwargs={'num':session.meeting.number, 'acronym': session.group.acronym}) r = self.client.get(url) - self.assertEqual(r.status_code, 200) - q = PyQuery(r.content) - self.assertIn('Revise', str(q("Title"))) - test_file = BytesIO(b'this is some different text for a test') - test_file.name = "also_some.txt" - r = self.client.post(url,dict(file=test_file,apply_to_all=True)) - self.assertEqual(r.status_code, 302) - doc = Document.objects.get(pk=doc.pk) - self.assertEqual(doc.rev,'02') - self.assertTrue(session2.sessionpresentation_set.filter(document__type_id=doctype)) - - # Test bad encoding - test_file = BytesIO('

Title

Some\x93text
'.encode('latin1')) - test_file.name = "some.html" - r = self.client.post(url,dict(file=test_file)) - self.assertContains(r, 'Could not identify the file encoding') - doc = Document.objects.get(pk=doc.pk) - self.assertEqual(doc.rev,'02') + self.assertTrue(r.status_code==200) + if future: + self.assertContains(r, "Session has not ended yet") + else: + self.assertNotContains(r, "Session has not ended yet") - # Verify that we don't have dead links - url = urlreverse('ietf.meeting.views.session_details', kwargs={'num':session.meeting.number, 'acronym': session.group.acronym}) - top = '/meeting/%s/' % session.meeting.number - self.requests_mock.get(f'{session.notes_url()}/download', text='markdown notes') - self.requests_mock.get(f'{session.notes_url()}/info', text=json.dumps({'title': 'title', 'updatetime': '2021-12-01T17:11:00z'})) - self.crawl_materials(url=url, top=top) + def test_upload_minutes_agenda(self): + for doctype in ('minutes','agenda'): + for future in (True, False): + mtg_date = date_today()+datetime.timedelta(days=180 if future else -180) + session = SessionFactory(meeting__type_id='ietf', meeting__date=mtg_date) + if doctype == 'minutes': + url = urlreverse('ietf.meeting.views.upload_session_minutes',kwargs={'num':session.meeting.number,'session_id':session.id}) + else: + url = urlreverse('ietf.meeting.views.upload_session_agenda',kwargs={'num':session.meeting.number,'session_id':session.id}) + self.client.logout() + login_testing_unauthorized(self,"secretary",url) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertIn('Upload', str(q("Title"))) + self.assertFalse(session.presentations.exists()) + self.assertFalse(q('form input[type="checkbox"]')) + if future and doctype == "minutes": + self.assertContains(r, "Session has not ended yet") + else: + self.assertNotContains(r, "Session has not ended yet") + + session2 = SessionFactory(meeting=session.meeting,group=session.group) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertTrue(q('form input[type="checkbox"]')) + + # test not submitting a file + r = self.client.post(url, dict(submission_method="upload")) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertTrue(q("form .is-invalid")) + + test_file = BytesIO(b'this is some text for a test') + test_file.name = "not_really.json" + r = self.client.post(url,dict(submission_method="upload",file=test_file)) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertTrue(q('form .is-invalid')) + + test_file = BytesIO(b'this is some text for a test'*1510000) + test_file.name = "not_really.pdf" + r = self.client.post(url,dict(submission_method="upload",file=test_file)) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertTrue(q('form .is-invalid')) + + test_file = BytesIO(b'') + test_file.name = "not_really.html" + r = self.client.post(url,dict(submission_method="upload",file=test_file)) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertTrue(q('form .is-invalid')) + + # Test html sanitization + test_file = BytesIO(b'Title

Title

Some text
') + test_file.name = "some.html" + r = self.client.post(url,dict(submission_method="upload",file=test_file)) + self.assertEqual(r.status_code, 302) + doc = session.presentations.filter(document__type_id=doctype).first().document + self.assertEqual(doc.rev,'00') + text = doc.text() + self.assertIn('Some text', text) + self.assertNotIn('
', text) + text = retrieve_str(doctype, f"{doc.name}-{doc.rev}.html") + self.assertIn('Some text', text) + self.assertNotIn('
', text) + + # txt upload + test_bytes = b'This is some text for a test, with the word\nvirtual at the beginning of a line.' + test_file = BytesIO(test_bytes) + test_file.name = "some.txt" + r = self.client.post(url,dict(submission_method="upload",file=test_file,apply_to_all=False)) + self.assertEqual(r.status_code, 302) + doc = session.presentations.filter(document__type_id=doctype).first().document + self.assertEqual(doc.rev,'01') + self.assertFalse(session2.presentations.filter(document__type_id=doctype)) + retrieved_bytes = retrieve_bytes(doctype, f"{doc.name}-{doc.rev}.txt") + self.assertEqual(retrieved_bytes, test_bytes) + + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertIn('Revise', str(q("Title"))) + test_bytes = b'this is some different text for a test' + test_file = BytesIO(test_bytes) + test_file.name = "also_some.txt" + r = self.client.post(url,dict(submission_method="upload",file=test_file,apply_to_all=True)) + self.assertEqual(r.status_code, 302) + doc = Document.objects.get(pk=doc.pk) + self.assertEqual(doc.rev,'02') + self.assertTrue(session2.presentations.filter(document__type_id=doctype)) + retrieved_bytes = retrieve_bytes(doctype, f"{doc.name}-{doc.rev}.txt") + self.assertEqual(retrieved_bytes, test_bytes) + + # Test bad encoding + test_file = BytesIO('

Title

Some\x93text
'.encode('latin1')) + test_file.name = "some.html" + r = self.client.post(url,dict(submission_method="upload",file=test_file)) + self.assertContains(r, 'Could not identify the file encoding') + doc = Document.objects.get(pk=doc.pk) + self.assertEqual(doc.rev,'02') + + # Verify that we don't have dead links + url = urlreverse('ietf.meeting.views.session_details', kwargs={'num':session.meeting.number, 'acronym': session.group.acronym}) + top = '/meeting/%s/' % session.meeting.number + self.requests_mock.get(f'{session.notes_url()}/download', text='markdown notes') + self.requests_mock.get(f'{session.notes_url()}/info', text=json.dumps({'title': 'title', 'updatetime': '2021-12-01T17:11:00z'})) + self.crawl_materials(url=url, top=top) def test_upload_minutes_agenda_unscheduled(self): for doctype in ('minutes','agenda'): @@ -5873,35 +6776,74 @@ def test_upload_minutes_agenda_unscheduled(self): self.assertEqual(r.status_code, 200) q = PyQuery(r.content) self.assertIn('Upload', str(q("Title"))) - self.assertFalse(session.sessionpresentation_set.exists()) + self.assertFalse(session.presentations.exists()) self.assertFalse(q('form input[type="checkbox"]')) + self.assertNotContains(r, "Session has not ended yet") test_file = BytesIO(b'this is some text for a test') test_file.name = "not_really.txt" - r = self.client.post(url,dict(file=test_file,apply_to_all=False)) + r = self.client.post(url,dict(submission_method="upload",file=test_file,apply_to_all=False)) self.assertEqual(r.status_code, 410) @override_settings(MEETING_MATERIALS_SERVE_LOCALLY=True) def test_upload_minutes_agenda_interim(self): - session=SessionFactory(meeting__type_id='interim') for doctype in ('minutes','agenda'): - if doctype=='minutes': - url = urlreverse('ietf.meeting.views.upload_session_minutes',kwargs={'num':session.meeting.number,'session_id':session.id}) - else: - url = urlreverse('ietf.meeting.views.upload_session_agenda',kwargs={'num':session.meeting.number,'session_id':session.id}) + for future in (True, False): + session=SessionFactory(meeting__type_id='interim', meeting__date = date_today()+datetime.timedelta(days=180 if future else -180)) + if doctype=='minutes': + url = urlreverse('ietf.meeting.views.upload_session_minutes',kwargs={'num':session.meeting.number,'session_id':session.id}) + else: + url = urlreverse('ietf.meeting.views.upload_session_agenda',kwargs={'num':session.meeting.number,'session_id':session.id}) + self.client.logout() + login_testing_unauthorized(self,"secretary",url) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertIn('Upload', str(q("title"))) + self.assertFalse(session.presentations.filter(document__type_id=doctype)) + if future and doctype == "minutes": + self.assertContains(r, "Session has not ended yet") + else: + self.assertNotContains(r, "Session has not ended yet") + test_bytes = b'this is some text for a test' + test_file = BytesIO(test_bytes) + test_file.name = "not_really.txt" + r = self.client.post(url,dict(submission_method="upload",file=test_file)) + self.assertEqual(r.status_code, 302) + doc = session.presentations.filter(document__type_id=doctype).first().document + self.assertEqual(doc.rev,'00') + retrieved_bytes = retrieve_bytes(doctype, f"{doc.name}-{doc.rev}.txt") + self.assertEqual(retrieved_bytes, test_bytes) + + # Verify that we don't have dead links + url = urlreverse('ietf.meeting.views.session_details', kwargs={'num':session.meeting.number, 'acronym': session.group.acronym}) + top = '/meeting/%s/' % session.meeting.number + self.requests_mock.get(f'{session.notes_url()}/download', text='markdown notes') + self.requests_mock.get(f'{session.notes_url()}/info', text=json.dumps({'title': 'title', 'updatetime': '2021-12-01T17:11:00z'})) + self.crawl_materials(url=url, top=top) + + @override_settings(MEETING_MATERIALS_SERVE_LOCALLY=True) + def test_upload_narrativeminutes(self): + for type_id in ["interim","ietf"]: + session=SessionFactory(meeting__type_id=type_id,group__acronym='iesg') + doctype='narrativeminutes' + url = urlreverse('ietf.meeting.views.upload_session_narrativeminutes',kwargs={'num':session.meeting.number,'session_id':session.id}) self.client.logout() login_testing_unauthorized(self,"secretary",url) r = self.client.get(url) self.assertEqual(r.status_code, 200) q = PyQuery(r.content) self.assertIn('Upload', str(q("title"))) - self.assertFalse(session.sessionpresentation_set.filter(document__type_id=doctype)) - test_file = BytesIO(b'this is some text for a test') + self.assertFalse(session.presentations.filter(document__type_id=doctype)) + test_bytes = b'this is some text for a test' + test_file = BytesIO(test_bytes) test_file.name = "not_really.txt" - r = self.client.post(url,dict(file=test_file)) + r = self.client.post(url,dict(submission_method="upload",file=test_file)) self.assertEqual(r.status_code, 302) - doc = session.sessionpresentation_set.filter(document__type_id=doctype).first().document + doc = session.presentations.filter(document__type_id=doctype).first().document self.assertEqual(doc.rev,'00') + retrieved_bytes = retrieve_bytes(doctype, f"{doc.name}-{doc.rev}.txt") + self.assertEqual(retrieved_bytes, test_bytes) # Verify that we don't have dead links url = urlreverse('ietf.meeting.views.session_details', kwargs={'num':session.meeting.number, 'acronym': session.group.acronym}) @@ -5910,7 +6852,55 @@ def test_upload_minutes_agenda_interim(self): self.requests_mock.get(f'{session.notes_url()}/info', text=json.dumps({'title': 'title', 'updatetime': '2021-12-01T17:11:00z'})) self.crawl_materials(url=url, top=top) - def test_upload_slides(self): + def test_enter_agenda(self): + session = SessionFactory(meeting__type_id='ietf') + url = urlreverse('ietf.meeting.views.upload_session_agenda',kwargs={'num':session.meeting.number,'session_id':session.id}) + redirect_url = urlreverse('ietf.meeting.views.session_details', kwargs={'num':session.meeting.number,'acronym':session.group.acronym}) + login_testing_unauthorized(self,"secretary",url) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertIn('Upload', str(q("Title"))) + self.assertFalse(session.presentations.exists()) + + test_text = 'Enter agenda from scratch' + r = self.client.post(url,dict(submission_method="enter",content=test_text)) + self.assertRedirects(r, redirect_url) + doc = session.presentations.filter(document__type_id='agenda').first().document + self.assertEqual(doc.rev,'00') + self.assertEqual(retrieve_str("agenda",f"{doc.name}-{doc.rev}.md"), test_text) + + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertIn('Revise', str(q("Title"))) + + test_bytes = b'Upload after enter' + test_file = BytesIO(test_bytes) + test_file.name = "some.txt" + r = self.client.post(url,dict(submission_method="upload",file=test_file)) + self.assertRedirects(r, redirect_url) + doc = Document.objects.get(pk=doc.pk) + self.assertEqual(doc.rev,'01') + retrieved_bytes = retrieve_bytes("agenda", f"{doc.name}-{doc.rev}.txt") + self.assertEqual(retrieved_bytes, test_bytes) + + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertIn('Revise', str(q("Title"))) + + test_text = 'Enter after upload' + r = self.client.post(url,dict(submission_method="enter",content=test_text)) + self.assertRedirects(r, redirect_url) + doc = Document.objects.get(pk=doc.pk) + self.assertEqual(doc.rev,'02') + self.assertEqual(retrieve_str("agenda",f"{doc.name}-{doc.rev}.md"), test_text) + + + @override_settings(MEETECHO_API_CONFIG="fake settings") # enough to trigger API calls + @patch("ietf.meeting.views.SlidesManager") + def test_upload_slides(self, mock_slides_manager_cls): session1 = SessionFactory(meeting__type_id='ietf') session2 = SessionFactory(meeting=session1.meeting,group=session1.group) @@ -5918,46 +6908,80 @@ def test_upload_slides(self): login_testing_unauthorized(self,"secretary",url) r = self.client.get(url) self.assertEqual(r.status_code, 200) + self.assertFalse(mock_slides_manager_cls.called) q = PyQuery(r.content) self.assertIn('Upload', str(q("title"))) - self.assertFalse(session1.sessionpresentation_set.filter(document__type_id='slides')) - test_file = BytesIO(b'this is not really a slide') + self.assertFalse(session1.presentations.filter(document__type_id='slides')) + test_bytes = b'this is not really a slide' + test_file = BytesIO(test_bytes) test_file.name = 'not_really.txt' - r = self.client.post(url,dict(file=test_file,title='a test slide file',apply_to_all=True)) + r = self.client.post(url,dict(file=test_file,title='a test slide file',apply_to_all=True,approved=True)) self.assertEqual(r.status_code, 302) - self.assertEqual(session1.sessionpresentation_set.count(),1) - self.assertEqual(session2.sessionpresentation_set.count(),1) - sp = session2.sessionpresentation_set.first() + self.assertEqual(session1.presentations.count(),1) + self.assertEqual(session2.presentations.count(),1) + sp = session2.presentations.first() self.assertEqual(sp.document.name, 'slides-%s-%s-a-test-slide-file' % (session1.meeting.number,session1.group.acronym ) ) self.assertEqual(sp.order,1) + self.assertEqual(mock_slides_manager_cls.call_count, 1) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertEqual(mock_slides_manager_cls.return_value.add.call_count, 2) + self.assertEqual(retrieve_bytes("slides", f"{sp.document.name}-{sp.document.rev}.txt"), test_bytes) + # don't care which order they were called in, just that both sessions were updated + self.assertCountEqual( + mock_slides_manager_cls.return_value.add.call_args_list, + [ + call(session=session1, slides=sp.document, order=1), + call(session=session2, slides=sp.document, order=1), + ], + ) + mock_slides_manager_cls.reset_mock() url = urlreverse('ietf.meeting.views.upload_session_slides',kwargs={'num':session2.meeting.number,'session_id':session2.id}) - test_file = BytesIO(b'some other thing still not slidelike') + test_bytes = b'some other thing still not slidelike' + test_file = BytesIO(test_bytes) test_file.name = 'also_not_really.txt' - r = self.client.post(url,dict(file=test_file,title='a different slide file',apply_to_all=False)) + r = self.client.post(url,dict(file=test_file,title='a different slide file',apply_to_all=False,approved=True)) self.assertEqual(r.status_code, 302) - self.assertEqual(session1.sessionpresentation_set.count(),1) - self.assertEqual(session2.sessionpresentation_set.count(),2) - sp = session2.sessionpresentation_set.get(document__name__endswith='-a-different-slide-file') + self.assertEqual(session1.presentations.count(),1) + self.assertEqual(session2.presentations.count(),2) + sp = session2.presentations.get(document__name__endswith='-a-different-slide-file') self.assertEqual(sp.order,2) self.assertEqual(sp.rev,'00') self.assertEqual(sp.document.rev,'00') + self.assertEqual(mock_slides_manager_cls.call_count, 1) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertEqual(mock_slides_manager_cls.return_value.add.call_count, 1) + self.assertEqual(retrieve_bytes("slides", f"{sp.document.name}-{sp.document.rev}.txt"), test_bytes) + self.assertEqual( + mock_slides_manager_cls.return_value.add.call_args, + call(session=session2, slides=sp.document, order=2), + ) + mock_slides_manager_cls.reset_mock() - url = urlreverse('ietf.meeting.views.upload_session_slides',kwargs={'num':session2.meeting.number,'session_id':session2.id,'name':session2.sessionpresentation_set.get(order=2).document.name}) + url = urlreverse('ietf.meeting.views.upload_session_slides',kwargs={'num':session2.meeting.number,'session_id':session2.id,'name':session2.presentations.get(order=2).document.name}) r = self.client.get(url) self.assertTrue(r.status_code, 200) q = PyQuery(r.content) self.assertIn('Revise', str(q("title"))) - test_file = BytesIO(b'new content for the second slide deck') + test_bytes = b'new content for the second slide deck' + test_file = BytesIO(test_bytes) test_file.name = 'doesnotmatter.txt' - r = self.client.post(url,dict(file=test_file,title='rename the presentation',apply_to_all=False)) + r = self.client.post(url,dict(file=test_file,title='rename the presentation',apply_to_all=False, approved=True)) self.assertEqual(r.status_code, 302) - self.assertEqual(session1.sessionpresentation_set.count(),1) - self.assertEqual(session2.sessionpresentation_set.count(),2) - sp = session2.sessionpresentation_set.get(order=2) - self.assertEqual(sp.rev,'01') - self.assertEqual(sp.document.rev,'01') - + self.assertEqual(session1.presentations.count(),1) + self.assertEqual(session2.presentations.count(),2) + replacement_sp = session2.presentations.get(order=2) + self.assertEqual(replacement_sp.rev,'01') + self.assertEqual(replacement_sp.document.rev,'01') + self.assertEqual(mock_slides_manager_cls.call_count, 1) + self.assertEqual(retrieve_bytes("slides", f"{replacement_sp.document.name}-{replacement_sp.document.rev}.txt"), test_bytes) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertEqual(mock_slides_manager_cls.return_value.revise.call_count, 1) + self.assertEqual( + mock_slides_manager_cls.return_value.revise.call_args, + call(session=session2, slides=sp.document), + ) + def test_upload_slide_title_bad_unicode(self): session1 = SessionFactory(meeting__type_id='ietf') url = urlreverse('ietf.meeting.views.upload_session_slides',kwargs={'num':session1.meeting.number,'session_id':session1.id}) @@ -5966,7 +6990,7 @@ def test_upload_slide_title_bad_unicode(self): self.assertEqual(r.status_code, 200) q = PyQuery(r.content) self.assertIn('Upload', str(q("title"))) - self.assertFalse(session1.sessionpresentation_set.filter(document__type_id='slides')) + self.assertFalse(session1.presentations.filter(document__type_id='slides')) test_file = BytesIO(b'this is not really a slide') test_file.name = 'not_really.txt' r = self.client.post(url,dict(file=test_file,title='title with bad character \U0001fabc ')) @@ -5975,29 +6999,60 @@ def test_upload_slide_title_bad_unicode(self): self.assertTrue(q('form .is-invalid')) self.assertIn("Unicode BMP", q('form .is-invalid div').text()) - def test_remove_sessionpresentation(self): + @override_settings(MEETECHO_API_CONFIG="fake settings") # enough to trigger API calls + @patch("ietf.meeting.views.SlidesManager") + def test_remove_sessionpresentation(self, mock_slides_manager_cls): session = SessionFactory(meeting__type_id='ietf') + agenda = DocumentFactory(type_id='agenda') doc = DocumentFactory(type_id='slides') - session.sessionpresentation_set.create(document=doc) + session.presentations.create(document=agenda) + session.presentations.create(document=doc) url = urlreverse('ietf.meeting.views.remove_sessionpresentation',kwargs={'num':session.meeting.number,'session_id':session.id,'name':'no-such-doc'}) response = self.client.get(url) self.assertEqual(response.status_code, 404) + self.assertFalse(mock_slides_manager_cls.called) url = urlreverse('ietf.meeting.views.remove_sessionpresentation',kwargs={'num':session.meeting.number,'session_id':0,'name':doc.name}) response = self.client.get(url) self.assertEqual(response.status_code, 404) + self.assertFalse(mock_slides_manager_cls.called) url = urlreverse('ietf.meeting.views.remove_sessionpresentation',kwargs={'num':session.meeting.number,'session_id':session.id,'name':doc.name}) login_testing_unauthorized(self,"secretary",url) response = self.client.get(url) self.assertEqual(response.status_code, 200) + self.assertFalse(mock_slides_manager_cls.called) - self.assertEqual(1,session.sessionpresentation_set.count()) + # Removing slides should remove the materials and call MeetechoAPI + self.assertEqual(2, session.presentations.count()) response = self.client.post(url,{'remove_session':''}) self.assertEqual(response.status_code, 302) - self.assertEqual(0,session.sessionpresentation_set.count()) - self.assertEqual(2,doc.docevent_set.count()) + self.assertEqual(1, session.presentations.count()) + self.assertEqual(2, doc.docevent_set.count()) + self.assertEqual(mock_slides_manager_cls.call_count, 1) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertEqual(mock_slides_manager_cls.return_value.delete.call_count, 1) + self.assertEqual( + mock_slides_manager_cls.return_value.delete.call_args, + call(session=session, slides=doc), + ) + mock_slides_manager_cls.reset_mock() + + # Removing non-slides should only remove the materials + url = urlreverse( + "ietf.meeting.views.remove_sessionpresentation", + kwargs={ + "num": session.meeting.number, + "session_id": session.id, + "name": agenda.name, + }, + ) + response = self.client.post(url, {"remove_session" : ""}) + self.assertEqual(response.status_code, 302) + self.assertEqual(0, session.presentations.count()) + self.assertEqual(2, agenda.docevent_set.count()) + self.assertFalse(mock_slides_manager_cls.called) def test_propose_session_slides(self): for type_id in ['ietf','interim']: @@ -6007,7 +7062,7 @@ def test_propose_session_slides(self): newperson = PersonFactory() session_overview_url = urlreverse('ietf.meeting.views.session_details',kwargs={'num':session.meeting.number,'acronym':session.group.acronym}) - propose_url = urlreverse('ietf.meeting.views.propose_session_slides', kwargs={'session_id':session.pk, 'num': session.meeting.number}) + upload_url = urlreverse('ietf.meeting.views.upload_session_slides', kwargs={'session_id':session.pk, 'num': session.meeting.number}) r = self.client.get(session_overview_url) self.assertEqual(r.status_code,200) @@ -6022,17 +7077,22 @@ def test_propose_session_slides(self): self.assertTrue(q('.proposeslides')) self.client.logout() - login_testing_unauthorized(self,newperson.user.username,propose_url) - r = self.client.get(propose_url) + login_testing_unauthorized(self,newperson.user.username,upload_url) + r = self.client.get(upload_url) self.assertEqual(r.status_code,200) - test_file = BytesIO(b'this is not really a slide') + test_bytes = b'this is not really a slide' + test_file = BytesIO(test_bytes) test_file.name = 'not_really.txt' empty_outbox() - r = self.client.post(propose_url,dict(file=test_file,title='a test slide file',apply_to_all=True)) + r = self.client.post(upload_url,dict(file=test_file,title='a test slide file',apply_to_all=True,approved=False)) self.assertEqual(r.status_code, 302) session = Session.objects.get(pk=session.pk) self.assertEqual(session.slidesubmission_set.count(),1) self.assertEqual(len(outbox),1) + self.assertEqual( + retrieve_bytes("staging", session.slidesubmission_set.get().filename), + test_bytes + ) r = self.client.get(session_overview_url) self.assertEqual(r.status_code, 200) @@ -6049,6 +7109,32 @@ def test_propose_session_slides(self): self.assertEqual(len(q('.proposedslidelist p')), 2) self.client.logout() + login_testing_unauthorized(self,chair.user.username,upload_url) + r = self.client.get(upload_url) + self.assertEqual(r.status_code,200) + test_bytes = b'this is not really a slide either' + test_file = BytesIO(test_bytes) + test_file.name = 'again_not_really.txt' + empty_outbox() + r = self.client.post(upload_url,dict(file=test_file,title='a selfapproved test slide file',apply_to_all=True,approved=True)) + self.assertEqual(r.status_code, 302) + self.assertEqual(len(outbox),0) + self.assertEqual(session.slidesubmission_set.count(),2) + sp = session.presentations.get(document__title__contains="selfapproved") + self.assertFalse(exists_in_storage("staging", sp.document.uploaded_filename)) + self.assertEqual( + retrieve_bytes("slides", sp.document.uploaded_filename), + test_bytes + ) + self.client.logout() + + self.client.login(username=chair.user.username, password=chair.user.username+"+password") + r = self.client.get(session_overview_url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + self.assertEqual(len(q('.uploadslidelist p')), 0) + self.client.logout() + def test_disapprove_proposed_slides(self): submission = SlideSubmissionFactory() submission.session.meeting.importantdate_set.create(name_id='revsub',date=date_today() + datetime.timedelta(days=20)) @@ -6062,11 +7148,15 @@ def test_disapprove_proposed_slides(self): self.assertEqual(r.status_code,302) self.assertEqual(SlideSubmission.objects.filter(status__slug = 'rejected').count(), 1) self.assertEqual(SlideSubmission.objects.filter(status__slug = 'pending').count(), 0) + if submission.filename is not None and submission.filename != "": + self.assertFalse(exists_in_storage("staging", submission.filename)) r = self.client.get(url) self.assertEqual(r.status_code, 200) - self.assertRegex(r.content.decode(), r"These\s+slides\s+have\s+already\s+been\s+rejected") + self.assertRegex(r.content.decode(), r"These\s+slides\s+have\s+already\s+been\s+declined") - def test_approve_proposed_slides(self): + @override_settings(MEETECHO_API_CONFIG="fake settings") # enough to trigger API calls + @patch("ietf.meeting.views.SlidesManager") + def test_approve_proposed_slides(self, mock_slides_manager_cls): submission = SlideSubmissionFactory() session = submission.session session.meeting.importantdate_set.create(name_id='revsub',date=date_today() + datetime.timedelta(days=20)) @@ -6077,20 +7167,39 @@ def test_approve_proposed_slides(self): self.assertIsNone(submission.doc) r = self.client.get(url) self.assertEqual(r.status_code,200) + empty_outbox() + self.assertTrue(exists_in_storage("staging", submission.filename)) r = self.client.post(url,dict(title='different title',approve='approve')) self.assertEqual(r.status_code,302) self.assertEqual(SlideSubmission.objects.filter(status__slug = 'pending').count(), 0) self.assertEqual(SlideSubmission.objects.filter(status__slug = 'approved').count(), 1) - submission = SlideSubmission.objects.get(id = submission.id) + submission.refresh_from_db() self.assertEqual(submission.status_id, 'approved') self.assertIsNotNone(submission.doc) - self.assertEqual(session.sessionpresentation_set.count(),1) - self.assertEqual(session.sessionpresentation_set.first().document.title,'different title') + self.assertEqual(session.presentations.count(),1) + self.assertEqual(session.presentations.first().document.title,'different title') + self.assertTrue(exists_in_storage("slides", submission.doc.uploaded_filename)) + self.assertFalse(exists_in_storage("staging", submission.filename)) + self.assertEqual(mock_slides_manager_cls.call_count, 1) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertEqual(mock_slides_manager_cls.return_value.add.call_count, 1) + self.assertEqual( + mock_slides_manager_cls.return_value.add.call_args, + call(session=session, slides=submission.doc, order=1), + ) + mock_slides_manager_cls.reset_mock() r = self.client.get(url) self.assertEqual(r.status_code, 200) self.assertRegex(r.content.decode(), r"These\s+slides\s+have\s+already\s+been\s+approved") - - def test_approve_proposed_slides_multisession_apply_one(self): + self.assertFalse(mock_slides_manager_cls.called) + self.assertEqual(len(outbox), 1) + self.assertIn(submission.submitter.email_address(), outbox[0]['To']) + self.assertIn('Slides approved', outbox[0]['Subject']) + + @override_settings(MEETECHO_API_CONFIG="fake settings") # enough to trigger API calls + @patch("ietf.meeting.views.SlidesManager") + def test_approve_proposed_slides_multisession_apply_one(self, mock_slides_manager_cls): + TestBlobstoreManager().emptyTestBlobstores() submission = SlideSubmissionFactory(session__meeting__type_id='ietf') session1 = submission.session session2 = SessionFactory(group=submission.session.group, meeting=submission.session.meeting) @@ -6103,11 +7212,23 @@ def test_approve_proposed_slides_multisession_apply_one(self): q = PyQuery(r.content) self.assertTrue(q('#id_apply_to_all')) r = self.client.post(url,dict(title='yet another title',approve='approve')) + submission.refresh_from_db() + self.assertIsNotNone(submission.doc) self.assertEqual(r.status_code,302) - self.assertEqual(session1.sessionpresentation_set.count(),1) - self.assertEqual(session2.sessionpresentation_set.count(),0) + self.assertEqual(session1.presentations.count(),1) + self.assertEqual(session2.presentations.count(),0) + self.assertEqual(mock_slides_manager_cls.call_count, 1) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertEqual(mock_slides_manager_cls.return_value.add.call_count, 1) + self.assertEqual( + mock_slides_manager_cls.return_value.add.call_args, + call(session=session1, slides=submission.doc, order=1), + ) - def test_approve_proposed_slides_multisession_apply_all(self): + @override_settings(MEETECHO_API_CONFIG="fake settings") # enough to trigger API calls + @patch("ietf.meeting.views.SlidesManager") + def test_approve_proposed_slides_multisession_apply_all(self, mock_slides_manager_cls): + TestBlobstoreManager().emptyTestBlobstores() submission = SlideSubmissionFactory(session__meeting__type_id='ietf') session1 = submission.session session2 = SessionFactory(group=submission.session.group, meeting=submission.session.meeting) @@ -6118,68 +7239,174 @@ def test_approve_proposed_slides_multisession_apply_all(self): r = self.client.get(url) self.assertEqual(r.status_code,200) r = self.client.post(url,dict(title='yet another title',apply_to_all=1,approve='approve')) + submission.refresh_from_db() self.assertEqual(r.status_code,302) - self.assertEqual(session1.sessionpresentation_set.count(),1) - self.assertEqual(session2.sessionpresentation_set.count(),1) + self.assertEqual(session1.presentations.count(),1) + self.assertEqual(session2.presentations.count(),1) + self.assertEqual(mock_slides_manager_cls.call_count, 1) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertEqual(mock_slides_manager_cls.return_value.add.call_count, 2) + self.assertCountEqual( + mock_slides_manager_cls.return_value.add.call_args_list, + [ + call(session=session1, slides=submission.doc, order=1), + call(session=session2, slides=submission.doc, order=1), + ] + ) - def test_submit_and_approve_multiple_versions(self): + @override_settings(MEETECHO_API_CONFIG="fake settings") # enough to trigger API calls + @patch("ietf.meeting.views.SlidesManager") + def test_submit_and_approve_multiple_versions(self, mock_slides_manager_cls): session = SessionFactory(meeting__type_id='ietf') chair = RoleFactory(group=session.group,name_id='chair').person session.meeting.importantdate_set.create(name_id='revsub',date=date_today()+datetime.timedelta(days=20)) newperson = PersonFactory() - propose_url = urlreverse('ietf.meeting.views.propose_session_slides', kwargs={'session_id':session.pk, 'num': session.meeting.number}) + upload_url = urlreverse('ietf.meeting.views.upload_session_slides', kwargs={'session_id':session.pk, 'num': session.meeting.number}) - login_testing_unauthorized(self,newperson.user.username,propose_url) + login_testing_unauthorized(self,newperson.user.username,upload_url) test_file = BytesIO(b'this is not really a slide') test_file.name = 'not_really.txt' - r = self.client.post(propose_url,dict(file=test_file,title='a test slide file',apply_to_all=True)) + r = self.client.post(upload_url,dict(file=test_file,title='a test slide file',apply_to_all=True,approved=False)) self.assertEqual(r.status_code, 302) self.client.logout() - submission = SlideSubmission.objects.get(session = session) + submission = SlideSubmission.objects.get(session=session) + self.assertTrue(exists_in_storage("staging", submission.filename)) approve_url = urlreverse('ietf.meeting.views.approve_proposed_slides', kwargs={'slidesubmission_id':submission.pk,'num':submission.session.meeting.number}) login_testing_unauthorized(self, chair.user.username, approve_url) r = self.client.post(approve_url,dict(title=submission.title,approve='approve')) + submission.refresh_from_db() self.assertEqual(r.status_code,302) self.client.logout() + self.assertFalse(exists_in_storage("staging", submission.filename)) + self.assertTrue(exists_in_storage("slides", submission.doc.uploaded_filename)) + self.assertEqual(mock_slides_manager_cls.call_count, 1) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertEqual(mock_slides_manager_cls.return_value.add.call_count, 1) + self.assertEqual( + mock_slides_manager_cls.return_value.add.call_args, + call(session=session, slides=submission.doc, order=1), + ) + mock_slides_manager_cls.reset_mock() + + self.assertEqual(session.presentations.first().document.rev,'00') - self.assertEqual(session.sessionpresentation_set.first().document.rev,'00') - - login_testing_unauthorized(self,newperson.user.username,propose_url) + login_testing_unauthorized(self,newperson.user.username,upload_url) test_file = BytesIO(b'this is not really a slide, but it is another version of it') test_file.name = 'not_really.txt' - r = self.client.post(propose_url,dict(file=test_file,title='a test slide file',apply_to_all=True)) + r = self.client.post(upload_url,dict(file=test_file,title='a test slide file',apply_to_all=True)) self.assertEqual(r.status_code, 302) test_file = BytesIO(b'this is not really a slide, but it is third version of it') test_file.name = 'not_really.txt' - r = self.client.post(propose_url,dict(file=test_file,title='a test slide file',apply_to_all=True)) + r = self.client.post(upload_url,dict(file=test_file,title='a test slide file',apply_to_all=True)) self.assertEqual(r.status_code, 302) self.client.logout() (first_submission, second_submission) = SlideSubmission.objects.filter(session=session, status__slug = 'pending').order_by('id') + self.assertTrue(exists_in_storage("staging", first_submission.filename)) + self.assertTrue(exists_in_storage("staging", second_submission.filename)) approve_url = urlreverse('ietf.meeting.views.approve_proposed_slides', kwargs={'slidesubmission_id':second_submission.pk,'num':second_submission.session.meeting.number}) login_testing_unauthorized(self, chair.user.username, approve_url) r = self.client.post(approve_url,dict(title=submission.title,approve='approve')) + first_submission.refresh_from_db() + second_submission.refresh_from_db() + self.assertTrue(exists_in_storage("staging", first_submission.filename)) + self.assertFalse(exists_in_storage("staging", second_submission.filename)) + self.assertTrue(exists_in_storage("slides", second_submission.doc.uploaded_filename)) self.assertEqual(r.status_code,302) + self.assertEqual(mock_slides_manager_cls.call_count, 1) + self.assertEqual(mock_slides_manager_cls.call_args, call(api_config="fake settings")) + self.assertEqual(mock_slides_manager_cls.return_value.add.call_count, 0) + self.assertEqual(mock_slides_manager_cls.return_value.revise.call_count, 1) + self.assertEqual( + mock_slides_manager_cls.return_value.revise.call_args, + call(session=session, slides=second_submission.doc), + ) + mock_slides_manager_cls.reset_mock() disapprove_url = urlreverse('ietf.meeting.views.approve_proposed_slides', kwargs={'slidesubmission_id':first_submission.pk,'num':first_submission.session.meeting.number}) r = self.client.post(disapprove_url,dict(title='some title',disapprove="disapprove")) self.assertEqual(r.status_code,302) self.client.logout() + self.assertFalse(mock_slides_manager_cls.called) + self.assertFalse(exists_in_storage("staging", first_submission.filename)) self.assertEqual(SlideSubmission.objects.filter(status__slug = 'pending').count(),0) self.assertEqual(SlideSubmission.objects.filter(status__slug = 'rejected').count(),1) - self.assertEqual(session.sessionpresentation_set.first().document.rev,'01') + self.assertEqual(session.presentations.first().document.rev,'01') path = os.path.join(submission.session.meeting.get_materials_path(),'slides') - filename = os.path.join(path,session.sessionpresentation_set.first().document.name+'-01.txt') + filename = os.path.join(path,session.presentations.first().document.name+'-01.txt') self.assertTrue(os.path.exists(filename)) - contents = io.open(filename,'r').read() + fd = io.open(filename, 'r') + contents = fd.read() + fd.close() self.assertIn('third version', contents) + @override_settings( + MEETECHO_API_CONFIG="fake settings" + ) # enough to trigger API calls + @patch("ietf.meeting.views.SlidesManager") + def test_notify_meetecho_of_all_slides(self, mock_slides_manager_cls): + for meeting_type in ["ietf", "interim"]: + # Reset for the sake of the second iteration + self.client.logout() + mock_slides_manager_cls.reset_mock() + + session = SessionFactory(meeting__type_id=meeting_type) + meeting = session.meeting + + # bad meeting + url = urlreverse( + "ietf.meeting.views.notify_meetecho_of_all_slides", + kwargs={"num": 9999, "acronym": session.group.acronym}, + ) + login_testing_unauthorized(self, "secretary", url) + r = self.client.get(url) + self.assertEqual(r.status_code, 404) + r = self.client.post(url) + self.assertEqual(r.status_code, 404) + self.assertFalse(mock_slides_manager_cls.called) + self.client.logout() + + # good meeting + url = urlreverse( + "ietf.meeting.views.notify_meetecho_of_all_slides", + kwargs={"num": meeting.number, "acronym": session.group.acronym}, + ) + login_testing_unauthorized(self, "secretary", url) + r = self.client.get(url) + self.assertEqual(r.status_code, 405) + self.assertFalse(mock_slides_manager_cls.called) + mock_slides_manager = mock_slides_manager_cls.return_value + mock_slides_manager.send_update.return_value = True + r = self.client.post(url) + self.assertEqual(r.status_code, 302) + self.assertEqual(mock_slides_manager.send_update.call_count, 1) + self.assertEqual(mock_slides_manager.send_update.call_args, call(session)) + r = self.client.get(r["Location"]) + messages = list(r.context["messages"]) + self.assertEqual(len(messages), 1) + self.assertEqual( + str(messages[0]), f"Notified Meetecho about slides for {session}" + ) + + mock_slides_manager.send_update.reset_mock() + mock_slides_manager.send_update.return_value = False + r = self.client.post(url) + self.assertEqual(r.status_code, 302) + self.assertEqual(mock_slides_manager.send_update.call_count, 1) + self.assertEqual(mock_slides_manager.send_update.call_args, call(session)) + r = self.client.get(r["Location"]) + messages = list(r.context["messages"]) + self.assertEqual(len(messages), 1) + self.assertIn( + "No sessions were eligible for Meetecho slides update.", str(messages[0]) + ) + @override_settings(IETF_NOTES_URL='https://notes.ietf.org/') class ImportNotesTests(TestCase): @@ -6258,6 +7485,10 @@ def test_imports_previewed_text(self): minutes_path = Path(self.meeting.get_materials_path()) / 'minutes' with (minutes_path / self.session.minutes().uploaded_filename).open() as f: self.assertEqual(f.read(), 'original markdown text') + self.assertEqual( + retrieve_str("minutes", self.session.minutes().uploaded_filename), + 'original markdown text' + ) def test_refuses_identical_import(self): """Should not be able to import text identical to the current revision""" @@ -6279,8 +7510,8 @@ def test_refuses_identical_import(self): r = self.client.get(url) # try to import the same text self.assertContains(r, "This document is identical", status_code=200) q = PyQuery(r.content) - self.assertEqual(len(q('button:disabled[type="submit"]')), 1) - self.assertEqual(len(q('button:enabled[type="submit"]')), 0) + self.assertEqual(len(q('#content button:disabled[type="submit"]')), 1) + self.assertEqual(len(q('#content button:enabled[type="submit"]')), 0) def test_allows_import_on_existing_bad_unicode(self): """Should not be able to import text identical to the current revision""" @@ -6290,7 +7521,7 @@ def test_allows_import_on_existing_bad_unicode(self): self.client.login(username='secretary', password='secretary+password') r = self.client.post(url, {'markdown_text': 'replaced below'}) # create a rev with open( - self.session.sessionpresentation_set.filter(document__type="minutes").first().document.get_file_name(), + self.session.presentations.filter(document__type="minutes").first().document.get_file_name(), 'wb' ) as f: # Replace existing content with an invalid Unicode byte string. The particular invalid @@ -6304,8 +7535,8 @@ def test_allows_import_on_existing_bad_unicode(self): r = self.client.get(url) # try to import the same text self.assertNotContains(r, "This document is identical", status_code=200) q = PyQuery(r.content) - self.assertEqual(len(q('button:enabled[type="submit"]')), 1) - self.assertEqual(len(q('button:disabled[type="submit"]')), 0) + self.assertEqual(len(q('#content button:enabled[type="submit"]')), 1) + self.assertEqual(len(q('#content button:disabled[type="submit"]')), 0) def test_handles_missing_previous_revision_file(self): """Should still allow import if the file for the previous revision is missing""" @@ -6315,9 +7546,11 @@ def test_handles_missing_previous_revision_file(self): self.client.login(username='secretary', password='secretary+password') r = self.client.post(url, {'markdown_text': 'original markdown text'}) # create a rev # remove the file uploaded for the first rev - minutes_docs = self.session.sessionpresentation_set.filter(document__type='minutes') + minutes_docs = self.session.presentations.filter(document__type='minutes') self.assertEqual(minutes_docs.count(), 1) - Path(minutes_docs.first().document.get_file_name()).unlink() + to_remove = Path(minutes_docs.first().document.get_file_name()) + to_remove.unlink() + remove_from_storage("minutes", to_remove.name) self.assertEqual(r.status_code, 302) with requests_mock.Mocker() as mock: @@ -6358,13 +7591,35 @@ def test_handles_notes_server_failure(self): class SessionTests(TestCase): + def test_get_summary_by_area(self): + meeting = make_meeting_test_data(meeting=MeetingFactory(type_id='ietf', number='100')) + sessions = Session.objects.filter(meeting=meeting).with_current_status() + data = get_summary_by_area(sessions) + self.assertEqual(data[0][0], 'Duration') + self.assertGreater(len(data), 2) + self.assertEqual(data[-1][0], 'Total Hours') + + def test_get_summary_by_type(self): + meeting = make_meeting_test_data(meeting=MeetingFactory(type_id='ietf', number='100')) + sessions = Session.objects.filter(meeting=meeting).with_current_status() + data = get_summary_by_type(sessions) + self.assertEqual(data[0][0], 'Group Type') + self.assertGreater(len(data), 2) + + def test_get_summary_by_purpose(self): + meeting = make_meeting_test_data(meeting=MeetingFactory(type_id='ietf', number='100')) + sessions = Session.objects.filter(meeting=meeting).with_current_status() + data = get_summary_by_purpose(sessions) + self.assertEqual(data[0][0], 'Purpose') + self.assertGreater(len(data), 2) + def test_meeting_requests(self): meeting = MeetingFactory(type_id='ietf') # a couple non-wg group types, confirm that their has_meetings features are as expected group_type_with_meetings = 'adhoc' self.assertTrue(GroupFeatures.objects.get(pk=group_type_with_meetings).has_meetings) - group_type_without_meetings = 'editorial' + group_type_without_meetings = 'sdo' self.assertFalse(GroupFeatures.objects.get(pk=group_type_without_meetings).has_meetings) area = GroupFactory(type_id='area', acronym='area') @@ -6430,9 +7685,23 @@ def test_meeting_requests(self): status_id='schedw', add_to_schedule=False, ) + session_with_none_purpose = SessionFactory( + meeting=meeting, + group__parent=area, + purpose_id="none", + status_id="schedw", + add_to_schedule=False, + ) + tutorial_session = SessionFactory( + meeting=meeting, + group__parent=area, + purpose_id="tutorial", + status_id="schedw", + add_to_schedule=False, + ) def _sreq_edit_link(sess): return urlreverse( - 'ietf.secr.sreq.views.edit', + 'ietf.meeting.views_session_request.edit_request', kwargs={ 'num': meeting.number, 'acronym': sess.group.acronym, @@ -6468,6 +7737,8 @@ def _sreq_edit_link(sess): self.assertContains(r, _sreq_edit_link(proposed_wg_session)) # link to the session request self.assertContains(r, rg_session.group.acronym) self.assertContains(r, _sreq_edit_link(rg_session)) # link to the session request + self.assertContains(r, session_with_none_purpose.group.acronym) + self.assertContains(r, tutorial_session.group.acronym) # check headings - note that the special types (has_meetings, etc) do not have a group parent # so they show up in 'other' q = PyQuery(r.content) @@ -6475,6 +7746,22 @@ def _sreq_edit_link(sess): self.assertEqual(len(q('h2#other-groups')), 1) self.assertEqual(len(q('h2#irtf')), 1) # rg group has irtf group as parent + # check rounded pills + self.assertNotContains( # no rounded pill for sessions with regular purpose + r, + 'Regular', + html=True, + ) + self.assertNotContains( # no rounded pill for session with no purpose specified + r, + 'None', + html=True, + ) + self.assertContains( # rounded pill for session with non-regular purpose + r, + 'Tutorial', + html=True, + ) def test_request_minutes(self): meeting = MeetingFactory(type_id='ietf') @@ -6497,6 +7784,156 @@ def test_request_minutes(self): self.assertEqual(r.status_code,302) self.assertEqual(len(outbox),1) + @override_settings(YOUTUBE_DOMAINS=["youtube.com"]) + def test_add_session_recordings(self): + session = SessionFactory(meeting__type_id="ietf") + url = urlreverse( + "ietf.meeting.views.add_session_recordings", + kwargs={"session_id": session.pk, "num": session.meeting.number}, + ) + # does not fully validate authorization for non-secretariat users :-( + login_testing_unauthorized(self, "secretary", url) + r = self.client.get(url) + pq = PyQuery(r.content) + title_input = pq("input#id_title") + self.assertIsNotNone(title_input) + self.assertEqual( + title_input.attr.value, + "Video recording of {acro} for {timestamp}".format( + acro=session.group.acronym, + timestamp=session.official_timeslotassignment().timeslot.utc_start_time().strftime( + "%Y-%m-%d %H:%M" + ), + ), + ) + + with patch("ietf.meeting.views.create_recording") as mock_create: + r = self.client.post( + url, + data={ + "title": "This is my video title", + "url": "", + } + ) + self.assertFalse(mock_create.called) + + with patch("ietf.meeting.views.create_recording") as mock_create: + r = self.client.post( + url, + data={ + "title": "This is my video title", + "url": "https://yubtub.com/this-is-not-a-youtube-video", + } + ) + self.assertFalse(mock_create.called) + + with patch("ietf.meeting.views.create_recording") as mock_create: + r = self.client.post( + url, + data={ + "title": "This is my video title", + "url": "https://youtube.com/finally-a-video", + } + ) + self.assertTrue(mock_create.called) + self.assertEqual( + mock_create.call_args, + call( + session, + "https://youtube.com/finally-a-video", + title="This is my video title", + user=Person.objects.get(user__username="secretary"), + ), + ) + + # CAN delete session presentation for this session + sp = SessionPresentationFactory( + session=session, + document__type_id="recording", + document__external_url="https://example.com/some-video", + ) + with patch("ietf.meeting.views.delete_recording") as mock_delete: + r = self.client.post( + url, + data={ + "delete": str(sp.pk), + } + ) + self.assertEqual(r.status_code, 200) + self.assertTrue(mock_delete.called) + self.assertEqual(mock_delete.call_args, call(sp)) + + # ValueError message from delete_recording does not reach the user + sp = SessionPresentationFactory( + session=session, + document__type_id="recording", + document__external_url="https://example.com/some-video", + ) + with patch("ietf.meeting.views.delete_recording", side_effect=ValueError("oh joy!")) as mock_delete: + r = self.client.post( + url, + data={ + "delete": str(sp.pk), + } + ) + self.assertTrue(mock_delete.called) + self.assertNotContains(r, "oh joy!", status_code=200) + + # CANNOT delete session presentation for a different session + sp_for_other_session = SessionPresentationFactory( + document__type_id="recording", + document__external_url="https://example.com/some-other-video", + ) + with patch("ietf.meeting.views.delete_recording") as mock_delete: + r = self.client.post( + url, + data={ + "delete": str(sp_for_other_session.pk), + } + ) + self.assertEqual(r.status_code, 404) + self.assertFalse(mock_delete.called) + + def test_show_chatlog_links(self): + meeting = MeetingFactory(type_id='ietf', number='122') + session = SessionFactory(meeting=meeting) + doc_name = 'chatlog-72-mars-197001010000' + SessionPresentation.objects.create(session=session,document=DocumentFactory(type_id='chatlog', name=doc_name)) + + session_url = urlreverse('ietf.meeting.views.session_details', + kwargs={'num':meeting.number, 'acronym':session.group.acronym}) + + r = self.client.get(session_url) + + self.assertEqual(r.status_code, 200) + + q = PyQuery(r.content) + + # Find the chatlog link in the desktop view + link_chatlog_box = q(f'a[title="Chat logs for {session.group.acronym}"]') + self.assertTrue(link_chatlog_box, 'Expected element with title "Chat logs for {group.acronym}" not found.') + self.assertEqual(link_chatlog_box.attr('href'), '/doc/'+ doc_name) + + # Find the chatlog link in the mobile view + link_chatlog_list = q('li:contains("Chat logs")') + self.assertTrue(link_chatlog_list, 'Expected
  • element containing "Chat logs" not found.') + self.assertEqual(link_chatlog_list.find('a').attr('href'), '/doc/'+ doc_name) + + def test_hide_chatlog_links(self): + # mock meeting and session, but no chatlog document + meeting = MeetingFactory(type_id='ietf', number='122') + session = SessionFactory(meeting=meeting) + + session_url = urlreverse('ietf.meeting.views.session_details', + kwargs={'num':meeting.number, 'acronym':session.group.acronym}) + + r = self.client.get(session_url) + + self.assertEqual(r.status_code, 200) + # validate no links for chat logs exist + self.assertNotContains(r, 'Chat logs') + + class HasMeetingsTests(TestCase): settings_temp_path_overrides = TestCase.settings_temp_path_overrides + ['AGENDA_PATH'] @@ -6591,10 +8028,7 @@ def test_cannot_request_interim(self): for gf in GroupFeatures.objects.filter(has_meetings=True): for role_name in all_role_names - set(gf.groupman_roles): role = RoleFactory(group__type_id=gf.type_id,name_id=role_name) - self.client.login(username=role.person.user.username, password=role.person.user.username+'+password') - r = self.client.get(url) - self.assertEqual(r.status_code, 403) - self.client.logout() + self.assertFalse(can_request_interim_meeting(role.person.user)) def test_appears_on_upcoming(self): url = urlreverse('ietf.meeting.views.upcoming') @@ -7278,8 +8712,7 @@ def _proceedings_file(): path = Path(settings.BASE_DIR) / 'meeting/test_procmat.pdf' return path.open('rb') - def _assertMeetingHostsDisplayed(self, response, meeting): - pq = PyQuery(response.content) + def _assertMeetingHostsDisplayed(self, pq: PyQuery, meeting): host_divs = pq('div.host-logo') self.assertEqual(len(host_divs), meeting.meetinghosts.count(), 'Should have a logo for every meeting host') self.assertEqual( @@ -7295,12 +8728,11 @@ def _assertMeetingHostsDisplayed(self, response, meeting): 'Correct image and name for each host should appear in the correct order' ) - def _assertProceedingsMaterialsDisplayed(self, response, meeting): + def _assertProceedingsMaterialsDisplayed(self, pq: PyQuery, meeting): """Checks that all (and only) active materials are linked with correct href and title""" expected_materials = [ m for m in meeting.proceedings_materials.order_by('type__order') if m.active() ] - pq = PyQuery(response.content) links = pq('div.proceedings-material a') self.assertEqual(len(links), len(expected_materials), 'Should have an entry for each active ProceedingsMaterial') self.assertEqual( @@ -7309,28 +8741,109 @@ def _assertProceedingsMaterialsDisplayed(self, response, meeting): 'Correct title and link for each ProceedingsMaterial should appear in the correct order' ) + def _assertGroupSessions(self, pq: PyQuery): + """Checks that group/sessions are present""" + sections = ["plenaries", "gen", "iab", "editorial", "irtf", "training"] + for section in sections: + self.assertEqual(len(pq(f"#{section}")), 1, f"{section} section should exists in proceedings") + def test_proceedings(self): - """Proceedings should be displayed correctly""" + """Proceedings should be displayed correctly + + Proceedings contents are tested in detail when testing generate_proceedings_content. + """ + # number must be >97 (settings.PROCEEDINGS_VERSION_CHANGES) meeting = make_meeting_test_data(meeting=MeetingFactory(type_id='ietf', number='100')) session = Session.objects.filter(meeting=meeting, group__acronym="mars").first() GroupEventFactory(group=session.group,type='status_update') SessionPresentationFactory(document__type_id='recording',session=session) SessionPresentationFactory(document__type_id='recording',session=session,document__title="Audio recording for tests") + # Add various group sessions + groups = [] + parent_groups = [ + GroupFactory.create(type_id="area", acronym="gen"), + GroupFactory.create(acronym="iab"), + GroupFactory.create(acronym="irtf"), + ] + for parent in parent_groups: + groups.append(GroupFactory.create(parent=parent)) + for acronym in ["rsab", "edu"]: + groups.append(GroupFactory.create(acronym=acronym)) + for group in groups: + SessionFactory(meeting=meeting, group=group) + self.write_materials_files(meeting, session) self._create_proceedings_materials(meeting) url = urlreverse("ietf.meeting.views.proceedings", kwargs=dict(num=meeting.number)) - r = self.client.get(url) + cached_content = mark_safe("

    Fake proceedings content

    ") + with patch("ietf.meeting.views.generate_proceedings_content") as mock_gpc: + mock_gpc.return_value = cached_content + r = self.client.get(url) self.assertEqual(r.status_code, 200) + self.assertIn(cached_content, r.content.decode()) + self.assertTemplateUsed(r, "meeting/proceedings_wrapper.html") + self.assertTemplateNotUsed(r, "meeting/proceedings.html") + # These are rendered in proceedings_wrapper.html, so test them here if len(meeting.city) > 0: self.assertContains(r, meeting.city) if len(meeting.venue_name) > 0: self.assertContains(r, meeting.venue_name) + self._assertMeetingHostsDisplayed(PyQuery(r.content), meeting) + + @patch("ietf.meeting.utils.caches") + def test_generate_proceedings_content(self, mock_caches): + # number must be >97 (settings.PROCEEDINGS_VERSION_CHANGES) + meeting = make_meeting_test_data(meeting=MeetingFactory(type_id='ietf', number='100')) + + # First, check that by default a value in the cache is used without doing any other computation + mock_default_cache = mock_caches["default"] + mock_default_cache.get.return_value = "a cached value" + result = generate_proceedings_content(meeting) + self.assertEqual(result, "a cached value") + self.assertFalse(mock_default_cache.set.called) + self.assertTrue(mock_default_cache.get.called) + cache_key = mock_default_cache.get.call_args.args[0] + mock_default_cache.get.reset_mock() + + # Now set up for actual computation of the proceedings content. + session = Session.objects.filter(meeting=meeting, group__acronym="mars").first() + GroupEventFactory(group=session.group,type='status_update') + SessionPresentationFactory(document__type_id='recording',session=session) + SessionPresentationFactory(document__type_id='recording',session=session,document__title="Audio recording for tests") + + # Add various group sessions + groups = [] + parent_groups = [ + GroupFactory.create(type_id="area", acronym="gen"), + GroupFactory.create(acronym="iab"), + GroupFactory.create(acronym="irtf"), + ] + for parent in parent_groups: + groups.append(GroupFactory.create(parent=parent)) + for acronym in ["rsab", "edu"]: + groups.append(GroupFactory.create(acronym=acronym)) + for group in groups: + SessionFactory(meeting=meeting, group=group) + + self.write_materials_files(meeting, session) + self._create_proceedings_materials(meeting) + + # Now "empty" the mock cache and see that we compute the expected proceedings content. + mock_default_cache.get.return_value = None + proceedings_content = generate_proceedings_content(meeting) + self.assertTrue(mock_default_cache.get.called) + self.assertEqual(mock_default_cache.get.call_args.args[0], cache_key, "same cache key each time") + self.assertTrue(mock_default_cache.set.called) + self.assertEqual(mock_default_cache.set.call_args.args, (cache_key, proceedings_content)) + self.assertGreater(mock_default_cache.set.call_args.kwargs["timeout"], 86400) + mock_default_cache.get.reset_mock() + mock_default_cache.set.reset_mock() # standard items on every proceedings - pq = PyQuery(r.content) + pq = PyQuery(proceedings_content) self.assertNotEqual( pq('a[href="{}"]'.format( urlreverse('ietf.meeting.views.proceedings_overview', kwargs=dict(num=meeting.number))) @@ -7347,7 +8860,7 @@ def test_proceedings(self): ) self.assertNotEqual( pq('a[href="{}"]'.format( - urlreverse('ietf.meeting.views.proceedings_progress_report', kwargs=dict(num=meeting.number))) + urlreverse('ietf.meeting.views.proceedings_activity_report', kwargs=dict(num=meeting.number))) ), [], 'Should have a link to activity report', @@ -7361,8 +8874,77 @@ def test_proceedings(self): ) # configurable contents - self._assertMeetingHostsDisplayed(r, meeting) - self._assertProceedingsMaterialsDisplayed(r, meeting) + self._assertProceedingsMaterialsDisplayed(pq, meeting) + self._assertGroupSessions(pq) + + # Finally, repeat the first cache test, but now with force_refresh=True. The cached value + # should be ignored and we should recompute the proceedings as before. + mock_default_cache.get.return_value = "a cached value" + result = generate_proceedings_content(meeting, force_refresh=True) + self.assertEqual(result, proceedings_content) # should have recomputed the same thing + self.assertFalse(mock_default_cache.get.called, "don't bother reading cache when force_refresh is True") + self.assertTrue(mock_default_cache.set.called) + self.assertEqual(mock_default_cache.set.call_args.args, (cache_key, proceedings_content)) + self.assertGreater(mock_default_cache.set.call_args.kwargs["timeout"], 86400) + + def test_named_session(self): + """Session with a name should appear separately in the proceedings""" + meeting = MeetingFactory(type_id='ietf', number='100', proceedings_final=True) + group = GroupFactory() + plain_session = SessionFactory(meeting=meeting, group=group) + named_session = SessionFactory(meeting=meeting, group=group, name='I Got a Name') + for doc_type_id in ('agenda', 'minutes', 'bluesheets', 'recording', 'slides', 'draft'): + # Set up sessions materials that will have distinct URLs for each session. + # This depends on settings.MEETING_DOC_HREFS and may need updating if that changes. + SessionPresentationFactory( + session=plain_session, + document__type_id=doc_type_id, + document__uploaded_filename=f'upload-{doc_type_id}-plain', + document__external_url=f'external_url-{doc_type_id}-plain', + ) + SessionPresentationFactory( + session=named_session, + document__type_id=doc_type_id, + document__uploaded_filename=f'upload-{doc_type_id}-named', + document__external_url=f'external_url-{doc_type_id}-named', + ) + + url = urlreverse('ietf.meeting.views.proceedings', kwargs={'num': meeting.number}) + r = self.client.get(url) + self.assertEqual(r.status_code, 200) + q = PyQuery(r.content) + + plain_label = q(f'div#{group.acronym}') + self.assertEqual(plain_label.text(), group.acronym) + plain_row = plain_label.closest('tr') + self.assertTrue(plain_row) + + named_label = q(f'div#{slugify(named_session.name)}') + self.assertEqual(named_label.text(), named_session.name) + named_row = named_label.closest('tr') + self.assertTrue(named_row) + + for material in (sp.document for sp in plain_session.presentations.all()): + if material.type_id == 'draft': + expected_url = urlreverse( + 'ietf.doc.views_doc.document_main', + kwargs={'name': material.name}, + ) + else: + expected_url = material.get_href(meeting) + self.assertTrue(plain_row.find(f'a[href="{expected_url}"]')) + self.assertFalse(named_row.find(f'a[href="{expected_url}"]')) + + for material in (sp.document for sp in named_session.presentations.all()): + if material.type_id == 'draft': + expected_url = urlreverse( + 'ietf.doc.views_doc.document_main', + kwargs={'name': material.name}, + ) + else: + expected_url = material.get_href(meeting) + self.assertFalse(plain_row.find(f'a[href="{expected_url}"]')) + self.assertTrue(named_row.find(f'a[href="{expected_url}"]')) def test_proceedings_no_agenda(self): # Meeting number must be larger than the last special-cased proceedings (currently 96) @@ -7421,47 +9003,69 @@ def test_proceedings_acknowledgements_link(self): 0, ) - @override_settings(STATS_REGISTRATION_ATTENDEES_JSON_URL='https://ietf.example.com/{number}') - @requests_mock.Mocker() - def test_proceedings_attendees(self, mock): - make_meeting_test_data() - meeting = MeetingFactory(type_id='ietf', date=datetime.date(2016,7,14), number="97") - mock.get( - settings.STATS_REGISTRATION_ATTENDEES_JSON_URL.format(number=meeting.number), - text=json.dumps([{"LastName": "Smith", "FirstName": "John", "Company": "ABC", "Country": "US"}]), - ) - finalize(meeting) - url = urlreverse('ietf.meeting.views.proceedings_attendees',kwargs={'num':97}) + def test_proceedings_attendees(self): + """Test proceedings attendee list. Check the following: + - assert onsite checkedin=True appears, not onsite checkedin=False + - assert remote attended appears, not remote not attended + - prefer onsite checkedin=True to remote attended when same person has both + - summary stats row shows correct counts + - chart data JSON is embedded with correct values + """ + + m = MeetingFactory(type_id='ietf', date=datetime.date(2023, 11, 4), number="118") + person_a = PersonFactory(name='Person A') + person_b = PersonFactory(name='Person B') + person_c = PersonFactory(name='Person C') + person_d = PersonFactory(name='Person D') + areg = RegistrationFactory(meeting=m, person=person_a, checkedin=True, with_ticket={'attendance_type_id': 'onsite'}) + RegistrationFactory(meeting=m, person=person_b, checkedin=False, with_ticket={'attendance_type_id': 'onsite'}) + creg = RegistrationFactory(meeting=m, person=person_c, with_ticket={'attendance_type_id': 'remote'}) + RegistrationFactory(meeting=m, person=person_d, with_ticket={'attendance_type_id': 'remote'}) + AttendedFactory(session__meeting=m, session__type_id='plenary', person=person_a) + AttendedFactory(session__meeting=m, session__type_id='plenary', person=person_c) + url = urlreverse('ietf.meeting.views.proceedings_attendees',kwargs={'num': 118}) response = self.client.get(url) self.assertContains(response, 'Attendee list') q = PyQuery(response.content) - self.assertEqual(1,len(q("#id_attendees tbody tr"))) - - @override_settings(STATS_REGISTRATION_ATTENDEES_JSON_URL='https://ietf.example.com/{number}') - @requests_mock.Mocker() - def test_proceedings_overview(self, mock): + self.assertEqual(2, len(q("#id_attendees tbody tr"))) + text = q('#id_attendees tbody tr').text().replace('\n', ' ') + self.assertEqual(text, f"A Person {areg.affiliation} {areg.country_code} onsite C Person {creg.affiliation} {creg.country_code} remote") + + # Summary stats row: Onsite / Remote / Total (matches registration.ietf.org) + self.assertContains(response, 'Onsite:') + self.assertContains(response, 'Remote:') + self.assertContains(response, 'Total:') + self.assertContains(response, '1') # onsite and remote + self.assertContains(response, '2') # total + + # Chart data embedded in page + chart_json = json.loads(q('#attendees-chart-data').text()) + self.assertEqual(chart_json['type'], [['Onsite', 1], ['Remote', 1]]) + + def test_proceedings_overview(self): '''Test proceedings IETF Overview page. Note: old meetings aren't supported so need to add a new meeting then test. ''' - make_meeting_test_data() - meeting = MeetingFactory(type_id='ietf', date=datetime.date(2016,7,14), number="97") - mock.get( - settings.STATS_REGISTRATION_ATTENDEES_JSON_URL.format(number=meeting.number), - text=json.dumps([{"LastName": "Smith", "FirstName": "John", "Company": "ABC", "Country": "US"}]), - ) - finalize(meeting) + meeting = make_meeting_test_data(meeting=MeetingFactory(type_id='ietf', date=datetime.date(2016,7,14), number="97")) + + # finalize meeting + url = urlreverse('ietf.meeting.views.finalize_proceedings',kwargs={'num':meeting.number}) + login_testing_unauthorized(self,"secretary",url) + r = self.client.post(url,{'finalize':1}) + self.assertEqual(r.status_code, 302) + url = urlreverse('ietf.meeting.views.proceedings_overview',kwargs={'num':97}) response = self.client.get(url) self.assertContains(response, 'The Internet Engineering Task Force') - def test_proceedings_progress_report(self): + def test_proceedings_activity_report(self): make_meeting_test_data() MeetingFactory(type_id='ietf', date=datetime.date(2016,4,3), number="96") MeetingFactory(type_id='ietf', date=datetime.date(2016,7,14), number="97") - url = urlreverse('ietf.meeting.views.proceedings_progress_report',kwargs={'num':97}) + url = urlreverse('ietf.meeting.views.proceedings_activity_report',kwargs={'num':97}) response = self.client.get(url) - self.assertContains(response, 'Progress Report') + self.assertContains(response, 'Activity Report') def test_feed(self): meeting = make_meeting_test_data() @@ -7626,12 +9230,13 @@ def test_add_proceedings_material_doc(self): """Upload proceedings materials document""" meeting = self._procmat_test_meeting() for mat_type in ProceedingsMaterialTypeName.objects.filter(used=True): - mat = self.upload_proceedings_material_test( - meeting, - mat_type, - {'file': self._proceedings_file(), 'external_url': ''}, - ) - self.assertEqual(mat.get_href(), f'{mat.document.name}:00') + with self._proceedings_file() as fd: + mat = self.upload_proceedings_material_test( + meeting, + mat_type, + {'file': fd, 'external_url': ''}, + ) + self.assertEqual(mat.get_href(), f'{mat.document.name}:00') def test_add_proceedings_material_doc_invalid_ext(self): """Upload proceedings materials document with disallowed extension""" @@ -7647,7 +9252,7 @@ def test_add_proceedings_material_doc_invalid_ext(self): invalid_file.seek(0) # read the file contents again r = self.client.post(url, {'file': invalid_file, 'external_url': ''}) self.assertEqual(r.status_code, 200) - self.assertFormError(r, 'form', 'file', 'Found an unexpected extension: .png. Expected one of .pdf') + self.assertFormError(r.context["form"], 'file', 'Found an unexpected extension: .png. Expected one of .pdf') def test_add_proceedings_material_doc_empty(self): """Upload proceedings materials document without specifying a file""" @@ -7660,7 +9265,7 @@ def test_add_proceedings_material_doc_empty(self): ) r = self.client.post(url, {'external_url': ''}) self.assertEqual(r.status_code, 200) - self.assertFormError(r, 'form', 'file', 'This field is required') + self.assertFormError(r.context["form"], 'file', 'This field is required') def test_add_proceedings_material_url(self): """Add a URL as proceedings material""" @@ -7684,7 +9289,7 @@ def test_add_proceedings_material_url_invalid(self): ) r = self.client.post(url, {'use_url': 'on', 'external_url': "Ceci n'est pas une URL"}) self.assertEqual(r.status_code, 200) - self.assertFormError(r, 'form', 'external_url', 'Enter a valid URL.') + self.assertFormError(r.context["form"], 'external_url', 'Enter a valid URL.') def test_add_proceedings_material_url_empty(self): """Add proceedings materials URL without specifying the URL""" @@ -7697,7 +9302,7 @@ def test_add_proceedings_material_url_empty(self): ) r = self.client.post(url, {'use_url': 'on', 'external_url': ''}) self.assertEqual(r.status_code, 200) - self.assertFormError(r, 'form', 'external_url', 'This field is required') + self.assertFormError(r.context["form"], 'external_url', 'This field is required') @override_settings(MEETING_DOC_HREFS={'procmaterials': '{doc.name}:{doc.rev}'}) def test_replace_proceedings_material(self): @@ -7718,12 +9323,13 @@ def test_replace_proceedings_material(self): kwargs=dict(num=meeting.number, material_type=pm_doc.type.slug), ) self.client.login(username='secretary', password='secretary+password') - r = self.client.post(pm_doc_url, {'file': self._proceedings_file(), 'external_url': ''}) - self.assertRedirects(r, success_url) - self.assertEqual(meeting.proceedings_materials.count(), 2) - pm_doc = meeting.proceedings_materials.get(pk=pm_doc.pk) # refresh from DB - self.assertEqual(pm_doc.document.rev, '01') - self.assertEqual(pm_doc.get_href(), f'{pm_doc.document.name}:01') + with self._proceedings_file() as fd: + r = self.client.post(pm_doc_url, {'file': fd, 'external_url': ''}) + self.assertRedirects(r, success_url) + self.assertEqual(meeting.proceedings_materials.count(), 2) + pm_doc = meeting.proceedings_materials.get(pk=pm_doc.pk) # refresh from DB + self.assertEqual(pm_doc.document.rev, '01') + self.assertEqual(pm_doc.get_href(), f'{pm_doc.document.name}:01') # Replace the uploaded document with a URL r = self.client.post(pm_doc_url, {'use_url': 'on', 'external_url': 'https://example.com/second'}) @@ -7746,12 +9352,13 @@ def test_replace_proceedings_material(self): self.assertEqual(pm_url.get_href(), 'https://example.com/third') # Now replace the URL doc with an uploaded file - r = self.client.post(pm_url_url, {'file': self._proceedings_file(), 'external_url': ''}) - self.assertRedirects(r, success_url) - self.assertEqual(meeting.proceedings_materials.count(), 2) - pm_url = meeting.proceedings_materials.get(pk=pm_url.pk) # refresh from DB - self.assertEqual(pm_url.document.rev, '02') - self.assertEqual(pm_url.get_href(), f'{pm_url.document.name}:02') + with self._proceedings_file() as fd: + r = self.client.post(pm_url_url, {'file': fd, 'external_url': ''}) + self.assertRedirects(r, success_url) + self.assertEqual(meeting.proceedings_materials.count(), 2) + pm_url = meeting.proceedings_materials.get(pk=pm_url.pk) # refresh from DB + self.assertEqual(pm_url.document.rev, '02') + self.assertEqual(pm_url.get_href(), f'{pm_url.document.name}:02') def test_remove_proceedings_material(self): """Proceedings material can be removed""" @@ -7815,3 +9422,158 @@ def test_rename_proceedings_material(self): pm = meeting.proceedings_materials.get(pk=pm.pk) self.assertEqual(str(pm), 'This Is Not the Default Name') self.assertEqual(pm.document.rev, orig_rev, 'Renaming should not change document revision') + + def test_create_recording(self): + session = SessionFactory(meeting__type_id='ietf', meeting__number=72, group__acronym='mars') + filename = 'ietf42-testroomt-20000101-0800.mp3' + url = settings.IETF_AUDIO_URL + 'ietf{}/{}'.format(session.meeting.number, filename) + doc = create_recording(session, url) + self.assertEqual(doc.name,'recording-72-mars-1') + self.assertEqual(doc.group,session.group) + self.assertEqual(doc.external_url,url) + self.assertTrue(doc in session.materials.all()) + + def test_get_next_sequence(self): + session = SessionFactory(meeting__type_id='ietf', meeting__number=72, group__acronym='mars') + meeting = session.meeting + group = session.group + sequence = get_next_sequence(group,meeting,'recording') + self.assertEqual(sequence,1) + + def test_participants_for_meeting(self): + m = MeetingFactory.create(type_id='ietf') + areg = RegistrationFactory(meeting=m, checkedin=True, with_ticket={'attendance_type_id': 'onsite'}) + breg = RegistrationFactory(meeting=m, checkedin=False, with_ticket={'attendance_type_id': 'onsite'}) + creg = RegistrationFactory(meeting=m, with_ticket={'attendance_type_id': 'remote'}) + dreg = RegistrationFactory(meeting=m, with_ticket={'attendance_type_id': 'remote'}) + AttendedFactory(session__meeting=m, session__type_id='plenary', person=creg.person) + checked_in, attended = participants_for_meeting(m) + self.assertIn(areg.person.pk, checked_in) + self.assertNotIn(breg.person.pk, checked_in) + self.assertNotIn(areg.person.pk, attended) + self.assertNotIn(breg.person.pk, attended) + self.assertIn(creg.person.pk, attended) + self.assertNotIn(dreg.person.pk, attended) + + def test_session_attendance(self): + meeting = MeetingFactory(type_id='ietf', date=datetime.date(2023, 11, 4), number='118') + make_meeting_test_data(meeting=meeting) + session = Session.objects.filter(meeting=meeting, group__acronym='mars').first() + regs = RegistrationFactory.create_batch(3, meeting=meeting) + persons = [reg.person for reg in regs] + self.assertEqual(session.attended_set.count(), 0) + + # If there are no attendees, the link isn't offered, and getting + # the page directly returns an empty list. + session_url = urlreverse('ietf.meeting.views.session_details', kwargs={'num':meeting.number, 'acronym':session.group.acronym}) + attendance_url = urlreverse('ietf.meeting.views.session_attendance', kwargs={'num':meeting.number, 'session_id':session.id}) + r = self.client.get(session_url) + self.assertNotContains(r, attendance_url) + r = self.client.get(attendance_url) + self.assertEqual(r.status_code, 200) + self.assertContains(r, '0 attendees') + + # Add some attendees + add_attendees_url = urlreverse('ietf.meeting.views.api_add_session_attendees') + recmanrole = RoleFactory(group__type_id='ietf', name_id='recman', person__user__last_login=timezone.now()) + recman = recmanrole.person + apikey = PersonalApiKeyFactory(endpoint=add_attendees_url, person=recman) + attendees = [person.user.pk for person in persons] + self.client.login(username='recman', password='recman+password') + r = self.client.post(add_attendees_url, {'apikey':apikey.hash(), 'attended':f'{{"session_id":{session.pk},"attendees":{attendees}}}'}) + self.assertEqual(r.status_code, 200) + self.assertEqual(session.attended_set.count(), 3) + + # Before a meeting is finalized, session_attendance renders a live + # view of the Attended records for the session. + r = self.client.get(session_url) + self.assertContains(r, attendance_url) + r = self.client.get(attendance_url) + self.assertEqual(r.status_code, 200) + self.assertContains(r, '3 attendees') + for person in persons: + self.assertContains(r, escape(person.plain_name())) + + # Test for the "I was there" button. + def _test_button(person, expected): + username = person.user.username + self.client.login(username=username, password=f'{username}+password') + r = self.client.get(attendance_url) + self.assertEqual(b"I was there" in r.content, expected) + # recman isn't registered for the meeting + _test_button(recman, False) + # person0 is already on the bluesheet + _test_button(persons[0], False) + # person3 attests he was there + persons.append(RegistrationFactory(meeting=meeting).person) + # button isn't shown if we're outside the corrections windows + meeting.importantdate_set.create(name_id='revsub',date=date_today() - datetime.timedelta(days=20)) + _test_button(persons[3], False) + # attempt to POST anyway is ignored + r = self.client.post(attendance_url) + self.assertEqual(r.status_code, 200) + self.assertNotContains(r, escape(persons[3].plain_name())) + self.assertEqual(session.attended_set.count(), 3) + # button is shown, and POST is accepted + meeting.importantdate_set.update(name_id='revsub',date=date_today() + datetime.timedelta(days=20)) + _test_button(persons[3], True) + r = self.client.post(attendance_url) + self.assertEqual(r.status_code, 200) + self.assertContains(r, escape(persons[3].plain_name())) + self.assertEqual(session.attended_set.count(), 4) + + # When the meeting is finalized, a bluesheet file is generated, + # and session_attendance redirects to the file. + self.client.login(username='secretary',password='secretary+password') + finalize_url = urlreverse('ietf.meeting.views.finalize_proceedings', kwargs={'num':meeting.number}) + r = self.client.post(finalize_url, {'finalize':1}) + self.assertRedirects(r, urlreverse('ietf.meeting.views.proceedings', kwargs={'num':meeting.number})) + doc = session.presentations.filter(document__type_id='bluesheets').first().document + self.assertEqual(doc.rev,'00') + text = doc.text() + self.assertIn('4 attendees', text) + for person in persons: + self.assertIn(person.plain_name(), text) + r = self.client.get(session_url) + self.assertContains(r, doc.get_href()) + self.assertNotContains(r, attendance_url) + r = self.client.get(attendance_url) + self.assertEqual(r.status_code,302) + self.assertEqual(r['Location'],doc.get_href()) + + # An interim meeting is considered finalized immediately. + meeting = make_interim_meeting(group=GroupFactory(acronym='mars'), date=date_today()) + session = Session.objects.filter(meeting=meeting, group__acronym='mars').first() + attendance_url = urlreverse('ietf.meeting.views.session_attendance', kwargs={'num':meeting.number, 'session_id':session.id}) + self.assertEqual(session.attended_set.count(), 0) + self.client.login(username='recman', password='recman+password') + attendees = [person.user.pk for person in persons] + r = self.client.post(add_attendees_url, {'apikey':apikey.hash(), 'attended':f'{{"session_id":{session.pk},"attendees":{attendees}}}'}) + self.assertEqual(r.status_code, 200) + self.assertEqual(session.attended_set.count(), 4) + doc = session.presentations.filter(document__type_id='bluesheets').first().document + self.assertEqual(doc.rev,'00') + session_url = urlreverse('ietf.meeting.views.session_details', kwargs={'num':meeting.number, 'acronym':session.group.acronym}) + r = self.client.get(session_url) + self.assertContains(r, doc.get_href()) + self.assertNotContains(r, attendance_url) + r = self.client.get(attendance_url) + self.assertEqual(r.status_code,302) + self.assertEqual(r['Location'],doc.get_href()) + + def test_bluesheet_data(self): + session = SessionFactory(meeting__type_id="ietf") + attended_with_affil = RegistrationFactory(meeting=session.meeting, affiliation="Somewhere") + AttendedFactory(session=session, person=attended_with_affil.person, time="2023-03-13T01:24:00Z") # joined 2nd + attended_no_affil = RegistrationFactory(meeting=session.meeting, affiliation="") + AttendedFactory(session=session, person=attended_no_affil.person, time="2023-03-13T01:23:00Z") # joined 1st + RegistrationFactory(meeting=session.meeting) # did not attend + + data = bluesheet_data(session) + self.assertEqual( + data, + [ + {"name": attended_no_affil.person.plain_name(), "affiliation": ""}, + {"name": attended_with_affil.person.plain_name(), "affiliation": "Somewhere"}, + ] + ) diff --git a/ietf/meeting/urls.py b/ietf/meeting/urls.py index 3beb23b0be..a038e1cfe6 100644 --- a/ietf/meeting/urls.py +++ b/ietf/meeting/urls.py @@ -1,27 +1,29 @@ -# Copyright The IETF Trust 2007-2020, All Rights Reserved +# Copyright The IETF Trust 2007-2025, All Rights Reserved -from django.conf.urls import include -from django.views.generic import RedirectView from django.conf import settings +from django.urls import include +from django.views.generic import RedirectView -from ietf.meeting import views, views_proceedings +from ietf.meeting import views, views_proceedings, views_session_request from ietf.utils.urls import url class AgendaRedirectView(RedirectView): ignore_kwargs = ('owner', 'name') def get_redirect_url(self, *args, **kwargs): - for kwarg in self.ignore_kwargs: - kwargs.pop(kwarg, None) + kwargs = {k: v for k, v in kwargs.items() if v is not None and k not in self.ignore_kwargs} return super().get_redirect_url(*args, **kwargs) safe_for_all_meeting_types = [ url(r'^session/(?P[-a-z0-9]+)/?$', views.session_details), + url(r'^session/(?P[-a-z0-9]+)/send_slide_notifications$', views.notify_meetecho_of_all_slides), url(r'^session/(?P\d+)/drafts$', views.add_session_drafts), + url(r'^session/(?P\d+)/recordings$', views.add_session_recordings), + url(r'^session/(?P\d+)/attendance$', views.session_attendance), url(r'^session/(?P\d+)/bluesheets$', views.upload_session_bluesheets), url(r'^session/(?P\d+)/minutes$', views.upload_session_minutes), + url(r'^session/(?P\d+)/narrativeminutes$', views.upload_session_narrativeminutes), url(r'^session/(?P\d+)/agenda$', views.upload_session_agenda), url(r'^session/(?P\d+)/import/minutes$', views.import_session_minutes), - url(r'^session/(?P\d+)/propose_slides$', views.propose_session_slides), url(r'^session/(?P\d+)/slides(?:/%(name)s)?$' % settings.URL_REGEXPS, views.upload_session_slides), url(r'^session/(?P\d+)/add_to_session$', views.ajax_add_slides_to_session), url(r'^session/(?P\d+)/remove_from_session$', views.ajax_remove_slides_from_session), @@ -29,7 +31,7 @@ def get_redirect_url(self, *args, **kwargs): url(r'^session/(?P\d+)/doc/%(name)s/remove$' % settings.URL_REGEXPS, views.remove_sessionpresentation), url(r'^session/(?P\d+)\.ics$', views.agenda_ical), url(r'^sessions/(?P[-a-z0-9]+)\.ics$', views.agenda_ical), - url(r'^slidesubmission/(?P\d+)$', views.approve_proposed_slides) + url(r'^slidesubmission/(?P\d+)$', views.approve_proposed_slides), ] @@ -63,14 +65,14 @@ def get_redirect_url(self, *args, **kwargs): type_interim_patterns = [ url(r'^agenda/(?P[A-Za-z0-9-]+)-drafts.pdf$', views.session_draft_pdf), url(r'^agenda/(?P[A-Za-z0-9-]+)-drafts.tgz$', views.session_draft_tarfile), - url(r'^materials/%(document)s((?P\.[a-z0-9]+)|/)?$' % settings.URL_REGEXPS, views.materials_document), + url(r'^materials/%(document)s(?P\.[A-Za-z0-9]+)$' % settings.URL_REGEXPS, views.materials_document), + url(r'^materials/%(document)s/?$' % settings.URL_REGEXPS, views.materials_document), url(r'^agenda.json$', views.agenda_json) ] type_ietf_only_patterns_id_optional = [ url(r'^agenda(?P-utc)?(?P\.html)?/?$', views.agenda, name='agenda'), - url(r'^agenda(?P\.txt)$', views.agenda_plain), - url(r'^agenda(?P\.csv)$', views.agenda_plain), + url(r'^agenda(?P-utc)?(?P\.txt|\.csv)$', views.agenda_plain), url(r'^agenda/edit$', RedirectView.as_view(pattern_name='ietf.meeting.views.edit_meeting_schedule', permanent=True), name='ietf.meeting.views.edit_meeting_schedule'), @@ -79,20 +81,19 @@ def get_redirect_url(self, *args, **kwargs): url(r'^agenda/agenda\.ics$', views.agenda_ical), url(r'^agenda\.ics$', views.agenda_ical), url(r'^agenda.json$', views.agenda_json), - url(r'^agenda/week-view(?:.html)?/?$', RedirectView.as_view(pattern_name='agenda', permanent=True)), + url(r'^agenda/week-view(?:.html)?/?$', AgendaRedirectView.as_view(pattern_name='agenda', permanent=True)), url(r'^floor-plan/?$', views.agenda, name='floor-plan'), - url(r'^floor-plan/(?P[-a-z0-9_]+)/?$', RedirectView.as_view(pattern_name='floor-plan', permanent=True)), - url(r'^week-view(?:.html)?/?$', RedirectView.as_view(pattern_name='agenda', permanent=True)), + url(r'^week-view(?:.html)?/?$', AgendaRedirectView.as_view(pattern_name='agenda', permanent=True)), url(r'^materials(?:.html)?/?$', views.materials), url(r'^request_minutes/?$', views.request_minutes), - url(r'^materials/%(document)s((?P\.[a-z0-9]+)|/)?$' % settings.URL_REGEXPS, views.materials_document), + url(r'^materials/%(document)s(?P\.[A-Za-z0-9]+)?/?$' % settings.URL_REGEXPS, views.materials_document), url(r'^session/?$', views.materials_editable_groups), url(r'^proceedings(?:.html)?/?$', views.proceedings), url(r'^proceedings(?:.html)?/finalize/?$', views.finalize_proceedings), url(r'^proceedings/acknowledgements/$', views.proceedings_acknowledgements), url(r'^proceedings/attendees/$', views.proceedings_attendees), url(r'^proceedings/overview/$', views.proceedings_overview), - url(r'^proceedings/progress-report/$', views.proceedings_progress_report), + url(r'^proceedings/activity-report/$', views.proceedings_activity_report), url(r'^proceedings/materials/$', views_proceedings.material_details), url(r'^proceedings/materials/(?P[a-z_]+)/$', views_proceedings.edit_material), url(r'^proceedings/materials/(?P[a-z_]+)/new/$', views_proceedings.upload_material), @@ -108,12 +109,13 @@ def get_redirect_url(self, *args, **kwargs): url(r'^important-dates.(?Pics)$', views.important_dates), url(r'^proceedings/meetinghosts/edit/', views_proceedings.edit_meetinghosts), url(r'^proceedings/meetinghosts/(?P\d+)/logo/$', views_proceedings.meetinghost_logo), + url(r'^session/request/%(acronym)s/edit/$' % settings.URL_REGEXPS, views_session_request.edit_request), + url(r'^session/request/%(acronym)s/view/$' % settings.URL_REGEXPS, views_session_request.view_request), ] urlpatterns = [ # First patterns which start with unique strings url(r'^$', views.current_materials), - url(r'^ajax/get-utc/?$', views.ajax_get_utc), url(r'^interim/announce/?$', views.interim_announce), url(r'^interim/announce/(?P[A-Za-z0-9._+-]+)/?$', views.interim_send_announcement), url(r'^interim/skip_announce/(?P[A-Za-z0-9._+-]+)/?$', views.interim_skip_announcement), @@ -128,6 +130,13 @@ def get_redirect_url(self, *args, **kwargs): url(r'^upcoming/?$', views.upcoming), url(r'^upcoming\.ics/?$', views.upcoming_ical), url(r'^upcoming\.json/?$', views.upcoming_json), + url(r'^session/request/$', views_session_request.list_view), + url(r'^session/request/%(acronym)s/new/$' % settings.URL_REGEXPS, views_session_request.new_request), + url(r'^session/request/%(acronym)s/approve/$' % settings.URL_REGEXPS, views_session_request.approve_request), + url(r'^session/request/%(acronym)s/no_session/$' % settings.URL_REGEXPS, views_session_request.no_session), + url(r'^session/request/%(acronym)s/cancel/$' % settings.URL_REGEXPS, views_session_request.cancel_request), + url(r'^session/request/%(acronym)s/confirm/$' % settings.URL_REGEXPS, views_session_request.confirm), + url(r'^session/request/status/$', views_session_request.status), url(r'^session/(?P\d+)/agenda_materials$', views.session_materials), url(r'^session/(?P\d+)/cancel/?', views.cancel_session), url(r'^session/(?P\d+)/edit/?', views.edit_session), @@ -141,4 +150,3 @@ def get_redirect_url(self, *args, **kwargs): url(r'^(?P\d+)/', include(safe_for_all_meeting_types)), url(r'^(?Pinterim-[a-z0-9-]+)/', include(safe_for_all_meeting_types)), ] - diff --git a/ietf/meeting/utils.py b/ietf/meeting/utils.py index e8efb92ad0..10ae0d3667 100644 --- a/ietf/meeting/utils.py +++ b/ietf/meeting/utils.py @@ -1,38 +1,71 @@ -# Copyright The IETF Trust 2016-2020, All Rights Reserved +# Copyright The IETF Trust 2016-2024, All Rights Reserved # -*- coding: utf-8 -*- import datetime import itertools -import pytz +from contextlib import suppress +from dataclasses import dataclass + +import jsonschema +import os import requests + +import pytz import subprocess from collections import defaultdict from pathlib import Path -from urllib.error import HTTPError from django.conf import settings from django.contrib import messages +from django.core.cache import caches +from django.core.files.base import ContentFile +from django.db import IntegrityError +from django.db.models import OuterRef, Subquery, TextField, Q, Value, Max +from django.db.models.functions import Coalesce from django.template.loader import render_to_string from django.utils import timezone -from django.utils.encoding import smart_text +from django.utils.encoding import smart_str import debug # pyflakes:ignore from ietf.dbtemplate.models import DBTemplate -from ietf.meeting.models import Session, SchedulingEvent, TimeSlot, Constraint, SchedTimeSessAssignment -from ietf.doc.models import Document, DocAlias, State, NewRevisionDocEvent +from ietf.doc.storage_utils import store_bytes, store_str, AlreadyExistsError +from ietf.meeting.models import ( + Session, + SchedulingEvent, + TimeSlot, + Constraint, + SchedTimeSessAssignment, + SessionPresentation, + Attended, + Registration, + Meeting, + RegistrationTicket, +) +from ietf.blobdb.models import ResolvedMaterial +from ietf.doc.models import ( + Document, + State, + NewRevisionDocEvent, + StateDocEvent, + StoredObject, +) +from ietf.doc.models import DocEvent from ietf.group.models import Group from ietf.group.utils import can_manage_materials from ietf.name.models import SessionStatusName, ConstraintName, DocTypeName from ietf.person.models import Person -from ietf.secr.proceedings.proc_utils import import_audio_files -from ietf.utils.html import sanitize_document +from ietf.utils import markdown +from ietf.utils.html import clean_html from ietf.utils.log import log from ietf.utils.timezone import date_today def session_time_for_sorting(session, use_meeting_date): - official_timeslot = TimeSlot.objects.filter(sessionassignments__session=session, sessionassignments__schedule__in=[session.meeting.schedule, session.meeting.schedule.base if session.meeting.schedule else None]).first() + if hasattr(session, "_otsa"): + official_timeslot=session._otsa.timeslot + else: + official_timeslot = TimeSlot.objects.filter(sessionassignments__session=session, sessionassignments__schedule__in=[session.meeting.schedule, session.meeting.schedule.base if session.meeting.schedule else None]).first() if official_timeslot: return official_timeslot.time elif use_meeting_date and session.meeting.date: @@ -75,13 +108,14 @@ def group_sessions(sessions): in_progress = [] recent = [] past = [] + for s in sessions: today = date_today(s.meeting.tz()) if s.meeting.date > today: future.append(s) elif s.meeting.end_date() >= today: in_progress.append(s) - elif not s.is_material_submission_cutoff(): + elif not getattr(s, "cached_is_cutoff", lambda: s.is_material_submission_cutoff): recent.append(s) else: past.append(s) @@ -91,6 +125,7 @@ def group_sessions(sessions): recent.reverse() past.reverse() + return future, in_progress, recent, past def get_upcoming_manageable_sessions(user): @@ -124,31 +159,7 @@ def sort_sessions(sessions): return sorted(sessions, key=lambda s: (s.meeting.number, s.group.acronym, session_time_for_sorting(s, use_meeting_date=False))) def create_proceedings_templates(meeting): - '''Create DBTemplates for meeting proceedings''' - # Get meeting attendees from registration system - url = settings.STATS_REGISTRATION_ATTENDEES_JSON_URL.format(number=meeting.number) - try: - attendees = requests.get(url, timeout=settings.DEFAULT_REQUESTS_TIMEOUT).json() - except (ValueError, HTTPError, requests.Timeout) as exc: - attendees = [] - log(f'Failed to retrieve meeting attendees from [{url}]: {exc}') - - if attendees: - attendees = sorted(attendees, key = lambda a: a['LastName']) - content = render_to_string('meeting/proceedings_attendees_table.html', { - 'attendees':attendees}) - try: - template = DBTemplate.objects.get(path='/meeting/proceedings/%s/attendees.html' % (meeting.number, )) - template.title='IETF %s Attendee List' % meeting.number - template.type_id='django' - template.content=content - template.save() - except DBTemplate.DoesNotExist: - DBTemplate.objects.create( - path='/meeting/proceedings/%s/attendees.html' % (meeting.number, ), - title='IETF %s Attendee List' % meeting.number, - type_id='django', - content=content) + '''Create DBTemplates for meeting proceedings''' # Make copy of default IETF Overview template if not meeting.overview: path = '/meeting/proceedings/%s/overview.rst' % (meeting.number, ) @@ -163,7 +174,88 @@ def create_proceedings_templates(meeting): meeting.overview = template meeting.save() -def finalize(meeting): + +def bluesheet_data(session): + attendance = ( + Attended.objects.filter(session=session) + .annotate( + affiliation=Coalesce( + Subquery( + Registration.objects.filter( + Q(meeting=session.meeting), + Q(person=OuterRef("person")) | Q(email=OuterRef("person__email")), + ).values("affiliation")[:1] + ), + Value(""), + output_field=TextField(), + ) + ).distinct() + .order_by("time") + ) + + return [ + { + "name": attended.person.plain_name(), + "affiliation": attended.affiliation, + } + for attended in attendance + ] + + +def save_bluesheet(request, session, file, encoding='utf-8'): + bluesheet_sp = session.presentations.filter(document__type='bluesheets').first() + _, ext = os.path.splitext(file.name) + + if bluesheet_sp: + doc = bluesheet_sp.document + doc.rev = '%02d' % (int(doc.rev)+1) + bluesheet_sp.rev = doc.rev + bluesheet_sp.save() + else: + ota = session.official_timeslotassignment() + sess_time = ota and ota.timeslot.time + + if session.meeting.type_id=='ietf': + name = 'bluesheets-%s-%s-%s' % (session.meeting.number, + session.group.acronym, + sess_time.strftime("%Y%m%d%H%M")) + title = 'Bluesheets IETF%s: %s : %s' % (session.meeting.number, + session.group.acronym, + sess_time.strftime("%a %H:%M")) + else: + name = 'bluesheets-%s-%s' % (session.meeting.number, sess_time.strftime("%Y%m%d%H%M")) + title = 'Bluesheets %s: %s' % (session.meeting.number, sess_time.strftime("%a %H:%M")) + doc = Document.objects.create( + name = name, + type_id = 'bluesheets', + title = title, + group = session.group, + rev = '00', + ) + doc.states.add(State.objects.get(type_id='bluesheets',slug='active')) + session.presentations.create(document=doc,rev='00') + filename = '%s-%s%s'% ( doc.name, doc.rev, ext) + doc.uploaded_filename = filename + e = NewRevisionDocEvent.objects.create(doc=doc, rev=doc.rev, by=request.user.person, type='new_revision', desc='New revision available: %s'%doc.rev) + save_error = handle_upload_file(file, filename, session.meeting, 'bluesheets', request=request, encoding=encoding) + if not save_error: + doc.save_with_history([e]) + resolve_uploaded_material(meeting=session.meeting, doc=doc) + return save_error + + +def generate_bluesheet(request, session): + data = bluesheet_data(session) + if not data: + return + text = render_to_string('meeting/bluesheet.txt', { + 'session': session, + 'data': data, + }) + return save_bluesheet(request, session, ContentFile(text.encode("utf-8"), name="unusednamepartsothereisanextension.txt")) + + +def finalize(request, meeting): end_date = meeting.end_date() end_time = meeting.tz().localize( datetime.datetime.combine( @@ -172,15 +264,20 @@ def finalize(meeting): ) ).astimezone(pytz.utc) + datetime.timedelta(days=1) for session in meeting.session_set.all(): - for sp in session.sessionpresentation_set.filter(document__type='draft',rev=None): + for sp in session.presentations.filter(document__type='draft',rev=None): rev_before_end = [e for e in sp.document.docevent_set.filter(newrevisiondocevent__isnull=False).order_by('-time') if e.time <= end_time ] if rev_before_end: sp.rev = rev_before_end[-1].newrevisiondocevent.rev else: sp.rev = '00' sp.save() + + # Don't try to generate a bluesheet if it's before we had Attended records. + if int(meeting.number) >= 108: + save_error = generate_bluesheet(request, session) + if save_error: + messages.error(request, save_error) - import_audio_files(meeting) create_proceedings_templates(meeting) meeting.proceedings_final = True meeting.save() @@ -205,7 +302,7 @@ def sort_accept_tuple(accept): return tup def condition_slide_order(session): - qs = session.sessionpresentation_set.filter(document__type_id='slides').order_by('order') + qs = session.presentations.filter(document__type_id='slides').order_by('order') order_list = qs.values_list('order',flat=True) if list(order_list) != list(range(1,qs.count()+1)): for num, sp in enumerate(qs, start=1): @@ -543,7 +640,8 @@ def bulk_create_timeslots(meeting, times, locations, other_props): def preprocess_meeting_important_dates(meetings): for m in meetings: - m.cached_updated = m.updated() + # cached_updated must be present, set it to 1970-01-01 if necessary + m.cached_updated = m.updated() or pytz.utc.localize(datetime.datetime(1970, 1, 1, 0, 0, 0)) m.important_dates = m.importantdate_set.prefetch_related("name") for d in m.important_dates: d.midnight_cutoff = "UTC 23:59" in d.name.name @@ -575,7 +673,7 @@ class SaveMaterialsError(Exception): pass -def save_session_minutes_revision(session, file, ext, request, encoding=None, apply_to_all=False): +def save_session_minutes_revision(session, file, ext, request, encoding=None, apply_to_all=False, narrative=False): """Creates or updates session minutes records This updates the database models to reflect a new version. It does not handle @@ -588,7 +686,8 @@ def save_session_minutes_revision(session, file, ext, request, encoding=None, ap Returns (Document, [DocEvents]), which should be passed to doc.save_with_history() if the file contents are stored successfully. """ - minutes_sp = session.sessionpresentation_set.filter(document__type='minutes').first() + document_type = DocTypeName.objects.get(slug= 'narrativeminutes' if narrative else 'minutes') + minutes_sp = session.presentations.filter(document__type=document_type).first() if minutes_sp: doc = minutes_sp.document doc.rev = '%02d' % (int(doc.rev)+1) @@ -600,40 +699,37 @@ def save_session_minutes_revision(session, file, ext, request, encoding=None, ap if not sess_time: raise SessionNotScheduledError if session.meeting.type_id=='ietf': - name = 'minutes-%s-%s' % (session.meeting.number, - session.group.acronym) - title = 'Minutes IETF%s: %s' % (session.meeting.number, - session.group.acronym) + name = f"{document_type.prefix}-{session.meeting.number}-{session.group.acronym}" + title = f"{document_type.name} IETF{session.meeting.number}: {session.group.acronym}" if not apply_to_all: name += '-%s' % (sess_time.strftime("%Y%m%d%H%M"),) title += ': %s' % (sess_time.strftime("%a %H:%M"),) else: - name = 'minutes-%s-%s' % (session.meeting.number, sess_time.strftime("%Y%m%d%H%M")) - title = 'Minutes %s: %s' % (session.meeting.number, sess_time.strftime("%a %H:%M")) + name =f"{document_type.prefix}-{session.meeting.number}-{sess_time.strftime('%Y%m%d%H%M')}" + title = f"{document_type.name} {session.meeting.number}: {sess_time.strftime('%a %H:%M')}" if Document.objects.filter(name=name).exists(): doc = Document.objects.get(name=name) doc.rev = '%02d' % (int(doc.rev)+1) else: doc = Document.objects.create( name = name, - type_id = 'minutes', + type = document_type, title = title, group = session.group, rev = '00', ) - DocAlias.objects.create(name=doc.name).docs.add(doc) - doc.states.add(State.objects.get(type_id='minutes',slug='active')) - if session.sessionpresentation_set.filter(document=doc).exists(): - sp = session.sessionpresentation_set.get(document=doc) + doc.states.add(State.objects.get(type_id=document_type.slug,slug='active')) + if session.presentations.filter(document=doc).exists(): + sp = session.presentations.get(document=doc) sp.rev = doc.rev sp.save() else: - session.sessionpresentation_set.create(document=doc,rev=doc.rev) + session.presentations.create(document=doc,rev=doc.rev) if apply_to_all: for other_session in get_meeting_sessions(session.meeting.number, session.group.acronym): if other_session != session: - other_session.sessionpresentation_set.filter(document__type='minutes').delete() - other_session.sessionpresentation_set.create(document=doc,rev=doc.rev) + other_session.presentations.filter(document__type=document_type).delete() + other_session.presentations.create(document=doc,rev=doc.rev) filename = f'{doc.name}-{doc.rev}{ext}' doc.uploaded_filename = filename e = NewRevisionDocEvent.objects.create( @@ -649,7 +745,7 @@ def save_session_minutes_revision(session, file, ext, request, encoding=None, ap file=file, filename=doc.uploaded_filename, meeting=session.meeting, - subdir='minutes', + subdir=document_type.slug, request=request, encoding=encoding, ) @@ -662,31 +758,27 @@ def save_session_minutes_revision(session, file, ext, request, encoding=None, ap def handle_upload_file(file, filename, meeting, subdir, request=None, encoding=None): """Accept an uploaded materials file - This function takes a file object, a filename and a meeting object and subdir as string. + This function takes a _binary mode_ file object, a filename and a meeting object and subdir as string. It saves the file to the appropriate directory, get_materials_path() + subdir. - If the file is a zip file, it creates a new directory in 'slides', which is the basename of the - zip file and unzips the file in the new directory. """ filename = Path(filename) - is_zipfile = filename.suffix == '.zip' path = Path(meeting.get_materials_path()) / subdir - if is_zipfile: - path = path / filename.stem path.mkdir(parents=True, exist_ok=True) - # agendas and minutes can only have one file instance so delete file if it already exists - if subdir in ('agenda', 'minutes'): - for f in path.glob(f'{filename.stem}.*'): + with (path / filename).open('wb+') as destination: + # prep file for reading + if hasattr(file, "chunks"): + chunks = file.chunks() + else: try: - f.unlink() - except FileNotFoundError: - pass # if the file is already gone, so be it + file.seek(0) + except AttributeError: + pass + chunks = [file.read()] # pretend we have chunks - with (path / filename).open('wb+') as destination: if filename.suffix in settings.MEETING_VALID_MIME_TYPE_EXTENSIONS['text/html']: - file.open() - text = file.read() + text = b"".join(chunks) if encoding: try: text = text.decode(encoding) @@ -698,13 +790,18 @@ def handle_upload_file(file, filename, meeting, subdir, request=None, encoding=N ) else: try: - text = smart_text(text) + text = smart_str(text) except UnicodeDecodeError as e: return "Failure trying to save '%s'. Hint: Try to upload as UTF-8: %s..." % (filename, str(e)[:120]) # Whole file sanitization; add back what's missing from a complete # document (sanitize will remove these). - clean = sanitize_document(text) - destination.write(clean.encode('utf8')) + clean = clean_html(text) + clean_bytes = clean.encode('utf8') + destination.write(clean_bytes) + # Assumes contents of subdir are always document type ids + # TODO-BLOBSTORE: see if we can refactor this so that the connection to the document isn't lost + # In the meantime, consider faking it by parsing filename (shudder). + store_bytes(subdir, filename.name, clean_bytes) if request and clean != text: messages.warning(request, ( @@ -713,15 +810,13 @@ def handle_upload_file(file, filename, meeting, subdir, request=None, encoding=N f"please check the resulting content. " )) else: - if hasattr(file, 'chunks'): - for chunk in file.chunks(): - destination.write(chunk) - else: - destination.write(file.read()) - - # unzip zipfile - if is_zipfile: - subprocess.call(['unzip', filename], cwd=path) + for chunk in chunks: + destination.write(chunk) + file.seek(0) + if hasattr(file, "chunks"): + chunks = file.chunks() + # TODO-BLOBSTORE: See above question about refactoring + store_bytes(subdir, filename.name, b"".join(chunks)) return None @@ -745,14 +840,1028 @@ def new_doc_for_session(type_id, session): rev = '00', ) doc.states.add(State.objects.get(type_id=type_id, slug='active')) - DocAlias.objects.create(name=doc.name).docs.add(doc) - session.sessionpresentation_set.create(document=doc,rev='00') + session.presentations.create(document=doc,rev='00') return doc +# TODO-BLOBSTORE - consider adding doc to this signature and factoring away type_id def write_doc_for_session(session, type_id, filename, contents): filename = Path(filename) path = Path(session.meeting.get_materials_path()) / type_id path.mkdir(parents=True, exist_ok=True) with open(path / filename, "wb") as file: file.write(contents.encode('utf-8')) - return + store_str(type_id, filename.name, contents) + return None + + +@dataclass +class BlobSpec: + bucket: str + name: str + + +def resolve_one_material( + doc: Document, rev: str | None, ext: str | None +) -> BlobSpec | None: + if doc.type_id is None: + log(f"Cannot resolve a doc with no type: {doc.name}") + return None + + # Get the Document's base name. It may or may not have an extension. + if rev is None: + basename = Path(doc.get_base_name()) + else: + basename = Path(f"{doc.name}-{int(rev):02d}") + + # If the document's file exists, the blob is _always_ named with this stem, + # even if it's different from the original. + blob_stem = Path(f"{doc.name}-{rev or doc.rev}") + + # If we have an extension, either from the URL or the Document's base name, look up + # the blob or file or return 404. N.b. the suffix check needs adjustment to handle + # a bare "." extension when we reach py3.14. + if ext or basename.suffix != "": + if ext: + blob_name = str(blob_stem.with_suffix(ext)) + else: + blob_name = str(blob_stem.with_suffix(basename.suffix)) + + # See if we have a stored object under that name + preferred_blob = ( + StoredObject.objects.exclude_deleted() + .filter(store=doc.type_id, name=blob_name) + .first() + ) + if preferred_blob is not None: + return BlobSpec( + bucket=preferred_blob.store, + name=preferred_blob.name, + ) + # No stored object, fall back to the file system. + filename = Path(doc.get_file_path()) / basename # use basename for file + if filename.is_file(): + return BlobSpec( + bucket=doc.type_id, + name=str(blob_stem.with_suffix(filename.suffix)), + ) + else: + return None + + # No extension has been specified so far, so look one up. + matching_stored_objects = ( + StoredObject.objects.exclude_deleted() + .filter( + store=doc.type_id, + name__startswith=f"{blob_stem}.", # anchor to end with trailing "." + ) + .order_by("name") + ) # orders by suffix + blob_ext_choices = { + Path(stored_obj.name).suffix: stored_obj + for stored_obj in matching_stored_objects + } + + # Short-circuit to return pdf if present + if ".pdf" in blob_ext_choices: + pdf_blob = blob_ext_choices[".pdf"] + return BlobSpec( + bucket=pdf_blob.store, + name=str(blob_stem.with_suffix(".pdf")), + ) + + # Now look for files + filename = Path(doc.get_file_path()) / basename + file_ext_choices = { + # Construct a map from suffix to full filename + fn.suffix: fn.name + for fn in sorted(filename.parent.glob(filename.stem + ".*")) + } + + # Short-circuit to return pdf if we have the file + if ".pdf" in file_ext_choices: + return BlobSpec( + bucket=doc.type_id, + name=str(blob_stem.with_suffix(".pdf")), + ) + + all_exts = set(blob_ext_choices.keys()).union(file_ext_choices.keys()) + if len(all_exts) > 0: + preferred_ext = sorted(all_exts)[0] + if preferred_ext in blob_ext_choices: + preferred_blob = blob_ext_choices[preferred_ext] + return BlobSpec( + bucket=preferred_blob.store, + name=preferred_blob.name, + ) + else: + return BlobSpec( + bucket=doc.type_id, + name=str(blob_stem.with_suffix(preferred_ext)), + ) + + return None + + +def resolve_materials_for_one_meeting(meeting: Meeting): + start_time = timezone.now() + meeting_documents = ( + Document.objects.filter( + type_id__in=settings.MATERIALS_TYPES_SERVED_BY_WORKER + ).filter( + Q(session__meeting=meeting) | Q(proceedingsmaterial__meeting=meeting) + ) + ).distinct() + + resolved = [] + for doc in meeting_documents: + # request by doc name with no rev + blob = resolve_one_material(doc, rev=None, ext=None) + if blob is not None: + resolved.append( + ResolvedMaterial( + name=doc.name, + meeting_number=meeting.number, + bucket=blob.bucket, + blob=blob.name, + ) + ) + # request by doc name + rev + blob = resolve_one_material(doc, rev=doc.rev, ext=None) + if blob is not None: + resolved.append( + ResolvedMaterial( + name=f"{doc.name}-{doc.rev:02}", + meeting_number=meeting.number, + bucket=blob.bucket, + blob=blob.name, + ) + ) + # for other revisions, only need request by doc name + rev + other_revisions = doc.revisions_by_newrevisionevent() + other_revisions.remove(doc.rev) + for rev in other_revisions: + blob = resolve_one_material(doc, rev=rev, ext=None) + if blob is not None: + resolved.append( + ResolvedMaterial( + name=f"{doc.name}-{rev:02}", + meeting_number=meeting.number, + bucket=blob.bucket, + blob=blob.name, + ) + ) + ResolvedMaterial.objects.bulk_create( + resolved, + update_conflicts=True, + unique_fields=["name", "meeting_number"], + update_fields=["bucket", "blob"], + ) + # Warn if any files were updated during the above process + last_update = meeting_documents.aggregate(Max("time"))["time__max"] + if last_update and last_update > start_time: + log( + f"Warning: materials for meeting {meeting.number} " + "changed during ResolvedMaterial update" + ) + +def resolve_uploaded_material(meeting: Meeting, doc: Document): + resolved: list[ResolvedMaterial] = [] + remove = ResolvedMaterial.objects.none() + blob = resolve_one_material(doc, rev=None, ext=None) + if blob is None: + # Versionless file does not exist. Remove the versionless ResolvedMaterial + # if it existed. This is to avoid leaving behind a stale link to a replaced + # version. This comes up e.g. if a ProceedingsMaterial is changed from having + # an uploaded file to being an external URL. + remove = ResolvedMaterial.objects.filter( + name=doc.name, meeting_number=meeting.number + ) + else: + resolved.append( + ResolvedMaterial( + name=doc.name, + meeting_number=meeting.number, + bucket=blob.bucket, + blob=blob.name, + ) + ) + # request by doc name + rev + blob = resolve_one_material(doc, rev=doc.rev, ext=None) + if blob is not None: + resolved.append( + ResolvedMaterial( + name=f"{doc.name}-{doc.rev:02}", + meeting_number=meeting.number, + bucket=blob.bucket, + blob=blob.name, + ) + ) + # Create the new record(s) + ResolvedMaterial.objects.bulk_create( + resolved, + update_conflicts=True, + unique_fields=["name", "meeting_number"], + update_fields=["bucket", "blob"], + ) + # and remove one if necessary (will be a none() queryset if not) + remove.delete() + + +def store_blob_for_one_material_file(doc: Document, rev: str, filepath: Path): + if not settings.ENABLE_BLOBSTORAGE: + raise RuntimeError("Cannot store blobs: ENABLE_BLOBSTORAGE is False") + + bucket = doc.type_id + if bucket not in settings.MATERIALS_TYPES_SERVED_BY_WORKER: + raise ValueError(f"Bucket {bucket} not found for doc {doc.name}.") + blob_stem = f"{doc.name}-{rev}" + suffix = filepath.suffix # includes leading "." + + # Store the file + try: + file_bytes = filepath.read_bytes() + except Exception as err: + log(f"Failed to read {filepath}: {err}") + raise + with suppress(AlreadyExistsError): + store_bytes( + kind=bucket, + name= blob_stem + suffix, + content=file_bytes, + mtime=datetime.datetime.fromtimestamp( + filepath.stat().st_mtime, + tz=datetime.UTC, + ), + allow_overwrite=False, + doc_name=doc.name, + doc_rev=rev, + ) + + # Special case: pre-render markdown into HTML as .md.html + if suffix == ".md": + try: + markdown_source = file_bytes.decode("utf-8") + except UnicodeDecodeError as err: + log(f"Unable to decode {filepath} as UTF-8, treating as latin-1: {err}") + markdown_source = file_bytes.decode("latin-1") + # render the markdown + try: + html = render_to_string( + "minimal.html", + { + "content": markdown.markdown(markdown_source), + "title": blob_stem, + "static_ietf_org": settings.STATIC_IETF_ORG, + }, + ) + except Exception as err: + log(f"Failed to render markdown for {filepath}: {err}") + else: + # Don't overwrite, but don't fail if the blob exists + with suppress(AlreadyExistsError): + store_str( + kind=bucket, + name=blob_stem + ".md.html", + content=html, + allow_overwrite=False, + doc_name=doc.name, + doc_rev=rev, + content_type="text/html;charset=utf-8", + ) + + +def store_blobs_for_one_material_doc(doc: Document): + """Ensure that all files related to a materials Document are in the blob store""" + if doc.type_id not in settings.MATERIALS_TYPES_SERVED_BY_WORKER: + log(f"This method does not handle docs of type {doc.name}") + return + + # Store files for current Document / rev + file_path = Path(doc.get_file_path()) + base_name = Path(doc.get_base_name()) + # .stem would remove directories, so use .with_suffix("") + base_name_stem = str(base_name.with_suffix("")) + if base_name_stem.endswith(".") and base_name.suffix == "": + # In Python 3.14, a trailing "." is a valid suffix, but in prior versions + # it is left as part of the stem. The suffix check ensures that either way, + # only a single "." will be removed. + base_name_stem = base_name_stem[:-1] + # Add any we find without the rev + for file_to_store in file_path.glob(base_name_stem + ".*"): + if not (file_to_store.is_file()): + continue + try: + store_blob_for_one_material_file(doc, doc.rev, file_to_store) + except Exception as err: + log( + f"Failed to store blob for {doc} rev {doc.rev} " + f"from {file_to_store}: {err}" + ) + + # Get other revisions + for rev in doc.revisions_by_newrevisionevent(): + if rev == doc.rev: + continue # already handled this + + # Add some that have the rev + for file_to_store in file_path.glob(doc.name + f"-{rev}.*"): + if not file_to_store.is_file(): + continue + try: + store_blob_for_one_material_file(doc, rev, file_to_store) + except Exception as err: + log( + f"Failed to store blob for {doc} rev {rev} " + f"from {file_to_store}: {err}" + ) + + +def store_blobs_for_one_meeting(meeting: Meeting): + meeting_documents = ( + Document.objects.filter( + type_id__in=settings.MATERIALS_TYPES_SERVED_BY_WORKER + ).filter( + Q(session__meeting=meeting) | Q(proceedingsmaterial__meeting=meeting) + ) + ).distinct() + + for doc in meeting_documents: + store_blobs_for_one_material_doc(doc) + + +def create_recording(session, url, title=None, user=None): + ''' + Creates the Document type=recording, setting external_url and creating + NewRevisionDocEvent + ''' + sequence = get_next_sequence(session.group,session.meeting,'recording') + name = 'recording-{}-{}-{}'.format(session.meeting.number,session.group.acronym,sequence) + time = session.official_timeslotassignment().timeslot.time.strftime('%Y-%m-%d %H:%M') + if not title: + if url.endswith('mp3'): + title = 'Audio recording for {}'.format(time) + else: + title = 'Video recording for {}'.format(time) + + doc = Document.objects.create(name=name, + title=title, + external_url=url, + group=session.group, + rev='00', + type_id='recording') + doc.set_state(State.objects.get(type='recording', slug='active')) + + # create DocEvent + NewRevisionDocEvent.objects.create(type='new_revision', + by=user or Person.objects.get(name='(System)'), + doc=doc, + rev=doc.rev, + desc='New revision available', + time=doc.time) + pres = SessionPresentation.objects.create(session=session,document=doc,rev=doc.rev) + session.presentations.add(pres) + + return doc + +def delete_recording(session_presentation, user=None): + """Delete a session recording""" + document = session_presentation.document + if document.type_id != "recording": + raise ValueError(f"Document {document.pk} is not a recording (type_id={document.type_id})") + recording_state = document.get_state("recording") + deleted_state = State.objects.get(type_id="recording", slug="deleted") + if recording_state != deleted_state: + # Update the recording state and create a history event + document.set_state(deleted_state) + StateDocEvent.objects.create( + type="changed_state", + by=user or Person.objects.get(name="(System)"), + doc=document, + rev=document.rev, + state_type=deleted_state.type, + state=deleted_state, + ) + session_presentation.delete() + +def get_next_sequence(group, meeting, type): + ''' + Returns the next sequence number to use for a document of type = type. + Takes a group=Group object, meeting=Meeting object, type = string + ''' + docs = Document.objects.filter(name__startswith='{}-{}-{}-'.format(type, meeting.number, group.acronym)) + if not docs: + return 1 + docs = docs.order_by('name') + sequence = int(docs.last().name.split('-')[-1]) + 1 + return sequence + +def get_activity_stats(sdate, edate): + ''' + This function takes a date range and produces a dictionary of statistics / objects for + use in an activity report. Generally the end date will be the date of the last meeting + and the start date will be the date of the meeting before that. + + Data between midnight UTC on the specified dates are included in the stats. + ''' + sdatetime = pytz.utc.localize(datetime.datetime.combine(sdate, datetime.time())) + edatetime = pytz.utc.localize(datetime.datetime.combine(edate, datetime.time())) + + data = {} + data['sdate'] = sdate + data['edate'] = edate + + events = DocEvent.objects.filter(doc__type='draft', time__gte=sdatetime, time__lt=edatetime) + + data['actions_count'] = events.filter(type='iesg_approved').count() + data['last_calls_count'] = events.filter(type='sent_last_call').count() + new_draft_events = events.filter(newrevisiondocevent__rev='00') + new_drafts = list(set([e.doc_id for e in new_draft_events])) + data['new_docs'] = list(set([e.doc for e in new_draft_events])) + data['new_drafts_count'] = len(new_drafts) + data['new_drafts_updated_count'] = events.filter(doc__id__in=new_drafts,newrevisiondocevent__rev='01').count() + data['new_drafts_updated_more_count'] = events.filter(doc__id__in=new_drafts,newrevisiondocevent__rev='02').count() + + update_events = events.filter(type='new_revision').exclude(doc__id__in=new_drafts) + data['updated_drafts_count'] = len(set([e.doc_id for e in update_events])) + + # Calculate Final Four Weeks stats (ffw) + ffwdate = edatetime - datetime.timedelta(days=28) + ffw_new_count = events.filter(time__gte=ffwdate, newrevisiondocevent__rev='00').count() + try: + ffw_new_percent = format(ffw_new_count / float(data['new_drafts_count']), '.0%') + except ZeroDivisionError: + ffw_new_percent = 0 + + data['ffw_new_count'] = ffw_new_count + data['ffw_new_percent'] = ffw_new_percent + + ffw_update_events = events.filter(time__gte=ffwdate, type='new_revision').exclude(doc__id__in=new_drafts) + ffw_update_count = len(set([e.doc_id for e in ffw_update_events])) + try: + ffw_update_percent = format(ffw_update_count / float(data['updated_drafts_count']),'.0%') + except ZeroDivisionError: + ffw_update_percent = 0 + + data['ffw_update_count'] = ffw_update_count + data['ffw_update_percent'] = ffw_update_percent + + rfcs_events = DocEvent.objects.filter(doc__type='rfc', time__gte=sdatetime, time__lt=edatetime) + rfcs = rfcs_events.filter(type='published_rfc') + data['rfcs'] = rfcs.select_related('doc').select_related('doc__group').select_related('doc__std_level') + + data['counts'] = {'std': rfcs.filter(doc__std_level__in=('ps', 'ds', 'std')).count(), + 'bcp': rfcs.filter(doc__std_level='bcp').count(), + 'exp': rfcs.filter(doc__std_level='exp').count(), + 'inf': rfcs.filter(doc__std_level='inf').count()} + + data['new_groups'] = Group.objects.filter( + type='wg', + groupevent__changestategroupevent__state='active', + groupevent__time__gte=sdatetime, + groupevent__time__lt=edatetime) + + data['concluded_groups'] = Group.objects.filter( + type='wg', + groupevent__changestategroupevent__state='conclude', + groupevent__time__gte=sdatetime, + groupevent__time__lt=edatetime) + + return data + +def is_powerpoint(doc): + ''' + Returns true if document is a Powerpoint presentation + ''' + return doc.file_extension() in ('ppt', 'pptx') + +def post_process(doc): + ''' + Does post processing on uploaded file. + - Convert PPT to PDF + ''' + if is_powerpoint(doc) and hasattr(settings, 'PPT2PDF_COMMAND'): + try: + cmd = list(settings.PPT2PDF_COMMAND) # Don't operate on the list actually in settings + cmd.append(doc.get_file_path()) # outdir + cmd.append(os.path.join(doc.get_file_path(), doc.uploaded_filename)) # filename + subprocess.check_call(cmd) + except (subprocess.CalledProcessError, OSError) as error: + log("Error converting PPT: %s" % (error)) + return + # change extension + base, ext = os.path.splitext(doc.uploaded_filename) + doc.uploaded_filename = base + '.pdf' + + e = DocEvent.objects.create( + type='changed_document', + by=Person.objects.get(name="(System)"), + doc=doc, + rev=doc.rev, + desc='Converted document to PDF', + ) + doc.save_with_history([e]) + + +def participants_for_meeting(meeting): + """ Return a tuple (checked_in, attended) + checked_in = queryset of onsite, checkedin participants values_list('person') + attended = queryset of remote participants who attended a session values_list('person') + """ + checked_in = meeting.registration_set.onsite().filter(checkedin=True).values_list('person', flat=True).distinct() + sessions = meeting.session_set.filter(Q(type='plenary') | Q(group__type__in=['wg', 'rg'])) + attended = Attended.objects.filter(session__in=sessions).values_list('person', flat=True).distinct() + return (checked_in, attended) + + +def generate_proceedings_content(meeting, force_refresh=False): + """Render proceedings content for a meeting and update cache + + Caches its value for 25 hours to ensure that the cache never expires if + we recompute the value daily. + + :meeting: meeting whose proceedings should be rendered + :force_refresh: true to force regeneration and cache refresh + """ + cache = caches["proceedings"] + key_components = [ + "proceedings", + str(meeting.number), + ] + if meeting.proceedings_final: + # Freeze the cache key once proceedings are finalized. Further changes will + # not be picked up until the cache expires or is refreshed by the + # proceedings_content_refresh_task() + key_components.append("final") + else: + # Build a cache key that changes when materials are modified. For all but drafts, + # use the last modification time of the document. Exclude drafts from this because + # revisions long after the meeting ends will otherwise show up as changes and + # incorrectly invalidate the cache. Instead, include an ordered list of the + # drafts linked to the meeting so adding or removing drafts will trigger a + # recalculation. The list is long but that doesn't matter because we hash it into + # a fixed-length key. + meeting_docs = Document.objects.filter(session__meeting__number=meeting.number) + last_materials_update = ( + meeting_docs.exclude(type_id="draft") + .filter(session__meeting__number=meeting.number) + .aggregate(Max("time"))["time__max"] + ) + draft_names = ( + meeting_docs + .filter(type_id="draft") + .order_by("name") + .values_list("name", flat=True) + ) + key_components += [ + last_materials_update.isoformat() if last_materials_update else "-", + ",".join(draft_names), + ] + + # Key is potentially long, but the "proceedings" cache hashes it to a fixed + # length. If that changes, hash it separately here first. + cache_key = ".".join(key_components) + if not force_refresh: + cached_content = cache.get(cache_key, None) + if cached_content is not None: + return cached_content + + def area_and_group_acronyms_from_session(s): + area = s.group_parent_at_the_time() + if area == None: + area = s.group.parent + group = s.group_at_the_time() + return (area.acronym, group.acronym) + + schedule = meeting.schedule + sessions = ( + meeting.session_set.with_current_status() + .filter(Q(timeslotassignments__schedule__in=[schedule, schedule.base if schedule else None]) + | Q(current_status='notmeet')) + .select_related() + .order_by('-current_status') + ) + + plenaries, _ = organize_proceedings_sessions( + sessions.filter(name__icontains='plenary') + .exclude(current_status='notmeet') + ) + irtf_meeting, irtf_not_meeting = organize_proceedings_sessions( + sessions.filter(group__parent__acronym = 'irtf').order_by('group__acronym') + ) + # per Colin (datatracker #5010) - don't report not meeting rags + irtf_not_meeting = [item for item in irtf_not_meeting if item["group"].type_id != "rag"] + irtf = {"meeting_groups":irtf_meeting, "not_meeting_groups":irtf_not_meeting} + + training, _ = organize_proceedings_sessions( + sessions.filter(group__acronym__in=['edu','iaoc'], type_id__in=['regular', 'other',]) + .exclude(current_status='notmeet') + ) + iab, _ = organize_proceedings_sessions( + sessions.filter(group__parent__acronym = 'iab') + .exclude(current_status='notmeet') + ) + editorial, _ = organize_proceedings_sessions( + sessions.filter(group__acronym__in=['rsab','rswg']) + .exclude(current_status='notmeet') + ) + + ietf = sessions.filter(group__parent__type__slug = 'area').exclude(group__acronym__in=['edu','iepg','tools']) + ietf = list(ietf) + ietf.sort(key=lambda s: area_and_group_acronyms_from_session(s)) + ietf_areas = [] + for area, area_sessions in itertools.groupby(ietf, key=lambda s: s.group_parent_at_the_time()): + meeting_groups, not_meeting_groups = organize_proceedings_sessions(area_sessions) + ietf_areas.append((area, meeting_groups, not_meeting_groups)) + + with timezone.override(meeting.tz()): + rendered_content = render_to_string( + "meeting/proceedings.html", + { + 'meeting': meeting, + 'plenaries': plenaries, + 'training': training, + 'irtf': irtf, + 'iab': iab, + 'editorial': editorial, + 'ietf_areas': ietf_areas, + 'meetinghost_logo': { + 'max_height': settings.MEETINGHOST_LOGO_MAX_DISPLAY_HEIGHT, + 'max_width': settings.MEETINGHOST_LOGO_MAX_DISPLAY_WIDTH, + } + }, + ) + cache.set( + cache_key, + rendered_content, + timeout=3600 + 86400, # one day + one hour, in seconds + ) + return rendered_content + + +def organize_proceedings_sessions(sessions): + # Collect sessions by Group, then bin by session name (including sessions with blank names). + # If all of a group's sessions are 'notmeet', the processed data goes in not_meeting_sessions. + # Otherwise, the data goes in meeting_sessions. + meeting_groups = [] + not_meeting_groups = [] + for group_acronym, group_sessions in itertools.groupby(sessions, key=lambda s: s.group.acronym): + by_name = {} + is_meeting = False + all_canceled = True + group = None + for s in sorted( + group_sessions, + key=lambda gs: ( + gs.official_timeslotassignment().timeslot.time + if gs.official_timeslotassignment() else datetime.datetime(datetime.MAXYEAR, 1, 1) + ), + ): + group = s.group + if s.current_status != 'notmeet': + is_meeting = True + if s.current_status != 'canceled': + all_canceled = False + by_name.setdefault(s.name, []) + if s.current_status != 'notmeet' or s.presentations.exists(): + by_name[s.name].append(s) # for notmeet, only include sessions with materials + for sess_name, ss in by_name.items(): + session = ss[0] if ss else None + def _format_materials(items): + """Format session/material for template + + Input is a list of (session, materials) pairs. The materials value can be a single value or a list. + """ + material_times = {} # key is material, value is first timestamp it appeared + for s, mats in items: + tsa = s.official_timeslotassignment() + timestamp = tsa.timeslot.time if tsa else None + if not isinstance(mats, list): + mats = [mats] + for mat in mats: + if mat and mat not in material_times: + material_times[mat] = timestamp + n_mats = len(material_times) + result = [] + if n_mats == 1: + result.append({'material': list(material_times)[0]}) # no 'time' when only a single material + elif n_mats > 1: + for mat, timestamp in material_times.items(): + result.append({'material': mat, 'time': timestamp}) + return result + + entry = { + 'group': group, + 'name': sess_name, + 'session': session, + 'canceled': all_canceled, + 'has_materials': s.presentations.exists(), + 'agendas': _format_materials((s, s.agenda()) for s in ss), + 'minutes': _format_materials((s, s.minutes()) for s in ss), + 'bluesheets': _format_materials((s, s.bluesheets()) for s in ss), + 'recordings': _format_materials((s, s.recordings()) for s in ss), + 'meetecho_recordings': _format_materials((s, [s.session_recording_url()]) for s in ss), + 'chatlogs': _format_materials((s, s.chatlogs()) for s in ss), + 'slides': _format_materials((s, s.slides()) for s in ss), + 'drafts': _format_materials((s, s.drafts()) for s in ss), + 'last_update': session.last_update if hasattr(session, 'last_update') else None + } + if session and session.meeting.type_id == 'ietf' and not session.meeting.proceedings_final: + entry['attendances'] = _format_materials((s, s) for s in ss if Attended.objects.filter(session=s).exists()) + if is_meeting: + meeting_groups.append(entry) + else: + not_meeting_groups.append(entry) + return meeting_groups, not_meeting_groups + + +import_registration_json_validator = jsonschema.Draft202012Validator( + schema={ + "type": "object", + "properties": { + "objects": { + "type": "object", + "patternProperties": { + # Email address as key (simplified pattern or just allow any key) + ".*": { + "type": "object", + "properties": { + "first_name": {"type": "string"}, + "last_name": {"type": "string"}, + "email": {"type": "string", "format": "email"}, + "affiliation": {"type": "string"}, + "country_code": {"type": "string", "minLength": 2, "maxLength": 2}, + "meeting": {"type": "string"}, + "checkedin": {"type": "boolean"}, + "cancelled": {"type": "boolean"}, + "is_nomcom_volunteer": {"type": "boolean"}, + "tickets": { + "type": "array", + "items": { + "type": "object", + "properties": { + "attendance_type": {"type": "string"}, + "ticket_type": {"type": "string"} + }, + "required": ["attendance_type", "ticket_type"] + } + } + }, + "required": [ + "first_name", "last_name", "email", + "country_code", "meeting", 'affiliation', + "checkedin", "is_nomcom_volunteer", "tickets", + "cancelled", + ] + } + }, + "additionalProperties": False + } + }, + "required": ["objects"] + } +) + + +def get_registration_data(meeting): + '''Retrieve data from registation system for meeting''' + url = settings.REGISTRATION_PARTICIPANTS_API_URL + key = settings.REGISTRATION_PARTICIPANTS_API_KEY + params = {'meeting': meeting.number, 'apikey': key} + try: + response = requests.get(url, params=params, timeout=settings.DEFAULT_REQUESTS_TIMEOUT) + except requests.Timeout as e: + log(f'GET request timed out for [{url}]: {e}') + raise Exception("Timeout retrieving data from registration API") from e + if response.status_code == 200: + try: + decoded = response.json() + except ValueError as e: + raise ValueError(f'Could not decode response from registration API: {e}') + else: + raise Exception(f'Bad response from registration API: {response.status_code}, {response.content[:64]}') + + # validate registration data + import_registration_json_validator.validate(decoded) + return decoded + + +def sync_registration_data(meeting): + """"Sync meeting.Registration with registration system. + + Registration records are created in realtime as people register for a + meeting. This function serves as an audit / reconciliation. Most records are + expected to already exist. The function has been optimized with this in mind. + + - Creates new registrations if they don't exist + - Updates existing registrations if fields differ + - Updates tickets as needed + - Deletes registrations that exist in the database but not in the JSON data + + Returns: + dict: Summary of changes made (created, updated, deleted counts) + """ + reg_data = get_registration_data(meeting) + + # Get the meeting ID from the first registration, the API only deals with one meeting at a time + first_email = next(iter(reg_data['objects'])) + meeting_number = reg_data['objects'][first_email]['meeting'] + try: + Meeting.objects.get(number=meeting_number) + except Meeting.DoesNotExist: + raise Exception(f'meeting does not exist {meeting_number}') + + # Get all existing registrations for this meeting + existing_registrations = meeting.registration_set.all() + existing_emails = set(reg.email for reg in existing_registrations if reg.email) + + # Track changes for reporting + stats = { + 'created': 0, + 'updated': 0, + 'deleted': 0, + 'processed': 0, + } + + # Process registrations from reg_data + reg_emails = set() + for email, data in reg_data['objects'].items(): + stats['processed'] += 1 + reg_emails.add(email) + + # Process this registration + _, action_taken = process_single_registration(data, meeting) + + # Update stats + if action_taken == 'created': + stats['created'] += 1 + elif action_taken == 'updated': + stats['updated'] += 1 + + # Delete registrations that exist in the DB but not in registration data, they've been cancelled + emails_to_delete = existing_emails - reg_emails + if emails_to_delete: + log(f"sync_reg: emails marked for deletion: {emails_to_delete}") + result = Registration.objects.filter( + email__in=emails_to_delete, + meeting=meeting + ).delete() + if 'meeting.Registration' in result[1]: + deleted_count = result[1]['meeting.Registration'] + else: + deleted_count = 0 + stats['deleted'] = deleted_count + # set meeting.attendees + count = Registration.objects.onsite().filter(meeting=meeting, checkedin=True).count() + if meeting.attendees != count: + meeting.attendees = count + meeting.save() + + return stats + + +def process_single_registration(reg_data, meeting): + """ + Process a single registration record - create, update, or leave unchanged as needed. + + Args: + reg_data (dict): Registration data + meeting (obj): The IETF meeting + + Returns: + tuple: (registration, action_taken) + - registration: Registration object + - action_taken: String indicating 'created', 'updated', or None + """ + # import here to avoid circular imports + from ietf.nomcom.models import Volunteer, NomCom + + action_taken = None + fields_updated = False + tickets_modified = False + + # handle deleted + # should not see cancelled records during nightly sync but can see + # them from realtime notifications + if reg_data['cancelled']: + try: + registration = Registration.objects.get(meeting=meeting, email=reg_data['email']) + except Registration.DoesNotExist: + return (None, None) + for ticket in reg_data['tickets']: + target = registration.tickets.filter( + attendance_type__slug=ticket['attendance_type'], + ticket_type__slug=ticket['ticket_type']).first() + if target: + target.delete() + if registration.tickets.count() == 0: + registration.delete() + log(f"sync_reg: cancelled registration {reg_data['email']}") + return (None, 'deleted') + + person = Person.objects.filter(email__address=reg_data['email']).first() + if not person: + log(f"ERROR: meeting registration email unknown {reg_data['email']}") + + registration, created = Registration.objects.get_or_create( + email=reg_data['email'], + meeting=meeting, + defaults={ + 'first_name': reg_data['first_name'], + 'last_name': reg_data['last_name'], + 'person': person, + 'affiliation': reg_data['affiliation'], + 'country_code': reg_data['country_code'], + 'checkedin': reg_data['checkedin'], + } + ) + + # If not created, check if we need to update + if not created: + for field in ['first_name', 'last_name', 'affiliation', 'country_code', 'checkedin']: + if getattr(registration, field) != reg_data[field]: + log(f"sync_reg: found update {reg_data['email']}, {field} different, data from reg: {reg_data}") + setattr(registration, field, reg_data[field]) + fields_updated = True + + if fields_updated: + registration.save() + + # Process tickets - handle counting properly for multiple same-type tickets + # Build count dictionaries for existing and new tickets + existing_ticket_counts = {} + for ticket in registration.tickets.all(): + key = (ticket.attendance_type.slug, ticket.ticket_type.slug) + existing_ticket_counts[key] = existing_ticket_counts.get(key, 0) + 1 + + # Get new tickets from reg_data and count them + reg_data_ticket_counts = {} + for ticket_data in reg_data.get('tickets', []): + key = (ticket_data['attendance_type'], ticket_data['ticket_type']) + reg_data_ticket_counts[key] = reg_data_ticket_counts.get(key, 0) + 1 + + # Calculate tickets to add and remove + all_ticket_types = set(existing_ticket_counts.keys()) | set(reg_data_ticket_counts.keys()) + + for ticket_type in all_ticket_types: + existing_count = existing_ticket_counts.get(ticket_type, 0) + new_count = reg_data_ticket_counts.get(ticket_type, 0) + + # Delete excess tickets + if existing_count > new_count: + tickets_to_delete = existing_count - new_count + # Get all tickets of this type + matching_tickets = registration.tickets.filter( + attendance_type__slug=ticket_type[0], + ticket_type__slug=ticket_type[1] + ).order_by('id') # Use a consistent order for deterministic deletion + + # Delete the required number + log(f"sync_reg: deleting {tickets_to_delete} of {ticket_type[0]}:{ticket_type[1]} of {reg_data['email']}") + for ticket in matching_tickets[:tickets_to_delete]: + ticket.delete() + tickets_modified = True + + # Add missing tickets + elif new_count > existing_count: + tickets_to_add = new_count - existing_count + + # Create the new tickets + log(f"sync_reg: adding {tickets_to_add} of {ticket_type[0]}:{ticket_type[1]} of {reg_data['email']}") + for _ in range(tickets_to_add): + try: + RegistrationTicket.objects.create( + registration=registration, + attendance_type_id=ticket_type[0], + ticket_type_id=ticket_type[1], + ) + tickets_modified = True + except IntegrityError as e: + log(f"Error adding RegistrationTicket {e}") + # handle nomcom volunteer + if reg_data['is_nomcom_volunteer'] and person: + try: + nomcom = NomCom.objects.get(is_accepting_volunteers=True) + except (NomCom.DoesNotExist, NomCom.MultipleObjectsReturned): + nomcom = None + if nomcom: + Volunteer.objects.get_or_create( + nomcom=nomcom, + person=person, + defaults={ + "affiliation": reg_data["affiliation"], + "origin": "registration" + } + ) + + # set action_taken + if created: + log(f"sync_reg: created record. {reg_data['email']}") + action_taken = 'created' + elif fields_updated or tickets_modified: + action_taken = 'updated' + + return registration, action_taken + + +def fetch_attendance_from_meetings(meetings): + return [sync_registration_data(meeting) for meeting in meetings] diff --git a/ietf/meeting/views.py b/ietf/meeting/views.py index b90f2506a0..67a81305b4 100644 --- a/ietf/meeting/views.py +++ b/ietf/meeting/views.py @@ -1,60 +1,77 @@ -# Copyright The IETF Trust 2007-2022, All Rights Reserved +# Copyright The IETF Trust 2007-2024, All Rights Reserved # -*- coding: utf-8 -*- import csv import datetime -import glob import io import itertools import json import math import os + import pytz import re import tarfile import tempfile +import shutil from calendar import timegm from collections import OrderedDict, Counter, deque, defaultdict, namedtuple -from urllib.parse import unquote +from functools import partialmethod +import jsonschema +from pathlib import Path +from urllib.parse import parse_qs, unquote, urlencode, urlsplit, urlunsplit, urlparse from tempfile import mkstemp from wsgiref.handlers import format_date_time +from itertools import chain from django import forms +from django.core.cache import caches +from django.core.files.storage import storages from django.shortcuts import render, redirect, get_object_or_404 from django.http import (HttpResponse, HttpResponseRedirect, HttpResponseForbidden, HttpResponseNotFound, Http404, HttpResponseBadRequest, - JsonResponse, HttpResponseGone) + JsonResponse, HttpResponseGone, HttpResponseNotAllowed, + FileResponse) from django.conf import settings from django.contrib import messages from django.contrib.auth.decorators import login_required from django.core.exceptions import ValidationError +from django.core.files.uploadedfile import SimpleUploadedFile from django.core.validators import URLValidator -from django.urls import reverse,reverse_lazy +from django.urls import reverse, reverse_lazy, NoReverseMatch from django.db.models import F, Max, Q from django.forms.models import modelform_factory, inlineformset_factory from django.template import TemplateDoesNotExist from django.template.loader import render_to_string from django.utils import timezone from django.utils.encoding import force_str -from django.utils.functional import curry from django.utils.text import slugify from django.views.decorators.cache import cache_page from django.views.decorators.csrf import ensure_csrf_cookie, csrf_exempt from django.views.generic import RedirectView +from rest_framework.status import HTTP_404_NOT_FOUND import debug # pyflakes:ignore from ietf.doc.fields import SearchableDocumentsField -from ietf.doc.models import Document, State, DocEvent, NewRevisionDocEvent, DocAlias +from ietf.doc.models import Document, State, DocEvent, NewRevisionDocEvent +from ietf.doc.storage_utils import ( + remove_from_storage, + retrieve_bytes, + store_file, +) from ietf.group.models import Group from ietf.group.utils import can_manage_session_materials, can_manage_some_groups, can_manage_group from ietf.person.models import Person, User from ietf.ietfauth.utils import role_required, has_role, user_is_person from ietf.mailtrigger.utils import gather_address_lists -from ietf.meeting.models import Meeting, Session, Schedule, FloorPlan, SessionPresentation, TimeSlot, SlideSubmission -from ietf.meeting.models import SessionStatusName, SchedulingEvent, SchedTimeSessAssignment, Room, TimeSlotTypeName +from ietf.meeting.models import Meeting, Session, Schedule, FloorPlan, \ + SessionPresentation, TimeSlot, SlideSubmission, Attended +from ..blobdb.models import ResolvedMaterial +from ietf.meeting.models import ImportantDate, SessionStatusName, SchedulingEvent, SchedTimeSessAssignment, Room, TimeSlotTypeName +from ietf.meeting.models import Registration from ietf.meeting.forms import ( CustomDurationField, SwapDaysForm, SwapTimeslotsForm, ImportMinutesForm, TimeSlotCreateForm, TimeSlotEditForm, SessionCancelForm, SessionEditForm ) from ietf.meeting.helpers import get_person_by_email, get_schedule_by_name @@ -71,7 +88,14 @@ from ietf.meeting.helpers import send_interim_approval from ietf.meeting.helpers import send_interim_approval_request from ietf.meeting.helpers import send_interim_announcement_request, sessions_post_cancel -from ietf.meeting.utils import finalize, sort_accept_tuple, condition_slide_order +from ietf.meeting.utils import ( + condition_slide_order, + finalize, + generate_proceedings_content, + organize_proceedings_sessions, + resolve_uploaded_material, + sort_accept_tuple, store_blobs_for_one_material_doc, +) from ietf.meeting.utils import add_event_info_to_session_qs from ietf.meeting.utils import session_time_for_sorting from ietf.meeting.utils import session_requested_by, SaveMaterialsError @@ -82,14 +106,15 @@ from ietf.meeting.utils import swap_meeting_schedule_timeslot_assignments, bulk_create_timeslots from ietf.meeting.utils import preprocess_meeting_important_dates from ietf.meeting.utils import new_doc_for_session, write_doc_for_session +from ietf.meeting.utils import get_activity_stats, post_process, create_recording, delete_recording +from ietf.meeting.utils import participants_for_meeting, generate_bluesheet, bluesheet_data, save_bluesheet from ietf.message.utils import infer_message -from ietf.name.models import SlideSubmissionStatusName, ProceedingsMaterialTypeName, SessionPurposeName -from ietf.secr.proceedings.proc_utils import (get_progress_stats, post_process, import_audio_files, - create_recording) +from ietf.name.models import SlideSubmissionStatusName, ProceedingsMaterialTypeName, SessionPurposeName, CountryName from ietf.utils import markdown from ietf.utils.decorators import require_api_key from ietf.utils.hedgedoc import Note, NoteError -from ietf.utils.log import assertion +from ietf.utils.meetecho import MeetechoAPIError, SlidesManager +from ietf.utils.log import assertion, log from ietf.utils.mail import send_mail_message, send_mail_text from ietf.utils.mime import get_mime_type from ietf.utils.pipe import pipe @@ -97,12 +122,21 @@ from ietf.utils.response import permission_denied from ietf.utils.text import xslugify from ietf.utils.timezone import datetime_today, date_today +from ietf.settings import YOUTUBE_DOMAINS from .forms import (InterimMeetingModelForm, InterimAnnounceForm, InterimSessionModelForm, InterimCancelForm, InterimSessionInlineFormSet, RequestMinutesForm, - UploadAgendaForm, UploadBlueSheetForm, UploadMinutesForm, UploadSlidesForm) + UploadAgendaForm, UploadBlueSheetForm, UploadMinutesForm, UploadSlidesForm, + UploadNarrativeMinutesForm) + +from icalendar import Calendar, Event +from ietf.doc.templatetags.ietf_filters import absurl +from ..api.ietf_utils import requires_api_token +from ..blobdb.storage import BlobdbStorage, BlobFile +request_summary_exclude_group_types = ['team'] + def get_interim_menu_entries(request): '''Setup menu entries for interim meeting view tabs''' entries = [] @@ -119,6 +153,10 @@ def send_interim_change_notice(request, meeting): message.related_groups.add(group) send_mail_message(request, message) +def parse_ical_line_endings(ical): + """Parse icalendar line endings to ensure they are RFC 5545 compliant""" + return re.sub(r'\r(?!\n)|(? 0: + if ".pdf" in ext_choices: + filename = ext_choices[".pdf"] + else: + filename = list(ext_choices.values())[0] + if not filename.exists(): + raise Http404(f"File not found: {filename}") old_proceedings_format = meeting.number.isdigit() and int(meeting.number) <= 96 if settings.MEETING_MATERIALS_SERVE_LOCALLY or old_proceedings_format: - with io.open(filename, 'rb') as file: - bytes = file.read() - - mtype, chset = get_mime_type(bytes) + bytes = filename.read_bytes() + mtype, chset = get_mime_type(bytes) # chset does not consider entire file! content_type = "%s; charset=%s" % (mtype, chset) - file_ext = os.path.splitext(filename) - if len(file_ext) == 2 and file_ext[1] == '.md' and mtype == 'text/plain': - sorted_accept = sort_accept_tuple(request.META.get('HTTP_ACCEPT')) + if filename.suffix == ".md" and mtype == "text/plain": + sorted_accept = sort_accept_tuple(request.META.get("HTTP_ACCEPT")) for atype in sorted_accept: - if atype[0] == 'text/markdown': - content_type = content_type.replace('plain', 'markdown', 1) - break; - elif atype[0] == 'text/html': - bytes = "\n\n\n%s\n\n\n" % markdown.markdown(bytes.decode(encoding=chset)) - content_type = content_type.replace('plain', 'html', 1) - break; - elif atype[0] == 'text/plain': - break; + if atype[0] == "text/markdown": + content_type = content_type.replace("plain", "markdown", 1) + break + elif atype[0] == "text/html": + # Render markdown, allowing that charset may be inaccurate. + try: + md_src = bytes.decode( + "utf-8" if chset in ["ascii", "us-ascii"] else chset + ) + except UnicodeDecodeError: + # latin-1, aka iso8859-1, accepts all 8-bit code points + md_src = bytes.decode("latin-1") + content = markdown.markdown(md_src) # a string + bytes = render_to_string( + "minimal.html", + { + "content": content, + "title": filename.name, + "static_ietf_org": settings.STATIC_IETF_ORG, + }, + ).encode("utf-8") + content_type = "text/html; charset=utf-8" + break + elif atype[0] == "text/plain": + break response = HttpResponse(bytes, content_type=content_type) - response['Content-Disposition'] = 'inline; filename="%s"' % basename + response["Content-Disposition"] = f'inline; filename="{filename.name}"' return response else: return HttpResponseRedirect(redirect_to=doc.get_href(meeting=meeting)) + +@requires_api_token("ietf.meeting.views.api_resolve_materials_name") +def api_resolve_materials_name_cached(request, document, num=None, ext=None): + """Resolve materials name into document to a blob spec + + Returns the bucket/name of a blob in the blob store that corresponds to the named + document. Handles resolution of revision if it is not specified and determines the + best extension if one is not provided. Response is JSON. + + As of 2025-10-10 we do not have blobs for all materials documents or for every + format of every document. This API still returns the bucket/name as if the blob + exists. Another API will allow the caller to obtain the file contents using that + name if it cannot be retrieved from the blob store. + """ + + def _error_response(status: int, detail: str): + return JsonResponse( + { + "status": status, + "title": "Error", + "detail": detail, + }, + status=status, + ) + + def _response(bucket: str, name: str): + return JsonResponse( + { + "bucket": bucket, + "name": name, + } + ) + + try: + resolved = ResolvedMaterial.objects.get( + meeting_number=num, name=document + ) + except ResolvedMaterial.DoesNotExist: + return _error_response( + HTTP_404_NOT_FOUND, f"No suitable file for {document} for meeting {num}" + ) + return _response(bucket=resolved.bucket, name=resolved.blob) + + +@requires_api_token +def api_retrieve_materials_blob(request, bucket, name): + """Retrieve contents of a meeting materials blob + + This is intended as a fallback if the web worker cannot retrieve a blob from + the blobstore itself. The most likely cause is retrieving an old materials document + that has not been backfilled. + + If a blob is requested that does not exist, this checks for it on the filesystem + and if found, adds it to the blobstore, creates a StoredObject record, and returns + the contents as it would have done if the blob was already present. + + As a special case, if a requested file with extension `.md.html` does not exist + but a file with the same name but extension `.md` does, `.md` file will be rendered + from markdown to html and returned / stored. + """ + DEFAULT_CONTENT_TYPES = { + ".html": "text/html;charset=utf-8", + ".md": "text/markdown;charset=utf-8", + ".pdf": "application/pdf", + ".txt": "text/plain;charset=utf-8", + } + + def _default_content_type(blob_name: str): + return DEFAULT_CONTENT_TYPES.get(Path(name).suffix, "application/octet-stream") + + if not ( + settings.ENABLE_BLOBSTORAGE + and bucket in settings.MATERIALS_TYPES_SERVED_BY_WORKER + ): + return HttpResponseNotFound(f"Bucket {bucket} not found.") + storage = storages[bucket] # if not configured, a server error will result + assert isinstance(storage, BlobdbStorage) + try: + blob = storage.open(name, "rb") + except FileNotFoundError: + pass + else: + # found the blob - return it + assert isinstance(blob, BlobFile) + log(f"Materials blob: directly returning {bucket}:{name}") + return FileResponse( + blob, + filename=name, + content_type=blob.content_type or _default_content_type(name), + ) + + # Did not find the blob. Create it if we can + name_as_path = Path(name) + if name_as_path.suffixes == [".md", ".html"]: + # special case: .md.html means we want to create the .md and the .md.html + # will come along as a bonus + name_to_store = name_as_path.stem # removes the .html + else: + name_to_store = name + + # See if we have a meeting-related document that matches the requested bucket and + # name. + try: + doc, rev = _get_materials_doc(Path(name_to_store).stem) + if doc.type_id != bucket: + raise Document.DoesNotExist + except Document.DoesNotExist: + log(f"Materials blob: no doc for {bucket}:{name}") + return HttpResponseNotFound( + f"Document corresponding to {bucket}:{name} not found." + ) + else: + # create all missing blobs for the doc while we're at it + log(f"Materials blob: storing blobs for {doc.name}-{doc.rev}") + store_blobs_for_one_material_doc(doc) + + # If we can make the blob at all, it now exists, so return it or a 404 + try: + blob = storage.open(name, "rb") + except FileNotFoundError: + log(f"Materials blob: no blob for {bucket}:{name}") + return HttpResponseNotFound(f"Object {bucket}:{name} not found.") + else: + # found the blob - return it + assert isinstance(blob, BlobFile) + return FileResponse( + blob, + filename=name, + content_type=blob.content_type or _default_content_type(name), + ) + + @login_required def materials_editable_groups(request, num=None): meeting = get_meeting(num) @@ -281,8 +508,13 @@ def materials_editable_groups(request, num=None): @role_required('Secretariat') def edit_timeslots(request, num=None): - meeting = get_meeting(num) + if 'sched' in request.GET: + schedule = Schedule.objects.filter(pk=request.GET.get('sched', None)).first() + schedule_edit_url = _schedule_edit_url(meeting, schedule) + else: + schedule_edit_url = None + with timezone.override(meeting.tz()): if request.method == 'POST': # handle AJAX requests @@ -333,6 +565,7 @@ def edit_timeslots(request, num=None): "slot_slices": slots, "date_slices":date_slices, "meeting":meeting, + "schedule_edit_url": schedule_edit_url, "ts_list":ts_list, "ts_with_official_assignments": ts_with_official_assignments, "ts_with_any_assignments": ts_with_any_assignments, @@ -659,7 +892,9 @@ def prepare_timeslots_for_display(timeslots, rooms): sorted_rooms = sorted( rooms_with_timeslots, key=lambda room: ( - # First, sort regular session rooms ahead of others - these will usually + # Sort lower capacity rooms first. + room.capacity if room.capacity is not None else math.inf, # sort rooms with capacity = None at end + # Sort regular session rooms ahead of others - these will usually # have more timeslots than other room types. 0 if room_data[room.pk]['timeslot_count'] == max_timeslots else 1, # Sort rooms with earlier timeslots ahead of later @@ -669,8 +904,6 @@ def prepare_timeslots_for_display(timeslots, rooms): # Sort by list of starting time and duration so that groups with identical # timeslot structure will be neighbors. The grouping algorithm relies on this! room_data[room.pk]['start_and_duration'], - # Within each group, sort higher capacity rooms first. - -room.capacity if room.capacity is not None else 1, # sort rooms with capacity = None at end # Finally, sort alphabetically by name room.name ) @@ -933,6 +1166,7 @@ def cubehelix(i, total, hue=1.2, start_angle=0.5): 'rtg' : { 'dark' : (222, 219, 124) , 'light' : (247, 247, 233) }, 'sec' : { 'dark' : (0, 114, 178) , 'light' : (245, 252, 248) }, 'tsv' : { 'dark' : (117,201,119) , 'light' : (251, 252, 255) }, + 'wit' : { 'dark' : (117,201,119) , 'light' : (251, 252, 255) }, # intentionally the same as tsv } for i, p in enumerate(session_parents): if p.acronym in liz_preferred_colors: @@ -1441,6 +1675,11 @@ def list_schedules(request, num): class DiffSchedulesForm(forms.Form): from_schedule = forms.ChoiceField() to_schedule = forms.ChoiceField() + show_room_changes = forms.BooleanField( + initial=False, + required=False, + help_text="Include changes to room without a date or time change", + ) def __init__(self, meeting, user, *args, **kwargs): super().__init__(*args, **kwargs) @@ -1473,6 +1712,14 @@ def diff_schedules(request, num): raw_diffs = diff_meeting_schedules(from_schedule, to_schedule) diffs = prefetch_schedule_diff_objects(raw_diffs) + if not form.cleaned_data["show_room_changes"]: + # filter out room-only changes + diffs = [ + d + for d in diffs + if (d["change"] != "move") or (d["from"].time != d["to"].time) + ] + for d in diffs: s = d['session'] s.session_label = s.short_name @@ -1510,7 +1757,7 @@ def get_assignments_for_agenda(schedule): @ensure_csrf_cookie -def agenda_plain(request, num=None, name=None, base=None, ext=None, owner=None, utc=""): +def agenda_plain(request, num=None, name=None, base=None, ext=None, owner=None, utc=None): base = base if base else 'agenda' ext = ext if ext else '.txt' mimetype = { @@ -1539,7 +1786,7 @@ def agenda_plain(request, num=None, name=None, base=None, ext=None, owner=None, person = get_person_by_email(owner) schedule = get_schedule_by_name(meeting, person, name) - if schedule == None: + if schedule is None: base = base.replace("-utc", "") return render(request, "meeting/no-"+base+ext, {'meeting':meeting }, content_type=mimetype[ext]) @@ -1554,13 +1801,13 @@ def agenda_plain(request, num=None, name=None, base=None, ext=None, owner=None, # Done processing for CSV output if ext == ".csv": - return agenda_csv(schedule, filtered_assignments) + return agenda_csv(schedule, filtered_assignments, utc=utc is not None) filter_organizer = AgendaFilterOrganizer(assignments=filtered_assignments) is_current_meeting = (num is None) or (num == get_current_ietf_meeting_num()) - display_timezone = 'UTC' if utc else meeting.time_zone + display_timezone = meeting.time_zone if utc is None else 'UTC' with timezone.override(display_timezone): rendered_page = render( request, @@ -1575,7 +1822,6 @@ def agenda_plain(request, num=None, name=None, base=None, ext=None, owner=None, "now": timezone.now().astimezone(meeting.tz()), "display_timezone": display_timezone, "is_current_meeting": is_current_meeting, - "use_notes": meeting.uses_notes(), "cache_time": 150 if is_current_meeting else 3600, }, content_type=mimetype[ext], @@ -1606,15 +1852,33 @@ def agenda(request, num=None, name=None, base=None, ext=None, owner=None, utc="" } }) -@cache_page(5 * 60) -def api_get_agenda_data (request, num=None): + +def generate_agenda_data(num=None, force_refresh=False): + """Generate data for the api_get_agenda_data endpoint + + :num: meeting number + :force_refresh: True to force a refresh of the cache + """ meeting = get_ietf_meeting(num) if meeting is None: raise Http404("No such full IETF meeting") elif int(meeting.number) <= 64: - return Http404("Pre-IETF 64 meetings are not available through this API") - else: - pass + raise Http404("Pre-IETF 64 meetings are not available through this API") + is_current_meeting = meeting.number == get_current_ietf_meeting_num() + + cache = caches["agenda"] + cache_timeout = ( + settings.AGENDA_CACHE_TIMEOUT_CURRENT_MEETING + if is_current_meeting + else settings.AGENDA_CACHE_TIMEOUT_DEFAULT + ) + cache_format = "1" # bump this on backward-incompatible data format changes + + cache_key = f"generate_agenda_data:{meeting.number}:v{cache_format}" + if not force_refresh: + cached_value = cache.get(cache_key) + if cached_value is not None: + return cached_value # Select the schedule to show schedule = get_schedule(meeting, None) @@ -1630,14 +1894,13 @@ def api_get_agenda_data (request, num=None): filter_organizer = AgendaFilterOrganizer(assignments=filtered_assignments) - is_current_meeting = (num is None) or (num == get_current_ietf_meeting_num()) - # Get Floor Plans floors = FloorPlan.objects.filter(meeting=meeting).order_by('order') + + # Get Preliminary Agenda Date + prelimAgendaDate = ImportantDate.objects.filter(name_id="prelimagenda", meeting=meeting).first() - #debug.show('all([(item.acronym,item.session.order_number,item.session.order_in_meeting()) for item in filtered_assignments])') - - return JsonResponse({ + result = { "meeting": { "number": schedule.meeting.number, "city": schedule.meeting.city, @@ -1646,83 +1909,109 @@ def api_get_agenda_data (request, num=None): "updated": updated, "timezone": meeting.time_zone, "infoNote": schedule.meeting.agenda_info_note, - "warningNote": schedule.meeting.agenda_warning_note + "warningNote": schedule.meeting.agenda_warning_note, + "prelimAgendaDate": prelimAgendaDate.date.isoformat() if prelimAgendaDate else "" }, "categories": filter_organizer.get_filter_categories(), "isCurrentMeeting": is_current_meeting, - "useNotes": meeting.uses_notes(), + "usesNotes": meeting.uses_notes(), "schedule": list(map(agenda_extract_schedule, filtered_assignments)), "floors": list(map(agenda_extract_floorplan, floors)) - }) + } + cache.set(cache_key, result, timeout=cache_timeout) + return result -def api_get_session_materials (request, session_id=None): - session = get_object_or_404(Session,pk=session_id) + +def api_get_agenda_data(request, num=None): + return JsonResponse(generate_agenda_data(num, force_refresh=False)) + + +def api_get_session_materials(request, session_id=None): + session = get_object_or_404(Session, pk=session_id) minutes = session.minutes() slides_actions = [] - if can_manage_session_materials(request.user, session.group, session): - slides_actions.append({ - 'label': 'Upload slides', - 'url': reverse( - 'ietf.meeting.views.upload_session_slides', - kwargs={'num': session.meeting.number, 'session_id': session.pk}, - ), - }) - elif not session.is_material_submission_cutoff(): - slides_actions.append({ - 'label': 'Propose slides', - 'url': reverse( - 'ietf.meeting.views.propose_session_slides', - kwargs={'num': session.meeting.number, 'session_id': session.pk}, - ), - }) + if can_manage_session_materials(request.user, session.group, session) or not session.is_material_submission_cutoff(): + slides_actions.append( + { + "label": "Upload slides", + "url": reverse( + "ietf.meeting.views.upload_session_slides", + kwargs={"num": session.meeting.number, "session_id": session.pk}, + ), + } + ) else: pass # no action available if it's past cutoff - return JsonResponse({ - "url": session.agenda().get_href(), - "slides": { - "decks": list(map(agenda_extract_slide, session.slides())), - "actions": slides_actions, - }, - "minutes": { - "id": minutes.id, - "title": minutes.title, - "url": minutes.get_href(), - "ext": minutes.file_extension() - } if minutes is not None else None - }) + agenda = session.agenda() + agenda_url = agenda.get_href() if agenda is not None else None + return JsonResponse( + { + "url": agenda_url, + "slides": { + "decks": [ + agenda_extract_slide(slide) | {"order": order} # add "order" field + for order, slide in enumerate(session.slides()) + ], + "actions": slides_actions, + }, + "minutes": { + "id": minutes.id, + "title": minutes.title, + "url": minutes.get_href(), + "ext": minutes.file_extension(), + } + if minutes is not None + else None, + } + ) -def agenda_extract_schedule (item): + +def agenda_extract_schedule(item): + if item.session.current_status == "resched": + resched_to = item.session.tombstone_for.official_timeslotassignment() + else: + resched_to = None return { "id": item.id, + "slug": item.slug(), "sessionId": item.session.id, - "room": item.room_name if item.timeslot.show_location else None, + "room": (item.timeslot.get_location() or None) if item.timeslot else None, "location": { "short": item.timeslot.location.floorplan.short, "name": item.timeslot.location.floorplan.name, } if (item.timeslot.show_location and item.timeslot.location and item.timeslot.location.floorplan) else {}, "acronym": item.acronym, - "duration": item.timeslot.duration.seconds, - "name": item.timeslot.name, + "duration": item.timeslot.duration.total_seconds(), + "name": item.session.name, + "slotId": item.timeslot.id, + "slotName": item.timeslot.name, + "slotModified": item.timeslot.modified.isoformat(), "startDateTime": item.timeslot.time.isoformat(), "status": item.session.current_status, + "rescheduledTo": { + "startDateTime": resched_to.timeslot.time.isoformat(), + "duration": resched_to.timeslot.duration.total_seconds(), + } if resched_to is not None else {}, "type": item.session.type.slug, + "purpose": item.session.purpose.slug, "isBoF": item.session.group_at_the_time().state_id == "bof", + "isProposed": item.session.group_at_the_time().state_id == "proposed", "filterKeywords": item.filter_keywords, "groupAcronym": item.session.group_at_the_time().acronym, "groupName": item.session.group_at_the_time().name, - "groupParent": { + "groupParent": ({ "acronym": item.session.group_parent_at_the_time().acronym - } if item.session.group_parent_at_the_time() else {}, + } if item.session.group_parent_at_the_time() else {}), "note": item.session.agenda_note, "remoteInstructions": item.session.remote_instructions, "flags": { "agenda": True if item.session.agenda() is not None else False, - "showAgenda": True if (item.session.agenda() is not None or item.session.remote_instructions or item.session.agenda_note) else False + "showAgenda": True if (item.session.agenda() is not None or item.session.remote_instructions) else False }, "agenda": { - "url": item.session.agenda().get_href() + "url": item.session.agenda().get_versionless_href() } if item.session.agenda() is not None else { "url": None }, @@ -1733,10 +2022,10 @@ def agenda_extract_schedule (item): "chat" : item.session.chat_room_url(), "chatArchive" : item.session.chat_archive_url(), "recordings": list(map(agenda_extract_recording, item.session.recordings())), - "videoStream": item.timeslot.location.video_stream_url() if item.timeslot.location else "", - "audioStream": item.timeslot.location.audio_stream_url() if item.timeslot.location else "", + "videoStream": item.session.video_stream_url() or "", + "audioStream": item.session.audio_stream_url() or "", "webex": item.timeslot.location.webex_url() if item.timeslot.location else "", - "onsiteTool": item.timeslot.location.onsite_tool_url() if item.timeslot.location else "", + "onsiteTool": item.session.onsite_tool_url() or "", "calendar": reverse( 'ietf.meeting.views.agenda_ical', kwargs={'num': item.schedule.meeting.number, 'session_id': item.session.id}, @@ -1747,7 +2036,8 @@ def agenda_extract_schedule (item): # } } -def agenda_extract_floorplan (item): + +def agenda_extract_floorplan(item): try: item.image.width except FileNotFoundError: @@ -1760,10 +2050,11 @@ def agenda_extract_floorplan (item): "short": item.short, "width": item.image.width, "height": item.image.height, - "rooms": list(map(agenda_extract_room, item.room_set.all())) + "rooms": list(map(agenda_extract_room, item.room_set.all())), } -def agenda_extract_room (item): + +def agenda_extract_room(item): return { "id": item.id, "name": item.name, @@ -1775,7 +2066,8 @@ def agenda_extract_room (item): "bottom": item.bottom() } -def agenda_extract_recording (item): + +def agenda_extract_recording(item): return { "id": item.id, "name": item.name, @@ -1783,27 +2075,30 @@ def agenda_extract_recording (item): "url": item.external_url } -def agenda_extract_slide (item): + +def agenda_extract_slide(item): return { "id": item.id, "title": item.title, - "url": item.get_versionless_href(), - "ext": item.file_extension() + "rev": item.rev, + "url": item.get_href(), + "ext": item.file_extension(), } -def agenda_csv(schedule, filtered_assignments): - response = HttpResponse(content_type="text/csv; charset=%s"%settings.DEFAULT_CHARSET) + +def agenda_csv(schedule, filtered_assignments, utc=False): + encoding = 'utf-8' + response = HttpResponse(content_type=f"text/csv; charset={encoding}") writer = csv.writer(response, delimiter=str(','), quoting=csv.QUOTE_ALL) headings = ["Date", "Start", "End", "Session", "Room", "Area", "Acronym", "Type", "Description", "Session ID", "Agenda", "Slides"] def write_row(row): - encoded_row = [v.encode('utf-8') if isinstance(v, str) else v for v in row] - - while len(encoded_row) < len(headings): - encoded_row.append(None) # produce empty entries at the end as necessary - - writer.writerow(encoded_row) + if len(row) < len(headings): + padding = [None] * (len(headings) - len(row)) # produce empty entries at the end as necessary + else: + padding = [] + writer.writerow(row + padding) def agenda_field(item): agenda_doc = item.session.agenda() @@ -1817,11 +2112,12 @@ def slides_field(item): write_row(headings) + tz = datetime.UTC if utc else schedule.meeting.tz() for item in filtered_assignments: row = [] - row.append(item.timeslot.time.strftime("%Y-%m-%d")) - row.append(item.timeslot.time.strftime("%H%M")) - row.append(item.timeslot.end_time().strftime("%H%M")) + row.append(item.timeslot.time.astimezone(tz).strftime("%Y-%m-%d")) + row.append(item.timeslot.time.astimezone(tz).strftime("%H%M")) + row.append(item.timeslot.end_time().astimezone(tz).strftime("%H%M")) if item.slot_type().slug == "break": row.append(item.slot_type().name) @@ -1884,8 +2180,10 @@ def agenda_by_type_ics(request,num=None,type=None): ).order_by('session__type__slug','timeslot__time') if type: assignments = assignments.filter(session__type__slug=type) - updated = meeting.updated() - return render(request,"meeting/agenda.ics",{"schedule":schedule,"updated":updated,"assignments":assignments},content_type="text/calendar") + + return render_icalendar(schedule, assignments) + + def session_draft_list(num, acronym): try: @@ -2005,6 +2303,246 @@ def ical_session_status(assignment): else: return "CONFIRMED" + +def render_icalendar_precomp(agenda_data): + ical_content = generate_agenda_ical_precomp(agenda_data) + return HttpResponse(ical_content, content_type="text/calendar") + + +def render_icalendar(schedule, assignments): + ical_content = generate_agenda_ical(schedule, assignments) + return HttpResponse(ical_content, content_type="text/calendar") + + +def generate_agenda_ical_precomp(agenda_data): + """Generate iCalendar from precomputed data using the icalendar library""" + + cal = Calendar() + cal.add("prodid", "-//IETF//datatracker.ietf.org ical agenda//EN") + cal.add("version", "2.0") + cal.add("method", "PUBLISH") + + meeting_data = agenda_data["meeting"] + for item in agenda_data["schedule"]: + event = Event() + + uid = f"ietf-{meeting_data["number"]}-{item["slotId"]}-{item["acronym"]}" + event.add("uid", uid) + + # add custom field with meeting's local TZ + event.add("x-meeting-tz", meeting_data["timezone"]) + + if item["name"]: + summary = item["name"] + else: + summary = f"{item["groupAcronym"]} - {item["groupName"]}" + + if item["note"]: + summary += f" ({item["note"]})" + + event.add("summary", summary) + + if item["room"]: + event.add("location", item["room"]) # room name + + if item["status"] == "canceled": + status = "CANCELLED" + elif item["status"] == "resched": + resched_to = item["rescheduledTo"] + if resched_to is None: + status = "RESCHEDULED" + else: + resched_start = datetime.datetime.fromisoformat( + resched_to["startDateTime"] + ) + dur = datetime.timedelta(seconds=resched_to["duration"]) + resched_end = resched_start + dur + formatted_start = resched_start.strftime("%A %H:%M").upper() + formatted_end = resched_end.strftime("%H:%M") + status = f"RESCHEDULED TO {formatted_start}-{formatted_end}" + else: + status = "CONFIRMED" + event.add("status", status) + + event.add("class", "PUBLIC") + + start_time = datetime.datetime.fromisoformat(item["startDateTime"]) + duration = datetime.timedelta(seconds=item["duration"]) + event.add("dtstart", start_time) + event.add("dtend", start_time + duration) + + # DTSTAMP: when the event was created or last modified (in UTC) + # n.b. timeslot.modified may not be an accurate measure of this + event.add("dtstamp", datetime.datetime.fromisoformat(item["slotModified"])) + + description_parts = [item["slotName"]] + + if item["note"]: + description_parts.append(f"Note: {item["note"]}") + + links = item["links"] + if links["onsiteTool"]: + description_parts.append(f"Onsite tool: {links["onsiteTool"]}") + + if links["videoStream"]: + description_parts.append(f"Meetecho: {links["videoStream"]}") + + if links["webex"]: + description_parts.append(f"Webex: {links["webex"]}") + + if item["remoteInstructions"]: + description_parts.append( + f"Remote instructions: {item["remoteInstructions"]}" + ) + + try: + materials_url = absurl( + "ietf.meeting.views.session_details", + num=meeting_data["number"], + acronym=item["acronym"], + ) + except NoReverseMatch: + pass + else: + description_parts.append(f"Session materials: {materials_url}") + event.add("url", materials_url) + + if meeting_data["number"].isdigit(): + try: + agenda_url = absurl("agenda", num=meeting_data["number"]) + except NoReverseMatch: + pass + else: + description_parts.append(f"See in schedule: {agenda_url}#row-{item["slug"]}") + + if item["agenda"] and item["agenda"]["url"]: + description_parts.append(f"Agenda {item["agenda"]["url"]}") + + # Join all description parts with 2 newlines + description = "\n\n".join(description_parts) + event.add("description", description) + + # Add event to calendar + cal.add_component(event) + + return cal.to_ical().decode("utf-8") + + +def generate_agenda_ical(schedule, assignments): + """Generate iCalendar using the icalendar library""" + + cal = Calendar() + cal.add("prodid", "-//IETF//datatracker.ietf.org ical agenda//EN") + cal.add("version", "2.0") + cal.add("method", "PUBLISH") + + for item in assignments: + event = Event() + + uid = f"ietf-{schedule.meeting.number}-{item.timeslot.pk}-{item.session.group.acronym}" + event.add("uid", uid) + + # add custom field with meeting's local TZ + event.add("x-meeting-tz", schedule.meeting.time_zone) + + if item.session.name: + summary = item.session.name + else: + group = item.session.group_at_the_time() + summary = f"{group.acronym} - {group.name}" + + if item.session.agenda_note: + summary += f" ({item.session.agenda_note})" + + event.add("summary", summary) + + if item.timeslot.show_location and item.timeslot.get_location(): + event.add("location", item.timeslot.get_location()) + + if item.session and hasattr(item.session, "current_status"): + status = ical_session_status(item) + else: + status = "" + event.add("status", status) + + event.add("class", "PUBLIC") + + event.add("dtstart", item.timeslot.utc_start_time()) + event.add("dtend", item.timeslot.utc_end_time()) + + # DTSTAMP: when the event was created or last modified (in UTC) + dtstamp = item.timeslot.modified.astimezone(pytz.UTC) + event.add("dtstamp", dtstamp) + + description_parts = [item.timeslot.name] + + if item.session.agenda_note: + description_parts.append(f"Note: {item.session.agenda_note}") + + if hasattr(item.session, "onsite_tool_url") and callable( + item.session.onsite_tool_url + ): + onsite_url = item.session.onsite_tool_url() + if onsite_url: + description_parts.append(f"Onsite tool: {onsite_url}") + + if hasattr(item.session, "video_stream_url") and callable( + item.session.video_stream_url + ): + video_url = item.session.video_stream_url() + if video_url: + description_parts.append(f"Meetecho: {video_url}") + + if ( + item.timeslot.location + and hasattr(item.timeslot.location, "webex_url") + and callable(item.timeslot.location.webex_url) + and item.timeslot.location.webex_url() is not None + ): + description_parts.append(f"Webex: {item.timeslot.location.webex_url()}") + + if item.session.remote_instructions: + description_parts.append( + f"Remote instructions: {item.session.remote_instructions}" + ) + + try: + materials_url = absurl( + "ietf.meeting.views.session_details", + num=schedule.meeting.number, + acronym=item.session.group.acronym, + ) + description_parts.append(f"Session materials: {materials_url}") + event.add("url", materials_url) + except: + pass + + if ( + hasattr(schedule.meeting, "get_number") + and schedule.meeting.get_number() is not None + ): + try: + agenda_url = absurl("agenda", num=schedule.meeting.number) + description_parts.append( + f"See in schedule: {agenda_url}#row-{item.slug()}" + ) + except: + pass + + agenda = item.session.agenda() + if agenda and hasattr(agenda, "get_versionless_href"): + agenda_url = agenda.get_versionless_href() + description_parts.append(f"{agenda.type} {agenda_url}") + + # Join all description parts with 2 newlines + description = "\n\n".join(description_parts) + event.add("description", description) + + # Add event to calendar + cal.add_component(event) + + return cal.to_ical().decode("utf-8") + def parse_agenda_filter_params(querydict): """Parse agenda filter parameters from a request""" if len(querydict) == 0: @@ -2024,31 +2562,43 @@ def parse_agenda_filter_params(querydict): def should_include_assignment(filter_params, assignment): """Decide whether to include an assignment""" - shown = len(set(filter_params['show']).intersection(assignment.filter_keywords)) > 0 - hidden = len(set(filter_params['hide']).intersection(assignment.filter_keywords)) > 0 + if hasattr(assignment, "filter_keywords"): + kw = assignment.filter_keywords + elif isinstance(assignment, dict): + kw = assignment.get("filterKeywords", []) + else: + raise ValueError("Unsupported assignment instance") + shown = len(set(filter_params['show']).intersection(kw)) > 0 + hidden = len(set(filter_params['hide']).intersection(kw)) > 0 return shown and not hidden -def agenda_ical(request, num=None, name=None, acronym=None, session_id=None): - """Agenda ical view - - By default, all agenda items will be shown. A filter can be specified in - the querystring. It has the format - - ?show=...&hide=...&showtypes=...&hidetypes=... - where any of the parameters can be omitted. The right-hand side of each - '=' is a comma separated list, which can be empty. If none of the filter - parameters are specified, no filtering will be applied, even if the query - string is not empty. +def agenda_ical_ietf(meeting, filt_params, acronym=None, session_id=None): + agenda_data = generate_agenda_data(meeting.number, force_refresh=False) + if acronym: + agenda_data["schedule"] = [ + item + for item in agenda_data["schedule"] + if item["groupAcronym"] == acronym + ] + elif session_id: + agenda_data["schedule"] = [ + item + for item in agenda_data["schedule"] + if item["sessionId"] == session_id + ] + if filt_params is not None: + # Apply the filter + agenda_data["schedule"] = [ + item + for item in agenda_data["schedule"] + if should_include_assignment(filt_params, item) + ] + return render_icalendar_precomp(agenda_data) - The show and hide parameters each take a list of working group (wg) acronyms. - The showtypes and hidetypes parameters take a list of session types. - Hiding (by wg or type) takes priority over showing. - """ - meeting = get_meeting(num, type_in=None) - schedule = get_schedule(meeting, name) - updated = meeting.updated() +def agenda_ical_interim(meeting, filt_params, acronym=None, session_id=None): + schedule = get_schedule(meeting) if schedule is None and acronym is None and session_id is None: raise Http404 @@ -2060,11 +2610,6 @@ def agenda_ical(request, num=None, name=None, acronym=None, session_id=None): assignments = preprocess_assignments_for_agenda(assignments, meeting) AgendaKeywordTagger(assignments=assignments).apply() - try: - filt_params = parse_agenda_filter_params(request.GET) - except ValueError as e: - return HttpResponseBadRequest(str(e)) - if filt_params is not None: # Apply the filter assignments = [a for a in assignments if should_include_assignment(filt_params, a)] @@ -2074,43 +2619,83 @@ def agenda_ical(request, num=None, name=None, acronym=None, session_id=None): elif session_id: assignments = [ a for a in assignments if a.session_id == int(session_id) ] - for a in assignments: - if a.session: - a.session.ical_status = ical_session_status(a) + return render_icalendar(schedule, assignments) - return render(request, "meeting/agenda.ics", { - "schedule": schedule, - "assignments": assignments, - "updated": updated - }, content_type="text/calendar") -@cache_page(15 * 60) -def agenda_json(request, num=None): - meeting = get_meeting(num, type_in=['ietf','interim']) +def agenda_ical(request, num=None, acronym=None, session_id=None): + """Agenda ical view - sessions = [] - locations = set() - parent_acronyms = set() - assignments = SchedTimeSessAssignment.objects.filter( - schedule__in=[meeting.schedule, meeting.schedule.base if meeting.schedule else None], - session__on_agenda=True, - ).exclude( - session__type__in=['break', 'reg'] - ) - # Update the assignments with historic information, i.e., valid at the - # time of the meeting - assignments = preprocess_assignments_for_agenda(assignments, meeting, extra_prefetches=[ - "session__materials__docevent_set", - "session__sessionpresentation_set", - "timeslot__meeting" - ]) - for asgn in assignments: - sessdict = dict() - sessdict['objtype'] = 'session' - sessdict['id'] = asgn.pk - sessdict['is_bof'] = False - if asgn.session.group_at_the_time(): - sessdict['group'] = { + If num is None, looks for the next IETF meeting. Otherwise, uses the requested meeting + regardless of its type. + + By default, all agenda items will be shown. A filter can be specified in + the querystring. It has the format + + ?show=...&hide=...&showtypes=...&hidetypes=... + + where any of the parameters can be omitted. The right-hand side of each + '=' is a comma separated list, which can be empty. If none of the filter + parameters are specified, no filtering will be applied, even if the query + string is not empty. + + The show and hide parameters each take a list of working group (wg) acronyms. + The showtypes and hidetypes parameters take a list of session types. + + Hiding (by wg or type) takes priority over showing. + """ + if num is None: + meeting = get_ietf_meeting() + if meeting is None: + raise Http404 + else: + meeting = get_meeting(num, type_in=None) # get requested meeting, whatever its type + + if isinstance(session_id, str) and session_id.isdigit(): + session_id = int(session_id) + + try: + filt_params = parse_agenda_filter_params(request.GET) + except ValueError as e: + return HttpResponseBadRequest(str(e)) + + if meeting.type_id == "ietf": + return agenda_ical_ietf(meeting, filt_params, acronym, session_id) + else: + return agenda_ical_interim(meeting, filt_params, acronym, session_id) + + +@cache_page(15 * 60) +def agenda_json(request, num=None): + if num is None: + meeting = get_ietf_meeting() + if meeting is None: + raise Http404 + else: + meeting = get_meeting(num, type_in=None) # get requested meeting, whatever its type + + sessions = [] + locations = set() + parent_acronyms = set() + assignments = SchedTimeSessAssignment.objects.filter( + schedule__in=[meeting.schedule, meeting.schedule.base if meeting.schedule else None], + session__on_agenda=True, + ).exclude( + session__type__in=['break', 'reg'] + ) + # Update the assignments with historic information, i.e., valid at the + # time of the meeting + assignments = preprocess_assignments_for_agenda(assignments, meeting, extra_prefetches=[ + "session__materials__docevent_set", + "session__presentations", + "timeslot__meeting" + ]) + for asgn in assignments: + sessdict = dict() + sessdict['objtype'] = 'session' + sessdict['id'] = asgn.pk + sessdict['is_bof'] = False + if asgn.session.group_at_the_time(): + sessdict['group'] = { "acronym": asgn.session.group_at_the_time().acronym, "name": asgn.session.group_at_the_time().name, "type": asgn.session.group_at_the_time().type_id, @@ -2212,6 +2797,70 @@ def agenda_json(request, num=None): response['Last-Modified'] = format_date_time(timegm(last_modified.timetuple())) return response +def request_summary_filter(session): + if (session.group.area is None + or session.group.type.slug in request_summary_exclude_group_types + or session.current_status == 'notmeet'): + return False + return True + +def get_area_column(area): + if area is None: + return '' + if area.type.slug in ['rfcedtyp']: + name = 'OTHER' + else: + name = area.acronym.upper() + return name + +def get_summary_by_area(sessions): + """Returns summary by area for list of session requests. + Summary is a two dimensional array[row=session duration][col=session area count] + It also includes row and column headers as well as a totals row. + """ + + # first build a dictionary of counts, key=(duration,area) + durations = set() + areas = set() + duration_totals = defaultdict(int) + data = defaultdict(int) + for session in sessions: + area_column = get_area_column(session.group.area) + duration = session.requested_duration.seconds / 3600 + key = (duration, area_column) + data[key] = data[key] + 1 + durations.add(duration) + areas.add(area_column) + duration_totals[duration] = duration_totals[duration] + 1 + + # build two dimensional array for use in template + rows = [] + sorted_areas = sorted(areas) + # move "other" to end + if 'OTHER' in sorted_areas: + sorted_areas.remove('OTHER') + sorted_areas.append('OTHER') + # add header row + rows.append(['Duration'] + sorted_areas + ['TOTAL SLOTS', 'TOTAL HOURS']) + for duration in sorted(durations): + rows.append([duration] + [data[(duration, a)] for a in sorted_areas] + [duration_totals[duration]] + [duration_totals[duration] * duration]) + # add total row + rows.append(['Total Slots'] + [sum([rows[r][c] for r in range(1, len(rows))]) for c in range(1, len(rows[0]))]) + rows.append(['Total Hours'] + [sum([d * data[(d, area)] for d in durations]) for area in sorted_areas]) + return rows + +def get_summary_by_type(sessions): + counter = Counter([s.group.type.name for s in sessions]) + data = counter.most_common() + data.insert(0, ('Group Type', 'Count')) + return data + +def get_summary_by_purpose(sessions): + counter = Counter([s.purpose.name for s in sessions]) + data = counter.most_common() + data.insert(0, ('Purpose', 'Count')) + return data + def meeting_requests(request, num=None): meeting = get_meeting(num) groups_to_show = Group.objects.filter( @@ -2227,7 +2876,7 @@ def meeting_requests(request, num=None): ).with_current_status().with_requested_by().exclude( requested_by=0 ).prefetch_related( - "group","group__ad_role__person" + "group", "group__ad_role__person", "group__type" ) ) @@ -2250,12 +2899,14 @@ def meeting_requests(request, num=None): ) groups_not_meeting = groups_to_show.exclude( - acronym__in = [session.group.acronym for session in sessions] + acronym__in=[session.group.acronym for session in sessions] ).order_by( "parent__acronym", "acronym", ).prefetch_related("parent") + summary_sessions = list(filter(request_summary_filter, sessions)) + return render( request, "meeting/requests.html", @@ -2263,6 +2914,9 @@ def meeting_requests(request, num=None): "meeting": meeting, "sessions": sessions, "groups_not_meeting": groups_not_meeting, + "summary_by_area": get_summary_by_area(summary_sessions), + "summary_by_group_type": get_summary_by_type(summary_sessions), + "summary_by_purpose": get_summary_by_purpose(summary_sessions), }, ) @@ -2302,12 +2956,28 @@ def session_details(request, num, acronym): session.cancelled = session.current_status in Session.CANCELED_STATUSES session.status = status_names.get(session.current_status, session.current_status) - session.filtered_artifacts = list(session.sessionpresentation_set.filter(document__type__slug__in=['agenda','minutes','bluesheets'])) - session.filtered_artifacts.sort(key=lambda d:['agenda','minutes','bluesheets'].index(d.document.type.slug)) - session.filtered_slides = session.sessionpresentation_set.filter(document__type__slug='slides').order_by('order') - session.filtered_drafts = session.sessionpresentation_set.filter(document__type__slug='draft') - session.filtered_chatlog_and_polls = session.sessionpresentation_set.filter(document__type__slug__in=('chatlog', 'polls')).order_by('document__type__slug') - # TODO FIXME Deleted materials shouldn't be in the sessionpresentation_set + if session.meeting.type_id == 'ietf' and not session.meeting.proceedings_final: + artifact_types = ['agenda','minutes','narrativeminutes'] + if Attended.objects.filter(session=session).exists(): + session.type_counter.update(['bluesheets']) + ota = session.official_timeslotassignment() + sess_time = ota and ota.timeslot.time + session.bluesheet_title = 'Attendance IETF%s: %s : %s' % (session.meeting.number, + session.group.acronym, + sess_time.strftime("%a %H:%M")) + else: + artifact_types = ['agenda','minutes','narrativeminutes','bluesheets'] + session.filtered_artifacts = list(session.presentations.filter(document__type__slug__in=artifact_types)) + session.filtered_artifacts.sort(key=lambda d:artifact_types.index(d.document.type.slug)) + session.filtered_slides = session.presentations.filter(document__type__slug='slides').order_by('order') + session.filtered_drafts = session.presentations.filter(document__type__slug='draft') + + filtered_polls = session.presentations.filter(document__type__slug=('polls')) + filtered_chatlogs = session.presentations.filter(document__type__slug=('chatlog')) + session.filtered_chatlog_and_polls = chain(filtered_chatlogs, filtered_polls) + session.chatlog = filtered_chatlogs.first() + + # TODO FIXME Deleted materials shouldn't be in the presentations for qs in [session.filtered_artifacts,session.filtered_slides,session.filtered_drafts]: qs = [p for p in qs if p.document.get_state_slug(p.document.type_id)!='deleted'] session.type_counter.update([p.document.type.slug for p in qs]) @@ -2316,31 +2986,35 @@ def session_details(request, num, acronym): # we somewhat arbitrarily use the group of the last session we get from # get_sessions() above when checking can_manage_session_materials() - can_manage = can_manage_session_materials(request.user, session.group, session) + group = session.group + can_manage = can_manage_session_materials(request.user, group, session) can_view_request = can_view_interim_request(meeting, request.user) scheduled_sessions = [s for s in sessions if s.current_status == 'sched'] unscheduled_sessions = [s for s in sessions if s.current_status != 'sched'] - pending_suggestions = None - if request.user.is_authenticated: - if can_manage: - pending_suggestions = session.slidesubmission_set.filter(status__slug='pending') - else: - pending_suggestions = session.slidesubmission_set.filter(status__slug='pending', submitter=request.user.person) + # Start with all the pending suggestions for all the group's sessions + pending_suggestions = SlideSubmission.objects.filter(session__in=sessions, status__slug='pending') + if can_manage: + pass # keep the full set + elif hasattr(request.user, "person"): + pending_suggestions = pending_suggestions.filter(submitter=request.user.person) + else: + pending_suggestions = SlideSubmission.objects.none() + tsa = session.official_timeslotassignment() + future = tsa is not None and timezone.now() < tsa.timeslot.end_time() return render(request, "meeting/session_details.html", { 'scheduled_sessions':scheduled_sessions , 'unscheduled_sessions':unscheduled_sessions , 'pending_suggestions' : pending_suggestions, 'meeting' :meeting , - 'acronym' :acronym, + 'group': group, 'is_materials_manager' : session.group.has_role(request.user, session.group.features.matman_roles), 'can_manage_materials' : can_manage, 'can_view_request': can_view_request, 'thisweek': datetime_today()-datetime.timedelta(days=7), - 'now': timezone.now(), - 'use_notes': meeting.uses_notes(), + 'future': future, }) class SessionDraftsForm(forms.Form): @@ -2365,7 +3039,7 @@ def add_session_drafts(request, session_id, num): if session.is_material_submission_cutoff() and not has_role(request.user, "Secretariat"): raise Http404 - already_linked = [sp.document for sp in session.sessionpresentation_set.filter(document__type_id='draft')] + already_linked = [sp.document for sp in session.presentations.filter(document__type_id='draft')] session_number = None sessions = get_sessions(session.meeting.number,session.group.acronym) @@ -2376,7 +3050,7 @@ def add_session_drafts(request, session_id, num): form = SessionDraftsForm(request.POST,already_linked=already_linked) if form.is_valid(): for draft in form.cleaned_data['drafts']: - session.sessionpresentation_set.create(document=draft,rev=None) + session.presentations.create(document=draft,rev=None) c = DocEvent(type="added_comment", doc=draft, rev=draft.rev, by=request.user.person) c.desc = "Added to session: %s" % session c.save() @@ -2387,10 +3061,153 @@ def add_session_drafts(request, session_id, num): return render(request, "meeting/add_session_drafts.html", { 'session': session, 'session_number': session_number, - 'already_linked': session.sessionpresentation_set.filter(document__type_id='draft'), + 'already_linked': session.presentations.filter(document__type_id='draft'), 'form': form, }) +class SessionRecordingsForm(forms.Form): + title = forms.CharField(max_length=255) + url = forms.URLField(label="URL of the recording (YouTube only)") + + def clean_url(self): + url = self.cleaned_data['url'] + parsed_url = urlparse(url) + if parsed_url.hostname not in YOUTUBE_DOMAINS: + raise forms.ValidationError("Must be a YouTube URL") + return url + + +def add_session_recordings(request, session_id, num): + # num is redundant, but we're dragging it along an artifact of where we are in the current URL structure + session = get_object_or_404(Session, pk=session_id) + if not session.can_manage_materials(request.user): + permission_denied( + request, "You don't have permission to manage recordings for this session." + ) + if session.is_material_submission_cutoff() and not has_role( + request.user, "Secretariat" + ): + raise Http404 + + session_number = None + official_timeslotassignment = session.official_timeslotassignment() + assertion("official_timeslotassignment is not None") + initial = { + "title": "Video recording of {acronym} for {timestamp}".format( + acronym=session.group.acronym, + timestamp=official_timeslotassignment.timeslot.utc_start_time().strftime( + "%Y-%m-%d %H:%M" + ), + ) + } + + # find session number if WG has more than one session at the meeting + sessions = get_sessions(session.meeting.number, session.group.acronym) + if len(sessions) > 1: + session_number = 1 + sessions.index(session) + + presentations = session.presentations.filter( + document__in=session.get_material("recording", only_one=False), + ).order_by("document__title", "document__external_url") + + if request.method == "POST": + pk_to_delete = request.POST.get("delete", None) + if pk_to_delete is not None: + session_presentation = get_object_or_404(presentations, pk=pk_to_delete) + try: + delete_recording(session_presentation) + except ValueError as err: + log(f"Error deleting recording from session {session.pk}: {err}") + messages.error( + request, + "Unable to delete this recording. Please contact the secretariat for assistance.", + ) + form = SessionRecordingsForm(initial=initial) + else: + form = SessionRecordingsForm(request.POST) + if form.is_valid(): + title = form.cleaned_data["title"] + url = form.cleaned_data["url"] + create_recording(session, url, title=title, user=request.user.person) + return redirect( + "ietf.meeting.views.session_details", + num=session.meeting.number, + acronym=session.group.acronym, + ) + else: + form = SessionRecordingsForm(initial=initial) + + return render( + request, + "meeting/add_session_recordings.html", + { + "session": session, + "session_number": session_number, + "already_linked": presentations, + "form": form, + }, + ) + + +def session_attendance(request, session_id, num): + """Session attendance view + + GET - retrieve the current session attendance or redirect to the published bluesheet if finalized + + POST - self-attest attendance for logged-in user; falls through to GET for AnonymousUser or invalid request + """ + # num is redundant, but we're dragging it along as an artifact of where we are in the current URL structure + session = get_object_or_404(Session, pk=session_id) + if session.meeting.type_id != "ietf" or session.meeting.proceedings_final: + bluesheets = session.presentations.filter( + document__type_id="bluesheets" + ) + if bluesheets: + bluesheet = bluesheets[0].document + return redirect(bluesheet.get_href(session.meeting)) + else: + raise Http404("Bluesheets not found") + + cor_cut_off_date = session.meeting.get_submission_correction_date() + today_utc = date_today(datetime.UTC) + was_there = False + can_add = False + if request.user.is_authenticated: + # use getattr() instead of request.user.person because it's a reverse OneToOne field + person = getattr(request.user, "person", None) + # Consider allowing self-declared attendance if we have a person and at least one Attended instance exists. + # The latter condition will be satisfied when Meetecho pushes their attendee records - assuming that at least + # one person will have accessed the meeting tool. This prevents people from self-declaring before they are + # marked as attending if they did log in to the meeting tool (except for a tiny window while records are + # being processed). + if person is not None and Attended.objects.filter(session=session).exists(): + was_there = Attended.objects.filter(session=session, person=person).exists() + can_add = ( + today_utc <= cor_cut_off_date + and Registration.objects.filter( + meeting=session.meeting, person=person + ).exists() + and not was_there + ) + if can_add and request.method == "POST": + session.attended_set.get_or_create( + person=person, defaults={"origin": "self declared"} + ) + can_add = False + was_there = True + + data = bluesheet_data(session) + return render( + request, + "meeting/attendance.html", + { + "session": session, + "data": data, + "can_add": can_add, + "was_there": was_there, + }, + ) + def upload_session_bluesheets(request, session_id, num): # num is redundant, but we're dragging it along an artifact of where we are in the current URL structure @@ -2417,7 +3234,10 @@ def upload_session_bluesheets(request, session_id, num): ota = session.official_timeslotassignment() sess_time = ota and ota.timeslot.time if not sess_time: - return HttpResponseGone("Cannot receive uploads for an unscheduled session. Please check the session ID.", content_type="text/plain") + return HttpResponseGone( + "Cannot receive uploads for an unscheduled session. Please check the session ID.", + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) save_error = save_bluesheet(request, session, file, encoding=form.file_encoding[file.name]) @@ -2429,7 +3249,7 @@ def upload_session_bluesheets(request, session_id, num): else: form = UploadBlueSheetForm() - bluesheet_sp = session.sessionpresentation_set.filter(document__type='bluesheets').first() + bluesheet_sp = session.presentations.filter(document__type='bluesheets').first() return render(request, "meeting/upload_session_bluesheets.html", {'session': session, @@ -2439,48 +3259,6 @@ def upload_session_bluesheets(request, session_id, num): }) -def save_bluesheet(request, session, file, encoding='utf-8'): - bluesheet_sp = session.sessionpresentation_set.filter(document__type='bluesheets').first() - _, ext = os.path.splitext(file.name) - - if bluesheet_sp: - doc = bluesheet_sp.document - doc.rev = '%02d' % (int(doc.rev)+1) - bluesheet_sp.rev = doc.rev - bluesheet_sp.save() - else: - ota = session.official_timeslotassignment() - sess_time = ota and ota.timeslot.time - - if session.meeting.type_id=='ietf': - name = 'bluesheets-%s-%s-%s' % (session.meeting.number, - session.group.acronym, - sess_time.strftime("%Y%m%d%H%M")) - title = 'Bluesheets IETF%s: %s : %s' % (session.meeting.number, - session.group.acronym, - sess_time.strftime("%a %H:%M")) - else: - name = 'bluesheets-%s-%s' % (session.meeting.number, sess_time.strftime("%Y%m%d%H%M")) - title = 'Bluesheets %s: %s' % (session.meeting.number, sess_time.strftime("%a %H:%M")) - doc = Document.objects.create( - name = name, - type_id = 'bluesheets', - title = title, - group = session.group, - rev = '00', - ) - doc.states.add(State.objects.get(type_id='bluesheets',slug='active')) - DocAlias.objects.create(name=doc.name).docs.add(doc) - session.sessionpresentation_set.create(document=doc,rev='00') - filename = '%s-%s%s'% ( doc.name, doc.rev, ext) - doc.uploaded_filename = filename - e = NewRevisionDocEvent.objects.create(doc=doc, rev=doc.rev, by=request.user.person, type='new_revision', desc='New revision available: %s'%doc.rev) - save_error = handle_upload_file(file, filename, session.meeting, 'bluesheets', request=request, encoding=encoding) - if not save_error: - doc.save_with_history([e]) - return save_error - - def upload_session_minutes(request, session_id, num): # num is redundant, but we're dragging it along an artifact of where we are in the current URL structure session = get_object_or_404(Session,pk=session_id) @@ -2496,7 +3274,7 @@ def upload_session_minutes(request, session_id, num): if len(sessions) > 1: session_number = 1 + sessions.index(session) - minutes_sp = session.sessionpresentation_set.filter(document__type='minutes').first() + minutes_sp = session.presentations.filter(document__type='minutes').first() if request.method == 'POST': form = UploadMinutesForm(show_apply_to_all_checkbox,request.POST,request.FILES) @@ -2520,24 +3298,130 @@ def upload_session_minutes(request, session_id, num): except SessionNotScheduledError: return HttpResponseGone( "Cannot receive uploads for an unscheduled session. Please check the session ID.", - content_type="text/plain", + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", ) except SaveMaterialsError as err: form.add_error(None, str(err)) else: # no exception -- success! + resolve_uploaded_material(meeting=session.meeting, doc=session.minutes()) messages.success(request, f'Successfully uploaded minutes as revision {session.minutes().rev}.') return redirect('ietf.meeting.views.session_details', num=num, acronym=session.group.acronym) else: form = UploadMinutesForm(show_apply_to_all_checkbox) + tsa = session.official_timeslotassignment() + future = tsa is not None and timezone.now() < tsa.timeslot.end_time() return render(request, "meeting/upload_session_minutes.html", {'session': session, 'session_number': session_number, 'minutes_sp' : minutes_sp, 'form': form, + 'future': future, }) +@role_required("Secretariat") +def upload_session_narrativeminutes(request, session_id, num): + # num is redundant, but we're dragging it along an artifact of where we are in the current URL structure + session = get_object_or_404(Session,pk=session_id) + if session.group.acronym != "iesg": + raise Http404() + + session_number = None + sessions = get_sessions(session.meeting.number,session.group.acronym) + show_apply_to_all_checkbox = len(sessions) > 1 if session.type_id == 'regular' else False + if len(sessions) > 1: + session_number = 1 + sessions.index(session) + + narrativeminutes_sp = session.presentations.filter(document__type='narrativeminutes').first() + + if request.method == 'POST': + form = UploadNarrativeMinutesForm(show_apply_to_all_checkbox,request.POST,request.FILES) + if form.is_valid(): + file = request.FILES['file'] + _, ext = os.path.splitext(file.name) + apply_to_all = session.type_id == 'regular' + if show_apply_to_all_checkbox: + apply_to_all = form.cleaned_data['apply_to_all'] + + # Set up the new revision + try: + save_session_minutes_revision( + session=session, + apply_to_all=apply_to_all, + file=file, + ext=ext, + encoding=form.file_encoding[file.name], + request=request, + narrative=True + ) + except SessionNotScheduledError: + return HttpResponseGone( + "Cannot receive uploads for an unscheduled session. Please check the session ID.", + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) + except SaveMaterialsError as err: + form.add_error(None, str(err)) + else: + # no exception -- success! + resolve_uploaded_material(meeting=session.meeting, doc=session.narrative_minutes()) + messages.success(request, f'Successfully uploaded narrative minutes as revision {session.narrative_minutes().rev}.') + return redirect('ietf.meeting.views.session_details', num=num, acronym=session.group.acronym) + else: + form = UploadMinutesForm(show_apply_to_all_checkbox) + + return render(request, "meeting/upload_session_narrativeminutes.html", + {'session': session, + 'session_number': session_number, + 'minutes_sp' : narrativeminutes_sp, + 'form': form, + }) + +class UploadOrEnterAgendaForm(UploadAgendaForm): + ACTIONS = [ + ("upload", "Upload agenda"), + ("enter", "Enter agenda"), + ] + submission_method = forms.ChoiceField(choices=ACTIONS, widget=forms.RadioSelect) + + content = forms.CharField(widget=forms.Textarea, required=False, strip=False, label="Agenda text") + + def __init__(self, *args, **kwargs): + super().__init__(*args, **kwargs) + self.fields["file"].required=False + self.order_fields(["submission_method", "file", "content"]) + + def clean_content(self): + return self.cleaned_data["content"].replace("\r", "") + + def clean_file(self): + submission_method = self.cleaned_data.get("submission_method") + if submission_method == "upload": + if self.cleaned_data.get("file", None) is not None: + return super().clean_file() + return None + + def clean(self): + def require_field(f): + if not self.cleaned_data.get(f): + self.add_error(f, ValidationError("You must fill in this field.")) + + submission_method = self.cleaned_data.get("submission_method") + if submission_method == "upload": + require_field("file") + elif submission_method == "enter": + require_field("content") + + def get_file(self): + """Get content as a file-like object""" + if self.cleaned_data.get("submission_method") == "upload": + return self.cleaned_data["file"] + else: + return SimpleUploadedFile( + name="uploaded.md", + content=self.cleaned_data["content"].encode("utf-8"), + content_type="text/markdown;charset=utf-8", + ) def upload_session_agenda(request, session_id, num): # num is redundant, but we're dragging it along an artifact of where we are in the current URL structure @@ -2554,12 +3438,12 @@ def upload_session_agenda(request, session_id, num): if len(sessions) > 1: session_number = 1 + sessions.index(session) - agenda_sp = session.sessionpresentation_set.filter(document__type='agenda').first() + agenda_sp = session.presentations.filter(document__type='agenda').first() if request.method == 'POST': - form = UploadAgendaForm(show_apply_to_all_checkbox,request.POST,request.FILES) + form = UploadOrEnterAgendaForm(show_apply_to_all_checkbox,request.POST,request.FILES) if form.is_valid(): - file = request.FILES['file'] + file = form.get_file() _, ext = os.path.splitext(file.name) apply_to_all = session.type.slug == 'regular' if show_apply_to_all_checkbox: @@ -2573,7 +3457,10 @@ def upload_session_agenda(request, session_id, num): ota = session.official_timeslotassignment() sess_time = ota and ota.timeslot.time if not sess_time: - return HttpResponseGone("Cannot receive uploads for an unscheduled session. Please check the session ID.", content_type="text/plain") + return HttpResponseGone( + "Cannot receive uploads for an unscheduled session. Please check the session ID.", + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) if session.meeting.type_id=='ietf': name = 'agenda-%s-%s' % (session.meeting.number, session.group.acronym) @@ -2599,32 +3486,40 @@ def upload_session_agenda(request, session_id, num): group = session.group, rev = '00', ) - DocAlias.objects.create(name=doc.name).docs.add(doc) doc.states.add(State.objects.get(type_id='agenda',slug='active')) - if session.sessionpresentation_set.filter(document=doc).exists(): - sp = session.sessionpresentation_set.get(document=doc) + if session.presentations.filter(document=doc).exists(): + sp = session.presentations.get(document=doc) sp.rev = doc.rev sp.save() else: - session.sessionpresentation_set.create(document=doc,rev=doc.rev) + session.presentations.create(document=doc,rev=doc.rev) if apply_to_all: for other_session in sessions: if other_session != session: - other_session.sessionpresentation_set.filter(document__type='agenda').delete() - other_session.sessionpresentation_set.create(document=doc,rev=doc.rev) + other_session.presentations.filter(document__type='agenda').delete() + other_session.presentations.create(document=doc,rev=doc.rev) filename = '%s-%s%s'% ( doc.name, doc.rev, ext) doc.uploaded_filename = filename e = NewRevisionDocEvent.objects.create(doc=doc,by=request.user.person,type='new_revision',desc='New revision available: %s'%doc.rev,rev=doc.rev) # The way this function builds the filename it will never trigger the file delete in handle_file_upload. - save_error = handle_upload_file(file, filename, session.meeting, 'agenda', request=request, encoding=form.file_encoding[file.name]) + try: + encoding=form.file_encoding[file.name] + except AttributeError: + encoding=None + save_error = handle_upload_file(file, filename, session.meeting, 'agenda', request=request, encoding=encoding) if save_error: form.add_error(None, save_error) else: doc.save_with_history([e]) + resolve_uploaded_material(meeting=session.meeting, doc=doc) messages.success(request, f'Successfully uploaded agenda as revision {doc.rev}.') return redirect('ietf.meeting.views.session_details',num=num,acronym=session.group.acronym) else: - form = UploadAgendaForm(show_apply_to_all_checkbox, initial={'apply_to_all':session.type_id=='regular'}) + initial={'apply_to_all':session.type_id=='regular', 'submission_method':'upload'} + if agenda_sp: + doc = agenda_sp.document + initial['content'] = doc.text() + form = UploadOrEnterAgendaForm(show_apply_to_all_checkbox, initial=initial) return render(request, "meeting/upload_session_agenda.html", {'session': session, @@ -2634,299 +3529,495 @@ def upload_session_agenda(request, session_id, num): }) -def upload_session_slides(request, session_id, num, name): +@login_required +def upload_session_slides(request, session_id, num, name=None): + """Upload new or replacement slides for a session + + If name is None or "", expects a new set of slides. Otherwise, replaces the named slides with a new rev. + """ # num is redundant, but we're dragging it along an artifact of where we are in the current URL structure - session = get_object_or_404(Session,pk=session_id) - if not session.can_manage_materials(request.user): - permission_denied(request, "You don't have permission to upload slides for this session.") - if session.is_material_submission_cutoff() and not has_role(request.user, "Secretariat"): - permission_denied(request, "The materials cutoff for this session has passed. Contact the secretariat for further action.") + session = get_object_or_404(Session, pk=session_id) + can_manage = session.can_manage_materials(request.user) + if session.is_material_submission_cutoff() and not has_role( + request.user, "Secretariat" + ): + permission_denied( + request, + "The materials cutoff for this session has passed. Contact the secretariat for further action.", + ) session_number = None - sessions = get_sessions(session.meeting.number,session.group.acronym) - show_apply_to_all_checkbox = len(sessions) > 1 if session.type_id == 'regular' else False + sessions = get_sessions(session.meeting.number, session.group.acronym) + show_apply_to_all_checkbox = ( + len(sessions) > 1 if session.type_id == "regular" else False + ) if len(sessions) > 1: - session_number = 1 + sessions.index(session) + session_number = 1 + sessions.index(session) - slides = None - slides_sp = None + doc = None if name: - slides = Document.objects.filter(name=name).first() - if not (slides and slides.type_id=='slides'): - raise Http404 - slides_sp = session.sessionpresentation_set.filter(document=slides).first() - - if request.method == 'POST': - form = UploadSlidesForm(session, show_apply_to_all_checkbox,request.POST,request.FILES) + doc = get_object_or_404( + session.presentations, document__name=name, document__type_id="slides" + ).document + + if request.method == "POST": + form = UploadSlidesForm( + session, show_apply_to_all_checkbox, can_manage, request.POST, request.FILES + ) if form.is_valid(): - file = request.FILES['file'] + file = request.FILES["file"] _, ext = os.path.splitext(file.name) - apply_to_all = session.type_id == 'regular' + apply_to_all = session.type_id == "regular" if show_apply_to_all_checkbox: - apply_to_all = form.cleaned_data['apply_to_all'] - if slides_sp: - doc = slides_sp.document - doc.rev = '%02d' % (int(doc.rev)+1) - doc.title = form.cleaned_data['title'] - slides_sp.rev = doc.rev - slides_sp.save() + apply_to_all = form.cleaned_data["apply_to_all"] + if can_manage: + approved = form.cleaned_data["approved"] else: + approved = False + + # Propose slides if not auto-approved + if not approved: title = form.cleaned_data['title'] + submission = SlideSubmission.objects.create(session = session, title = title, filename = '', apply_to_all = apply_to_all, submitter=request.user.person) + if session.meeting.type_id=='ietf': name = 'slides-%s-%s' % (session.meeting.number, - session.group.acronym) + session.group.acronym) if not apply_to_all: name += '-%s' % (session.docname_token(),) else: name = 'slides-%s-%s' % (session.meeting.number, session.docname_token()) name = name + '-' + slugify(title).replace('_', '-')[:128] + filename = '%s-ss%d%s'% (name, submission.id, ext) + destination = io.open(os.path.join(settings.SLIDE_STAGING_PATH, filename),'wb+') + for chunk in file.chunks(): + destination.write(chunk) + destination.close() + file.seek(0) + store_file("staging", filename, file) + + submission.filename = filename + submission.save() + + (to, cc) = gather_address_lists('slides_proposed', group=session.group, proposer=request.user.person).as_strings() + msg_txt = render_to_string("meeting/slides_proposed.txt", { + "to": to, + "cc": cc, + "submission": submission, + "settings": settings, + }) + msg = infer_message(msg_txt) + msg.by = request.user.person + msg.save() + send_mail_message(request, msg) + messages.success(request, 'Successfully submitted proposed slides.') + return redirect('ietf.meeting.views.session_details',num=num,acronym=session.group.acronym) + + # Handle creation / update of the Document (but do not save yet) + if doc is not None: + # This is a revision - bump the version and update the title. + doc.rev = "%02d" % (int(doc.rev) + 1) + doc.title = form.cleaned_data["title"] + else: + # This is a new slide deck - create a new doc unless one exists with that name + title = form.cleaned_data["title"] + if session.meeting.type_id == "ietf": + name = "slides-%s-%s" % ( + session.meeting.number, + session.group.acronym, + ) + if not apply_to_all: + name += "-%s" % (session.docname_token(),) + else: + name = "slides-%s-%s" % ( + session.meeting.number, + session.docname_token(), + ) + name = name + "-" + slugify(title).replace("_", "-")[:128] if Document.objects.filter(name=name).exists(): - doc = Document.objects.get(name=name) - doc.rev = '%02d' % (int(doc.rev)+1) - doc.title = form.cleaned_data['title'] + doc = Document.objects.get(name=name) + doc.rev = "%02d" % (int(doc.rev) + 1) + doc.title = form.cleaned_data["title"] else: doc = Document.objects.create( - name = name, - type_id = 'slides', - title = title, - group = session.group, - rev = '00', - ) - DocAlias.objects.create(name=doc.name).docs.add(doc) - doc.states.add(State.objects.get(type_id='slides',slug='active')) - doc.states.add(State.objects.get(type_id='reuse_policy',slug='single')) - if session.sessionpresentation_set.filter(document=doc).exists(): - sp = session.sessionpresentation_set.get(document=doc) - sp.rev = doc.rev - sp.save() - else: - max_order = session.sessionpresentation_set.filter(document__type='slides').aggregate(Max('order'))['order__max'] or 0 - session.sessionpresentation_set.create(document=doc,rev=doc.rev,order=max_order+1) - if apply_to_all: - for other_session in sessions: - if other_session != session and not other_session.sessionpresentation_set.filter(document=doc).exists(): - max_order = other_session.sessionpresentation_set.filter(document__type='slides').aggregate(Max('order'))['order__max'] or 0 - other_session.sessionpresentation_set.create(document=doc,rev=doc.rev,order=max_order+1) - filename = '%s-%s%s'% ( doc.name, doc.rev, ext) + name=name, + type_id="slides", + title=title, + group=session.group, + rev="00", + ) + doc.states.add(State.objects.get(type_id="slides", slug="active")) + doc.states.add(State.objects.get(type_id="reuse_policy", slug="single")) + + # Now handle creation / update of the SessionPresentation(s) + sessions_to_apply = sessions if apply_to_all else [session] + added_presentations = [] + revised_presentations = [] + for sess in sessions_to_apply: + sp = sess.presentations.filter(document=doc).first() + if sp is not None: + sp.rev = doc.rev + sp.save() + revised_presentations.append(sp) + else: + max_order = ( + sess.presentations.filter(document__type="slides").aggregate( + Max("order") + )["order__max"] + or 0 + ) + sp = sess.presentations.create( + document=doc, rev=doc.rev, order=max_order + 1 + ) + added_presentations.append(sp) + + # Now handle the uploaded file + filename = "%s-%s%s" % (doc.name, doc.rev, ext) doc.uploaded_filename = filename - e = NewRevisionDocEvent.objects.create(doc=doc,by=request.user.person,type='new_revision',desc='New revision available: %s'%doc.rev,rev=doc.rev) + e = NewRevisionDocEvent.objects.create( + doc=doc, + by=request.user.person, + type="new_revision", + desc="New revision available: %s" % doc.rev, + rev=doc.rev, + ) # The way this function builds the filename it will never trigger the file delete in handle_file_upload. - save_error = handle_upload_file(file, filename, session.meeting, 'slides', request=request, encoding=form.file_encoding[file.name]) + save_error = handle_upload_file( + file, + filename, + session.meeting, + "slides", + request=request, + encoding=form.file_encoding[file.name], + ) if save_error: form.add_error(None, save_error) else: doc.save_with_history([e]) post_process(doc) + resolve_uploaded_material(meeting=session.meeting, doc=doc) + + # Send MeetEcho updates even if we had a problem saving - that will keep it in sync with the + # SessionPresentation, which was already saved regardless of problems saving the file. + if hasattr(settings, "MEETECHO_API_CONFIG"): + sm = SlidesManager(api_config=settings.MEETECHO_API_CONFIG) + for sp in added_presentations: + try: + sm.add(session=sp.session, slides=doc, order=sp.order) + except MeetechoAPIError as err: + log(f"Error in SlidesManager.add(): {err}") + for sp in revised_presentations: + try: + sm.revise(session=sp.session, slides=doc) + except MeetechoAPIError as err: + log(f"Error in SlidesManager.revise(): {err}") + + if not save_error: messages.success( request, - f'Successfully uploaded slides as revision {doc.rev} of {doc.name}.') - return redirect('ietf.meeting.views.session_details',num=num,acronym=session.group.acronym) - else: + f"Successfully uploaded slides as revision {doc.rev} of {doc.name}.", + ) + return redirect( + "ietf.meeting.views.session_details", + num=num, + acronym=session.group.acronym, + ) + else: initial = {} - if slides: - initial = {'title':slides.title} - form = UploadSlidesForm(session, show_apply_to_all_checkbox, initial=initial) + if doc is not None: + initial = {"title": doc.title} + form = UploadSlidesForm(session, show_apply_to_all_checkbox, can_manage, initial=initial) - return render(request, "meeting/upload_session_slides.html", - {'session': session, - 'session_number': session_number, - 'slides_sp' : slides_sp, - 'form': form, - }) -@login_required -def propose_session_slides(request, session_id, num): - session = get_object_or_404(Session,pk=session_id) - if session.is_material_submission_cutoff() and not has_role(request.user, "Secretariat"): - permission_denied(request, "The materials cutoff for this session has passed. Contact the secretariat for further action.") - - session_number = None - sessions = get_sessions(session.meeting.number,session.group.acronym) - show_apply_to_all_checkbox = len(sessions) > 1 if session.type_id == 'regular' else False - if len(sessions) > 1: - session_number = 1 + sessions.index(session) - - - if request.method == 'POST': - form = UploadSlidesForm(session, show_apply_to_all_checkbox,request.POST,request.FILES) - if form.is_valid(): - file = request.FILES['file'] - _, ext = os.path.splitext(file.name) - apply_to_all = session.type_id == 'regular' - if show_apply_to_all_checkbox: - apply_to_all = form.cleaned_data['apply_to_all'] - title = form.cleaned_data['title'] - - submission = SlideSubmission.objects.create(session = session, title = title, filename = '', apply_to_all = apply_to_all, submitter=request.user.person) - - if session.meeting.type_id=='ietf': - name = 'slides-%s-%s' % (session.meeting.number, - session.group.acronym) - if not apply_to_all: - name += '-%s' % (session.docname_token(),) - else: - name = 'slides-%s-%s' % (session.meeting.number, session.docname_token()) - name = name + '-' + slugify(title).replace('_', '-')[:128] - filename = '%s-ss%d%s'% (name, submission.id, ext) - destination = io.open(os.path.join(settings.SLIDE_STAGING_PATH, filename),'wb+') - for chunk in file.chunks(): - destination.write(chunk) - destination.close() - - submission.filename = filename - submission.save() - - (to, cc) = gather_address_lists('slides_proposed', group=session.group).as_strings() - msg_txt = render_to_string("meeting/slides_proposed.txt", { - "to": to, - "cc": cc, - "submission": submission, - "settings": settings, - }) - msg = infer_message(msg_txt) - msg.by = request.user.person - msg.save() - send_mail_message(request, msg) - messages.success(request, 'Successfully submitted proposed slides.') - return redirect('ietf.meeting.views.session_details',num=num,acronym=session.group.acronym) - else: - initial = {} - form = UploadSlidesForm(session, show_apply_to_all_checkbox, initial=initial) + return render( + request, + "meeting/upload_session_slides.html", + { + "session": session, + "session_number": session_number, + "slides_sp": session.presentations.filter(document=doc).first() if doc else None, + "manage": session.can_manage_materials(request.user), + "form": form, + }, + ) - return render(request, "meeting/propose_session_slides.html", - {'session': session, - 'session_number': session_number, - 'form': form, - }) def remove_sessionpresentation(request, session_id, num, name): - sp = get_object_or_404(SessionPresentation,session_id=session_id,document__name=name) + sp = get_object_or_404( + SessionPresentation, session_id=session_id, document__name=name + ) session = sp.session if not session.can_manage_materials(request.user): - permission_denied(request, "You don't have permission to manage materials for this session.") - if session.is_material_submission_cutoff() and not has_role(request.user, "Secretariat"): - permission_denied(request, "The materials cutoff for this session has passed. Contact the secretariat for further action.") - if request.method == 'POST': - session.sessionpresentation_set.filter(pk=sp.pk).delete() - c = DocEvent(type="added_comment", doc=sp.document, rev=sp.document.rev, by=request.user.person) + permission_denied( + request, "You don't have permission to manage materials for this session." + ) + if session.is_material_submission_cutoff() and not has_role( + request.user, "Secretariat" + ): + permission_denied( + request, + "The materials cutoff for this session has passed. Contact the secretariat for further action.", + ) + if request.method == "POST": + session.presentations.filter(pk=sp.pk).delete() + c = DocEvent( + type="added_comment", + doc=sp.document, + rev=sp.document.rev, + by=request.user.person, + ) c.desc = "Removed from session: %s" % (session) c.save() - messages.success(request, f'Successfully removed {name}.') - return redirect('ietf.meeting.views.session_details', num=session.meeting.number, acronym=session.group.acronym) + messages.success(request, f"Successfully removed {name}.") + if sp.document.type_id == "slides" and hasattr(settings, "MEETECHO_API_CONFIG"): + sm = SlidesManager(api_config=settings.MEETECHO_API_CONFIG) + try: + sm.delete(session=session, slides=sp.document) + except MeetechoAPIError as err: + log(f"Error in SlidesManager.delete(): {err}") + + return redirect( + "ietf.meeting.views.session_details", + num=session.meeting.number, + acronym=session.group.acronym, + ) + + return render(request, "meeting/remove_sessionpresentation.html", {"sp": sp}) - return render(request,'meeting/remove_sessionpresentation.html', {'sp': sp }) def ajax_add_slides_to_session(request, session_id, num): - session = get_object_or_404(Session,pk=session_id) + session = get_object_or_404(Session, pk=session_id) if not session.can_manage_materials(request.user): - permission_denied(request, "You don't have permission to upload slides for this session.") - if session.is_material_submission_cutoff() and not has_role(request.user, "Secretariat"): - permission_denied(request, "The materials cutoff for this session has passed. Contact the secretariat for further action.") + permission_denied( + request, "You don't have permission to upload slides for this session." + ) + if session.is_material_submission_cutoff() and not has_role( + request.user, "Secretariat" + ): + permission_denied( + request, + "The materials cutoff for this session has passed. Contact the secretariat for further action.", + ) - if request.method != 'POST' or not request.POST: - return HttpResponse(json.dumps({ 'success' : False, 'error' : 'No data submitted or not POST' }),content_type='application/json') + if request.method != "POST" or not request.POST: + return HttpResponse( + json.dumps({"success": False, "error": "No data submitted or not POST"}), + content_type="application/json", + ) - order_str = request.POST.get('order', None) + order_str = request.POST.get("order", None) try: order = int(order_str) except (ValueError, TypeError): - return HttpResponse(json.dumps({ 'success' : False, 'error' : 'Supplied order is not valid' }),content_type='application/json') - if order < 1 or order > session.sessionpresentation_set.filter(document__type_id='slides').count() + 1 : - return HttpResponse(json.dumps({ 'success' : False, 'error' : 'Supplied order is not valid' }),content_type='application/json') + return HttpResponse( + json.dumps({"success": False, "error": "Supplied order is not valid"}), + content_type="application/json", + ) + if ( + order < 1 + or order > session.presentations.filter(document__type_id="slides").count() + 1 + ): + return HttpResponse( + json.dumps({"success": False, "error": "Supplied order is not valid"}), + content_type="application/json", + ) - name = request.POST.get('name', None) + name = request.POST.get("name", None) doc = Document.objects.filter(name=name).first() if not doc: - return HttpResponse(json.dumps({ 'success' : False, 'error' : 'Supplied name is not valid' }),content_type='application/json') + return HttpResponse( + json.dumps({"success": False, "error": "Supplied name is not valid"}), + content_type="application/json", + ) - if not session.sessionpresentation_set.filter(document=doc).exists(): + if not session.presentations.filter(document=doc).exists(): condition_slide_order(session) - session.sessionpresentation_set.filter(document__type_id='slides', order__gte=order).update(order=F('order')+1) - session.sessionpresentation_set.create(document=doc,rev=doc.rev,order=order) - DocEvent.objects.create(type="added_comment", doc=doc, rev=doc.rev, by=request.user.person, desc="Added to session: %s" % session) + session.presentations.filter( + document__type_id="slides", order__gte=order + ).update(order=F("order") + 1) + session.presentations.create(document=doc, rev=doc.rev, order=order) + DocEvent.objects.create( + type="added_comment", + doc=doc, + rev=doc.rev, + by=request.user.person, + desc="Added to session: %s" % session, + ) + + # Notify Meetecho of new slides if the API is configured + if hasattr(settings, "MEETECHO_API_CONFIG"): + sm = SlidesManager(api_config=settings.MEETECHO_API_CONFIG) + try: + sm.add(session=session, slides=doc, order=order) + except MeetechoAPIError as err: + log(f"Error in SlidesManager.add(): {err}") - return HttpResponse(json.dumps({'success':True}), content_type='application/json') + return HttpResponse(json.dumps({"success": True}), content_type="application/json") def ajax_remove_slides_from_session(request, session_id, num): - session = get_object_or_404(Session,pk=session_id) + session = get_object_or_404(Session, pk=session_id) if not session.can_manage_materials(request.user): - permission_denied(request, "You don't have permission to upload slides for this session.") - if session.is_material_submission_cutoff() and not has_role(request.user, "Secretariat"): - permission_denied(request, "The materials cutoff for this session has passed. Contact the secretariat for further action.") + permission_denied( + request, "You don't have permission to upload slides for this session." + ) + if session.is_material_submission_cutoff() and not has_role( + request.user, "Secretariat" + ): + permission_denied( + request, + "The materials cutoff for this session has passed. Contact the secretariat for further action.", + ) - if request.method != 'POST' or not request.POST: - return HttpResponse(json.dumps({ 'success' : False, 'error' : 'No data submitted or not POST' }),content_type='application/json') + if request.method != "POST" or not request.POST: + return HttpResponse( + json.dumps({"success": False, "error": "No data submitted or not POST"}), + content_type="application/json", + ) - oldIndex_str = request.POST.get('oldIndex', None) + oldIndex_str = request.POST.get("oldIndex", None) try: oldIndex = int(oldIndex_str) except (ValueError, TypeError): - return HttpResponse(json.dumps({ 'success' : False, 'error' : 'Supplied index is not valid' }),content_type='application/json') - if oldIndex < 1 or oldIndex > session.sessionpresentation_set.filter(document__type_id='slides').count() : - return HttpResponse(json.dumps({ 'success' : False, 'error' : 'Supplied index is not valid' }),content_type='application/json') + return HttpResponse( + json.dumps({"success": False, "error": "Supplied index is not valid"}), + content_type="application/json", + ) + if ( + oldIndex < 1 + or oldIndex > session.presentations.filter(document__type_id="slides").count() + ): + return HttpResponse( + json.dumps({"success": False, "error": "Supplied index is not valid"}), + content_type="application/json", + ) - name = request.POST.get('name', None) + name = request.POST.get("name", None) doc = Document.objects.filter(name=name).first() if not doc: - return HttpResponse(json.dumps({ 'success' : False, 'error' : 'Supplied name is not valid' }),content_type='application/json') + return HttpResponse( + json.dumps({"success": False, "error": "Supplied name is not valid"}), + content_type="application/json", + ) condition_slide_order(session) - affected_presentations = session.sessionpresentation_set.filter(document=doc).first() + affected_presentations = session.presentations.filter(document=doc).first() if affected_presentations: if affected_presentations.order == oldIndex: affected_presentations.delete() - session.sessionpresentation_set.filter(document__type_id='slides', order__gt=oldIndex).update(order=F('order')-1) - DocEvent.objects.create(type="added_comment", doc=doc, rev=doc.rev, by=request.user.person, desc="Removed from session: %s" % session) - return HttpResponse(json.dumps({'success':True}), content_type='application/json') + session.presentations.filter( + document__type_id="slides", order__gt=oldIndex + ).update(order=F("order") - 1) + DocEvent.objects.create( + type="added_comment", + doc=doc, + rev=doc.rev, + by=request.user.person, + desc="Removed from session: %s" % session, + ) + # Notify Meetecho of removed slides if the API is configured + if hasattr(settings, "MEETECHO_API_CONFIG"): + sm = SlidesManager(api_config=settings.MEETECHO_API_CONFIG) + try: + sm.delete(session=session, slides=doc) + except MeetechoAPIError as err: + log(f"Error in SlidesManager.delete(): {err}") + # Report success + return HttpResponse( + json.dumps({"success": True}), content_type="application/json" + ) else: - return HttpResponse(json.dumps({ 'success' : False, 'error' : 'Name does not match index' }),content_type='application/json') + return HttpResponse( + json.dumps({"success": False, "error": "Name does not match index"}), + content_type="application/json", + ) else: - return HttpResponse(json.dumps({ 'success' : False, 'error' : 'SessionPresentation not found' }),content_type='application/json') + return HttpResponse( + json.dumps({"success": False, "error": "SessionPresentation not found"}), + content_type="application/json", + ) def ajax_reorder_slides_in_session(request, session_id, num): - session = get_object_or_404(Session,pk=session_id) + session = get_object_or_404(Session, pk=session_id) if not session.can_manage_materials(request.user): - permission_denied(request, "You don't have permission to upload slides for this session.") - if session.is_material_submission_cutoff() and not has_role(request.user, "Secretariat"): - permission_denied(request, "The materials cutoff for this session has passed. Contact the secretariat for further action.") + permission_denied( + request, "You don't have permission to upload slides for this session." + ) + if session.is_material_submission_cutoff() and not has_role( + request.user, "Secretariat" + ): + permission_denied( + request, + "The materials cutoff for this session has passed. Contact the secretariat for further action.", + ) - if request.method != 'POST' or not request.POST: - return HttpResponse(json.dumps({ 'success' : False, 'error' : 'No data submitted or not POST' }),content_type='application/json') + if request.method != "POST" or not request.POST: + return HttpResponse( + json.dumps({"success": False, "error": "No data submitted or not POST"}), + content_type="application/json", + ) - num_slides_in_session = session.sessionpresentation_set.filter(document__type_id='slides').count() - oldIndex_str = request.POST.get('oldIndex', None) + session_slides = session.presentations.filter(document__type_id="slides") + num_slides_in_session = session_slides.count() + oldIndex_str = request.POST.get("oldIndex", None) try: oldIndex = int(oldIndex_str) except (ValueError, TypeError): - return HttpResponse(json.dumps({ 'success' : False, 'error' : 'Supplied index is not valid' }),content_type='application/json') - if oldIndex < 1 or oldIndex > num_slides_in_session : - return HttpResponse(json.dumps({ 'success' : False, 'error' : 'Supplied index is not valid' }),content_type='application/json') + return HttpResponse( + json.dumps({"success": False, "error": "Supplied index is not valid"}), + content_type="application/json", + ) + if oldIndex < 1 or oldIndex > num_slides_in_session: + return HttpResponse( + json.dumps({"success": False, "error": "Supplied index is not valid"}), + content_type="application/json", + ) - newIndex_str = request.POST.get('newIndex', None) + newIndex_str = request.POST.get("newIndex", None) try: newIndex = int(newIndex_str) except (ValueError, TypeError): - return HttpResponse(json.dumps({ 'success' : False, 'error' : 'Supplied index is not valid' }),content_type='application/json') - if newIndex < 1 or newIndex > num_slides_in_session : - return HttpResponse(json.dumps({ 'success' : False, 'error' : 'Supplied index is not valid' }),content_type='application/json') + return HttpResponse( + json.dumps({"success": False, "error": "Supplied index is not valid"}), + content_type="application/json", + ) + if newIndex < 1 or newIndex > num_slides_in_session: + return HttpResponse( + json.dumps({"success": False, "error": "Supplied index is not valid"}), + content_type="application/json", + ) if newIndex == oldIndex: - return HttpResponse(json.dumps({ 'success' : False, 'error' : 'Supplied index is not valid' }),content_type='application/json') + return HttpResponse( + json.dumps({"success": False, "error": "Supplied index is not valid"}), + content_type="application/json", + ) condition_slide_order(session) - sp = session.sessionpresentation_set.get(order=oldIndex) + sp = session_slides.get(order=oldIndex) if oldIndex < newIndex: - session.sessionpresentation_set.filter(order__gt=oldIndex, order__lte=newIndex).update(order=F('order')-1) + session_slides.filter(order__gt=oldIndex, order__lte=newIndex).update( + order=F("order") - 1 + ) else: - session.sessionpresentation_set.filter(order__gte=newIndex, order__lt=oldIndex).update(order=F('order')+1) + session_slides.filter(order__gte=newIndex, order__lt=oldIndex).update( + order=F("order") + 1 + ) sp.order = newIndex sp.save() - return HttpResponse(json.dumps({'success':True}), content_type='application/json') + # Update slide order with Meetecho if the API is configured + if hasattr(settings, "MEETECHO_API_CONFIG"): + sm = SlidesManager(api_config=settings.MEETECHO_API_CONFIG) + try: + sm.send_update(session) + except MeetechoAPIError as err: + log(f"Error in SlidesManager.send_update(): {err}") + + return HttpResponse(json.dumps({"success": True}), content_type="application/json") @role_required('Secretariat') @@ -3004,43 +4095,6 @@ def delete_schedule(request, num, owner, name): # ------------------------------------------------- # Interim Views # ------------------------------------------------- - - -def ajax_get_utc(request): - '''Ajax view that takes arguments time, timezone, date and returns UTC data''' - time = request.GET.get('time') - timezone = request.GET.get('timezone') - date = request.GET.get('date') - time_re = re.compile(r'^\d{2}:\d{2}$') - # validate input - if not time_re.match(time) or not date: - return HttpResponse(json.dumps({'error': True}), - content_type='application/json') - hour, minute = time.split(':') - if not (int(hour) <= 23 and int(minute) <= 59): - return HttpResponse(json.dumps({'error': True}), - content_type='application/json') - year, month, day = date.split('-') - dt = datetime.datetime(int(year), int(month), int(day), int(hour), int(minute)) - tz = pytz.timezone(timezone) - aware_dt = tz.localize(dt, is_dst=None) - utc_dt = aware_dt.astimezone(pytz.utc) - utc = utc_dt.strftime('%H:%M') - # calculate utc day offset - naive_utc_dt = utc_dt.replace(tzinfo=None) - utc_day_offset = (naive_utc_dt.date() - dt.date()).days - html = "{utc} UTC".format(utc=utc) - if utc_day_offset != 0: - html = html + ' {0:+d} Day'.format(utc_day_offset) - context_data = {'timezone': timezone, - 'time': time, - 'utc': utc, - 'utc_day_offset': utc_day_offset, - 'html': html} - return HttpResponse(json.dumps(context_data), - content_type='application/json') - - def interim_announce(request): '''View which shows interim meeting requests awaiting announcement''' meetings = data_for_meetings_overview(Meeting.objects.filter(type='interim').order_by('date'), interim_status='scheda') @@ -3151,8 +4205,8 @@ def interim_request(request): if meeting_type in ('single', 'multi-day'): meeting = form.save(date=get_earliest_session_date(formset)) - # need to use curry here to pass custom variable to form init - SessionFormset.form.__init__ = curry( + # need to use partialmethod here to pass custom variable to form init + SessionFormset.form.__init__ = partialmethod( InterimSessionModelForm.__init__, user=request.user, group=group, @@ -3174,7 +4228,7 @@ def interim_request(request): # subsequently dealt with individually elif meeting_type == 'series': series = [] - SessionFormset.form.__init__ = curry( + SessionFormset.form.__init__ = partialmethod( InterimSessionModelForm.__init__, user=request.user, group=group, @@ -3394,7 +4448,7 @@ def interim_request_edit(request, number): group = Group.objects.get(pk=form.data['group']) is_approved = is_interim_meeting_approved(meeting) - SessionFormset.form.__init__ = curry( + SessionFormset.form.__init__ = partialmethod( InterimSessionModelForm.__init__, user=request.user, group=group, @@ -3424,7 +4478,6 @@ def interim_request_edit(request, number): "form": form, "formset": formset}) -@cache_page(60*60) def past(request): '''List of past meetings''' today = timezone.now() @@ -3554,21 +4607,137 @@ def upcoming_ical(request): else: ietfs = [] - meeting_vtz = {meeting.vtimezone() for meeting in meetings} - meeting_vtz.discard(None) - - # icalendar response file should have '\r\n' line endings per RFC5545 - response = render_to_string('meeting/upcoming.ics', { - 'vtimezones': ''.join(sorted(meeting_vtz)), - 'assignments': assignments, - 'ietfs': ietfs, - }, request=request) - response = re.sub("\r(?!\n)|(? begin_date, - 'cache_version': cache_version, 'attendance': meeting.get_attendance(), - 'meetinghost_logo': { - 'max_height': settings.MEETINGHOST_LOGO_MAX_DISPLAY_HEIGHT, - 'max_width': settings.MEETINGHOST_LOGO_MAX_DISPLAY_WIDTH, - } + 'proceedings_content': generate_proceedings_content(meeting), }) @role_required('Secretariat') def finalize_proceedings(request, num=None): meeting = get_meeting(num) - if (meeting.number.isdigit() and int(meeting.number) <= 64) or not meeting.schedule or not meeting.schedule.assignments.exists() or meeting.proceedings_final: raise Http404 if request.method=='POST': - finalize(meeting) + finalize(request, meeting) return HttpResponseRedirect(reverse('ietf.meeting.views.proceedings',kwargs={'num':meeting.number})) return render(request, "meeting/finalize.html", {'meeting':meeting,}) @@ -3680,14 +4808,53 @@ def proceedings_attendees(request, num=None): meeting = get_meeting(num) if meeting.proceedings_format_version == 1: return HttpResponseRedirect(f'{settings.PROCEEDINGS_V1_BASE_URL.format(meeting=meeting)}/attendee.html') - overview_template = '/meeting/proceedings/%s/attendees.html' % meeting.number - try: - template = render_to_string(overview_template, {}) - except TemplateDoesNotExist: - raise Http404 + + template = None + registrations = None + + stats = None + chart_data = None + + if int(meeting.number) >= 118: + checked_in, attended = participants_for_meeting(meeting) + regs = list(Registration.objects.onsite().filter(meeting__number=num, checkedin=True)) + onsite_count = len(regs) + regs += [ + reg + for reg in Registration.objects.remote().filter(meeting__number=num).select_related('person') + if reg.person.pk in attended and reg.person.pk not in checked_in + ] + remote_count = len(regs) - onsite_count + + registrations = sorted(regs, key=lambda x: (x.last_name, x.first_name)) + + country_codes = [r.country_code for r in registrations if r.country_code] + stats = { + 'total': onsite_count + remote_count, + 'onsite': onsite_count, + 'remote': remote_count, + } + + code_to_name = dict(CountryName.objects.values_list('slug', 'name')) + country_counts = Counter(code_to_name.get(c, c) for c in country_codes).most_common() + + chart_data = { + 'type': [['Onsite', onsite_count], ['Remote', remote_count]], + 'countries': country_counts, + } + else: + overview_template = "/meeting/proceedings/%s/attendees.html" % meeting.number + try: + template = render_to_string(overview_template, {}) + except TemplateDoesNotExist: + raise Http404 + return render(request, "meeting/proceedings_attendees.html", { 'meeting': meeting, + 'registrations': registrations, 'template': template, + 'stats': stats, + 'chart_data': chart_data, }) def proceedings_overview(request, num=None): @@ -3707,9 +4874,8 @@ def proceedings_overview(request, num=None): 'template': template, }) -@cache_page( 60 * 60 ) -def proceedings_progress_report(request, num=None): - '''Display Progress Report (stats since last meeting)''' +def proceedings_activity_report(request, num=None): + '''Display Activity Report (stats since last meeting)''' if not (num and num.isdigit()): raise Http404 meeting = get_meeting(num) @@ -3717,30 +4883,143 @@ def proceedings_progress_report(request, num=None): return HttpResponseRedirect(f'{settings.PROCEEDINGS_V1_BASE_URL.format(meeting=meeting)}/progress-report.html') sdate = meeting.previous_meeting().date edate = meeting.date - context = get_progress_stats(sdate,edate) + context = get_activity_stats(sdate,edate) context['meeting'] = meeting - return render(request, "meeting/proceedings_progress_report.html", context) + context['is_meeting_report'] = True + return render(request, "meeting/proceedings_activity_report.html", context) class OldUploadRedirect(RedirectView): def get_redirect_url(self, **kwargs): return reverse_lazy('ietf.meeting.views.session_details',kwargs=self.kwargs) + +@require_api_key +@role_required("Recording Manager") @csrf_exempt -def api_import_recordings(request, number): - '''REST API to check for recording files and import''' - if request.method == 'POST': - meeting = get_meeting(number) - import_audio_files(meeting) - return HttpResponse(status=201) - else: - return HttpResponse(status=405) +def api_set_meetecho_recording_name(request): + """Set name for meetecho recording + + parameters: + apikey: the poster's personal API key + session_id: id of the session to update + name: the name to use for the recording at meetecho player + """ + def err(code, text): + return HttpResponse( + text, + status=code, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) + + if request.method != "POST": + return HttpResponseNotAllowed( + content="Method not allowed", + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + permitted_methods=('POST',), + ) + + session_id = request.POST.get('session_id', None) + if session_id is None: + return err(400, 'Missing session_id parameter') + name = request.POST.get('name', None) + if name is None: + return err(400, 'Missing name parameter') + + try: + session = Session.objects.get(pk=session_id) + except Session.DoesNotExist: + return err(400, f"Session not found with session_id '{session_id}'") + except ValueError: + return err(400, "Invalid session_id: {session_id}") + + session.meetecho_recording_name = name + session.save() + + return HttpResponse( + "Done", + status=200, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) @require_api_key @role_required('Recording Manager') @csrf_exempt def api_set_session_video_url(request): + """Set video URL for session + + parameters: + apikey: the poster's personal API key + session_id: id of session to update + url: The recording url (on YouTube, or whatever) + """ + def err(code, text): + return HttpResponse( + text, + status=code, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) + + if request.method != 'POST': + return HttpResponseNotAllowed( + content="Method not allowed", + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + permitted_methods=('POST',), + ) + + # Temporary: fall back to deprecated interface if we have old-style parameters. + # Do away with this once meetecho is using the new pk-based interface. + if any(k in request.POST for k in ['meeting', 'group', 'item']): + return deprecated_api_set_session_video_url(request) + + session_id = request.POST.get('session_id', None) + if session_id is None: + return err(400, 'Missing session_id parameter') + incoming_url = request.POST.get('url', None) + if incoming_url is None: + return err(400, 'Missing url parameter') + + try: + session = Session.objects.get(pk=session_id) + except Session.DoesNotExist: + return err(400, f"Session not found with session_id '{session_id}'") + except ValueError: + return err(400, "Invalid session_id: {session_id}") + + try: + URLValidator()(incoming_url) + except ValidationError: + return err(400, f"Invalid url value: '{incoming_url}'") + + recordings = [(r.name, r.title, r) for r in session.recordings() if 'video' in r.title.lower()] + if recordings: + r = recordings[-1][-1] + if r.external_url != incoming_url: + e = DocEvent.objects.create(doc=r, rev=r.rev, type="added_comment", by=request.user.person, + desc="External url changed from %s to %s" % (r.external_url, incoming_url)) + r.external_url = incoming_url + r.save_with_history([e]) + else: + time = session.official_timeslotassignment().timeslot.time + title = 'Video recording for %s on %s at %s' % (session.group.acronym, time.date(), time.time()) + create_recording(session, incoming_url, title=title, user=request.user.person) + return HttpResponse( + "Done", + status=200, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) + + +def deprecated_api_set_session_video_url(request): + """Set video URL for session (deprecated) + + Uses meeting/group/item to identify session. + """ def err(code, text): - return HttpResponse(text, status=code, content_type='text/plain') + return HttpResponse( + text, + status=code, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) if request.method == 'POST': # parameters: # apikey: the poster's personal API key @@ -3794,46 +5073,136 @@ def err(code, text): else: return err(405, "Method not allowed") - return HttpResponse("Done", status=200, content_type='text/plain') + return HttpResponse( + "Done", + status=200, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) + @require_api_key @role_required('Recording Manager') # TODO : Rework how Meetecho interacts via APIs. There may be better paths to pursue than Personal API keys as they are currently defined. @csrf_exempt def api_add_session_attendees(request): + """Upload attendees for one or more sessions + + parameters: + apikey: the poster's personal API key + attended: json blob with + { + "session_id": session pk, + "attendees": [ + {"user_id": user-pk-1, "join_time": "2024-02-21T18:00:00Z"}, + {"user_id": user-pk-2, "join_time": "2024-02-21T18:00:01Z"}, + {"user_id": user-pk-3, "join_time": "2024-02-21T18:00:02Z"}, + ... + ] + } + """ + json_validator = jsonschema.Draft202012Validator( + schema={ + "type": "object", + "properties": { + "session_id": {"type": "integer"}, + "attendees": { + # Allow either old or new format until after IETF 119 + "anyOf": [ + {"type": "array", "items": {"type": "integer"}}, # old: array of user PKs + { + # new: array of user_id / join_time objects + "type": "array", + "items": { + "type": "object", + "properties": { + "user_id": {"type": "integer", }, + "join_time": {"type": "string", "format": "date-time"} + }, + "required": ["user_id", "join_time"], + }, + }, + ], + } + }, + "required": ["session_id", "attendees"], + }, + format_checker=jsonschema.Draft202012Validator.FORMAT_CHECKER, # format-checks disabled by default + ) def err(code, text): - return HttpResponse(text, status=code, content_type='text/plain') + return HttpResponse( + text, + status=code, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) - if request.method != 'POST': + if request.method != "POST": return err(405, "Method not allowed") - attended_post = request.POST.get('attended') + attended_post = request.POST.get("attended") if not attended_post: return err(400, "Missing attended parameter") + + # Validate the request payload try: - attended = json.loads(attended_post) - except json.decoder.JSONDecodeError: - return err(400, "Malformed post") - if not ( 'session_id' in attended and type(attended['session_id']) is int ): - return err(400, "Malformed post") - session_id = attended['session_id'] - if not ( 'attendees' in attended and type(attended['attendees']) is list and all([type(el) is int for el in attended['attendees']]) ): + payload = json.loads(attended_post) + json_validator.validate(payload) + except (json.decoder.JSONDecodeError, jsonschema.exceptions.ValidationError): return err(400, "Malformed post") + + session_id = payload["session_id"] session = Session.objects.filter(pk=session_id).first() if not session: return err(400, "Invalid session") - users = User.objects.filter(pk__in=attended['attendees']) - if users.count() != len(attended['attendees']): - return err(400, "Invalid attendee") - for user in users: - session.attended_set.get_or_create(person=user.person) - return HttpResponse("Done", status=200, content_type='text/plain') + + attendees = payload["attendees"] + if len(attendees) > 0: + # Check whether we have old or new format + if type(attendees[0]) == int: + # it's the old format + users = User.objects.filter(pk__in=attendees) + if users.count() != len(payload["attendees"]): + return err(400, "Invalid attendee") + for user in users: + session.attended_set.get_or_create(person=user.person) + else: + # it's the new format + join_time_by_pk = { + att["user_id"]: datetime.datetime.fromisoformat( + att["join_time"].replace("Z", "+00:00") # Z not understood until py311 + ) + for att in attendees + } + persons = list(Person.objects.filter(user__pk__in=join_time_by_pk)) + if len(persons) != len(join_time_by_pk): + return err(400, "Invalid attendee") + to_create = [ + Attended(session=session, person=person, time=join_time_by_pk[person.user_id]) + for person in persons + ] + # Create in bulk, ignoring any that already exist + Attended.objects.bulk_create(to_create, ignore_conflicts=True) + + if session.meeting.type_id == "interim": + save_error = generate_bluesheet(request, session) + if save_error: + return err(400, save_error) + + return HttpResponse( + "Done", + status=200, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) + @require_api_key @role_required('Recording Manager') @csrf_exempt def api_upload_chatlog(request): def err(code, text): - return HttpResponse(text, status=code, content_type='text/plain') + return HttpResponse( + text, + status=code, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) if request.method != 'POST': return err(405, "Method not allowed") apidata_post = request.POST.get('apidata') @@ -3851,7 +5220,7 @@ def err(code, text): session = Session.objects.filter(pk=session_id).first() if not session: return err(400, "Invalid session") - chatlog_sp = session.sessionpresentation_set.filter(document__type='chatlog').first() + chatlog_sp = session.presentations.filter(document__type='chatlog').first() if chatlog_sp: doc = chatlog_sp.document doc.rev = f"{(int(doc.rev)+1):02d}" @@ -3866,14 +5235,23 @@ def err(code, text): write_doc_for_session(session, 'chatlog', filename, json.dumps(apidata['chatlog'])) e = NewRevisionDocEvent.objects.create(doc=doc, rev=doc.rev, by=request.user.person, type='new_revision', desc='New revision available: %s'%doc.rev) doc.save_with_history([e]) - return HttpResponse("Done", status=200, content_type='text/plain') + resolve_uploaded_material(meeting=session.meeting, doc=doc) + return HttpResponse( + "Done", + status=200, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) @require_api_key @role_required('Recording Manager') @csrf_exempt def api_upload_polls(request): def err(code, text): - return HttpResponse(text, status=code, content_type='text/plain') + return HttpResponse( + text, + status=code, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) if request.method != 'POST': return err(405, "Method not allowed") apidata_post = request.POST.get('apidata') @@ -3891,7 +5269,7 @@ def err(code, text): session = Session.objects.filter(pk=session_id).first() if not session: return err(400, "Invalid session") - polls_sp = session.sessionpresentation_set.filter(document__type='polls').first() + polls_sp = session.presentations.filter(document__type='polls').first() if polls_sp: doc = polls_sp.document doc.rev = f"{(int(doc.rev)+1):02d}" @@ -3906,68 +5284,76 @@ def err(code, text): write_doc_for_session(session, 'polls', filename, json.dumps(apidata['polls'])) e = NewRevisionDocEvent.objects.create(doc=doc, rev=doc.rev, by=request.user.person, type='new_revision', desc='New revision available: %s'%doc.rev) doc.save_with_history([e]) - return HttpResponse("Done", status=200, content_type='text/plain') + resolve_uploaded_material(meeting=session.meeting, doc=doc) + return HttpResponse( + "Done", + status=200, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) @require_api_key @role_required('Recording Manager', 'Secretariat') @csrf_exempt def api_upload_bluesheet(request): + """Upload bluesheet for a session + + parameters: + apikey: the poster's personal API key + session_id: id of session to update + bluesheet: json blob with + [{'name': 'Name', 'affiliation': 'Organization', }, ...] + """ def err(code, text): - return HttpResponse(text, status=code, content_type='text/plain') - if request.method == 'POST': - # parameters: - # apikey: the poster's personal API key - # meeting: number as string, i.e., '101', or 'interim-2018-quic-02' - # group: acronym or special, i.e., 'quic' or 'plenary' - # item: '1', '2', '3' (the group's first, second, third etc. - # session during the week) - # bluesheet: json blob with - # [{'name': 'Name', 'affiliation': 'Organization', }, ...] - for item in ['meeting', 'group', 'item', 'bluesheet',]: - value = request.POST.get(item) - if not value: - return err(400, "Missing %s parameter" % item) - number = request.POST.get('meeting') - sessions = Session.objects.filter(meeting__number=number) - if not sessions.exists(): - return err(400, "No sessions found for meeting '%s'" % (number, )) - acronym = request.POST.get('group') - sessions = sessions.filter(group__acronym=acronym) - if not sessions.exists(): - return err(400, "No sessions found in meeting '%s' for group '%s'" % (number, acronym)) - session_times = [ (s.official_timeslotassignment().timeslot.time, s.id, s) for s in sessions if s.official_timeslotassignment() ] - session_times.sort() - item = request.POST.get('item') - if not item.isdigit(): - return err(400, "Expected a numeric value for 'item', found '%s'" % (item, )) - n = int(item)-1 # change 1-based to 0-based - try: - time, __, session = session_times[n] - except IndexError: - return err(400, "No item '%s' found in list of sessions for group" % (item, )) - bjson = request.POST.get('bluesheet') - try: - data = json.loads(bjson) - except json.decoder.JSONDecodeError: - return err(400, "Invalid json value: '%s'" % (bjson, )) + return HttpResponse( + text, + status=code, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) - text = render_to_string('meeting/bluesheet.txt', { - 'data': data, - 'session': session, - }) + if request.method != 'POST': + return HttpResponseNotAllowed( + content="Method not allowed", + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + permitted_methods=('POST',), + ) - fd, name = tempfile.mkstemp(suffix=".txt", text=True) - os.close(fd) - with open(name, "w") as file: - file.write(text) - with open(name, "br") as file: - save_err = save_bluesheet(request, session, file) - if save_err: - return err(400, save_err) - else: - return err(405, "Method not allowed") + session_id = request.POST.get('session_id', None) + if session_id is None: + return err(400, 'Missing session_id parameter') + bjson = request.POST.get('bluesheet', None) + if bjson is None: + return err(400, 'Missing bluesheet parameter') + + try: + session = Session.objects.get(pk=session_id) + except Session.DoesNotExist: + return err(400, f"Session not found with session_id '{session_id}'") + except ValueError: + return err(400, f"Invalid session_id '{session_id}'") + + try: + data = json.loads(bjson) + except json.decoder.JSONDecodeError: + return err(400, f"Invalid json value: '{bjson}'") + + text = render_to_string('meeting/bluesheet.txt', { + 'data': data, + 'session': session, + }) - return HttpResponse("Done", status=200, content_type='text/plain') + fd, name = tempfile.mkstemp(suffix=".txt", text=True) + os.close(fd) + with open(name, "w") as file: + file.write(text) + with open(name, "br") as file: + save_err = save_bluesheet(request, session, file) + if save_err: + return err(400, save_err) + return HttpResponse( + "Done", + status=200, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) def important_dates(request, num=None, output_format=None): @@ -3993,11 +5379,8 @@ def important_dates(request, num=None, output_format=None): if output_format == 'ics': preprocess_meeting_important_dates(meetings) - ics = render_to_string('meeting/important_dates.ics', { - 'meetings': meetings, - }, request=request) - # icalendar response file should have '\r\n' line endings per RFC5545 - response = HttpResponse(re.sub("\r(?!\n)|(? 0: + messages.success( + request, + f"Notified Meetecho about slides for {','.join(str(s) for s in updated)}", + ) + elif sm.slides_notify_time is not None: + messages.warning( + request, + "No sessions were eligible for Meetecho slides update. Updates are " + f"only sent within {sm.slides_notify_time} before or after the session.", + ) + else: + messages.warning( + request, + "No sessions were eligible for Meetecho slides update. Updates are " + "currently disabled.", + ) + return redirect( + "ietf.meeting.views.session_details", num=meeting.number, acronym=acronym + ) + + def import_session_minutes(request, session_id, num): """Import session minutes from the ietf.notes.org site @@ -4338,11 +5821,12 @@ def import_session_minutes(request, session_id, num): except SessionNotScheduledError: return HttpResponseGone( "Cannot import minutes for an unscheduled session. Please check the session ID.", - content_type="text/plain", + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", ) except SaveMaterialsError as err: form.add_error(None, str(err)) else: + resolve_uploaded_material(meeting=session.meeting, doc=session.minutes()) messages.success(request, f'Successfully imported minutes as revision {session.minutes().rev}.') return redirect('ietf.meeting.views.session_details', num=num, acronym=session.group.acronym) else: @@ -4378,3 +5862,4 @@ def import_session_minutes(request, session_id, num): 'contents_unchanged': not contents_changed, }, ) + diff --git a/ietf/meeting/views_proceedings.py b/ietf/meeting/views_proceedings.py index 87b7ffea35..639efa1da4 100644 --- a/ietf/meeting/views_proceedings.py +++ b/ietf/meeting/views_proceedings.py @@ -8,13 +8,13 @@ import debug # pyflakes:ignore from ietf.doc.utils import add_state_change_event -from ietf.doc.models import DocAlias, DocEvent, Document, NewRevisionDocEvent, State +from ietf.doc.models import DocEvent, Document, NewRevisionDocEvent, State from ietf.ietfauth.utils import role_required from ietf.meeting.forms import FileUploadForm from ietf.meeting.models import Meeting, MeetingHost from ietf.meeting.helpers import get_meeting from ietf.name.models import ProceedingsMaterialTypeName -from ietf.meeting.utils import handle_upload_file +from ietf.meeting.utils import handle_upload_file, resolve_uploaded_material from ietf.utils.text import xslugify class UploadProceedingsMaterialForm(FileUploadForm): @@ -98,10 +98,6 @@ def save_proceedings_material_doc(meeting, material_type, title, request, file=N ) created = True - # do this even if we did not create the document, just to be sure the alias exists - alias, _ = DocAlias.objects.get_or_create(name=doc.name) - alias.docs.add(doc) - if file: if not created: doc.rev = '{:02}'.format(int(doc.rev) + 1) @@ -154,7 +150,7 @@ def save_proceedings_material_doc(meeting, material_type, title, request, file=N if events: doc.save_with_history(events) - + resolve_uploaded_material(meeting, doc) return doc diff --git a/ietf/meeting/views_session_request.py b/ietf/meeting/views_session_request.py new file mode 100644 index 0000000000..a1ef74f1b8 --- /dev/null +++ b/ietf/meeting/views_session_request.py @@ -0,0 +1,931 @@ +# Copyright The IETF Trust 2007-2025, All Rights Reserved +# -*- coding: utf-8 -*- + +import datetime +import inflect +from collections import defaultdict, OrderedDict + +from django.conf import settings +from django.contrib import messages +from django.core.exceptions import ObjectDoesNotExist +from django.db.models import Q +from django.shortcuts import render, get_object_or_404, redirect +from django.http import Http404 + +from ietf.group.models import Group, GroupFeatures +from ietf.ietfauth.utils import has_role, role_required +from ietf.meeting.helpers import get_meeting +from ietf.meeting.models import Session, Meeting, Constraint, ResourceAssociation, SchedulingEvent +from ietf.meeting.utils import add_event_info_to_session_qs +from ietf.meeting.forms import (SessionRequestStatusForm, SessionRequestForm, allowed_conflicting_groups, + JOINT_FOR_SESSION_CHOICES) +from ietf.name.models import SessionStatusName, ConstraintName +from ietf.secr.utils.decorators import check_permissions +from ietf.utils.mail import send_mail +from ietf.mailtrigger.utils import gather_address_lists + +# ------------------------------------------------- +# Globals +# ------------------------------------------------- +# TODO: This needs to be replaced with something that pays attention to groupfeatures +AUTHORIZED_ROLES = ( + 'WG Chair', + 'WG Secretary', + 'RG Chair', + 'IAB Group Chair', + 'Area Director', + 'Secretariat', + 'Team Chair', + 'IRTF Chair', + 'Program Chair', + 'Program Lead', + 'Program Secretary', + 'EDWG Chair') + +# ------------------------------------------------- +# Helper Functions +# ------------------------------------------------- + + +def check_app_locked(meeting=None): + ''' + This function returns True if the application is locked to non-secretariat users. + ''' + if not meeting: + meeting = get_meeting(days=14) + return bool(meeting.session_request_lock_message) + + +def get_lock_message(meeting=None): + ''' + Returns the message to display to non-secretariat users when the tool is locked. + ''' + if not meeting: + meeting = get_meeting(days=14) + return meeting.session_request_lock_message + + +def get_my_groups(user, conclude=False): + ''' + Takes a Django user object (from request) + Returns a list of groups the user has access to. Rules are as follows + secretariat - has access to all groups + area director - has access to all groups in their area + wg chair or secretary - has access to their own group + chair of irtf has access to all irtf groups + + If user=None than all groups are returned. + concluded=True means include concluded groups. Need this to upload materials for groups + after they've been concluded. it happens. + ''' + my_groups = set() + states = ['bof', 'proposed', 'active'] + if conclude: + states.extend(['conclude', 'bof-conc']) + + all_groups = Group.objects.filter(type__features__has_meetings=True, state__in=states).order_by('acronym') + if user is None or has_role(user, 'Secretariat'): + return all_groups + + try: + person = user.person + except ObjectDoesNotExist: + return list() + + for group in all_groups: + if group.role_set.filter(person=person, name__in=('chair', 'secr', 'ad')): + my_groups.add(group) + continue + if group.parent and group.parent.role_set.filter(person=person, name__in=('ad', 'chair')): + my_groups.add(group) + continue + + return list(my_groups) + + +def get_initial_session(sessions, prune_conflicts=False): + ''' + This function takes a queryset of sessions ordered by 'id' for consistency. It returns + a dictionary to be used as the initial for a legacy session form + ''' + initial = {} + if len(sessions) == 0: + return initial + + meeting = sessions[0].meeting + group = sessions[0].group + + constraints = group.constraint_source_set.filter(meeting=meeting) # all constraints with this group as source + conflicts = constraints.filter(name__is_group_conflict=True) # only the group conflict constraints + + if group.features.acts_like_wg: + # even if there are three sessions requested, the old form has 2 in this field + initial['num_session'] = min(sessions.count(), 2) + initial['third_session'] = sessions.count() > 2 + else: + initial['num_session'] = sessions.count() + initial['third_session'] = False + initial['attendees'] = sessions[0].attendees + + def valid_conflict(conflict): + return conflict.target != sessions[0].group and allowed_conflicting_groups().filter(pk=conflict.target_id).exists() + if prune_conflicts: + conflicts = [c for c in conflicts if valid_conflict(c)] + + conflict_name_ids = set(c.name_id for c in conflicts) + for name_id in conflict_name_ids: + target_acros = [c.target.acronym for c in conflicts if c.name_id == name_id] + initial['constraint_{}'.format(name_id)] = ' '.join(target_acros) + + initial['comments'] = sessions[0].comments + initial['resources'] = sessions[0].resources.all() + initial['bethere'] = [x.person for x in sessions[0].constraints().filter(name='bethere').select_related("person")] + wg_adjacent = constraints.filter(name__slug='wg_adjacent') + initial['adjacent_with_wg'] = wg_adjacent[0].target.acronym if wg_adjacent else None + time_relation = constraints.filter(name__slug='time_relation') + initial['session_time_relation'] = time_relation[0].time_relation if time_relation else None + initial['session_time_relation_display'] = time_relation[0].get_time_relation_display if time_relation else None + timeranges = constraints.filter(name__slug='timerange') + initial['timeranges'] = timeranges[0].timeranges.all() if timeranges else [] + initial['timeranges_display'] = [t.desc for t in initial['timeranges']] + for idx, session in enumerate(sessions): + if session.joint_with_groups.count(): + initial['joint_with_groups'] = ' '.join(session.joint_with_groups_acronyms()) + initial['joint_for_session'] = str(idx + 1) + initial['joint_for_session_display'] = dict(JOINT_FOR_SESSION_CHOICES)[initial['joint_for_session']] + return initial + + +def inbound_session_conflicts_as_string(group, meeting): + ''' + Takes a Group object and Meeting object and returns a string of other groups which have + a conflict with this one + ''' + constraints = group.constraint_target_set.filter(meeting=meeting, name__is_group_conflict=True) + group_set = set(constraints.values_list('source__acronym', flat=True)) # set to de-dupe + group_list = sorted(group_set) # give a consistent order + return ', '.join(group_list) + + +def get_outbound_conflicts(form: SessionRequestForm): + """extract wg conflict constraint data from a SessionForm""" + outbound_conflicts = [] + for conflictname, cfield_id in form.wg_constraint_field_ids(): + conflict_groups = form.cleaned_data[cfield_id] + if len(conflict_groups) > 0: + outbound_conflicts.append(dict(name=conflictname, groups=conflict_groups)) + return outbound_conflicts + + +def save_conflicts(group, meeting, conflicts, name): + ''' + This function takes a Group, Meeting a string which is a list of Groups acronyms (conflicts), + and the constraint name (conflict|conflic2|conflic3) and creates Constraint records + ''' + constraint_name = ConstraintName.objects.get(slug=name) + acronyms = conflicts.replace(',',' ').split() + for acronym in acronyms: + target = Group.objects.get(acronym=acronym) + + constraint = Constraint(source=group, + target=target, + meeting=meeting, + name=constraint_name) + constraint.save() + + +def get_requester_text(person, group): + """ + This function takes a Person object and a Group object and returns the text to use + in the session request notification email, ie. Joe Smith, a Chair of the ancp + working group + """ + roles = group.role_set.filter(name__in=("chair", "secr", "ad"), person=person) + if roles: + rolename = str(roles[0].name) + return "%s, %s of the %s %s" % ( + person.name, + inflect.engine().a(rolename), + group.acronym.upper(), + group.type.verbose_name, + ) + if person.role_set.filter(name="secr", group__acronym="secretariat"): + return "%s, on behalf of the %s %s" % ( + person.name, + group.acronym.upper(), + group.type.verbose_name, + ) + + +def send_notification(group, meeting, login, sreq_data, session_data, action): + ''' + This function generates email notifications for various session request activities. + sreq_data argument is a dictionary of fields from the session request form + session_data is an array of data from individual session subforms + action argument is a string [new|update]. + ''' + (to_email, cc_list) = gather_address_lists('session_requested', group=group, person=login) + from_email = (settings.SESSION_REQUEST_FROM_EMAIL) + subject = '%s - New Meeting Session Request for IETF %s' % (group.acronym, meeting.number) + template = 'meeting/session_request_notification.txt' + + # send email + context = {} + context['session'] = sreq_data + context['group'] = group + context['meeting'] = meeting + context['login'] = login + context['header'] = 'A new' + context['requester'] = get_requester_text(login, group) + + # update overrides + if action == 'update': + subject = '%s - Update to a Meeting Session Request for IETF %s' % (group.acronym, meeting.number) + context['header'] = 'An update to a' + + # if third session requested approval is required + # change headers TO=ADs, CC=session-request, submitter and cochairs + if len(session_data) > 2: + (to_email, cc_list) = gather_address_lists('session_requested_long', group=group, person=login) + subject = '%s - Request for meeting session approval for IETF %s' % (group.acronym, meeting.number) + template = 'meeting/session_approval_notification.txt' + # status_text = 'the %s Directors for approval' % group.parent + + context['session_lengths'] = [sd['requested_duration'] for sd in session_data] + + send_mail(None, + to_email, + from_email, + subject, + template, + context, + cc=cc_list) + + +def session_changed(session): + latest_event = SchedulingEvent.objects.filter(session=session).order_by('-time', '-id').first() + + if latest_event and latest_event.status_id == "schedw" and session.meeting.schedule is not None: + # send an email to iesg-secretariat to alert to change + pass + + +def status_slug_for_new_session(session, session_number): + if session.group.features.acts_like_wg and session_number == 2: + return 'apprw' + return 'schedw' + +# ------------------------------------------------- +# View Functions +# ------------------------------------------------- + + +@role_required(*AUTHORIZED_ROLES) +def list_view(request): + ''' + Display list of groups the user has access to. + ''' + meeting = get_meeting(days=14) + + # check for locked flag + is_locked = check_app_locked() + if is_locked and not has_role(request.user, 'Secretariat'): + message = get_lock_message() + return render(request, 'meeting/session_request_locked.html', { + 'message': message, + 'meeting': meeting}) + + scheduled_groups = [] + unscheduled_groups = [] + + group_types = GroupFeatures.objects.filter(has_meetings=True).values_list('type', flat=True) + + my_groups = [g for g in get_my_groups(request.user, conclude=True) if g.type_id in group_types] + + sessions_by_group = defaultdict(list) + for s in add_event_info_to_session_qs(Session.objects.filter(meeting=meeting, group__in=my_groups)).filter(current_status__in=['schedw', 'apprw', 'appr', 'sched']): + sessions_by_group[s.group_id].append(s) + + for group in my_groups: + group.meeting_sessions = sessions_by_group.get(group.pk, []) + + if group.pk in sessions_by_group: + # include even if concluded as we need to to see that the + # sessions are there + scheduled_groups.append(group) + else: + if group.state_id not in ['conclude', 'bof-conc']: + # too late for unscheduled if concluded + unscheduled_groups.append(group) + + # warn if there are no associated groups + if not scheduled_groups and not unscheduled_groups: + messages.warning(request, 'The account %s is not associated with any groups. If you have multiple Datatracker accounts you may try another or report a problem to %s' % (request.user, settings.SECRETARIAT_ACTION_EMAIL)) + + # add session status messages for use in template + for group in scheduled_groups: + if not group.features.acts_like_wg or (len(group.meeting_sessions) < 3): + group.status_message = group.meeting_sessions[0].current_status + else: + group.status_message = 'First two sessions: %s, Third session: %s' % (group.meeting_sessions[0].current_status, group.meeting_sessions[2].current_status) + + # add not meeting indicators for use in template + for group in unscheduled_groups: + if any(s.current_status == 'notmeet' for s in group.meeting_sessions): + group.not_meeting = True + + return render(request, 'meeting/session_request_list.html', { + 'is_locked': is_locked, + 'meeting': meeting, + 'scheduled_groups': scheduled_groups, + 'unscheduled_groups': unscheduled_groups}, + ) + + +@role_required('Secretariat') +def status(request): + ''' + This view handles locking and unlocking of the session request tool to the public. + ''' + meeting = get_meeting(days=14) + is_locked = check_app_locked(meeting=meeting) + + if request.method == 'POST': + button_text = request.POST.get('submit', '') + if button_text == 'Back': + return redirect('ietf.meeting.views_session_request.list_view') + + form = SessionRequestStatusForm(request.POST) + + if button_text == 'Lock': + if form.is_valid(): + meeting.session_request_lock_message = form.cleaned_data['message'] + meeting.save() + messages.success(request, 'Session Request Tool is now Locked') + return redirect('ietf.meeting.views_session_request.list_view') + + elif button_text == 'Unlock': + meeting.session_request_lock_message = '' + meeting.save() + messages.success(request, 'Session Request Tool is now Unlocked') + return redirect('ietf.meeting.views_session_request.list_view') + + else: + if is_locked: + message = get_lock_message() + initial = {'message': message} + form = SessionRequestStatusForm(initial=initial) + else: + form = SessionRequestStatusForm() + + return render(request, 'meeting/session_request_status.html', { + 'is_locked': is_locked, + 'form': form}, + ) + + +@check_permissions +def new_request(request, acronym): + ''' + This view gathers details for a new session request. The user proceeds to confirm() + to create the request. + ''' + group = get_object_or_404(Group, acronym=acronym) + if len(group.features.session_purposes) == 0: + raise Http404(f'Cannot request sessions for group "{acronym}"') + meeting = get_meeting(days=14) + session_conflicts = dict(inbound=inbound_session_conflicts_as_string(group, meeting)) + + # check if app is locked + is_locked = check_app_locked() + if is_locked and not has_role(request.user, 'Secretariat'): + messages.warning(request, "The Session Request Tool is closed") + return redirect('ietf.meeting.views_session_request.list_view') + + if request.method == 'POST': + button_text = request.POST.get('submit', '') + if button_text == 'Cancel': + return redirect('ietf.meeting.views_session_request.list_view') + + form = SessionRequestForm(group, meeting, request.POST, notifications_optional=has_role(request.user, "Secretariat")) + if form.is_valid(): + return confirm(request, acronym) + + # the "previous" querystring causes the form to be returned + # pre-populated with data from last meeeting's session request + elif request.method == 'GET' and 'previous' in request.GET: + latest_session = add_event_info_to_session_qs(Session.objects.filter(meeting__type_id='ietf', group=group)).exclude(current_status__in=['notmeet', 'deleted', 'canceled',]).order_by('-meeting__date').first() + if latest_session: + previous_meeting = Meeting.objects.get(number=latest_session.meeting.number) + previous_sessions = add_event_info_to_session_qs(Session.objects.filter(meeting=previous_meeting, group=group)).exclude(current_status__in=['notmeet', 'deleted']).order_by('id') + if not previous_sessions: + messages.warning(request, 'This group did not meet at %s' % previous_meeting) + return redirect('ietf.meeting.views_session_request.new_request', acronym=acronym) + else: + messages.info(request, 'Fetched session info from %s' % previous_meeting) + else: + messages.warning(request, 'Did not find any previous meeting') + return redirect('ietf.meeting.views_session_request.new_request', acronym=acronym) + + initial = get_initial_session(previous_sessions, prune_conflicts=True) + if 'resources' in initial: + initial['resources'] = [x.pk for x in initial['resources']] + form = SessionRequestForm(group, meeting, initial=initial, notifications_optional=has_role(request.user, "Secretariat")) + + else: + initial = {} + form = SessionRequestForm(group, meeting, initial=initial, notifications_optional=has_role(request.user, "Secretariat")) + + return render(request, 'meeting/session_request_form.html', { + 'meeting': meeting, + 'form': form, + 'group': group, + 'is_create': True, + 'session_conflicts': session_conflicts}, + ) + + +@role_required(*AUTHORIZED_ROLES) +def confirm(request, acronym): + ''' + This view displays details of the new session that has been requested for the user + to confirm for submission. + ''' + # FIXME: this should be using form.is_valid/form.cleaned_data - invalid input will make it crash + group = get_object_or_404(Group, acronym=acronym) + if len(group.features.session_purposes) == 0: + raise Http404(f'Cannot request sessions for group "{acronym}"') + meeting = get_meeting(days=14) + form = SessionRequestForm(group, meeting, request.POST, hidden=True, notifications_optional=has_role(request.user, "Secretariat")) + form.is_valid() + + login = request.user.person + + # check if request already exists for this group + if add_event_info_to_session_qs(Session.objects.filter(group=group, meeting=meeting)).filter(Q(current_status__isnull=True) | ~Q(current_status__in=['deleted', 'notmeet'])): + messages.warning(request, 'Sessions for working group %s have already been requested once.' % group.acronym) + return redirect('ietf.meeting.views_session_request.list_view') + + session_data = form.data.copy() + # use cleaned_data for the 'bethere' field so we get the Person instances + session_data['bethere'] = form.cleaned_data['bethere'] if 'bethere' in form.cleaned_data else [] + if session_data.get('session_time_relation'): + session_data['session_time_relation_display'] = dict(Constraint.TIME_RELATION_CHOICES)[session_data['session_time_relation']] + if session_data.get('joint_for_session'): + session_data['joint_for_session_display'] = dict(JOINT_FOR_SESSION_CHOICES)[session_data['joint_for_session']] + if form.cleaned_data.get('timeranges'): + session_data['timeranges_display'] = [t.desc for t in form.cleaned_data['timeranges']] + session_data['resources'] = [ResourceAssociation.objects.get(pk=pk) for pk in request.POST.getlist('resources')] + + # extract wg conflict constraint data for the view / notifications + outbound_conflicts = get_outbound_conflicts(form) + + button_text = request.POST.get('submit', '') + if button_text == 'Cancel': + messages.success(request, 'Session Request has been cancelled') + return redirect('ietf.meeting.views_session_request.list_view') + + if request.method == 'POST' and button_text == 'Submit': + # delete any existing session records with status = canceled or notmeet + add_event_info_to_session_qs(Session.objects.filter(group=group, meeting=meeting)).filter(current_status__in=['canceled', 'notmeet']).delete() + num_sessions = int(form.cleaned_data['num_session']) + (1 if form.cleaned_data['third_session'] else 0) + # Create new session records + form.session_forms.save() + for count, sess_form in enumerate(form.session_forms[:num_sessions]): + new_session = sess_form.instance + SchedulingEvent.objects.create( + session=new_session, + status=SessionStatusName.objects.get(slug=status_slug_for_new_session(new_session, count)), + by=login, + ) + if 'resources' in form.data: + new_session.resources.set(session_data['resources']) + jfs = form.data.get('joint_for_session', '-1') + if not jfs: # jfs might be '' + jfs = '-1' + if int(jfs) == count + 1: # count is zero-indexed + groups_split = form.cleaned_data.get('joint_with_groups').replace(',', ' ').split() + joint = Group.objects.filter(acronym__in=groups_split) + new_session.joint_with_groups.set(joint) + new_session.save() + session_changed(new_session) + + # write constraint records + for conflictname, cfield_id in form.wg_constraint_field_ids(): + save_conflicts(group, meeting, form.data.get(cfield_id, ''), conflictname.slug) + save_conflicts(group, meeting, form.data.get('adjacent_with_wg', ''), 'wg_adjacent') + + if form.cleaned_data.get('session_time_relation'): + cn = ConstraintName.objects.get(slug='time_relation') + Constraint.objects.create(source=group, meeting=meeting, name=cn, + time_relation=form.cleaned_data['session_time_relation']) + + if form.cleaned_data.get('timeranges'): + cn = ConstraintName.objects.get(slug='timerange') + constraint = Constraint.objects.create(source=group, meeting=meeting, name=cn) + constraint.timeranges.set(form.cleaned_data['timeranges']) + + if 'bethere' in form.data: + bethere_cn = ConstraintName.objects.get(slug='bethere') + for p in session_data['bethere']: + Constraint.objects.create(name=bethere_cn, source=group, person=p, meeting=new_session.meeting) + + # clear not meeting + add_event_info_to_session_qs(Session.objects.filter(group=group, meeting=meeting)).filter(current_status='notmeet').delete() + + # send notification + if form.cleaned_data.get("send_notifications"): + session_data['outbound_conflicts'] = [f"{d['name']}: {d['groups']}" for d in outbound_conflicts] + send_notification( + group, + meeting, + login, + session_data, + [sf.cleaned_data for sf in form.session_forms[:num_sessions]], + 'new', + ) + + status_text = 'IETF Agenda to be scheduled' + messages.success(request, 'Your request has been sent to %s' % status_text) + return redirect('ietf.meeting.views_session_request.list_view') + + # POST from request submission + session_conflicts = dict( + outbound=outbound_conflicts, # each is a dict with name and groups as keys + inbound=inbound_session_conflicts_as_string(group, meeting), + ) + if form.cleaned_data.get('third_session'): + messages.warning(request, 'Note: Your request for a third session must be approved by an area director before being submitted to agenda@ietf.org. Click "Submit" below to email an approval request to the area directors') + + return render(request, 'meeting/session_request_confirm.html', { + 'form': form, + 'session': session_data, + 'group': group, + 'meeting': meeting, + 'session_conflicts': session_conflicts}, + ) + + +@role_required(*AUTHORIZED_ROLES) +def view_request(request, acronym, num=None): + ''' + This view displays the session request info + ''' + meeting = get_meeting(num, days=14) + group = get_object_or_404(Group, acronym=acronym) + query = Session.objects.filter(meeting=meeting, group=group) + status_is_null = Q(current_status__isnull=True) + status_allowed = ~Q(current_status__in=("canceled", "notmeet", "deleted")) + sessions = ( + add_event_info_to_session_qs(query) + .filter(status_is_null | status_allowed) + .order_by("id") + ) + + # check if app is locked + is_locked = check_app_locked() + if is_locked: + messages.warning(request, "The Session Request Tool is closed") + + # if there are no session requests yet, redirect to new session request page + if not sessions: + if is_locked: + return redirect('ietf.meeting.views_session_request.list_view') + else: + return redirect('ietf.meeting.views_session_request.new_request', acronym=acronym) + + activities = [{ + 'act_date': e.time.strftime('%b %d, %Y'), + 'act_time': e.time.strftime('%H:%M:%S'), + 'activity': e.status.name, + 'act_by': e.by, + } for e in sessions[0].schedulingevent_set.select_related('status', 'by')] + + # gather outbound conflicts + outbound_dict = OrderedDict() + for obc in group.constraint_source_set.filter(meeting=meeting, name__is_group_conflict=True): + if obc.name.slug not in outbound_dict: + outbound_dict[obc.name.slug] = [] + outbound_dict[obc.name.slug].append(obc.target.acronym) + + session_conflicts = dict( + inbound=inbound_session_conflicts_as_string(group, meeting), + outbound=[dict(name=ConstraintName.objects.get(slug=slug), groups=' '.join(groups)) + for slug, groups in outbound_dict.items()], + ) + + show_approve_button = False + + # if sessions include a 3rd session waiting approval and the user is a secretariat or AD of the group + # display approve button + if any(s.current_status == 'apprw' for s in sessions): + if has_role(request.user, 'Secretariat') or group.parent.role_set.filter(name='ad', person=request.user.person): + show_approve_button = True + + # build session dictionary (like querydict from new session request form) for use in template + session = get_initial_session(sessions) + + return render(request, 'meeting/session_request_view.html', { + 'can_edit': (not is_locked) or has_role(request.user, 'Secretariat'), + 'can_cancel': (not is_locked) or has_role(request.user, 'Secretariat'), + 'session': session, # legacy processed data + 'sessions': sessions, # actual session instances + 'activities': activities, + 'meeting': meeting, + 'group': group, + 'session_conflicts': session_conflicts, + 'show_approve_button': show_approve_button}, + ) + + +@check_permissions +def edit_request(request, acronym, num=None): + ''' + This view allows the user to edit details of the session request + ''' + meeting = get_meeting(num, days=14) + group = get_object_or_404(Group, acronym=acronym) + if len(group.features.session_purposes) == 0: + raise Http404(f'Cannot request sessions for group "{acronym}"') + sessions = add_event_info_to_session_qs( + Session.objects.filter(group=group, meeting=meeting) + ).filter( + Q(current_status__isnull=True) | ~Q(current_status__in=['canceled', 'notmeet', 'deleted']) + ).order_by('id') + initial = get_initial_session(sessions) + + if 'resources' in initial: + initial['resources'] = [x.pk for x in initial['resources']] + + # check if app is locked + is_locked = check_app_locked(meeting=meeting) + if is_locked: + messages.warning(request, "The Session Request Tool is closed") + + # Only need the inbound conflicts here, the form itself renders the outbound + session_conflicts = dict( + inbound=inbound_session_conflicts_as_string(group, meeting), + ) + login = request.user.person + + first_session = Session() + if (len(sessions) > 0): + first_session = sessions[0] + + if request.method == 'POST': + button_text = request.POST.get('submit', '') + if button_text == 'Cancel': + return redirect('ietf.meeting.views_session_request.view_request', acronym=acronym) + + form = SessionRequestForm(group, meeting, request.POST, initial=initial, notifications_optional=has_role(request.user, "Secretariat")) + if form.is_valid(): + if form.has_changed(): + changed_session_forms = [sf for sf in form.session_forms.forms_to_keep if sf.has_changed()] + form.session_forms.save() + for n, subform in enumerate(form.session_forms): + if subform.instance in form.session_forms.new_objects: + SchedulingEvent.objects.create( + session=subform.instance, + status_id=status_slug_for_new_session(subform.instance, n), + by=request.user.person, + ) + for sf in changed_session_forms: + session_changed(sf.instance) + + # New sessions may have been created, refresh the sessions list + sessions = add_event_info_to_session_qs( + Session.objects.filter(group=group, meeting=meeting)).filter( + Q(current_status__isnull=True) | ~Q( + current_status__in=['canceled', 'notmeet'])).order_by('id') + + if 'joint_with_groups' in form.changed_data or 'joint_for_session' in form.changed_data: + joint_with_groups_list = form.cleaned_data.get('joint_with_groups').replace(',', ' ').split() + new_joint_with_groups = Group.objects.filter(acronym__in=joint_with_groups_list) + new_joint_for_session_idx = int(form.data.get('joint_for_session', '-1')) - 1 + current_joint_for_session_idx = None + current_joint_with_groups = None + for idx, sess in enumerate(sessions): + if sess.joint_with_groups.count(): + current_joint_for_session_idx = idx + current_joint_with_groups = sess.joint_with_groups.all() + + if current_joint_with_groups != new_joint_with_groups or current_joint_for_session_idx != new_joint_for_session_idx: + if current_joint_for_session_idx is not None: + sessions[current_joint_for_session_idx].joint_with_groups.clear() + session_changed(sessions[current_joint_for_session_idx]) + sessions[new_joint_for_session_idx].joint_with_groups.set(new_joint_with_groups) + session_changed(sessions[new_joint_for_session_idx]) + + # Update sessions to match changes to shared form fields + if 'attendees' in form.changed_data: + sessions.update(attendees=form.cleaned_data['attendees']) + if 'comments' in form.changed_data: + sessions.update(comments=form.cleaned_data['comments']) + + # Handle constraints + for cname, cfield_id in form.wg_constraint_field_ids(): + if cfield_id in form.changed_data: + Constraint.objects.filter(meeting=meeting, source=group, name=cname.slug).delete() + save_conflicts(group, meeting, form.cleaned_data[cfield_id], cname.slug) + + # see if any inactive constraints should be deleted + for cname, field_id in form.inactive_wg_constraint_field_ids(): + if form.cleaned_data[field_id]: + Constraint.objects.filter(meeting=meeting, source=group, name=cname.slug).delete() + + if 'adjacent_with_wg' in form.changed_data: + Constraint.objects.filter(meeting=meeting, source=group, name='wg_adjacent').delete() + save_conflicts(group, meeting, form.cleaned_data['adjacent_with_wg'], 'wg_adjacent') + + if 'resources' in form.changed_data: + new_resource_ids = form.cleaned_data['resources'] + new_resources = [ResourceAssociation.objects.get(pk=a) + for a in new_resource_ids] + first_session.resources = new_resources + + if 'bethere' in form.changed_data and set(form.cleaned_data['bethere']) != set(initial['bethere']): + first_session.constraints().filter(name='bethere').delete() + bethere_cn = ConstraintName.objects.get(slug='bethere') + for p in form.cleaned_data['bethere']: + Constraint.objects.create(name=bethere_cn, source=group, person=p, meeting=first_session.meeting) + + if 'session_time_relation' in form.changed_data: + Constraint.objects.filter(meeting=meeting, source=group, name='time_relation').delete() + if form.cleaned_data['session_time_relation']: + cn = ConstraintName.objects.get(slug='time_relation') + Constraint.objects.create(source=group, meeting=meeting, name=cn, + time_relation=form.cleaned_data['session_time_relation']) + + if 'timeranges' in form.changed_data: + Constraint.objects.filter(meeting=meeting, source=group, name='timerange').delete() + if form.cleaned_data['timeranges']: + cn = ConstraintName.objects.get(slug='timerange') + constraint = Constraint.objects.create(source=group, meeting=meeting, name=cn) + constraint.timeranges.set(form.cleaned_data['timeranges']) + + # deprecated + # log activity + # add_session_activity(group,'Session Request was updated',meeting,user) + + # send notification + if form.cleaned_data.get("send_notifications"): + outbound_conflicts = get_outbound_conflicts(form) + session_data = form.cleaned_data.copy() # do not add things to the original cleaned_data + session_data['outbound_conflicts'] = [f"{d['name']}: {d['groups']}" for d in outbound_conflicts] + send_notification( + group, + meeting, + login, + session_data, + [sf.cleaned_data for sf in form.session_forms.forms_to_keep], + 'update', + ) + + messages.success(request, 'Session Request updated') + return redirect('ietf.meeting.views_session_request.view_request', acronym=acronym) + + else: # method is not POST + # gather outbound conflicts for initial value + outbound_constraints = defaultdict(list) + for obc in group.constraint_source_set.filter(meeting=meeting, name__is_group_conflict=True): + outbound_constraints[obc.name.slug].append(obc.target.acronym) + for slug, groups in outbound_constraints.items(): + initial['constraint_{}'.format(slug)] = ' '.join(groups) + + if not sessions: + return redirect('ietf.meeting.views_session_request.new_request', acronym=acronym) + form = SessionRequestForm(group, meeting, initial=initial, notifications_optional=has_role(request.user, "Secretariat")) + + return render(request, 'meeting/session_request_form.html', { + 'is_locked': is_locked and not has_role(request.user, 'Secretariat'), + 'meeting': meeting, + 'form': form, + 'group': group, + 'is_create': False, + 'session_conflicts': session_conflicts}, + ) + + +@check_permissions +def approve_request(request, acronym): + ''' + This view approves the third session. For use by ADs or Secretariat. + ''' + meeting = get_meeting(days=14) + group = get_object_or_404(Group, acronym=acronym) + + session = add_event_info_to_session_qs(Session.objects.filter(meeting=meeting, group=group)).filter(current_status='apprw').first() + if session is None: + raise Http404 + + if has_role(request.user, 'Secretariat') or group.parent.role_set.filter(name='ad', person=request.user.person): + SchedulingEvent.objects.create( + session=session, + status=SessionStatusName.objects.get(slug='appr'), + by=request.user.person, + ) + session_changed(session) + + messages.success(request, 'Third session approved') + return redirect('ietf.meeting.views_session_request.view_request', acronym=acronym) + else: + # if an unauthorized user gets here return error + messages.error(request, 'Not authorized to approve the third session') + return redirect('ietf.meeting.views_session_request.view_request', acronym=acronym) + + +@check_permissions +def no_session(request, acronym): + ''' + The user has indicated that the named group will not be having a session this IETF meeting. + Actions: + - send notification + - update session_activity log + ''' + meeting = get_meeting(days=14) + group = get_object_or_404(Group, acronym=acronym) + login = request.user.person + + # delete canceled record if there is one + add_event_info_to_session_qs(Session.objects.filter(group=group, meeting=meeting)).filter(current_status='canceled').delete() + + # skip if state is already notmeet + if add_event_info_to_session_qs(Session.objects.filter(group=group, meeting=meeting)).filter(current_status='notmeet'): + messages.info(request, 'The group %s is already marked as not meeting' % group.acronym) + return redirect('ietf.meeting.views_session_request.list_view') + + session = Session.objects.create( + group=group, + meeting=meeting, + requested_duration=datetime.timedelta(0), + type_id='regular', + purpose_id='regular', + has_onsite_tool=group.features.acts_like_wg, + ) + SchedulingEvent.objects.create( + session=session, + status=SessionStatusName.objects.get(slug='notmeet'), + by=login, + ) + session_changed(session) + + # send notification + (to_email, cc_list) = gather_address_lists('session_request_not_meeting', group=group, person=login) + from_email = (settings.SESSION_REQUEST_FROM_EMAIL) + subject = '%s - Not having a session at IETF %s' % (group.acronym, meeting.number) + send_mail(request, to_email, from_email, subject, 'meeting/session_not_meeting_notification.txt', + {'login': login, + 'group': group, + 'meeting': meeting}, cc=cc_list) + + # deprecated? + # log activity + # text = 'A message was sent to notify not having a session at IETF %d' % meeting.meeting_num + # add_session_activity(group,text,meeting,request.person) + + # redirect + messages.success(request, 'A message was sent to notify not having a session at IETF %s' % meeting.number) + return redirect('ietf.meeting.views_session_request.list_view') + + +@check_permissions +def cancel_request(request, acronym): + ''' + This view cancels a session request and sends a notification. + To cancel, or withdraw the request set status = deleted. + "canceled" status is used by the secretariat. + + NOTE: this function can also be called after a session has been + scheduled during the period when the session request tool is + reopened. In this case be sure to clear the timeslot assignment as well. + ''' + meeting = get_meeting(days=14) + group = get_object_or_404(Group, acronym=acronym) + sessions = Session.objects.filter(meeting=meeting, group=group).order_by('id') + login = request.user.person + + # delete conflicts + Constraint.objects.filter(meeting=meeting, source=group).delete() + + # mark sessions as deleted + for session in sessions: + SchedulingEvent.objects.create( + session=session, + status=SessionStatusName.objects.get(slug='deleted'), + by=request.user.person, + ) + session_changed(session) + + # clear schedule assignments if already scheduled + session.timeslotassignments.all().delete() + + # send notifitcation + (to_email, cc_list) = gather_address_lists('session_request_cancelled', group=group, person=login) + from_email = (settings.SESSION_REQUEST_FROM_EMAIL) + subject = '%s - Cancelling a meeting request for IETF %s' % (group.acronym, meeting.number) + send_mail(request, to_email, from_email, subject, 'meeting/session_cancel_notification.txt', + {'requester': get_requester_text(login, group), + 'meeting': meeting}, cc=cc_list) + + messages.success(request, 'The %s Session Request has been cancelled' % group.acronym) + return redirect('ietf.meeting.views_session_request.list_view') diff --git a/ietf/message/admin.py b/ietf/message/admin.py index c2564c04b9..6a876cdc70 100644 --- a/ietf/message/admin.py +++ b/ietf/message/admin.py @@ -1,32 +1,104 @@ -from django.contrib import admin +# Copyright The IETF Trust 2012-2025, All Rights Reserved +from django.contrib import admin, messages +from django.db.models import QuerySet +from rangefilter.filters import DateRangeQuickSelectListFilterBuilder from ietf.message.models import Message, MessageAttachment, SendQueue, AnnouncementFrom +from ietf.message.tasks import retry_send_messages_by_pk_task + + +class MessageSentStatusListFilter(admin.SimpleListFilter): + """Filter Messages by whether or not they were sent""" + + title = "status" + parameter_name = "status" + + def lookups(self, request, model_admin): + return [ + ("sent", "Sent"), + ("unsent", "Not sent"), + ] + + def queryset(self, request, queryset): + if self.value() == "unsent": + return queryset.filter(sent__isnull=True) + elif self.value() == "sent": + return queryset.filter(sent__isnull=False) + class MessageAdmin(admin.ModelAdmin): - list_display = ["subject", "by", "time", "groups"] + list_display = ["sent_status", "display_subject", "by", "time", "groups"] + list_display_links = ["display_subject"] search_fields = ["subject", "body"] raw_id_fields = ["by", "related_groups", "related_docs"] + list_filter = [ + MessageSentStatusListFilter, + ("time", DateRangeQuickSelectListFilterBuilder()), + ] ordering = ["-time"] + actions = ["retry_send"] + + @admin.display(description="Subject", empty_value="(no subject)") + def display_subject(self, instance): + return instance.subject or None # None triggers the empty_value def groups(self, instance): return ", ".join(g.acronym for g in instance.related_groups.all()) + + @admin.display(description="Sent", boolean=True) + def sent_status(self, instance): + return instance.sent is not None + + @admin.action(description="Send selected messages if unsent") + def retry_send(self, request, queryset: QuerySet[Message]): + try: + retry_send_messages_by_pk_task.delay( + message_pks=list(queryset.values_list("pk", flat=True)), + resend=False, + ) + except Exception as err: + self.message_user( + request, + f"Error: {repr(err)}", + messages.ERROR, + ) + else: + self.message_user(request, "Messages queued for delivery", messages.SUCCESS) + + admin.site.register(Message, MessageAdmin) + class MessageAttachmentAdmin(admin.ModelAdmin): - list_display = ['id', 'message', 'filename', 'removed',] - raw_id_fields = ['message'] + list_display = [ + "id", + "message", + "filename", + "removed", + ] + raw_id_fields = ["message"] + + admin.site.register(MessageAttachment, MessageAttachmentAdmin) + class SendQueueAdmin(admin.ModelAdmin): list_display = ["time", "by", "message", "send_at", "sent_at"] list_filter = ["time", "send_at", "sent_at"] search_fields = ["message__body"] raw_id_fields = ["by", "message"] ordering = ["-time"] + + admin.site.register(SendQueue, SendQueueAdmin) + class AnnouncementFromAdmin(admin.ModelAdmin): - list_display = ['name', 'group', 'address', ] -admin.site.register(AnnouncementFrom, AnnouncementFromAdmin) + list_display = [ + "name", + "group", + "address", + ] +admin.site.register(AnnouncementFrom, AnnouncementFromAdmin) diff --git a/ietf/message/factories.py b/ietf/message/factories.py new file mode 100644 index 0000000000..72781512e4 --- /dev/null +++ b/ietf/message/factories.py @@ -0,0 +1,27 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +import factory + +from ietf.person.models import Person +from .models import Message, SendQueue + + +class MessageFactory(factory.django.DjangoModelFactory): + class Meta: + model = Message + + by = factory.LazyFunction(lambda: Person.objects.get(name="(System)")) + subject = factory.Faker("sentence") + to = factory.Faker("email") + frm = factory.Faker("email") + cc = factory.Faker("email") + bcc = factory.Faker("email") + body = factory.Faker("paragraph") + content_type = "text/plain" + + +class SendQueueFactory(factory.django.DjangoModelFactory): + class Meta: + model = SendQueue + + by = factory.LazyFunction(lambda: Person.objects.get(name="(System)")) + message = factory.SubFactory(MessageFactory) diff --git a/ietf/message/migrations/0001_initial.py b/ietf/message/migrations/0001_initial.py index c9e850dfb6..108e9d2db5 100644 --- a/ietf/message/migrations/0001_initial.py +++ b/ietf/message/migrations/0001_initial.py @@ -1,12 +1,10 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-20 10:52 +# Generated by Django 2.2.28 on 2023-03-30 19:14 - -import datetime from django.db import migrations, models import django.db.models.deletion +import django.utils.timezone import email.utils +import ietf.message.models import ietf.utils.models @@ -16,41 +14,45 @@ class Migration(migrations.Migration): dependencies = [ ('group', '0001_initial'), - ('name', '0001_initial'), ('person', '0001_initial'), + ('name', '0001_initial'), ('doc', '0001_initial'), ] operations = [ migrations.CreateModel( - name='AnnouncementFrom', + name='Message', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('address', models.CharField(max_length=255)), - ('group', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='group.Group')), - ('name', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.RoleName')), + ('time', models.DateTimeField(default=django.utils.timezone.now)), + ('subject', ietf.message.models.HeaderField()), + ('frm', ietf.message.models.HeaderField()), + ('to', ietf.message.models.HeaderField()), + ('cc', ietf.message.models.HeaderField(blank=True)), + ('bcc', ietf.message.models.HeaderField(blank=True)), + ('reply_to', ietf.message.models.HeaderField(blank=True)), + ('body', models.TextField()), + ('content_type', models.CharField(blank=True, default='text/plain', max_length=255)), + ('msgid', ietf.message.models.HeaderField(blank=True, default=email.utils.make_msgid, null=True)), + ('sent', models.DateTimeField(blank=True, null=True)), + ('by', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), + ('related_docs', models.ManyToManyField(blank=True, to='doc.Document')), + ('related_groups', models.ManyToManyField(blank=True, to='group.Group')), ], options={ - 'verbose_name_plural': 'Announcement From addresses', + 'ordering': ['time'], }, ), migrations.CreateModel( - name='Message', + name='SendQueue', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('time', models.DateTimeField(default=datetime.datetime.now)), - ('subject', models.CharField(max_length=255)), - ('frm', models.CharField(max_length=255)), - ('to', models.CharField(max_length=1024)), - ('cc', models.CharField(blank=True, max_length=1024)), - ('bcc', models.CharField(blank=True, max_length=255)), - ('reply_to', models.CharField(blank=True, max_length=255)), - ('body', models.TextField()), - ('content_type', models.CharField(blank=True, default='text/plain', max_length=255)), - ('msgid', models.CharField(blank=True, default=email.utils.make_msgid, max_length=255, null=True)), + ('time', models.DateTimeField(default=django.utils.timezone.now)), + ('send_at', models.DateTimeField(blank=True, null=True)), + ('sent_at', models.DateTimeField(blank=True, null=True)), + ('note', models.TextField(blank=True)), ('by', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), - ('related_docs', models.ManyToManyField(blank=True, to='doc.Document')), - ('related_groups', models.ManyToManyField(blank=True, to='group.Group')), + ('message', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='message.Message')), ], options={ 'ordering': ['time'], @@ -69,18 +71,23 @@ class Migration(migrations.Migration): ], ), migrations.CreateModel( - name='SendQueue', + name='AnnouncementFrom', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('time', models.DateTimeField(default=datetime.datetime.now)), - ('send_at', models.DateTimeField(blank=True, null=True)), - ('sent_at', models.DateTimeField(blank=True, null=True)), - ('note', models.TextField(blank=True)), - ('by', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), - ('message', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='message.Message')), + ('address', models.CharField(max_length=255)), + ('group', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='group.Group')), + ('name', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.RoleName')), ], options={ - 'ordering': ['time'], + 'verbose_name_plural': 'Announcement From addresses', }, ), + migrations.AddIndex( + model_name='sendqueue', + index=models.Index(fields=['time'], name='message_sen_time_07ab31_idx'), + ), + migrations.AddIndex( + model_name='message', + index=models.Index(fields=['time'], name='message_mes_time_20eabf_idx'), + ), ] diff --git a/ietf/message/migrations/0002_add_message_docs2_m2m.py b/ietf/message/migrations/0002_add_message_docs2_m2m.py deleted file mode 100644 index c6a138ba6d..0000000000 --- a/ietf/message/migrations/0002_add_message_docs2_m2m.py +++ /dev/null @@ -1,31 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-21 14:23 - - -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('message', '0001_initial'), - ] - - operations = [ - migrations.CreateModel( - name='MessageDocs', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('message', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='message.Message')), - ('document', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field=b'id')), - ], - ), - migrations.AddField( - model_name='message', - name='related_docs2', - field=models.ManyToManyField(blank=True, related_name='messages', through='message.MessageDocs', to='doc.Document'), - ), - ] diff --git a/ietf/message/migrations/0003_set_document_m2m_keys.py b/ietf/message/migrations/0003_set_document_m2m_keys.py deleted file mode 100644 index 6429af4432..0000000000 --- a/ietf/message/migrations/0003_set_document_m2m_keys.py +++ /dev/null @@ -1,47 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-21 14:27 - - -import sys - -from tqdm import tqdm - -from django.db import migrations - - -def forward(apps, schema_editor): - - Document = apps.get_model('doc','Document') - Message = apps.get_model('message', 'Message') - MessageDocs = apps.get_model('message', 'MessageDocs') - - - # Document id fixup ------------------------------------------------------------ - - objs = Document.objects.in_bulk() - nameid = { o.name: o.id for id, o in objs.items() } - - sys.stderr.write('\n') - - sys.stderr.write(' %s.%s:\n' % (Message.__name__, 'related_docs')) - count = 0 - for m in tqdm(Message.objects.all()): - for d in m.related_docs.all(): - count += 1; - MessageDocs.objects.get_or_create(message=m, document_id=nameid[d.name]) - sys.stderr.write(' %s MessageDocs objects created\n' % (count, )) - -def reverse(apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('message', '0002_add_message_docs2_m2m'), - ('doc', '0014_set_document_docalias_id'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/message/migrations/0004_1_del_docs_m2m_table.py b/ietf/message/migrations/0004_1_del_docs_m2m_table.py deleted file mode 100644 index 7b116789d5..0000000000 --- a/ietf/message/migrations/0004_1_del_docs_m2m_table.py +++ /dev/null @@ -1,32 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-22 08:01 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('message', '0003_set_document_m2m_keys'), - ] - - # The implementation of AlterField in Django 1.11 applies - # 'ALTER TABLE
  • MODIFY ...;' in order to fix foregn keys - # to the altered field, but as it seems does _not_ fix up m2m - # intermediary tables in an equivalent manner, so here we remove and - # then recreate the m2m tables so they will have the appropriate field - # types. - - operations = [ - migrations.RemoveField( - model_name='message', - name='related_docs', - ), - migrations.AddField( - model_name='message', - name='related_docs', - field=models.ManyToManyField(to='doc.Document'), - ), - ] diff --git a/ietf/message/migrations/0004_2_add_docs_m2m_table.py b/ietf/message/migrations/0004_2_add_docs_m2m_table.py deleted file mode 100644 index 4f24a26fe8..0000000000 --- a/ietf/message/migrations/0004_2_add_docs_m2m_table.py +++ /dev/null @@ -1,32 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-22 08:01 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('message', '0004_1_del_docs_m2m_table'), - ] - - # The implementation of AlterField in Django 1.11 applies - # 'ALTER TABLE
    MODIFY ...;' in order to fix foregn keys - # to the altered field, but as it seems does _not_ fix up m2m - # intermediary tables in an equivalent manner, so here we remove and - # then recreate the m2m tables so they will have the appropriate field - # types. - - operations = [ - migrations.RemoveField( - model_name='message', - name='related_docs', - ), - migrations.AddField( - model_name='message', - name='related_docs', - field=models.ManyToManyField(blank=True, to='doc.Document'), - ), - ] diff --git a/ietf/message/migrations/0005_copy_docs_m2m_table.py b/ietf/message/migrations/0005_copy_docs_m2m_table.py deleted file mode 100644 index 8f7c185886..0000000000 --- a/ietf/message/migrations/0005_copy_docs_m2m_table.py +++ /dev/null @@ -1,28 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-27 05:56 - - -import sys, time - -from django.db import migrations - - -def timestamp(apps, schema_editor): - sys.stderr.write('\n %s' % time.strftime('%Y-%m-%d %H:%M:%S')) - -class Migration(migrations.Migration): - - dependencies = [ - ('message', '0004_2_add_docs_m2m_table'), - ] - - operations = [ - #migrations.RunPython(forward, reverse), - migrations.RunPython(timestamp, timestamp), - migrations.RunSQL( - "INSERT INTO message_message_related_docs SELECT * FROM message_messagedocs;", - "" - ), - migrations.RunPython(timestamp, timestamp), - ] diff --git a/ietf/message/migrations/0006_remove_docs2_m2m.py b/ietf/message/migrations/0006_remove_docs2_m2m.py deleted file mode 100644 index 97e3ef5f0e..0000000000 --- a/ietf/message/migrations/0006_remove_docs2_m2m.py +++ /dev/null @@ -1,31 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-30 03:32 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('message', '0005_copy_docs_m2m_table'), - ] - - operations = [ - migrations.RemoveField( - model_name='messagedocs', - name='document', - ), - migrations.RemoveField( - model_name='messagedocs', - name='message', - ), - migrations.RemoveField( - model_name='message', - name='related_docs2', - ), - migrations.DeleteModel( - name='MessageDocs', - ), - ] diff --git a/ietf/message/migrations/0007_message_sent.py b/ietf/message/migrations/0007_message_sent.py deleted file mode 100644 index 9ab1a4041e..0000000000 --- a/ietf/message/migrations/0007_message_sent.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.27 on 2020-02-22 09:29 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('message', '0006_remove_docs2_m2m'), - ] - - operations = [ - migrations.AddField( - model_name='message', - name='sent', - field=models.DateTimeField(blank=True, null=True), - ), - ] diff --git a/ietf/message/migrations/0008_set_message_sent.py b/ietf/message/migrations/0008_set_message_sent.py deleted file mode 100644 index 281099305f..0000000000 --- a/ietf/message/migrations/0008_set_message_sent.py +++ /dev/null @@ -1,34 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-21 14:27 - - -from tqdm import tqdm - -from django.db import migrations - - -def forward(apps, schema_editor): - - Message = apps.get_model('message', 'Message') - - for m in tqdm(Message.objects.filter(sent=None)): - if m.sendqueue_set.exists(): - q = m.sendqueue_set.last() - m.sent = q.sent_at - else: - m.sent = m.time - m.save() - -def reverse(apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('message', '0007_message_sent'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/message/migrations/0009_fix_address_lists.py b/ietf/message/migrations/0009_fix_address_lists.py deleted file mode 100644 index e14b6511c3..0000000000 --- a/ietf/message/migrations/0009_fix_address_lists.py +++ /dev/null @@ -1,41 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-21 14:27 - - -from tqdm import tqdm - -from django.db import migrations - -import debug # pyflakes:ignore - - -def forward(apps, schema_editor): - - Message = apps.get_model('message', 'Message') - - for m in tqdm(Message.objects.all()): - dirty = False - for fieldname in ['to', 'cc', 'bcc', ]: - f = getattr(m, fieldname) - if f.startswith("['") or f.startswith('[]') or f.startswith("[u'"): - l = eval(f) - if isinstance(l, list): - f = ','.join(l) - setattr(m, fieldname, f) - dirty = True - if dirty: - m.save() - -def reverse(apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('message', '0008_set_message_sent'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/message/migrations/0010_fix_content_type.py b/ietf/message/migrations/0010_fix_content_type.py deleted file mode 100644 index 0741142905..0000000000 --- a/ietf/message/migrations/0010_fix_content_type.py +++ /dev/null @@ -1,32 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-21 14:27 - - -from tqdm import tqdm - -from django.db import migrations - -import debug # pyflakes:ignore - - -def forward(apps, schema_editor): - - Message = apps.get_model('message', 'Message') - - for m in tqdm(Message.objects.filter(content_type='')): - m.content_type = 'text/plain' - m.save() - -def reverse(apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('message', '0009_fix_address_lists'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/message/migrations/0011_auto_20201109_0439.py b/ietf/message/migrations/0011_auto_20201109_0439.py deleted file mode 100644 index 5134592cff..0000000000 --- a/ietf/message/migrations/0011_auto_20201109_0439.py +++ /dev/null @@ -1,21 +0,0 @@ -# Generated by Django 2.2.17 on 2020-11-09 04:39 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('message', '0010_fix_content_type'), - ] - - operations = [ - migrations.AddIndex( - model_name='message', - index=models.Index(fields=['time'], name='message_mes_time_20eabf_idx'), - ), - migrations.AddIndex( - model_name='sendqueue', - index=models.Index(fields=['time'], name='message_sen_time_07ab31_idx'), - ), - ] diff --git a/ietf/message/migrations/0012_use_timezone_now_for_message_models.py b/ietf/message/migrations/0012_use_timezone_now_for_message_models.py deleted file mode 100644 index dbef893ea4..0000000000 --- a/ietf/message/migrations/0012_use_timezone_now_for_message_models.py +++ /dev/null @@ -1,24 +0,0 @@ -# Generated by Django 2.2.28 on 2022-07-12 11:24 - -from django.db import migrations, models -import django.utils.timezone - - -class Migration(migrations.Migration): - - dependencies = [ - ('message', '0011_auto_20201109_0439'), - ] - - operations = [ - migrations.AlterField( - model_name='message', - name='time', - field=models.DateTimeField(default=django.utils.timezone.now), - ), - migrations.AlterField( - model_name='sendqueue', - name='time', - field=models.DateTimeField(default=django.utils.timezone.now), - ), - ] diff --git a/ietf/message/models.py b/ietf/message/models.py index fe2c8d325f..dfa23abd4d 100644 --- a/ietf/message/models.py +++ b/ietf/message/models.py @@ -4,6 +4,7 @@ import email.utils +from django.forms import TextInput from django.db import models from django.utils import timezone @@ -16,19 +17,26 @@ from ietf.utils.models import ForeignKey from ietf.utils.mail import get_email_addresses_from_text + +class HeaderField(models.TextField): + """TextField that defaults to a TextInput widget""" + def formfield(self, **kwargs): + return super().formfield(**{'widget': TextInput, **kwargs}) + + class Message(models.Model): time = models.DateTimeField(default=timezone.now) by = ForeignKey(Person) - subject = models.CharField(max_length=255) - frm = models.CharField(max_length=255) - to = models.CharField(max_length=1024) - cc = models.CharField(max_length=1024, blank=True) - bcc = models.CharField(max_length=255, blank=True) - reply_to = models.CharField(max_length=255, blank=True) + subject = HeaderField() + frm = HeaderField() + to = HeaderField() + cc = HeaderField(blank=True) + bcc = HeaderField(blank=True) + reply_to = HeaderField(blank=True) body = models.TextField() content_type = models.CharField(default="text/plain", max_length=255, blank=True) - msgid = models.CharField(max_length=255, blank=True, null=True, default=email.utils.make_msgid) + msgid = HeaderField(blank=True, null=True, default=email.utils.make_msgid) related_groups = models.ManyToManyField(Group, blank=True) related_docs = models.ManyToManyField(Document, blank=True) diff --git a/ietf/message/tasks.py b/ietf/message/tasks.py new file mode 100644 index 0000000000..1fdff7bea4 --- /dev/null +++ b/ietf/message/tasks.py @@ -0,0 +1,47 @@ +# Copyright The IETF Trust 2024 All Rights Reserved +# +# Celery task definitions +# +from celery import shared_task +from smtplib import SMTPException + +from ietf.message.utils import send_scheduled_message_from_send_queue, retry_send_messages +from ietf.message.models import SendQueue, Message +from ietf.utils import log +from ietf.utils.mail import log_smtp_exception, send_error_email + + +@shared_task +def send_scheduled_mail_task(): + """Send scheduled email + + This is equivalent to `ietf/bin/send-scheduled-mail all`, which was the only form used in the cron job. + """ + needs_sending = SendQueue.objects.filter(sent_at=None).select_related("message") + for s in needs_sending: + try: + send_scheduled_message_from_send_queue(s) + log.log('Sent scheduled message %s "%s"' % (s.id, s.message.subject)) + except SMTPException as e: + log_smtp_exception(e) + send_error_email(e) + + +@shared_task +def retry_send_messages_by_pk_task(message_pks: list, resend=False): + """Task to retry sending Messages by PK + + Sends Messages whose PK is included in the list. + Only previously unsent messages are sent unless `resend` is true. + """ + log.log( + "retry_send_messages_by_pk_task: " + "retrying send of Message PKs [{}] (resend={})".format( + ", ".join(str(pk) for pk in message_pks), + resend, + ) + ) + retry_send_messages( + messages=Message.objects.filter(pk__in=message_pks), + resend=resend, + ) diff --git a/ietf/message/tests.py b/ietf/message/tests.py index a027df4473..e1bad9a1e6 100644 --- a/ietf/message/tests.py +++ b/ietf/message/tests.py @@ -1,8 +1,9 @@ # Copyright The IETF Trust 2013-2020, All Rights Reserved # -*- coding: utf-8 -*- - - import datetime +from unittest import mock + +from smtplib import SMTPException from django.urls import reverse as urlreverse from django.utils import timezone @@ -10,8 +11,10 @@ import debug # pyflakes:ignore from ietf.group.factories import GroupFactory +from ietf.message.factories import MessageFactory, SendQueueFactory from ietf.message.models import Message, SendQueue -from ietf.message.utils import send_scheduled_message_from_send_queue +from ietf.message.tasks import send_scheduled_mail_task, retry_send_messages_by_pk_task +from ietf.message.utils import send_scheduled_message_from_send_queue, retry_send_messages from ietf.person.models import Person from ietf.utils.mail import outbox, send_mail_text, send_mail_message, get_payload_text from ietf.utils.test_utils import TestCase @@ -128,3 +131,75 @@ def test_send_mime_announcement(self): self.assertTrue("This is a test" in outbox[-1]["Subject"]) self.assertTrue("--NextPart" in outbox[-1].as_string()) self.assertTrue(SendQueue.objects.get(id=q.id).sent_at) + + +class UtilsTests(TestCase): + @mock.patch("ietf.message.utils.send_mail_message") + def test_retry_send_messages(self, mock_send_mail_message): + sent_message = MessageFactory(sent=timezone.now()) + unsent_messages = MessageFactory.create_batch(2, sent=None) + + # Send the sent message and one of the unsent messages + retry_send_messages( + Message.objects.filter(pk__in=[ + sent_message.pk, + unsent_messages[0].pk, + ]), + resend=False, + ) + self.assertEqual(mock_send_mail_message.call_count, 1) + self.assertEqual( + mock_send_mail_message.call_args.args[1], + unsent_messages[0], + ) + + mock_send_mail_message.reset_mock() + # Once again, send the sent message and one of the unsent messages + # (we can use the same one because our mock prevented it from having + # its status updated to sent) + retry_send_messages( + Message.objects.filter(pk__in=[ + sent_message.pk, + unsent_messages[0].pk, + ]), + resend=True, + ) + self.assertEqual(mock_send_mail_message.call_count, 2) + self.assertCountEqual( + [call_args.args[1] for call_args in mock_send_mail_message.call_args_list], + [sent_message, unsent_messages[0]], + ) + + +class TaskTests(TestCase): + @mock.patch("ietf.message.tasks.log_smtp_exception") + @mock.patch("ietf.message.tasks.send_scheduled_message_from_send_queue") + def test_send_scheduled_mail_task(self, mock_send_message, mock_log_smtp_exception): + not_yet_sent = SendQueueFactory() + SendQueueFactory(sent_at=timezone.now()) # already sent + send_scheduled_mail_task() + self.assertEqual(mock_send_message.call_count, 1) + self.assertEqual(mock_send_message.call_args[0], (not_yet_sent,)) + self.assertFalse(mock_log_smtp_exception.called) + + mock_send_message.reset_mock() + mock_send_message.side_effect = SMTPException + send_scheduled_mail_task() + self.assertEqual(mock_send_message.call_count, 1) + self.assertEqual(mock_send_message.call_args[0], (not_yet_sent,)) + self.assertTrue(mock_log_smtp_exception.called) + + @mock.patch("ietf.message.tasks.retry_send_messages") + def test_retry_send_messages_by_pk_task(self, mock_retry_send): + msgs = MessageFactory.create_batch(3) + MessageFactory() # an extra message that won't be resent + + retry_send_messages_by_pk_task([msg.pk for msg in msgs], resend=False) + called_with_messages = mock_retry_send.call_args.kwargs["messages"] + self.assertCountEqual(msgs, called_with_messages) + self.assertFalse(mock_retry_send.call_args.kwargs["resend"]) + + retry_send_messages_by_pk_task([msg.pk for msg in msgs], resend=True) + called_with_messages = mock_retry_send.call_args.kwargs["messages"] + self.assertCountEqual(msgs, called_with_messages) + self.assertTrue(mock_retry_send.call_args.kwargs["resend"]) diff --git a/ietf/message/utils.py b/ietf/message/utils.py index 2601eccab8..74448ca7c9 100644 --- a/ietf/message/utils.py +++ b/ietf/message/utils.py @@ -1,13 +1,17 @@ # Copyright The IETF Trust 2012-2020, All Rights Reserved # -*- coding: utf-8 -*- +import email +import email.utils +import re +import smtplib -import re, email - +from django.db.models import QuerySet from django.utils import timezone from django.utils.encoding import force_str -from ietf.utils.mail import send_mail_text, send_mail_mime +from ietf.utils import log +from ietf.utils.mail import send_mail_text, send_mail_mime, send_mail_message from ietf.message.models import Message first_dot_on_line_re = re.compile(r'^\.', re.MULTILINE) @@ -58,3 +62,29 @@ def send_scheduled_message_from_send_queue(queue_item): queue_item.message.sent = queue_item.sent_at queue_item.message.save() + + +def retry_send_messages(messages: QuerySet[Message], resend=False): + """Attempt delivery of Messages""" + if not resend: + # only include sent messages on explicit request + for already_sent in messages.filter(sent__isnull=False): + assert already_sent.sent is not None # appease mypy type checking + log.log( + f"retry_send_messages: skipping {already_sent.pk} " + f"(already sent {already_sent.sent.isoformat(timespec='milliseconds')})" + ) + messages = messages.filter(sent__isnull=True) + for msg in messages: + to = ",".join(a[1] for a in email.utils.getaddresses([msg.to])) + try: + send_mail_message(None, msg) + log.log( + f'retry_send_messages: ' + f'sent {msg.pk} {msg.frm} -> {to} "{msg.subject.strip()}"' + ) + except smtplib.SMTPException as e: + log.log( + f'retry_send_messages: ' + f'Failure {e}: {msg.pk} {msg.frm} -> {to} "{msg.subject.strip()}"' + ) diff --git a/ietf/message/views.py b/ietf/message/views.py index e4cca63017..355dcdd8d2 100644 --- a/ietf/message/views.py +++ b/ietf/message/views.py @@ -1,3 +1,4 @@ +# Copyright The IETF Trust 2013-2025, All Rights Reserved from django.shortcuts import render, get_object_or_404 from ietf.message.models import Message diff --git a/ietf/middleware.py b/ietf/middleware.py index 48146abf5e..fa2e8efd0c 100644 --- a/ietf/middleware.py +++ b/ietf/middleware.py @@ -8,6 +8,7 @@ from django.http import HttpResponsePermanentRedirect from ietf.utils.log import log, exc_parts from ietf.utils.mail import log_smtp_exception +from opentelemetry.propagate import inject import re import smtplib import unicodedata @@ -17,45 +18,61 @@ def sql_log_middleware(get_response): def sql_log(request): response = get_response(request) for q in connection.queries: - if re.match('(update|insert)', q['sql'], re.IGNORECASE): - log(q['sql']) + if re.match("(update|insert)", q["sql"], re.IGNORECASE): + log(q["sql"]) return response + return sql_log + class SMTPExceptionMiddleware(object): def __init__(self, get_response): self.get_response = get_response + def __call__(self, request): return self.get_response(request) + def process_exception(self, request, exception): if isinstance(exception, smtplib.SMTPException): (extype, value, tb) = log_smtp_exception(exception) - return render(request, 'email_failed.html', - {'exception': extype, 'args': value, 'traceback': "".join(tb)} ) + return render( + request, + "email_failed.html", + {"exception": extype, "args": value, "traceback": "".join(tb)}, + ) return None + class Utf8ExceptionMiddleware(object): def __init__(self, get_response): self.get_response = get_response + def __call__(self, request): return self.get_response(request) + def process_exception(self, request, exception): if isinstance(exception, OperationalError): extype, e, tb = exc_parts() if e.args[0] == 1366: log("Database 4-byte utf8 exception: %s: %s" % (extype, e)) - return render(request, 'utf8_4byte_failed.html', - {'exception': extype, 'args': e.args, 'traceback': "".join(tb)} ) + return render( + request, + "utf8_4byte_failed.html", + {"exception": extype, "args": e.args, "traceback": "".join(tb)}, + ) return None + def redirect_trailing_period_middleware(get_response): def redirect_trailing_period(request): response = get_response(request) if response.status_code == 404 and request.path.endswith("."): return HttpResponsePermanentRedirect(request.path.rstrip(".")) return response + return redirect_trailing_period + def unicode_nfkc_normalization_middleware(get_response): def unicode_nfkc_normalization(request): """Do Unicode NFKC normalization to turn ligatures into individual characters. @@ -65,9 +82,30 @@ def unicode_nfkc_normalization(request): There are probably other elements of a request which may need this normalization too, but let's put that in as it comes up, rather than guess ahead. """ - request.META["PATH_INFO"] = unicodedata.normalize('NFKC', request.META["PATH_INFO"]) - request.path_info = unicodedata.normalize('NFKC', request.path_info) + request.META["PATH_INFO"] = unicodedata.normalize( + "NFKC", request.META["PATH_INFO"] + ) + request.path_info = unicodedata.normalize("NFKC", request.path_info) response = get_response(request) return response + return unicode_nfkc_normalization - + + +def is_authenticated_header_middleware(get_response): + """Middleware to add an is-authenticated header to the response""" + def add_header(request): + response = get_response(request) + response["X-Datatracker-Is-Authenticated"] = "yes" if request.user.is_authenticated else "no" + return response + + return add_header + +def add_otel_traceparent_header(get_response): + """Middleware to add the OpenTelemetry traceparent id header to the response""" + def add_header(request): + response = get_response(request) + inject(response) + return response + + return add_header diff --git a/ietf/name/admin.py b/ietf/name/admin.py index dd659d1135..b89d6d141c 100644 --- a/ietf/name/admin.py +++ b/ietf/name/admin.py @@ -2,63 +2,145 @@ from django.contrib import admin from ietf.name.models import ( - AgendaTypeName, BallotPositionName, ConstraintName, ContinentName, CountryName, DBTemplateTypeName, - DocRelationshipName, DocReminderTypeName, DocTagName, DocTypeName, DraftSubmissionStateName, - FeedbackTypeName, FormalLanguageName, GroupMilestoneStateName, GroupStateName, GroupTypeName, - ImportantDateName, IntendedStdLevelName, IprDisclosureStateName, IprEventTypeName, - IprLicenseTypeName, LiaisonStatementEventTypeName, LiaisonStatementPurposeName, - LiaisonStatementState, LiaisonStatementTagName, MeetingTypeName, NomineePositionStateName, - ReviewRequestStateName, ReviewResultName, ReviewTypeName, RoleName, RoomResourceName, - SessionStatusName, StdLevelName, StreamName, TimeSlotTypeName, TopicAudienceName, - DocUrlTagName, ReviewAssignmentStateName, ReviewerQueuePolicyName, TimerangeName, - ExtResourceName, ExtResourceTypeName, SlideSubmissionStatusName, ProceedingsMaterialTypeName, - AgendaFilterTypeName, SessionPurposeName ) + AgendaTypeName, + AttendanceTypeName, + BallotPositionName, + ConstraintName, + ContinentName, + CountryName, + DBTemplateTypeName, + DocRelationshipName, + DocReminderTypeName, + DocTagName, + DocTypeName, + DraftSubmissionStateName, + FeedbackTypeName, + FormalLanguageName, + GroupMilestoneStateName, + GroupStateName, + GroupTypeName, + ImportantDateName, + IntendedStdLevelName, + IprDisclosureStateName, + IprEventTypeName, + IprLicenseTypeName, + LiaisonStatementEventTypeName, + LiaisonStatementPurposeName, + LiaisonStatementState, + LiaisonStatementTagName, + MeetingTypeName, + NomineePositionStateName, + RegistrationTicketTypeName, + ReviewRequestStateName, + ReviewResultName, + ReviewTypeName, + RoleName, + RoomResourceName, + SessionStatusName, + StdLevelName, + StreamName, + TimeSlotTypeName, + TopicAudienceName, + DocUrlTagName, + ReviewAssignmentStateName, + ReviewerQueuePolicyName, + TimerangeName, + ExtResourceName, + ExtResourceTypeName, + SlideSubmissionStatusName, + ProceedingsMaterialTypeName, + AgendaFilterTypeName, + SessionPurposeName, + TelechatAgendaSectionName, + AppealArtifactTypeName, +) from ietf.stats.models import CountryAlias +from ietf.utils.admin import SaferTabularInline + class NameAdmin(admin.ModelAdmin): list_display = ["slug", "name", "desc", "used", "order"] search_fields = ["slug", "name"] - prepopulate_from = { "slug": ("name",) } + prepopulate_from = {"slug": ("name",)} + class DocRelationshipNameAdmin(NameAdmin): list_display = ["slug", "name", "revname", "desc", "used"] + + admin.site.register(DocRelationshipName, DocRelationshipNameAdmin) - + + class DocTypeNameAdmin(NameAdmin): list_display = ["slug", "name", "prefix", "desc", "used"] + + admin.site.register(DocTypeName, DocTypeNameAdmin) + class GroupTypeNameAdmin(NameAdmin): list_display = ["slug", "name", "verbose_name", "desc", "used"] + + admin.site.register(GroupTypeName, GroupTypeNameAdmin) -class CountryAliasInline(admin.TabularInline): + +class CountryAliasInline(SaferTabularInline): model = CountryAlias extra = 1 + class CountryNameAdmin(NameAdmin): list_display = ["slug", "name", "continent", "in_eu"] list_filter = ["continent", "in_eu"] inlines = [CountryAliasInline] + + admin.site.register(CountryName, CountryNameAdmin) + class ImportantDateNameAdmin(NameAdmin): list_display = ["slug", "name", "desc", "used", "default_offset_days"] - ordering = ('-used','default_offset_days',) -admin.site.register(ImportantDateName,ImportantDateNameAdmin) + ordering = ( + "-used", + "default_offset_days", + ) + + +admin.site.register(ImportantDateName, ImportantDateNameAdmin) + class ExtResourceNameAdmin(NameAdmin): - list_display = ["slug", "name", "type", "desc", "used",] -admin.site.register(ExtResourceName,ExtResourceNameAdmin) + list_display = [ + "slug", + "name", + "type", + "desc", + "used", + ] + + +admin.site.register(ExtResourceName, ExtResourceNameAdmin) + class ProceedingsMaterialTypeNameAdmin(NameAdmin): - list_display = ["slug", "name", "desc", "used", "order",] + list_display = [ + "slug", + "name", + "desc", + "used", + "order", + ] + + admin.site.register(ProceedingsMaterialTypeName, ProceedingsMaterialTypeNameAdmin) admin.site.register(AgendaFilterTypeName, NameAdmin) admin.site.register(AgendaTypeName, NameAdmin) +admin.site.register(AppealArtifactTypeName, NameAdmin) +admin.site.register(AttendanceTypeName, NameAdmin) admin.site.register(BallotPositionName, NameAdmin) admin.site.register(ConstraintName, NameAdmin) admin.site.register(ContinentName, NameAdmin) @@ -80,6 +162,7 @@ class ProceedingsMaterialTypeNameAdmin(NameAdmin): admin.site.register(LiaisonStatementTagName, NameAdmin) admin.site.register(MeetingTypeName, NameAdmin) admin.site.register(NomineePositionStateName, NameAdmin) +admin.site.register(RegistrationTicketTypeName, NameAdmin) admin.site.register(ReviewRequestStateName, NameAdmin) admin.site.register(ReviewAssignmentStateName, NameAdmin) admin.site.register(ReviewResultName, NameAdmin) @@ -97,3 +180,4 @@ class ProceedingsMaterialTypeNameAdmin(NameAdmin): admin.site.register(ExtResourceTypeName, NameAdmin) admin.site.register(SlideSubmissionStatusName, NameAdmin) admin.site.register(SessionPurposeName, NameAdmin) +admin.site.register(TelechatAgendaSectionName, NameAdmin) diff --git a/ietf/name/factories.py b/ietf/name/factories.py new file mode 100644 index 0000000000..73399dcbb4 --- /dev/null +++ b/ietf/name/factories.py @@ -0,0 +1,13 @@ +# Copyright The IETF Trust 2023, All Rights Reserved +# -*- coding: utf-8 -*- + +import factory + +from .models import ( + AppealArtifactTypeName, +) + +class AppealArtifactTypeNameFactory(factory.django.DjangoModelFactory): + class Meta: + model = AppealArtifactTypeName + django_get_or_create = ("slug",) diff --git a/ietf/name/fixtures/names.json b/ietf/name/fixtures/names.json index 2f0523cea6..64e26e503a 100644 --- a/ietf/name/fixtures/names.json +++ b/ietf/name/fixtures/names.json @@ -147,6 +147,23 @@ "model": "doc.ballottype", "pk": 7 }, + { + "fields": { + "doc_type": "draft", + "name": "RSAB Approve", + "order": 0, + "positions": [ + "concern", + "yes", + "recuse" + ], + "question": "Is this draft ready for publication in the Editorial stream?", + "slug": "rsab-approve", + "used": true + }, + "model": "doc.ballottype", + "pk": 8 + }, { "fields": { "desc": "", @@ -295,7 +312,7 @@ "order": 42, "slug": "watching", "type": "draft-iesg", - "used": true + "used": false }, "model": "doc.state", "pk": 11 @@ -633,7 +650,7 @@ }, { "fields": { - "desc": "4.2.1. Call for Adoption by WG Issued\r\n\r\n The \"Call for Adoption by WG Issued\" state should be used to indicate when an I-D is being considered for adoption by an IETF WG. An I-D that is in this state is actively being considered for adoption and has not yet achieved consensus, preference, or selection in the WG.\r\n\r\n This state may be used to describe an I-D that someone has asked a WG to consider for adoption, if the WG Chair has agreed with the request. This state may also be used to identify an I-D that a WG Chair asked an author to write specifically for consideration as a candidate WG item [WGDTSPEC], and/or an I-D that is listed as a 'candidate draft' in the WG's charter.\r\n\r\n Under normal conditions, it should not be possible for an I-D to be in the \"Call for Adoption by WG Issued\" state in more than one working group at the same time. This said, it is not uncommon for authors to \"shop\" their I-Ds to more than one WG at a time, with the hope of getting their documents adopted somewhere.\r\n\r\n After this state is implemented in the Datatracker, an I-D that is in the \"Call for Adoption by WG Issued\" state will not be able to be \"shopped\" to any other WG without the consent of the WG Chairs and the responsible ADs impacted by the shopping.\r\n\r\n Note that Figure 1 includes an arc leading from this state to outside of the WG state machine. This illustrates that some I-Ds that are considered do not get adopted as WG drafts. An I-D that is not adopted as a WG draft will transition out of the WG state machine and revert back to having no stream-specific state; however, the status change history log of the I-D will record that the I-D was previously in the \"Call for Adoption by WG Issued\" state.", + "desc": "A call for adoption of the individual submission document has been issued by the Working Group (WG) chairs. This call is still running but the WG has not yet reached consensus for adoption.", "name": "Call For Adoption By WG Issued", "next_states": [ 36, @@ -649,7 +666,7 @@ }, { "fields": { - "desc": "4.2.2. Adopted by a WG\r\n\r\n The \"Adopted by a WG\" state describes an individual submission I-D that an IETF WG has agreed to adopt as one of its WG drafts.\r\n\r\n WG Chairs who use this state will be able to clearly indicate when their WGs adopt individual submission I-Ds. This will facilitate the Datatracker's ability to correctly capture \"Replaces\" information for WG drafts and correct \"Replaced by\" information for individual submission I-Ds that have been replaced by WG drafts.\r\n\r\n This state is needed because the Datatracker uses the filename of an I-D as a key to search its database for status information about the I-D, and because the filename of a WG I-D is supposed to be different from the filename of an individual submission I-D. The filename of an individual submission I-D will typically be formatted as 'draft-author-wgname-topic-nn'.\r\n\r\n The filename of a WG document is supposed to be formatted as 'draft- ietf-wgname-topic-nn'.\r\n\r\n An individual I-D that is adopted by a WG may take weeks or months to be resubmitted by the author as a new (version-00) WG draft. If the \"Adopted by a WG\" state is not used, the Datatracker has no way to determine that an I-D has been adopted until a new version of the I-D is submitted to the WG by the author and until the I-D is approved for posting by a WG Chair.", + "desc": "The individual submission document has been adopted by the Working Group (WG), but a WG document replacing this document with the typical naming convention of 'draft- ietf-wgname-topic-nn' has not yet been submitted.", "name": "Adopted by a WG", "next_states": [ 38 @@ -664,7 +681,7 @@ }, { "fields": { - "desc": "4.2.3. Adopted for WG Info Only\r\n\r\n The \"Adopted for WG Info Only\" state describes a document that contains useful information for the WG that adopted it, but the document is not intended to be published as an RFC. The WG will not actively develop the contents of the I-D or progress it for publication as an RFC. The only purpose of the I-D is to provide information for internal use by the WG.", + "desc": "The document is adopted by the Working Group (WG) for its internal use. The WG has decided that it will not pursue publication of it as an RFC.", "name": "Adopted for WG Info Only", "next_states": [], "order": 3, @@ -677,7 +694,7 @@ }, { "fields": { - "desc": "4.2.4. WG Document\r\n\r\n The \"WG Document\" state describes an I-D that has been adopted by an IETF WG and is being actively developed.\r\n\r\n A WG Chair may transition an I-D into the \"WG Document\" state at any time as long as the I-D is not being considered or developed in any other WG.\r\n\r\n Alternatively, WG Chairs may rely upon new functionality to be added to the Datatracker to automatically move version-00 drafts into the \"WG Document\" state as described in Section 4.1.\r\n\r\n Under normal conditions, it should not be possible for an I-D to be in the \"WG Document\" state in more than one WG at a time. This said, I-Ds may be transferred from one WG to another with the consent of the WG Chairs and the responsible ADs.", + "desc": "The document has been adopted by the Working Group (WG) and is under development. A document can only be adopted by one WG at a time. However, a document may be transferred between WGs.", "name": "WG Document", "next_states": [ 39, @@ -695,7 +712,7 @@ }, { "fields": { - "desc": "4.2.5. Parked WG Document\r\n\r\n A \"Parked WG Document\" is an I-D that has lost its author or editor, is waiting for another document to be written or for a review to be completed, or cannot be progressed by the working group for some other reason.\r\n\r\n Some of the annotation tags described in Section 4.3 may be used in conjunction with this state to indicate why an I-D has been parked, and/or what may need to happen for the I-D to be un-parked.\r\n\r\n Parking a WG draft will not prevent it from expiring; however, this state can be used to indicate why the I-D has stopped progressing in the WG.\r\n\r\n A \"Parked WG Document\" that is not expired may be transferred from one WG to another with the consent of the WG Chairs and the responsible ADs.", + "desc": "The Working Group (WG) document is in a temporary state where it will not be actively developed. The reason for the pause is explained via a datatracker comments section.", "name": "Parked WG Document", "next_states": [ 38 @@ -710,7 +727,7 @@ }, { "fields": { - "desc": "4.2.6. Dead WG Document\r\n\r\n A \"Dead WG Document\" is an I-D that has been abandoned. Note that 'Dead' is not always a final state for a WG I-D. If consensus is subsequently achieved, a \"Dead WG Document\" may be resurrected. A \"Dead WG Document\" that is not resurrected will eventually expire.\r\n\r\n Note that an I-D that is declared to be \"Dead\" in one WG and that is not expired may be transferred to a non-dead state in another WG with the consent of the WG Chairs and the responsible ADs.", + "desc": "The Working Group (WG) document has been abandoned by the WG. No further development is planned in this WG. A decision to resume work on this document and move it out of this state is possible.", "name": "Dead WG Document", "next_states": [ 38 @@ -725,7 +742,7 @@ }, { "fields": { - "desc": "4.2.7. In WG Last Call\r\n\r\n A document \"In WG Last Call\" is an I-D for which a WG Last Call (WGLC) has been issued and is in progress.\r\n\r\n Note that conducting a WGLC is an optional part of the IETF WG process, per Section 7.4 of RFC 2418 [RFC2418].\r\n\r\n If a WG Chair decides to conduct a WGLC on an I-D, the \"In WG Last Call\" state can be used to track the progress of the WGLC. The Chair may configure the Datatracker to send a WGLC message to one or more mailing lists when the Chair moves the I-D into this state. The WG Chair may also be able to select a different set of mailing lists for a different document undergoing a WGLC; some documents may deserve coordination with other WGs.\r\n\r\n A WG I-D in this state should remain \"In WG Last Call\" until the WG Chair moves it to another state. The WG Chair may configure the Datatracker to send an e-mail after a specified period of time to remind or 'nudge' the Chair to conclude the WGLC and to determine the next state for the document.\r\n\r\n It is possible for one WGLC to lead into another WGLC for the same document. For example, an I-D that completed a WGLC as an \"Informational\" document may need another WGLC if a decision is taken to convert the I-D into a Standards Track document.", + "desc": "The Working Group (WG) document is currently subject to an active WG Last Call (WGLC) review per Section 7.4 of RFC2418.", "name": "In WG Last Call", "next_states": [ 38, @@ -742,7 +759,7 @@ }, { "fields": { - "desc": "4.2.8. Waiting for WG Chair Go-Ahead\r\n\r\n A WG Chair may wish to place an I-D that receives a lot of comments during a WGLC into the \"Waiting for WG Chair Go-Ahead\" state. This state describes an I-D that has undergone a WGLC; however, the Chair is not yet ready to call consensus on the document.\r\n\r\n If comments from the WGLC need to be responded to, or a revision to the I-D is needed, the Chair may place an I-D into this state until all of the WGLC comments are adequately addressed and the (possibly revised) document is in the I-D repository.", + "desc": "The Working Group (WG) document has completed Working Group Last Call (WGLC), but the WG chair(s) are not yet ready to call consensus on the document. The reasons for this may include comments from the WGLC need to be responded to, or a revision to the document is needed", "name": "Waiting for WG Chair Go-Ahead", "next_states": [ 41, @@ -758,7 +775,7 @@ }, { "fields": { - "desc": "4.2.9. WG Consensus: Waiting for Writeup\r\n\r\n A document in the \"WG Consensus: Waiting for Writeup\" state has essentially completed its development within the working group, and is nearly ready to be sent to the IESG for publication. The last thing to be done is the preparation of a protocol writeup by a Document Shepherd. The IESG requires that a document shepherd writeup be completed before publication of the I-D is requested. The IETF document shepherding process and the role of a WG Document Shepherd is described in RFC 4858 [RFC4858]\r\n\r\n A WG Chair may call consensus on an I-D without a formal WGLC and transition an I-D that was in the \"WG Document\" state directly into this state.\r\n\r\n The name of this state includes the words \"Waiting for Writeup\" because a good document shepherd writeup takes time to prepare.", + "desc": "The Working Group (WG) document has consensus to proceed to publication. However, the document is waiting for a document shepherd write-up per RFC4858.", "name": "WG Consensus: Waiting for Write-Up", "next_states": [ 44 @@ -773,7 +790,7 @@ }, { "fields": { - "desc": "4.2.10. Submitted to IESG for Publication\r\n\r\n This state describes a WG document that has been submitted to the IESG for publication and that has not been sent back to the working group for revision.\r\n\r\n An I-D in this state may be under review by the IESG, it may have been approved and be in the RFC Editor's queue, or it may have been published as an RFC. Other possibilities exist too. The document may be \"Dead\" (in the IESG state machine) or in a \"Do Not Publish\" state.", + "desc": "The Working Group (WG) document has left the WG and been submitted to the Internet Engineering Steering Group (IESG) for evaluation and publication. See the “IESG State” or “RFC Editor State” for further details on the state of the document.", "name": "Submitted to IESG for Publication", "next_states": [ 38 @@ -2003,7 +2020,7 @@ }, { "fields": { - "desc": "The document has been marked as a candidate for WG adoption by the WG Chair. This state can be used before a call for adoption is issued (and the document is put in the \"Call For Adoption By WG Issued\" state), to indicate that the document is in the queue for a call for adoption, even if none has been issued yet.", + "desc": "The individual submission document has been marked by the Working Group (WG) chairs as a candidate for adoption by the WG, but no adoption call has been started.", "name": "Candidate for WG Adoption", "next_states": [ 35 @@ -2135,7 +2152,7 @@ }, { "fields": { - "desc": "In some areas, it can be desirable to wait for multiple interoperable implementations before progressing a draft to be an RFC, and in some WGs this is required. This state should be entered after WG Last Call has completed.", + "desc": "The progression of this Working Group (WG) document towards publication is paused as it awaits implementation. The process governing the approach to implementations is WG-specific.", "name": "Waiting for Implementation", "next_states": [], "order": 8, @@ -2148,7 +2165,7 @@ }, { "fields": { - "desc": "Held by WG, see document history for details.", + "desc": "Held by Working Group (WG) chairs for administrative reasons. See document history for details.", "name": "Held by WG", "next_states": [], "order": 9, @@ -2457,6 +2474,188 @@ "model": "doc.state", "pk": 168 }, + { + "fields": { + "desc": "", + "name": "Replaced editorial stream document", + "next_states": [], + "order": 0, + "slug": "repl", + "type": "draft-stream-editorial", + "used": true + }, + "model": "doc.state", + "pk": 170 + }, + { + "fields": { + "desc": "", + "name": "Active editorial stream document", + "next_states": [], + "order": 2, + "slug": "active", + "type": "draft-stream-editorial", + "used": true + }, + "model": "doc.state", + "pk": 171 + }, + { + "fields": { + "desc": "", + "name": "Editorial stream document under RSAB review", + "next_states": [], + "order": 3, + "slug": "rsabpoll", + "type": "draft-stream-editorial", + "used": true + }, + "model": "doc.state", + "pk": 172 + }, + { + "fields": { + "desc": "", + "name": "Published RFC", + "next_states": [], + "order": 4, + "slug": "pub", + "type": "draft-stream-editorial", + "used": true + }, + "model": "doc.state", + "pk": 173 + }, + { + "fields": { + "desc": "", + "name": "Dead editorial stream document", + "next_states": [], + "order": 5, + "slug": "dead", + "type": "draft-stream-editorial", + "used": true + }, + "model": "doc.state", + "pk": 174 + }, + { + "fields": { + "desc": "The statement is active", + "name": "Active", + "next_states": [], + "order": 0, + "slug": "active", + "type": "statement", + "used": true + }, + "model": "doc.state", + "pk": 175 + }, + { + "fields": { + "desc": "The statement has been replaced", + "name": "Replaced", + "next_states": [], + "order": 0, + "slug": "replaced", + "type": "statement", + "used": true + }, + "model": "doc.state", + "pk": 176 + }, + { + "fields": { + "desc": "", + "name": "Published", + "next_states": [], + "order": 1, + "slug": "published", + "type": "rfc", + "used": true + }, + "model": "doc.state", + "pk": 177 + }, + { + "fields": { + "desc": "", + "name": "Active", + "next_states": [], + "order": 0, + "slug": "active", + "type": "narrativeminutes", + "used": true + }, + "model": "doc.state", + "pk": 178 + }, + { + "fields": { + "desc": "", + "name": "Deleted", + "next_states": [], + "order": 1, + "slug": "deleted", + "type": "narrativeminutes", + "used": true + }, + "model": "doc.state", + "pk": 179 + }, + { + "fields": { + "desc": "The editorial stream processing of this document is complete and it has been sent to the RFC Editor for publication. The document may be in the RFC Editor's queue, or it may have been published as an RFC; this state doesn't distinguish between different states occurring after the document has left the RSAB.", + "name": "Sent to the RFC Editor", + "next_states": [], + "order": 10, + "slug": "rfc-edit", + "type": "draft-stream-editorial", + "used": true + }, + "model": "doc.state", + "pk": 180 + }, + { + "fields": { + "desc": "The BOF request is spam", + "name": "Spam", + "next_states": [], + "order": 5, + "slug": "spam", + "type": "bofreq", + "used": true + }, + "model": "doc.state", + "pk": 181 + }, + { + "fields": { + "desc": "The statement has been marked historic", + "name": "Historic", + "next_states": [], + "order": 0, + "slug": "historic", + "type": "statement", + "used": false + }, + "model": "doc.state", + "pk": 182 + }, + { + "fields": { + "desc": "The statement is no longer active", + "name": "Inactive", + "next_states": [], + "order": 0, + "slug": "inactive", + "type": "statement", + "used": true + }, + "model": "doc.state", + "pk": 183 + }, { "fields": { "label": "State" @@ -2464,6 +2663,13 @@ "model": "doc.statetype", "pk": "agenda" }, + { + "fields": { + "label": "bcp state" + }, + "model": "doc.statetype", + "pk": "bcp" + }, { "fields": { "label": "State" @@ -2506,13 +2712,6 @@ "model": "doc.statetype", "pk": "draft" }, - { - "fields": { - "label": "IANA state" - }, - "model": "doc.statetype", - "pk": "draft-iana" - }, { "fields": { "label": "IANA Action state" @@ -2548,6 +2747,13 @@ "model": "doc.statetype", "pk": "draft-rfceditor" }, + { + "fields": { + "label": "Editorial stream state" + }, + "model": "doc.statetype", + "pk": "draft-stream-editorial" + }, { "fields": { "label": "IAB state" @@ -2576,6 +2782,13 @@ "model": "doc.statetype", "pk": "draft-stream-ise" }, + { + "fields": { + "label": "fyi state" + }, + "model": "doc.statetype", + "pk": "fyi" + }, { "fields": { "label": "State" @@ -2597,6 +2810,13 @@ "model": "doc.statetype", "pk": "minutes" }, + { + "fields": { + "label": "State" + }, + "model": "doc.statetype", + "pk": "narrativeminutes" + }, { "fields": { "label": "State" @@ -2632,6 +2852,13 @@ "model": "doc.statetype", "pk": "review" }, + { + "fields": { + "label": "State" + }, + "model": "doc.statetype", + "pk": "rfc" + }, { "fields": { "label": "Shepherd's Writeup State" @@ -2653,11 +2880,27 @@ "model": "doc.statetype", "pk": "statchg" }, + { + "fields": { + "label": "Statement State" + }, + "model": "doc.statetype", + "pk": "statement" + }, + { + "fields": { + "label": "std state" + }, + "model": "doc.statetype", + "pk": "std" + }, { "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"chair\"\n]", + "admin_roles": [ + "chair" + ], "agenda_filter_type": "special", "agenda_type": "ietf", "create_wiki": true, @@ -2665,10 +2908,24 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"matman\",\n \"ad\",\n \"chair\",\n \"lead\",\n \"delegate\"\n]", - "docman_roles": "[\n \"chair\"\n]", - "groupman_authroles": "[\n \"Secretariat\"\n]", - "groupman_roles": "[\n \"chair\",\n \"lead\",\n \"delegate\"\n]", + "default_used_roles": [ + "matman", + "ad", + "chair", + "lead", + "delegate" + ], + "docman_roles": [ + "chair" + ], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [ + "chair", + "lead", + "delegate" + ], "has_chartering_process": false, "has_default_chat": true, "has_documents": false, @@ -2678,15 +2935,29 @@ "has_reviews": false, "has_session_materials": true, "is_schedulable": true, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"chair\",\n \"lead\",\n \"delegate\",\n \"matman\"\n]", + "material_types": [ + "slides" + ], + "matman_roles": [ + "chair", + "lead", + "delegate", + "matman" + ], "need_parent": false, "parent_types": [ "ietf" ], "req_subm_approval": true, - "role_order": "[\n \"chair\",\n \"lead\",\n \"delegate\",\n \"matman\"\n]", - "session_purposes": "[\n \"presentation\"\n]", + "role_order": [ + "chair", + "lead", + "delegate", + "matman" + ], + "session_purposes": [ + "presentation" + ], "show_on_agenda": true }, "model": "group.groupfeatures", @@ -2696,35 +2967,55 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"chair\"\n]", - "agenda_filter_type": "none", + "admin_roles": [ + "chair" + ], + "agenda_filter_type": "heading", "agenda_type": "ietf", "create_wiki": false, "custom_group_roles": false, "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"member\",\n \"chair\"\n]", - "docman_roles": "[\n \"chair\"\n]", - "groupman_authroles": "[\n \"Secretariat\"\n]", - "groupman_roles": "[\n \"chair\"\n]", + "default_used_roles": [ + "member", + "chair" + ], + "docman_roles": [ + "chair" + ], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [ + "chair" + ], "has_chartering_process": false, "has_default_chat": false, "has_documents": false, - "has_meetings": false, + "has_meetings": true, "has_milestones": false, "has_nonsession_materials": false, "has_reviews": false, "has_session_materials": false, - "is_schedulable": false, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"chair\"\n]", + "is_schedulable": true, + "material_types": [ + "slides" + ], + "matman_roles": [ + "chair" + ], "need_parent": false, "parent_types": [], "req_subm_approval": false, - "role_order": "[\n \"chair\"\n]", - "session_purposes": "[\n \"closed_meeting\",\n \"officehours\"\n]", - "show_on_agenda": false + "role_order": [ + "chair" + ], + "session_purposes": [ + "closed_meeting", + "officehours" + ], + "show_on_agenda": true }, "model": "group.groupfeatures", "pk": "adm" @@ -2733,7 +3024,9 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": true, - "admin_roles": "[\n \"chair\"\n]", + "admin_roles": [ + "chair" + ], "agenda_filter_type": "normal", "agenda_type": "ietf", "create_wiki": true, @@ -2741,10 +3034,26 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"ad\",\n \"chair\",\n \"secr\",\n \"delegate\"\n]", - "docman_roles": "[\n \"chair\",\n \"delegate\",\n \"secr\"\n]", - "groupman_authroles": "[\n \"Secretariat\",\n \"Area Director\"\n]", - "groupman_roles": "[\n \"ad\",\n \"chair\",\n \"delegate\"\n]", + "default_used_roles": [ + "ad", + "chair", + "secr", + "delegate" + ], + "docman_roles": [ + "chair", + "delegate", + "secr" + ], + "groupman_authroles": [ + "Secretariat", + "Area Director" + ], + "groupman_roles": [ + "ad", + "chair", + "delegate" + ], "has_chartering_process": false, "has_default_chat": false, "has_documents": true, @@ -2754,16 +3063,28 @@ "has_reviews": false, "has_session_materials": true, "is_schedulable": true, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"ad\",\n \"chair\",\n \"delegate\",\n \"secr\"\n]", + "material_types": [ + "slides" + ], + "matman_roles": [ + "ad", + "chair", + "delegate", + "secr" + ], "need_parent": false, "parent_types": [ "area", "ietf" ], "req_subm_approval": true, - "role_order": "[\n \"chair\",\n \"secr\"\n]", - "session_purposes": "[\n \"regular\"\n]", + "role_order": [ + "chair", + "secr" + ], + "session_purposes": [ + "regular" + ], "show_on_agenda": true }, "model": "group.groupfeatures", @@ -2773,7 +3094,9 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"ad\"\n]", + "admin_roles": [ + "ad" + ], "agenda_filter_type": "heading", "agenda_type": "ietf", "create_wiki": true, @@ -2781,10 +3104,22 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"ad\",\n \"liaison_contact\",\n \"liaison_cc_contact\"\n]", - "docman_roles": "[\n \"ad\",\n \"delegate\",\n \"secr\"\n]", - "groupman_authroles": "[\n \"Secretariat\"\n]", - "groupman_roles": "[\n \"ad\"\n]", + "default_used_roles": [ + "ad", + "liaison_contact", + "liaison_cc_contact" + ], + "docman_roles": [ + "ad", + "delegate", + "secr" + ], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [ + "ad" + ], "has_chartering_process": false, "has_default_chat": false, "has_documents": false, @@ -2794,15 +3129,27 @@ "has_reviews": false, "has_session_materials": false, "is_schedulable": false, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"ad\",\n \"chair\",\n \"delegate\",\n \"secr\"\n]", + "material_types": [ + "slides" + ], + "matman_roles": [ + "ad", + "chair", + "delegate", + "secr" + ], "need_parent": true, "parent_types": [ "ietf" ], "req_subm_approval": true, - "role_order": "[\n \"chair\",\n \"secr\"\n]", - "session_purposes": "[\n \"regular\"\n]", + "role_order": [ + "chair", + "secr" + ], + "session_purposes": [ + "regular" + ], "show_on_agenda": false }, "model": "group.groupfeatures", @@ -2812,7 +3159,10 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"chair\",\n \"secr\"\n]", + "admin_roles": [ + "chair", + "secr" + ], "agenda_filter_type": "special", "agenda_type": "ad", "create_wiki": true, @@ -2820,11 +3170,26 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"ad\",\n \"chair\",\n \"reviewer\",\n \"secr\",\n \"delegate\"\n]", - "docman_roles": "[\n \"chair\"\n]", - "groupman_authroles": "[\n \"Secretariat\"\n]", - "groupman_roles": "[\n \"ad\",\n \"secr\",\n \"delegate\",\n \"chair\"\n]", - "has_chartering_process": false, + "default_used_roles": [ + "ad", + "chair", + "reviewer", + "secr", + "delegate" + ], + "docman_roles": [ + "chair" + ], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [ + "ad", + "secr", + "delegate", + "chair" + ], + "has_chartering_process": false, "has_default_chat": false, "has_documents": false, "has_meetings": true, @@ -2833,15 +3198,31 @@ "has_reviews": false, "has_session_materials": true, "is_schedulable": false, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"ad\",\n \"chair\",\n \"delegate\",\n \"secr\"\n]", + "material_types": [ + "slides" + ], + "matman_roles": [ + "ad", + "chair", + "delegate", + "secr" + ], "need_parent": true, "parent_types": [ "area" ], "req_subm_approval": true, - "role_order": "[\n \"chair\",\n \"secr\"\n]", - "session_purposes": "[\n \"open_meeting\",\n \"presentation\",\n \"regular\",\n \"social\",\n \"tutorial\"\n]", + "role_order": [ + "chair", + "secr" + ], + "session_purposes": [ + "open_meeting", + "presentation", + "regular", + "social", + "tutorial" + ], "show_on_agenda": false }, "model": "group.groupfeatures", @@ -2851,44 +3232,124 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"chair\"\n]", - "agenda_filter_type": "none", - "agenda_type": "side", + "admin_roles": [ + "chair" + ], + "agenda_filter_type": "normal", + "agenda_type": "ietf", "create_wiki": false, - "custom_group_roles": true, + "custom_group_roles": false, "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"auth\",\n \"chair\"\n]", - "docman_roles": "[]", - "groupman_authroles": "[\n \"Secretariat\"\n]", - "groupman_roles": "[]", + "default_used_roles": [ + "chair", + "member" + ], + "docman_roles": [ + "chair" + ], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [ + "chair" + ], "has_chartering_process": false, - "has_default_chat": false, + "has_default_chat": true, "has_documents": false, - "has_meetings": false, + "has_meetings": true, "has_milestones": false, "has_nonsession_materials": false, "has_reviews": false, - "has_session_materials": false, - "is_schedulable": false, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[]", + "has_session_materials": true, + "is_schedulable": true, + "material_types": [ + "slides" + ], + "matman_roles": [ + "chair" + ], "need_parent": false, "parent_types": [], + "req_subm_approval": false, + "role_order": [ + "chair", + "member" + ], + "session_purposes": [ + "officehours", + "regular" + ], + "show_on_agenda": true + }, + "model": "group.groupfeatures", + "pk": "edappr" + }, + { + "fields": { + "about_page": "ietf.group.views.group_about", + "acts_like_wg": true, + "admin_roles": [ + "chair" + ], + "agenda_filter_type": "normal", + "agenda_type": "ietf", + "create_wiki": false, + "custom_group_roles": false, + "customize_workflow": true, + "default_parent": "", + "default_tab": "ietf.group.views.group_documents", + "default_used_roles": [ + "chair" + ], + "docman_roles": [ + "chair" + ], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [ + "chair" + ], + "has_chartering_process": false, + "has_default_chat": true, + "has_documents": true, + "has_meetings": true, + "has_milestones": false, + "has_nonsession_materials": false, + "has_reviews": false, + "has_session_materials": true, + "is_schedulable": true, + "material_types": [ + "slides" + ], + "matman_roles": [ + "chair" + ], + "need_parent": false, + "parent_types": [ + "rfcedtyp" + ], "req_subm_approval": true, - "role_order": "[\n \"chair\",\n \"secr\"\n]", - "session_purposes": "[\n \"officehours\"\n]", - "show_on_agenda": false + "role_order": [ + "chair" + ], + "session_purposes": [ + "regular" + ], + "show_on_agenda": true }, "model": "group.groupfeatures", - "pk": "editorial" + "pk": "edwg" }, { "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"chair\"\n]", + "admin_roles": [ + "chair" + ], "agenda_filter_type": "normal", "agenda_type": "ietf", "create_wiki": false, @@ -2896,10 +3357,16 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"chair\"\n]", - "docman_roles": "[\n \"chair\"\n]", - "groupman_authroles": "[\n \"Secretariat\"\n]", - "groupman_roles": "[]", + "default_used_roles": [ + "chair" + ], + "docman_roles": [ + "chair" + ], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [], "has_chartering_process": false, "has_default_chat": false, "has_documents": false, @@ -2909,15 +3376,26 @@ "has_reviews": false, "has_session_materials": false, "is_schedulable": false, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"chair\",\n \"delegate\"\n]", + "material_types": [ + "slides" + ], + "matman_roles": [ + "chair", + "delegate" + ], "need_parent": false, "parent_types": [ "ietf" ], "req_subm_approval": true, - "role_order": "[\n \"chair\",\n \"secr\"\n]", - "session_purposes": "[\n \"closed_meeting\",\n \"regular\"\n]", + "role_order": [ + "chair", + "secr" + ], + "session_purposes": [ + "closed_meeting", + "regular" + ], "show_on_agenda": true }, "model": "group.groupfeatures", @@ -2927,7 +3405,9 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"lead\"\n]", + "admin_roles": [ + "lead" + ], "agenda_filter_type": "none", "agenda_type": "ad", "create_wiki": false, @@ -2935,10 +3415,27 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"member\",\n \"chair\",\n \"lead\",\n \"delegate\"\n]", - "docman_roles": "[\n \"lead\",\n \"chair\",\n \"secr\"\n]", - "groupman_authroles": "[\n \"Secretariat\",\n \"IAB\"\n]", - "groupman_roles": "[\n \"lead\",\n \"chair\",\n \"secr\",\n \"delegate\"\n]", + "default_used_roles": [ + "member", + "chair", + "lead", + "delegate" + ], + "docman_roles": [ + "lead", + "chair", + "secr" + ], + "groupman_authroles": [ + "Secretariat", + "IAB" + ], + "groupman_roles": [ + "lead", + "chair", + "secr", + "delegate" + ], "has_chartering_process": false, "has_default_chat": false, "has_documents": true, @@ -2948,15 +3445,29 @@ "has_reviews": false, "has_session_materials": false, "is_schedulable": false, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"lead\",\n \"chair\",\n \"secr\"\n]", + "material_types": [ + "slides" + ], + "matman_roles": [ + "lead", + "chair", + "secr" + ], "need_parent": false, "parent_types": [ "ietf" ], "req_subm_approval": false, - "role_order": "[\n \"lead\",\n \"chair\",\n \"secr\"\n]", - "session_purposes": "[\n \"closed_meeting\",\n \"officehours\",\n \"open_meeting\"\n]", + "role_order": [ + "lead", + "chair", + "secr" + ], + "session_purposes": [ + "closed_meeting", + "officehours", + "open_meeting" + ], "show_on_agenda": false }, "model": "group.groupfeatures", @@ -2966,7 +3477,71 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"chair\"\n]", + "admin_roles": [ + "chair" + ], + "agenda_filter_type": "none", + "agenda_type": "ietf", + "create_wiki": false, + "custom_group_roles": false, + "customize_workflow": false, + "default_parent": "iab", + "default_tab": "ietf.group.views.group_about", + "default_used_roles": [], + "docman_roles": [ + "ad", + "chair", + "delegate", + "secr" + ], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [ + "ad", + "chair" + ], + "has_chartering_process": false, + "has_default_chat": true, + "has_documents": true, + "has_meetings": true, + "has_milestones": false, + "has_nonsession_materials": false, + "has_reviews": false, + "has_session_materials": true, + "is_schedulable": false, + "material_types": [ + "slides" + ], + "matman_roles": [ + "ad", + "chair", + "delegate", + "secr" + ], + "need_parent": true, + "parent_types": [ + "ietf" + ], + "req_subm_approval": false, + "role_order": [ + "chair", + "secr", + "member" + ], + "session_purposes": "[\"regular\"]", + "show_on_agenda": false + }, + "model": "group.groupfeatures", + "pk": "iabworkshop" + }, + { + "fields": { + "about_page": "ietf.group.views.group_about", + "acts_like_wg": false, + "admin_roles": [ + "chair" + ], "agenda_filter_type": "none", "agenda_type": "ietf", "create_wiki": false, @@ -2974,10 +3549,18 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"auth\"\n]", - "docman_roles": "[\n \"chair\"\n]", - "groupman_authroles": "[\n \"Secretariat\"\n]", - "groupman_roles": "[\n \"chair\"\n]", + "default_used_roles": [ + "auth" + ], + "docman_roles": [ + "chair" + ], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [ + "chair" + ], "has_chartering_process": false, "has_default_chat": false, "has_documents": false, @@ -2987,13 +3570,21 @@ "has_reviews": false, "has_session_materials": false, "is_schedulable": false, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"chair\"\n]", + "material_types": [ + "slides" + ], + "matman_roles": [ + "chair" + ], "need_parent": false, "parent_types": [], "req_subm_approval": false, - "role_order": "[\n \"chair\"\n]", - "session_purposes": "[\n \"officehours\"\n]", + "role_order": [ + "chair" + ], + "session_purposes": [ + "officehours" + ], "show_on_agenda": false }, "model": "group.groupfeatures", @@ -3003,7 +3594,9 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"chair\"\n]", + "admin_roles": [ + "chair" + ], "agenda_filter_type": "none", "agenda_type": "ad", "create_wiki": false, @@ -3011,10 +3604,19 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"delegate\"\n]", - "docman_roles": "[\n \"chair\"\n]", - "groupman_authroles": "[\n \"Secretariat\"\n]", - "groupman_roles": "[\n \"chair\",\n \"delegate\"\n]", + "default_used_roles": [ + "delegate" + ], + "docman_roles": [ + "chair" + ], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [ + "chair", + "delegate" + ], "has_chartering_process": false, "has_default_chat": false, "has_documents": false, @@ -3024,13 +3626,24 @@ "has_reviews": false, "has_session_materials": false, "is_schedulable": false, - "material_types": "\"[]\"", - "matman_roles": "[\n \"chair\",\n \"delegate\",\n \"member\"\n]", + "material_types": "[]", + "matman_roles": [ + "chair", + "delegate", + "member" + ], "need_parent": false, "parent_types": [], "req_subm_approval": true, - "role_order": "[\n \"chair\",\n \"delegate\",\n \"member\"\n]", - "session_purposes": "[\n \"closed_meeting\",\n \"open_meeting\"\n]", + "role_order": [ + "chair", + "delegate", + "member" + ], + "session_purposes": [ + "closed_meeting", + "open_meeting" + ], "show_on_agenda": false }, "model": "group.groupfeatures", @@ -3040,7 +3653,10 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"chair\",\n \"lead\"\n]", + "admin_roles": [ + "chair", + "lead" + ], "agenda_filter_type": "heading", "agenda_type": "ietf", "create_wiki": false, @@ -3048,10 +3664,26 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"ad\",\n \"member\",\n \"comdir\",\n \"delegate\",\n \"execdir\",\n \"recman\",\n \"secr\",\n \"trac-editor\",\n \"trac-admin\",\n \"chair\"\n]", - "docman_roles": "[\n \"chair\"\n]", - "groupman_authroles": "[\n \"Secretariat\"\n]", - "groupman_roles": "[\n \"chair\",\n \"delegate\"\n]", + "default_used_roles": [ + "ad", + "member", + "comdir", + "delegate", + "execdir", + "recman", + "secr", + "chair" + ], + "docman_roles": [ + "chair" + ], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [ + "chair", + "delegate" + ], "has_chartering_process": false, "has_default_chat": false, "has_documents": false, @@ -3061,15 +3693,29 @@ "has_reviews": false, "has_session_materials": true, "is_schedulable": false, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"chair\",\n \"delegate\"\n]", + "material_types": [ + "slides" + ], + "matman_roles": [ + "chair", + "delegate" + ], "need_parent": false, "parent_types": [ "ietf" ], "req_subm_approval": true, - "role_order": "[\n \"chair\",\n \"secr\"\n]", - "session_purposes": "[\n \"admin\",\n \"plenary\",\n \"presentation\",\n \"social\",\n \"officehours\"\n]", + "role_order": [ + "chair", + "secr" + ], + "session_purposes": [ + "admin", + "plenary", + "presentation", + "social", + "officehours" + ], "show_on_agenda": false }, "model": "group.groupfeatures", @@ -3079,7 +3725,9 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"chair\"\n]", + "admin_roles": [ + "chair" + ], "agenda_filter_type": "none", "agenda_type": "ad", "create_wiki": false, @@ -3087,10 +3735,16 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"ad\"\n]", - "docman_roles": "[\n \"auth\"\n]", - "groupman_authroles": "[\n \"Secretariat\"\n]", - "groupman_roles": "[]", + "default_used_roles": [ + "ad" + ], + "docman_roles": [ + "auth" + ], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [], "has_chartering_process": false, "has_default_chat": false, "has_documents": false, @@ -3100,15 +3754,20 @@ "has_reviews": false, "has_session_materials": false, "is_schedulable": false, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[]", + "material_types": [ + "slides" + ], + "matman_roles": [], "need_parent": true, "parent_types": [ "area" ], "req_subm_approval": false, - "role_order": "[\n \"chair\",\n \"secr\"\n]", - "session_purposes": "[]", + "role_order": [ + "chair", + "secr" + ], + "session_purposes": [], "show_on_agenda": false }, "model": "group.groupfeatures", @@ -3118,7 +3777,9 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"chair\"\n]", + "admin_roles": [ + "chair" + ], "agenda_filter_type": "heading", "agenda_type": "ietf", "create_wiki": false, @@ -3126,10 +3787,20 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"member\",\n \"atlarge\",\n \"chair\",\n \"delegate\"\n]", - "docman_roles": "[]", - "groupman_authroles": "[\n \"Secretariat\"\n]", - "groupman_roles": "[\n \"chair\",\n \"delegate\"\n]", + "default_used_roles": [ + "member", + "atlarge", + "chair", + "delegate" + ], + "docman_roles": [], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [ + "chair", + "delegate" + ], "has_chartering_process": false, "has_default_chat": false, "has_documents": false, @@ -3139,15 +3810,24 @@ "has_reviews": false, "has_session_materials": false, "is_schedulable": false, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"chair\",\n \"delegate\",\n \"secr\"\n]", + "material_types": [ + "slides" + ], + "matman_roles": [ + "chair", + "delegate", + "secr" + ], "need_parent": false, "parent_types": [ "irtf" ], "req_subm_approval": true, - "role_order": "[\n \"chair\",\n \"secr\"\n]", - "session_purposes": "[]", + "role_order": [ + "chair", + "secr" + ], + "session_purposes": [], "show_on_agenda": false }, "model": "group.groupfeatures", @@ -3157,35 +3837,59 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"chair\",\n \"lead\"\n]", - "agenda_filter_type": "none", - "agenda_type": "ad", + "admin_roles": [ + "chair", + "lead" + ], + "agenda_filter_type": "heading", + "agenda_type": "ietf", "create_wiki": false, "custom_group_roles": true, "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"chair\",\n \"delegate\"\n]", - "docman_roles": "[\n \"chair\"\n]", - "groupman_authroles": "[\n \"Secretariat\"\n]", - "groupman_roles": "[\n \"chair\",\n \"delegate\"\n]", + "default_used_roles": [ + "chair", + "delegate" + ], + "docman_roles": [ + "chair" + ], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [ + "chair", + "delegate" + ], "has_chartering_process": false, "has_default_chat": false, "has_documents": true, - "has_meetings": false, + "has_meetings": true, "has_milestones": false, "has_nonsession_materials": false, "has_reviews": false, "has_session_materials": false, - "is_schedulable": false, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"chair\",\n \"delegate\"\n]", + "is_schedulable": true, + "material_types": [ + "slides" + ], + "matman_roles": [ + "chair", + "delegate" + ], "need_parent": false, "parent_types": [], "req_subm_approval": true, - "role_order": "[\n \"chair\",\n \"delegate\"\n]", - "session_purposes": "[\n \"officehours\"\n]", - "show_on_agenda": false + "role_order": [ + "chair", + "delegate" + ], + "session_purposes": [ + "officehours", + "regular" + ], + "show_on_agenda": true }, "model": "group.groupfeatures", "pk": "ise" @@ -3194,7 +3898,9 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"chair\"\n]", + "admin_roles": [ + "chair" + ], "agenda_filter_type": "none", "agenda_type": null, "create_wiki": false, @@ -3202,10 +3908,17 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"chair\",\n \"ceo\"\n]", - "docman_roles": "[]", - "groupman_authroles": "[\n \"Secretariat\"\n]", - "groupman_roles": "[\n \"chair\"\n]", + "default_used_roles": [ + "chair", + "ceo" + ], + "docman_roles": [], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [ + "chair" + ], "has_chartering_process": false, "has_default_chat": false, "has_documents": false, @@ -3215,15 +3928,27 @@ "has_reviews": false, "has_session_materials": false, "is_schedulable": false, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"chair\",\n \"secr\"\n]", + "material_types": [ + "slides" + ], + "matman_roles": [ + "chair", + "secr" + ], "need_parent": false, "parent_types": [ "isoc" ], "req_subm_approval": true, - "role_order": "[\n \"chair\",\n \"secr\"\n]", - "session_purposes": "[\n \"officehours\",\n \"open_meeting\",\n \"presentation\"\n]", + "role_order": [ + "chair", + "secr" + ], + "session_purposes": [ + "officehours", + "open_meeting", + "presentation" + ], "show_on_agenda": false }, "model": "group.groupfeatures", @@ -3233,7 +3958,10 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"chair\",\n \"advisor\"\n]", + "admin_roles": [ + "chair", + "advisor" + ], "agenda_filter_type": "none", "agenda_type": "side", "create_wiki": true, @@ -3241,10 +3969,23 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"member\",\n \"advisor\",\n \"liaison\",\n \"chair\",\n \"techadv\"\n]", - "docman_roles": "[\n \"chair\"\n]", - "groupman_authroles": "[\n \"Secretariat\"\n]", - "groupman_roles": "[\n \"chair\",\n \"advisor\"\n]", + "default_used_roles": [ + "member", + "advisor", + "liaison", + "chair", + "techadv" + ], + "docman_roles": [ + "chair" + ], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [ + "chair", + "advisor" + ], "has_chartering_process": false, "has_default_chat": false, "has_documents": false, @@ -3254,15 +3995,26 @@ "has_reviews": false, "has_session_materials": false, "is_schedulable": false, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"chair\"\n]", + "material_types": [ + "slides" + ], + "matman_roles": [ + "chair" + ], "need_parent": false, "parent_types": [ "area" ], "req_subm_approval": true, - "role_order": "[\n \"chair\",\n \"member\",\n \"advisor\"\n]", - "session_purposes": "[\n \"closed_meeting\",\n \"officehours\"\n]", + "role_order": [ + "chair", + "member", + "advisor" + ], + "session_purposes": [ + "closed_meeting", + "officehours" + ], "show_on_agenda": false }, "model": "group.groupfeatures", @@ -3272,7 +4024,9 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"lead\"\n]", + "admin_roles": [ + "lead" + ], "agenda_filter_type": "normal", "agenda_type": "ad", "create_wiki": false, @@ -3280,10 +4034,27 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"member\",\n \"chair\",\n \"lead\",\n \"delegate\"\n]", - "docman_roles": "[\n \"lead\",\n \"chair\",\n \"secr\"\n]", - "groupman_authroles": "[\n \"Secretariat\",\n \"IAB\"\n]", - "groupman_roles": "[\n \"lead\",\n \"chair\",\n \"secr\",\n \"delegate\"\n]", + "default_used_roles": [ + "member", + "chair", + "lead", + "delegate" + ], + "docman_roles": [ + "lead", + "chair", + "secr" + ], + "groupman_authroles": [ + "Secretariat", + "IAB" + ], + "groupman_roles": [ + "lead", + "chair", + "secr", + "delegate" + ], "has_chartering_process": false, "has_default_chat": false, "has_documents": true, @@ -3293,15 +4064,28 @@ "has_reviews": false, "has_session_materials": true, "is_schedulable": false, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"lead\",\n \"chair\",\n \"secr\"\n]", + "material_types": [ + "slides" + ], + "matman_roles": [ + "lead", + "chair", + "secr" + ], "need_parent": false, "parent_types": [ "ietf" ], "req_subm_approval": false, - "role_order": "[\n \"lead\",\n \"chair\",\n \"secr\"\n]", - "session_purposes": "[\n \"regular\",\n \"tutorial\"\n]", + "role_order": [ + "lead", + "chair", + "secr" + ], + "session_purposes": [ + "regular", + "tutorial" + ], "show_on_agenda": false }, "model": "group.groupfeatures", @@ -3311,7 +4095,9 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": true, - "admin_roles": "[\n \"chair\"\n]", + "admin_roles": [ + "chair" + ], "agenda_filter_type": "normal", "agenda_type": "ietf", "create_wiki": true, @@ -3319,10 +4105,24 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"chair\",\n \"secr\",\n \"delegate\"\n]", - "docman_roles": "[\n \"chair\",\n \"delegate\",\n \"secr\"\n]", - "groupman_authroles": "[\n \"Secretariat\",\n \"IRTF Chair\"\n]", - "groupman_roles": "[\n \"chair\",\n \"delegate\"\n]", + "default_used_roles": [ + "chair", + "secr", + "delegate" + ], + "docman_roles": [ + "chair", + "delegate", + "secr" + ], + "groupman_authroles": [ + "Secretariat", + "IRTF Chair" + ], + "groupman_roles": [ + "chair", + "delegate" + ], "has_chartering_process": false, "has_default_chat": false, "has_documents": true, @@ -3332,15 +4132,26 @@ "has_reviews": false, "has_session_materials": true, "is_schedulable": true, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"chair\",\n \"delegate\",\n \"secr\"\n]", + "material_types": [ + "slides" + ], + "matman_roles": [ + "chair", + "delegate", + "secr" + ], "need_parent": false, "parent_types": [ "irtf" ], "req_subm_approval": true, - "role_order": "[\n \"chair\",\n \"secr\"\n]", - "session_purposes": "[\n \"regular\"\n]", + "role_order": [ + "chair", + "secr" + ], + "session_purposes": [ + "regular" + ], "show_on_agenda": true }, "model": "group.groupfeatures", @@ -3350,7 +4161,10 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"chair\",\n \"secr\"\n]", + "admin_roles": [ + "chair", + "secr" + ], "agenda_filter_type": "normal", "agenda_type": "ietf", "create_wiki": true, @@ -3358,10 +4172,24 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.review_requests", - "default_used_roles": "[\n \"ad\",\n \"chair\",\n \"reviewer\",\n \"secr\",\n \"delegate\"\n]", - "docman_roles": "[\n \"secr\"\n]", - "groupman_authroles": "[\n \"Secretariat\"\n]", - "groupman_roles": "[\n \"ad\",\n \"secr\",\n \"delegate\"\n]", + "default_used_roles": [ + "ad", + "chair", + "reviewer", + "secr", + "delegate" + ], + "docman_roles": [ + "secr" + ], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [ + "ad", + "secr", + "delegate" + ], "has_chartering_process": false, "has_default_chat": false, "has_documents": false, @@ -3371,15 +4199,26 @@ "has_reviews": true, "has_session_materials": true, "is_schedulable": false, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"ad\",\n \"secr\"\n]", + "material_types": [ + "slides" + ], + "matman_roles": [ + "ad", + "secr" + ], "need_parent": true, "parent_types": [ "area" ], "req_subm_approval": true, - "role_order": "[\n \"chair\",\n \"secr\"\n]", - "session_purposes": "[\n \"open_meeting\",\n \"social\"\n]", + "role_order": [ + "chair", + "secr" + ], + "session_purposes": [ + "open_meeting", + "social" + ], "show_on_agenda": false }, "model": "group.groupfeatures", @@ -3389,7 +4228,9 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"chair\"\n]", + "admin_roles": [ + "chair" + ], "agenda_filter_type": "normal", "agenda_type": "ietf", "create_wiki": false, @@ -3397,10 +4238,19 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"auth\",\n \"chair\"\n]", - "docman_roles": "[]", - "groupman_authroles": "[\n \"Secretariat\"\n]", - "groupman_roles": "[\n \"chair\"\n]", + "default_used_roles": [ + "auth", + "chair" + ], + "docman_roles": [ + "chair" + ], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [ + "chair" + ], "has_chartering_process": false, "has_default_chat": false, "has_documents": false, @@ -3410,13 +4260,23 @@ "has_reviews": false, "has_session_materials": false, "is_schedulable": false, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"chair\"\n]", + "material_types": [ + "slides" + ], + "matman_roles": [ + "chair" + ], "need_parent": false, "parent_types": [], "req_subm_approval": true, - "role_order": "[\n \"chair\",\n \"secr\"\n]", - "session_purposes": "[\n \"officehours\",\n \"regular\"\n]", + "role_order": [ + "chair", + "secr" + ], + "session_purposes": [ + "officehours", + "regular" + ], "show_on_agenda": false }, "model": "group.groupfeatures", @@ -3426,7 +4286,9 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": true, - "admin_roles": "[\n \"chair\"\n]", + "admin_roles": [ + "chair" + ], "agenda_filter_type": "normal", "agenda_type": "ietf", "create_wiki": true, @@ -3434,10 +4296,25 @@ "customize_workflow": true, "default_parent": "irtf", "default_tab": "ietf.group.views.group_documents", - "default_used_roles": "[\n \"chair\",\n \"techadv\",\n \"secr\",\n \"delegate\"\n]", - "docman_roles": "[\n \"chair\",\n \"delegate\",\n \"secr\"\n]", - "groupman_authroles": "[\n \"Secretariat\",\n \"IRTF Chair\"\n]", - "groupman_roles": "[\n \"chair\",\n \"delegate\"\n]", + "default_used_roles": [ + "chair", + "techadv", + "secr", + "delegate" + ], + "docman_roles": [ + "chair", + "delegate", + "secr" + ], + "groupman_authroles": [ + "Secretariat", + "IRTF Chair" + ], + "groupman_roles": [ + "chair", + "delegate" + ], "has_chartering_process": true, "has_default_chat": true, "has_documents": true, @@ -3447,15 +4324,27 @@ "has_reviews": false, "has_session_materials": true, "is_schedulable": true, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"chair\",\n \"delegate\",\n \"secr\"\n]", + "material_types": [ + "slides" + ], + "matman_roles": [ + "chair", + "delegate", + "secr" + ], "need_parent": true, "parent_types": [ "irtf" ], "req_subm_approval": true, - "role_order": "[\n \"chair\",\n \"delegate\",\n \"secr\"\n]", - "session_purposes": "[\n \"regular\"\n]", + "role_order": [ + "chair", + "delegate", + "secr" + ], + "session_purposes": [ + "regular" + ], "show_on_agenda": true }, "model": "group.groupfeatures", @@ -3465,7 +4354,9 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"chair\"\n]", + "admin_roles": [ + "chair" + ], "agenda_filter_type": "none", "agenda_type": null, "create_wiki": false, @@ -3473,10 +4364,23 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"liaiman\",\n \"ceo\",\n \"coord\",\n \"auth\",\n \"chair\",\n \"liaison_contact\",\n \"liaison_cc_contact\"\n]", - "docman_roles": "[\n \"liaiman\",\n \"matman\"\n]", - "groupman_authroles": "[\n \"Secretariat\"\n]", - "groupman_roles": "[]", + "default_used_roles": [ + "liaiman", + "ceo", + "coord", + "auth", + "chair", + "liaison_contact", + "liaison_cc_contact" + ], + "docman_roles": [ + "liaiman", + "matman" + ], + "groupman_authroles": [ + "Secretariat" + ], + "groupman_roles": [], "has_chartering_process": false, "has_default_chat": false, "has_documents": false, @@ -3486,16 +4390,20 @@ "has_reviews": false, "has_session_materials": false, "is_schedulable": false, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[]", + "material_types": [ + "slides" + ], + "matman_roles": [], "need_parent": false, "parent_types": [ "area", "sdo" ], "req_subm_approval": true, - "role_order": "[\n \"liaiman\"\n]", - "session_purposes": "[]", + "role_order": [ + "liaiman" + ], + "session_purposes": [], "show_on_agenda": false }, "model": "group.groupfeatures", @@ -3505,7 +4413,9 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": false, - "admin_roles": "[\n \"chair\"\n]", + "admin_roles": [ + "chair" + ], "agenda_filter_type": "special", "agenda_type": "ietf", "create_wiki": true, @@ -3513,10 +4423,28 @@ "customize_workflow": false, "default_parent": "", "default_tab": "ietf.group.views.group_about", - "default_used_roles": "[\n \"ad\",\n \"member\",\n \"delegate\",\n \"secr\",\n \"liaison\",\n \"atlarge\",\n \"chair\",\n \"matman\",\n \"techadv\"\n]", - "docman_roles": "[\n \"chair\"\n]", - "groupman_authroles": "[\n \"Secretariat\",\n \"Area Director\"\n]", - "groupman_roles": "[\n \"chair\",\n \"delegate\"\n]", + "default_used_roles": [ + "ad", + "member", + "delegate", + "secr", + "liaison", + "atlarge", + "chair", + "matman", + "techadv" + ], + "docman_roles": [ + "chair" + ], + "groupman_authroles": [ + "Secretariat", + "Area Director" + ], + "groupman_roles": [ + "chair", + "delegate" + ], "has_chartering_process": false, "has_default_chat": false, "has_documents": false, @@ -3526,15 +4454,30 @@ "has_reviews": false, "has_session_materials": true, "is_schedulable": true, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"chair\",\n \"matman\"\n]", + "material_types": [ + "slides" + ], + "matman_roles": [ + "chair", + "matman" + ], "need_parent": false, "parent_types": [ "area" ], "req_subm_approval": false, - "role_order": "[\n \"chair\",\n \"member\",\n \"matman\"\n]", - "session_purposes": "[\n \"coding\",\n \"presentation\",\n \"social\",\n \"tutorial\"\n]", + "role_order": [ + "chair", + "member", + "matman" + ], + "session_purposes": [ + "coding", + "open_meeting", + "presentation", + "social", + "tutorial" + ], "show_on_agenda": true }, "model": "group.groupfeatures", @@ -3544,7 +4487,9 @@ "fields": { "about_page": "ietf.group.views.group_about", "acts_like_wg": true, - "admin_roles": "[\n \"chair\"\n]", + "admin_roles": [ + "chair" + ], "agenda_filter_type": "normal", "agenda_type": "ietf", "create_wiki": true, @@ -3552,10 +4497,32 @@ "customize_workflow": true, "default_parent": "", "default_tab": "ietf.group.views.group_documents", - "default_used_roles": "[\n \"ad\",\n \"editor\",\n \"delegate\",\n \"secr\",\n \"chair\",\n \"matman\",\n \"techadv\",\n \"liaison_contact\",\n \"liaison_cc_contact\"\n]", - "docman_roles": "[\n \"chair\",\n \"delegate\",\n \"secr\"\n]", - "groupman_authroles": "[\n \"Secretariat\",\n \"Area Director\"\n]", - "groupman_roles": "[\n \"ad\",\n \"chair\",\n \"delegate\",\n \"secr\"\n]", + "default_used_roles": [ + "ad", + "editor", + "delegate", + "secr", + "chair", + "matman", + "techadv", + "liaison_contact", + "liaison_cc_contact" + ], + "docman_roles": [ + "chair", + "delegate", + "secr" + ], + "groupman_authroles": [ + "Secretariat", + "Area Director" + ], + "groupman_roles": [ + "ad", + "chair", + "delegate", + "secr" + ], "has_chartering_process": true, "has_default_chat": true, "has_documents": true, @@ -3565,15 +4532,28 @@ "has_reviews": false, "has_session_materials": true, "is_schedulable": true, - "material_types": "[\n \"slides\"\n]", - "matman_roles": "[\n \"ad\",\n \"chair\",\n \"delegate\",\n \"secr\"\n]", + "material_types": [ + "slides" + ], + "matman_roles": [ + "ad", + "chair", + "delegate", + "secr" + ], "need_parent": false, "parent_types": [ "area" ], "req_subm_approval": true, - "role_order": "[\n \"chair\",\n \"secr\",\n \"delegate\"\n]", - "session_purposes": "[\n \"regular\"\n]", + "role_order": [ + "chair", + "secr", + "delegate" + ], + "session_purposes": [ + "regular" + ], "show_on_agenda": true }, "model": "group.groupfeatures", @@ -3611,7 +4591,7 @@ ], "desc": "Recipients when a charter is approved", "to": [ - "ietf_announce" + "group_stream_announce" ] }, "model": "mailtrigger.mailtrigger", @@ -3972,8 +4952,8 @@ "doc_ad", "doc_group_chairs", "doc_group_delegates", - "doc_shepherd", - "doc_stream_manager" + "doc_non_ietf_stream_manager", + "doc_shepherd" ] }, "model": "mailtrigger.mailtrigger", @@ -4150,6 +5130,34 @@ "model": "mailtrigger.mailtrigger", "pk": "doc_telechat_details_changed" }, + { + "fields": { + "cc": [], + "desc": "Recipients when a working group call for adoption is issued", + "to": [ + "doc_authors", + "doc_group_chairs", + "doc_group_mail_list", + "doc_shepherd" + ] + }, + "model": "mailtrigger.mailtrigger", + "pk": "doc_wg_call_for_adoption_issued" + }, + { + "fields": { + "cc": [], + "desc": "Recipients when a working group last call is issued", + "to": [ + "doc_authors", + "doc_group_chairs", + "doc_group_mail_list", + "doc_shepherd" + ] + }, + "model": "mailtrigger.mailtrigger", + "pk": "doc_wg_last_call_issued" + }, { "fields": { "cc": [], @@ -4509,30 +5517,47 @@ }, { "fields": { - "cc": [], - "desc": "Recipients for a message requesting an updated list of authorized individuals", + "cc": [], + "desc": "Recipients for a message requesting an updated list of authorized individuals", + "to": [ + "liaison_manager" + ] + }, + "model": "mailtrigger.mailtrigger", + "pk": "liaison_manager_update_request" + }, + { + "fields": { + "cc": [ + "liaison_cc", + "liaison_coordinators", + "liaison_response_contacts", + "liaison_technical_contacts" + ], + "desc": "Recipients for a message when a new incoming liaison statement is posted", "to": [ - "liaison_manager" + "liaison_to_contacts" ] }, "model": "mailtrigger.mailtrigger", - "pk": "liaison_manager_update_request" + "pk": "liaison_statement_posted_incoming" }, { "fields": { "cc": [ "liaison_cc", "liaison_coordinators", + "liaison_from_contact", "liaison_response_contacts", "liaison_technical_contacts" ], - "desc": "Recipient for a message when a new liaison statement is posted", + "desc": "Recipients for a message when a new outgoing liaison statement is posted", "to": [ "liaison_to_contacts" ] }, "model": "mailtrigger.mailtrigger", - "pk": "liaison_statement_posted" + "pk": "liaison_statement_posted_outgoing" }, { "fields": { @@ -4879,6 +5904,50 @@ "model": "mailtrigger.mailtrigger", "pk": "review_completed_genart_telechat" }, + { + "fields": { + "cc": [ + "review_doc_all_parties", + "review_doc_group_mail_list" + ], + "desc": "Recipients when a httpdir Early review is completed", + "to": [ + "review_team_mail_list" + ] + }, + "model": "mailtrigger.mailtrigger", + "pk": "review_completed_httpdir_early" + }, + { + "fields": { + "cc": [ + "ietf_last_call", + "review_doc_all_parties", + "review_doc_group_mail_list" + ], + "desc": "Recipients when a httpdir Last Call review is completed", + "to": [ + "review_team_mail_list" + ] + }, + "model": "mailtrigger.mailtrigger", + "pk": "review_completed_httpdir_lc" + }, + { + "fields": { + "cc": [ + "ietf_last_call", + "review_doc_all_parties", + "review_doc_group_mail_list" + ], + "desc": "Recipients when a httpdir Telechat review is completed", + "to": [ + "review_team_mail_list" + ] + }, + "model": "mailtrigger.mailtrigger", + "pk": "review_completed_httpdir_telechat" + }, { "fields": { "cc": [ @@ -5055,6 +6124,36 @@ "model": "mailtrigger.mailtrigger", "pk": "review_completed_opsdir_telechat" }, + { + "fields": { + "cc": [ + "ietf_last_call", + "review_doc_all_parties", + "review_doc_group_mail_list" + ], + "desc": "Recipients when a perfmetrdir IETF Last Call review is completed", + "to": [ + "review_team_mail_list" + ] + }, + "model": "mailtrigger.mailtrigger", + "pk": "review_completed_perfmetrdir_lc" + }, + { + "fields": { + "cc": [ + "ietf_last_call", + "review_doc_all_parties", + "review_doc_group_mail_list" + ], + "desc": "Recipients when a perfmetrdir Telechat review is completed", + "to": [ + "review_team_mail_list" + ] + }, + "model": "mailtrigger.mailtrigger", + "pk": "review_completed_perfmetrdir_telechat" + }, { "fields": { "cc": [ @@ -5268,6 +6367,46 @@ "model": "mailtrigger.mailtrigger", "pk": "review_req_changed" }, + { + "fields": { + "cc": [ + "doc_affecteddoc_authors", + "doc_affecteddoc_group_chairs", + "doc_affecteddoc_notify", + "doc_authors", + "doc_group_chairs", + "doc_group_mail_list", + "doc_notify", + "doc_shepherd" + ], + "desc": "Recipients when a new RSAB ballot is issued", + "to": [ + "rsab" + ] + }, + "model": "mailtrigger.mailtrigger", + "pk": "rsab_ballot_issued" + }, + { + "fields": { + "cc": [ + "doc_affecteddoc_authors", + "doc_affecteddoc_group_chairs", + "doc_affecteddoc_notify", + "doc_authors", + "doc_group_chairs", + "doc_group_mail_list", + "doc_notify", + "doc_shepherd" + ], + "desc": "Recipients when a new RSAB ballot position with comments is saved", + "to": [ + "rsab" + ] + }, + "model": "mailtrigger.mailtrigger", + "pk": "rsab_ballot_saved" + }, { "fields": { "cc": [ @@ -5362,7 +6501,22 @@ }, { "fields": { - "cc": [], + "cc": [ + "group_chairs" + ], + "desc": "Recipients when slides are approved for a given session", + "to": [ + "slides_proposer" + ] + }, + "model": "mailtrigger.mailtrigger", + "pk": "slides_approved" + }, + { + "fields": { + "cc": [ + "slides_proposer" + ], "desc": "Recipients when slides are proposed for a given session", "to": [ "group_chairs", @@ -5604,7 +6758,7 @@ { "fields": { "desc": "The document's authors", - "template": "{% if doc.type_id == \"draft\" %}<{{doc.name}}@ietf.org>{% endif %}" + "template": "{% if doc.type_id == \"draft\" or doc.type_id == \"rfc\" %}<{{doc.name}}@ietf.org>{% endif %}" }, "model": "mailtrigger.recipient", "pk": "doc_authors" @@ -5859,7 +7013,7 @@ }, { "fields": { - "desc": "The internet drafts ticketing system", + "desc": "The Internet-Draft ticketing system", "template": "" }, "model": "mailtrigger.recipient", @@ -5945,6 +7099,14 @@ "model": "mailtrigger.recipient", "pk": "liaison_coordinators" }, + { + "fields": { + "desc": "Email address of the formal sender of the statement", + "template": "{{liaison.from_contact}}" + }, + "model": "mailtrigger.recipient", + "pk": "liaison_from_contact" + }, { "fields": { "desc": "The assigned liaison manager for an external group ", @@ -6129,6 +7291,14 @@ "model": "mailtrigger.recipient", "pk": "rfc_editor_if_doc_in_queue" }, + { + "fields": { + "desc": "The RFC Series Approval Board", + "template": "The RSAB " + }, + "model": "mailtrigger.recipient", + "pk": "rsab" + }, { "fields": { "desc": "The person that requested a meeting slot for a given group", @@ -6145,6 +7315,14 @@ "model": "mailtrigger.recipient", "pk": "session_requests" }, + { + "fields": { + "desc": "Person who proposed slides", + "template": "{{ proposer.email }}" + }, + "model": "mailtrigger.recipient", + "pk": "slides_proposer" + }, { "fields": { "desc": "The managers of any related streams", @@ -6369,6 +7547,116 @@ "model": "name.agendatypename", "pk": "workshop" }, + { + "fields": { + "desc": "The content of an appeal", + "name": "Appeal", + "order": 1, + "used": true + }, + "model": "name.appealartifacttypename", + "pk": "appeal" + }, + { + "fields": { + "desc": "The content of an appeal combined with the content of a response", + "name": "Response (with appeal included)", + "order": 2, + "used": true + }, + "model": "name.appealartifacttypename", + "pk": "appeal_with_response" + }, + { + "fields": { + "desc": "Other content related to an appeal", + "name": "Other content", + "order": 5, + "used": true + }, + "model": "name.appealartifacttypename", + "pk": "other_content" + }, + { + "fields": { + "desc": "The content of a reply to an appeal response", + "name": "Reply to response", + "order": 4, + "used": true + }, + "model": "name.appealartifacttypename", + "pk": "reply_to_response" + }, + { + "fields": { + "desc": "The content of a response to an appeal", + "name": "Response", + "order": 3, + "used": true + }, + "model": "name.appealartifacttypename", + "pk": "response" + }, + { + "fields": { + "desc": "", + "name": "ANRW Onsite", + "order": 0, + "used": true + }, + "model": "name.attendancetypename", + "pk": "anrw_onsite" + }, + { + "fields": { + "desc": "", + "name": "Hackathon Onsite", + "order": 0, + "used": true + }, + "model": "name.attendancetypename", + "pk": "hackathon_onsite" + }, + { + "fields": { + "desc": "", + "name": "Hackathon Remote", + "order": 0, + "used": true + }, + "model": "name.attendancetypename", + "pk": "hackathon_remote" + }, + { + "fields": { + "desc": "", + "name": "Onsite", + "order": 0, + "used": true + }, + "model": "name.attendancetypename", + "pk": "onsite" + }, + { + "fields": { + "desc": "", + "name": "Remote", + "order": 0, + "used": true + }, + "model": "name.attendancetypename", + "pk": "remote" + }, + { + "fields": { + "desc": "", + "name": "Unknown", + "order": 0, + "used": true + }, + "model": "name.attendancetypename", + "pk": "unknown" + }, { "fields": { "blocking": false, @@ -6391,6 +7679,17 @@ "model": "name.ballotpositionname", "pk": "block" }, + { + "fields": { + "blocking": true, + "desc": "", + "name": "Concern", + "order": 0, + "used": true + }, + "model": "name.ballotpositionname", + "pk": "concern" + }, { "fields": { "blocking": true, @@ -9676,6 +10975,17 @@ "model": "name.dbtemplatetypename", "pk": "rst" }, + { + "fields": { + "desc": "", + "name": "became RFC", + "order": 0, + "revname": "came from draft", + "used": true + }, + "model": "name.docrelationshipname", + "pk": "became_rfc" + }, { "fields": { "desc": "", @@ -9687,6 +10997,17 @@ "model": "name.docrelationshipname", "pk": "conflrev" }, + { + "fields": { + "desc": "This document contains other documents (e.g., STDs contain RFCs)", + "name": "Contains", + "order": 0, + "revname": "Is part of", + "used": true + }, + "model": "name.docrelationshipname", + "pk": "contains" + }, { "fields": { "desc": "Approval for downref", @@ -10203,6 +11524,17 @@ "model": "name.doctypename", "pk": "agenda" }, + { + "fields": { + "desc": "", + "name": "Best Current Practice", + "order": 0, + "prefix": "bcp", + "used": true + }, + "model": "name.doctypename", + "pk": "bcp" + }, { "fields": { "desc": "", @@ -10261,13 +11593,24 @@ { "fields": { "desc": "", - "name": "Draft", + "name": "Draft", + "order": 0, + "prefix": "draft", + "used": true + }, + "model": "name.doctypename", + "pk": "draft" + }, + { + "fields": { + "desc": "", + "name": "For Your Information", "order": 0, - "prefix": "draft", + "prefix": "fyi", "used": true }, "model": "name.doctypename", - "pk": "draft" + "pk": "fyi" }, { "fields": { @@ -10302,6 +11645,17 @@ "model": "name.doctypename", "pk": "minutes" }, + { + "fields": { + "desc": "", + "name": "Narrative Minutes", + "order": 0, + "prefix": "narrative-minutes", + "used": true + }, + "model": "name.doctypename", + "pk": "narrativeminutes" + }, { "fields": { "desc": "", @@ -10346,6 +11700,17 @@ "model": "name.doctypename", "pk": "review" }, + { + "fields": { + "desc": "", + "name": "RFC", + "order": 0, + "prefix": "rfc", + "used": true + }, + "model": "name.doctypename", + "pk": "rfc" + }, { "fields": { "desc": "", @@ -10379,6 +11744,28 @@ "model": "name.doctypename", "pk": "statchg" }, + { + "fields": { + "desc": "", + "name": "Statement", + "order": 0, + "prefix": "statement", + "used": true + }, + "model": "name.doctypename", + "pk": "statement" + }, + { + "fields": { + "desc": "", + "name": "Standard", + "order": 0, + "prefix": "std", + "used": true + }, + "model": "name.doctypename", + "pk": "std" + }, { "fields": { "desc": "", @@ -10713,6 +12100,17 @@ "model": "name.extresourcename", "pk": "mailing_list_archive" }, + { + "fields": { + "desc": "ORCID", + "name": "ORCID", + "order": 0, + "type": "url", + "used": true + }, + "model": "name.extresourcename", + "pk": "orcid" + }, { "fields": { "desc": "Related Implementations", @@ -10748,8 +12146,8 @@ }, { "fields": { - "desc": "Issuer Tracker", - "name": "Issuer Tracker", + "desc": "Issue Tracker", + "name": "Issue Tracker", "order": 0, "type": "url", "used": true @@ -10845,8 +12243,9 @@ { "fields": { "desc": "", + "legend": "C", "name": "Comment", - "order": 0, + "order": 1, "used": true }, "model": "name.feedbacktypename", @@ -10855,8 +12254,9 @@ { "fields": { "desc": "", + "legend": "J", "name": "Junk", - "order": 0, + "order": 5, "used": true }, "model": "name.feedbacktypename", @@ -10865,8 +12265,9 @@ { "fields": { "desc": "", + "legend": "N", "name": "Nomination", - "order": 0, + "order": 2, "used": true }, "model": "name.feedbacktypename", @@ -10875,8 +12276,20 @@ { "fields": { "desc": "", + "legend": "O", + "name": "Overcome by events", + "order": 4, + "used": true + }, + "model": "name.feedbacktypename", + "pk": "obe" + }, + { + "fields": { + "desc": "", + "legend": "Q", "name": "Questionnaire response", - "order": 0, + "order": 3, "used": true }, "model": "name.feedbacktypename", @@ -10885,8 +12298,9 @@ { "fields": { "desc": "", + "legend": "R", "name": "Read", - "order": 0, + "order": 6, "used": true }, "model": "name.feedbacktypename", @@ -11139,14 +12553,25 @@ }, { "fields": { - "desc": "Editorial Stream Group", - "name": "Editorial", + "desc": "Editorial Stream Approval Group", + "name": "Editorial Stream Approval Group", "order": 0, "used": true, - "verbose_name": "" + "verbose_name": "Editorial Stream Approval Group" }, "model": "name.grouptypename", - "pk": "editorial" + "pk": "edappr" + }, + { + "fields": { + "desc": "Editorial Stream Working Group", + "name": "Editorial Stream Working Group", + "order": 0, + "used": true, + "verbose_name": "Editorial Stream Working Group" + }, + "model": "name.grouptypename", + "pk": "edwg" }, { "fields": { @@ -11170,6 +12595,17 @@ "model": "name.grouptypename", "pk": "iabasg" }, + { + "fields": { + "desc": "IAB Workshop", + "name": "IAB Workshop", + "order": 0, + "used": true, + "verbose_name": "IAB Workshop" + }, + "model": "name.grouptypename", + "pk": "iabworkshop" + }, { "fields": { "desc": "", @@ -11264,7 +12700,7 @@ "name": "Program", "order": 0, "used": true, - "verbose_name": "" + "verbose_name": "IAB Program" }, "model": "name.grouptypename", "pk": "program" @@ -11286,7 +12722,7 @@ "name": "Directorate (with reviews)", "order": 0, "used": true, - "verbose_name": "" + "verbose_name": "Review Team" }, "model": "name.grouptypename", "pk": "review" @@ -11349,8 +12785,8 @@ { "fields": { "default_offset_days": -19, - "desc": "Internet Draft submission cut-off for -00 drafts by UTC 23:59", - "name": "00 ID Cutoff", + "desc": "Internet-Draft submission cut-off for -00 I-Ds by UTC 23:59", + "name": "00 I-D Cutoff", "order": 0, "used": false }, @@ -11360,8 +12796,8 @@ { "fields": { "default_offset_days": -12, - "desc": "Internet Draft submission cut-off for revised (-01 and above) drafts by UTC 23:59", - "name": "01 ID Cutoff", + "desc": "Internet-Draft submission cut-off for revised (-01 and above) I-Ds by UTC 23:59", + "name": "01 I-D Cutoff", "order": 0, "used": false }, @@ -11371,7 +12807,7 @@ { "fields": { "default_offset_days": -57, - "desc": "Cut-off date for BOF proposal requests. To request a BOF, please see instructions at https://www.ietf.org/how/bofs/bof-procedures on Requesting a BOF", + "desc": "Cut-off date for BOF proposal requests. To request a __BoF__ session use the [IETF BoF Request Tool](/doc/bof-requests).", "name": "Cut-off preliminary BOF requests", "order": 0, "used": true @@ -11382,7 +12818,7 @@ { "fields": { "default_offset_days": -57, - "desc": "Preliminary BOF proposals requested. To request a BOF, please see instructions on requesting a BOF at https://www.ietf.org/how/bofs/bof-procedures/", + "desc": "Preliminary BOF proposals requested. To request a __BoF__ session use the [IETF BoF Request Tool](/doc/bof-requests).", "name": "Preliminary BOF proposals requested", "order": 0, "used": false @@ -11415,7 +12851,7 @@ { "fields": { "default_offset_days": -43, - "desc": "Cut-off date for BOF proposal requests to Area Directors at UTC 23:59", + "desc": "Cut-off date for BOF proposal requests to Area Directors at UTC 23:59. To request a __BoF__ session use the [IETF BoF Request Tool](/doc/bof-requests).", "name": "Cut-off BOF scheduling Requests", "order": 0, "used": false @@ -11459,7 +12895,7 @@ { "fields": { "default_offset_days": -43, - "desc": "Cut-off date for requests to schedule Working Group Meetings at UTC 23:59", + "desc": "Cut-off date for requests to schedule Working Group Meetings at UTC 23:59. To request a __Working Group__ session, use the [IETF Meeting Session Request Tool](/secr/sreq/).", "name": "Cut-off WG scheduling Requests", "order": 0, "used": true @@ -11478,13 +12914,24 @@ "model": "name.importantdatename", "pk": "draftwgagenda" }, + { + "fields": { + "default_offset_days": -12, + "desc": "Early registration and payment cut-off at UTC 23:59", + "name": "Early cutoff", + "order": 0, + "used": true + }, + "model": "name.importantdatename", + "pk": "early" + }, { "fields": { "default_offset_days": -47, "desc": "Early Bird registration and payment cut-off at UTC 23:59", "name": "Earlybird cutoff", "order": 0, - "used": true + "used": false }, "model": "name.importantdatename", "pk": "earlybird" @@ -11503,8 +12950,8 @@ { "fields": { "default_offset_days": -12, - "desc": "Internet Draft submission cut-off (for all drafts, including -00) by UTC 23:59", - "name": "ID Cutoff", + "desc": "Internet-Draft submission cut-off (for all Internet-Drafts, including -00) by UTC 23:59. Upload using the [I-D Submission Tool](/submit/).", + "name": "I-D Cutoff", "order": 0, "used": true }, @@ -11536,7 +12983,7 @@ { "fields": { "default_offset_days": -82, - "desc": "IETF Online Registration Opens", + "desc": "IETF Online Registration Opens [Register Here](https://www.ietf.org/how/meetings/register/).", "name": "Registration Opens", "order": 0, "used": true @@ -11547,7 +12994,7 @@ { "fields": { "default_offset_days": -89, - "desc": "Working Group and BOF scheduling begins", + "desc": "Working Group and BOF scheduling begins. To request a Working Group session, use the [IETF Meeting Session Request Tool](/secr/sreq/). If you are working on a BOF request, it is highly recommended to tell the IESG now by sending an [email to iesg@ietf.org](mailtp:iesg@ietf.org) to get advance help with the request.", "name": "Scheduling Opens", "order": 0, "used": true @@ -11605,11 +13052,22 @@ "desc": "Standard rate registration and payment cut-off at UTC 23:59.", "name": "Standard rate registration ends", "order": 18, - "used": true + "used": false }, "model": "name.importantdatename", "pk": "stdratecutoff" }, + { + "fields": { + "default_offset_days": -47, + "desc": "Super Early registration cutoff at UTC 23:59", + "name": "Super Early cutoff", + "order": 0, + "used": true + }, + "model": "name.importantdatename", + "pk": "superearly" + }, { "fields": { "desc": "", @@ -11730,6 +13188,16 @@ "model": "name.iprdisclosurestatename", "pk": "removed" }, + { + "fields": { + "desc": "", + "name": "Removed Objectively False", + "order": 5, + "used": true + }, + "model": "name.iprdisclosurestatename", + "pk": "removed_objfalse" + }, { "fields": { "desc": "", @@ -11840,6 +13308,16 @@ "model": "name.ipreventtypename", "pk": "removed" }, + { + "fields": { + "desc": "", + "name": "Removed Objectively False", + "order": 0, + "used": true + }, + "model": "name.ipreventtypename", + "pk": "removed_objfalse" + }, { "fields": { "desc": "", @@ -12240,6 +13718,86 @@ "model": "name.proceedingsmaterialtypename", "pk": "wiki" }, + { + "fields": { + "desc": "", + "name": "ANRW Combo", + "order": 0, + "used": true + }, + "model": "name.registrationtickettypename", + "pk": "anrw_combo" + }, + { + "fields": { + "desc": "", + "name": "ANRW Only", + "order": 0, + "used": true + }, + "model": "name.registrationtickettypename", + "pk": "anrw_only" + }, + { + "fields": { + "desc": "", + "name": "Hackathon Combo", + "order": 0, + "used": true + }, + "model": "name.registrationtickettypename", + "pk": "hackathon_combo" + }, + { + "fields": { + "desc": "", + "name": "Hackathon Only", + "order": 0, + "used": true + }, + "model": "name.registrationtickettypename", + "pk": "hackathon_only" + }, + { + "fields": { + "desc": "", + "name": "One Day", + "order": 0, + "used": true + }, + "model": "name.registrationtickettypename", + "pk": "one_day" + }, + { + "fields": { + "desc": "", + "name": "Student", + "order": 0, + "used": true + }, + "model": "name.registrationtickettypename", + "pk": "student" + }, + { + "fields": { + "desc": "", + "name": "Unknown", + "order": 0, + "used": true + }, + "model": "name.registrationtickettypename", + "pk": "unknown" + }, + { + "fields": { + "desc": "", + "name": "Week Pass", + "order": 0, + "used": true + }, + "model": "name.registrationtickettypename", + "pk": "week_pass" + }, { "fields": { "desc": "The reviewer has accepted the assignment", @@ -12573,7 +14131,7 @@ { "fields": { "desc": "", - "name": "Last Call", + "name": "IETF Last Call", "order": 2, "used": true }, @@ -12750,6 +14308,16 @@ "model": "name.rolename", "pk": "lead" }, + { + "fields": { + "desc": "", + "name": "Lead Maintainer", + "order": 0, + "used": true + }, + "model": "name.rolename", + "pk": "leadmaintainer" + }, { "fields": { "desc": "", @@ -12775,7 +14343,7 @@ "desc": "", "name": "Liaison CC Contact", "order": 9, - "used": true + "used": false }, "model": "name.rolename", "pk": "liaison_cc_contact" @@ -12785,11 +14353,21 @@ "desc": "", "name": "Liaison Contact", "order": 8, - "used": true + "used": false }, "model": "name.rolename", "pk": "liaison_contact" }, + { + "fields": { + "desc": "Coordinates liaison handling for the IAB", + "name": "Liaison Coordinator", + "order": 14, + "used": true + }, + "model": "name.rolename", + "pk": "liaison_coordinator" + }, { "fields": { "desc": "", @@ -12875,7 +14453,7 @@ "desc": "Assigned permission TRAC_ADMIN in datatracker-managed Trac Wiki instances", "name": "Trac Admin", "order": 0, - "used": true + "used": false }, "model": "name.rolename", "pk": "trac-admin" @@ -12885,7 +14463,7 @@ "desc": "Provides log-in permission to restricted Trac instances. Used by the generate_apache_perms management command, called from ../../scripts/Cron-runner", "name": "Trac Editor", "order": 0, - "used": true + "used": false }, "model": "name.rolename", "pk": "trac-editor" @@ -12955,7 +14533,7 @@ "desc": "Flipchars", "name": "Flipcharts", "order": 0, - "used": true + "used": false }, "model": "name.roomresourcename", "pk": "flipcharts" @@ -13005,7 +14583,7 @@ "desc": "Experimental Room Setup (U-Shape and classroom, subject to availability)", "name": "Experimental Room Setup (U-Shape and classroom)", "order": 0, - "used": true + "used": false }, "model": "name.roomresourcename", "pk": "u-shape" @@ -13026,7 +14604,10 @@ "name": "Administrative", "on_agenda": true, "order": 5, - "timeslot_types": "[\n \"other\",\n \"reg\"\n]", + "timeslot_types": [ + "other", + "reg" + ], "used": true }, "model": "name.sessionpurposename", @@ -13038,7 +14619,10 @@ "name": "Closed meeting", "on_agenda": false, "order": 10, - "timeslot_types": "[\n \"other\",\n \"regular\"\n]", + "timeslot_types": [ + "other", + "regular" + ], "used": true }, "model": "name.sessionpurposename", @@ -13050,7 +14634,9 @@ "name": "Coding", "on_agenda": true, "order": 4, - "timeslot_types": "[\n \"other\"\n]", + "timeslot_types": [ + "other" + ], "used": true }, "model": "name.sessionpurposename", @@ -13062,7 +14648,7 @@ "name": "None", "on_agenda": true, "order": 0, - "timeslot_types": "[]", + "timeslot_types": [], "used": false }, "model": "name.sessionpurposename", @@ -13074,7 +14660,9 @@ "name": "Office hours", "on_agenda": true, "order": 3, - "timeslot_types": "[\n \"other\"\n]", + "timeslot_types": [ + "other" + ], "used": true }, "model": "name.sessionpurposename", @@ -13086,7 +14674,9 @@ "name": "Open meeting", "on_agenda": true, "order": 9, - "timeslot_types": "[\n \"other\"\n]", + "timeslot_types": [ + "other" + ], "used": true }, "model": "name.sessionpurposename", @@ -13098,7 +14688,9 @@ "name": "Plenary", "on_agenda": true, "order": 7, - "timeslot_types": "[\n \"plenary\"\n]", + "timeslot_types": [ + "plenary" + ], "used": true }, "model": "name.sessionpurposename", @@ -13110,7 +14702,10 @@ "name": "Presentation", "on_agenda": true, "order": 8, - "timeslot_types": "[\n \"other\",\n \"regular\"\n]", + "timeslot_types": [ + "other", + "regular" + ], "used": true }, "model": "name.sessionpurposename", @@ -13122,7 +14717,9 @@ "name": "Regular", "on_agenda": true, "order": 1, - "timeslot_types": "[\n \"regular\"\n]", + "timeslot_types": [ + "regular" + ], "used": true }, "model": "name.sessionpurposename", @@ -13134,7 +14731,10 @@ "name": "Social", "on_agenda": true, "order": 6, - "timeslot_types": "[\n \"break\",\n \"other\"\n]", + "timeslot_types": [ + "break", + "other" + ], "used": true }, "model": "name.sessionpurposename", @@ -13146,7 +14746,9 @@ "name": "Tutorial", "on_agenda": true, "order": 2, - "timeslot_types": "[\n \"other\"\n]", + "timeslot_types": [ + "other" + ], "used": true }, "model": "name.sessionpurposename", @@ -13424,7 +15026,7 @@ }, { "fields": { - "desc": "Legacy stream", + "desc": "Legacy", "name": "Legacy", "order": 6, "used": true @@ -13432,6 +15034,36 @@ "model": "name.streamname", "pk": "legacy" }, + { + "fields": { + "desc": "Action items section", + "name": "Action Items", + "order": 3, + "used": true + }, + "model": "name.telechatagendasectionname", + "pk": "action_items" + }, + { + "fields": { + "desc": "Minutes section", + "name": "Minutes", + "order": 2, + "used": true + }, + "model": "name.telechatagendasectionname", + "pk": "minutes" + }, + { + "fields": { + "desc": "Roll call section", + "name": "Roll Call", + "order": 1, + "used": true + }, + "model": "name.telechatagendasectionname", + "pk": "roll_call" + }, { "fields": { "desc": "Friday early afternoon", @@ -16125,49 +17757,5 @@ }, "model": "stats.countryalias", "pk": 303 - }, - { - "fields": { - "command": "xym", - "switch": "--version", - "time": "2022-12-14T08:09:37.183Z", - "used": true, - "version": "xym 0.6.2" - }, - "model": "utils.versioninfo", - "pk": 1 - }, - { - "fields": { - "command": "pyang", - "switch": "--version", - "time": "2022-12-14T08:09:37.496Z", - "used": true, - "version": "pyang 2.5.3" - }, - "model": "utils.versioninfo", - "pk": 2 - }, - { - "fields": { - "command": "yanglint", - "switch": "--version", - "time": "2022-12-14T08:09:37.549Z", - "used": true, - "version": "yanglint SO 1.9.2" - }, - "model": "utils.versioninfo", - "pk": 3 - }, - { - "fields": { - "command": "xml2rfc", - "switch": "--version", - "time": "2022-12-14T08:09:38.461Z", - "used": true, - "version": "xml2rfc 3.15.3" - }, - "model": "utils.versioninfo", - "pk": 4 } ] diff --git a/ietf/name/management/commands/generate_name_fixture.py b/ietf/name/management/commands/generate_name_fixture.py index 02dc08faf5..ef30e54c73 100644 --- a/ietf/name/management/commands/generate_name_fixture.py +++ b/ietf/name/management/commands/generate_name_fixture.py @@ -67,7 +67,7 @@ def output(seq): pprint(connection.queries) raise - objects = [] # type: List[object] + objects: List[object] = [] # type: ignore[annotation-unchecked] model_objects = {} import ietf.name.models @@ -77,7 +77,6 @@ def output(seq): from ietf.mailtrigger.models import MailTrigger, Recipient from ietf.meeting.models import BusinessConstraint from ietf.stats.models import CountryAlias - from ietf.utils.models import VersionInfo # Grab all ietf.name.models for n in dir(ietf.name.models): @@ -87,7 +86,7 @@ def output(seq): model_objects[model_name(item)] = list(item.objects.all().order_by('pk')) for m in ( BallotType, State, StateType, GroupFeatures, MailTrigger, Recipient, - CountryAlias, VersionInfo, BusinessConstraint ): + CountryAlias, BusinessConstraint ): model_objects[model_name(m)] = list(m.objects.all().order_by('pk')) for m in ( DBTemplate, ): diff --git a/ietf/name/migrations/0001_initial.py b/ietf/name/migrations/0001_initial.py index 0750545040..9781e65db1 100644 --- a/ietf/name/migrations/0001_initial.py +++ b/ietf/name/migrations/0001_initial.py @@ -1,32 +1,29 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-20 10:52 - - -from typing import List, Tuple # pyflakes:ignore +# Generated by Django 2.2.28 on 2023-03-20 19:22 +from typing import List, Tuple from django.db import migrations, models import django.db.models.deletion import ietf.utils.models +import ietf.utils.validators +import jsonfield.fields class Migration(migrations.Migration): initial = True - dependencies = [ - ] # type: List[Tuple[str]] + dependencies: List[Tuple[str, str]] = [ + ] operations = [ migrations.CreateModel( - name='BallotPositionName', + name='AgendaFilterTypeName', fields=[ ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), ('name', models.CharField(max_length=255)), ('desc', models.TextField(blank=True)), ('used', models.BooleanField(default=True)), ('order', models.IntegerField(default=0)), - ('blocking', models.BooleanField(default=False)), ], options={ 'ordering': ['order', 'name'], @@ -34,14 +31,13 @@ class Migration(migrations.Migration): }, ), migrations.CreateModel( - name='ConstraintName', + name='AgendaTypeName', fields=[ ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), ('name', models.CharField(max_length=255)), ('desc', models.TextField(blank=True)), ('used', models.BooleanField(default=True)), ('order', models.IntegerField(default=0)), - ('penalty', models.IntegerField(default=0, help_text='The penalty for violating this kind of constraint; for instance 10 (small penalty) or 10000 (large penalty)')), ], options={ 'ordering': ['order', 'name'], @@ -49,13 +45,14 @@ class Migration(migrations.Migration): }, ), migrations.CreateModel( - name='ContinentName', + name='BallotPositionName', fields=[ ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), ('name', models.CharField(max_length=255)), ('desc', models.TextField(blank=True)), ('used', models.BooleanField(default=True)), ('order', models.IntegerField(default=0)), + ('blocking', models.BooleanField(default=False)), ], options={ 'ordering': ['order', 'name'], @@ -63,15 +60,29 @@ class Migration(migrations.Migration): }, ), migrations.CreateModel( - name='CountryName', + name='ConstraintName', + fields=[ + ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), + ('name', models.CharField(max_length=255)), + ('desc', models.TextField(blank=True)), + ('used', models.BooleanField(default=True)), + ('order', models.IntegerField(default=0)), + ('penalty', models.IntegerField(default=0, help_text='The penalty for violating this kind of constraint; for instance 10 (small penalty) or 10000 (large penalty)')), + ('is_group_conflict', models.BooleanField(default=False, help_text='Does this constraint capture a conflict between groups?')), + ], + options={ + 'ordering': ['order', 'name'], + 'abstract': False, + }, + ), + migrations.CreateModel( + name='ContinentName', fields=[ ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), ('name', models.CharField(max_length=255)), ('desc', models.TextField(blank=True)), ('used', models.BooleanField(default=True)), ('order', models.IntegerField(default=0)), - ('in_eu', models.BooleanField(default=False, verbose_name='In EU')), - ('continent', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ContinentName')), ], options={ 'ordering': ['order', 'name'], @@ -165,14 +176,13 @@ class Migration(migrations.Migration): }, ), migrations.CreateModel( - name='DraftSubmissionStateName', + name='ExtResourceTypeName', fields=[ ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), ('name', models.CharField(max_length=255)), ('desc', models.TextField(blank=True)), ('used', models.BooleanField(default=True)), ('order', models.IntegerField(default=0)), - ('next_states', models.ManyToManyField(blank=True, related_name='previous_states', to='name.DraftSubmissionStateName')), ], options={ 'ordering': ['order', 'name'], @@ -405,6 +415,48 @@ class Migration(migrations.Migration): 'abstract': False, }, ), + migrations.CreateModel( + name='ProceedingsMaterialTypeName', + fields=[ + ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), + ('name', models.CharField(max_length=255)), + ('desc', models.TextField(blank=True)), + ('used', models.BooleanField(default=True)), + ('order', models.IntegerField(default=0)), + ], + options={ + 'ordering': ['order', 'name'], + 'abstract': False, + }, + ), + migrations.CreateModel( + name='ReviewAssignmentStateName', + fields=[ + ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), + ('name', models.CharField(max_length=255)), + ('desc', models.TextField(blank=True)), + ('used', models.BooleanField(default=True)), + ('order', models.IntegerField(default=0)), + ], + options={ + 'ordering': ['order', 'name'], + 'abstract': False, + }, + ), + migrations.CreateModel( + name='ReviewerQueuePolicyName', + fields=[ + ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), + ('name', models.CharField(max_length=255)), + ('desc', models.TextField(blank=True)), + ('used', models.BooleanField(default=True)), + ('order', models.IntegerField(default=0)), + ], + options={ + 'ordering': ['order', 'name'], + 'abstract': False, + }, + ), migrations.CreateModel( name='ReviewRequestStateName', fields=[ @@ -475,6 +527,22 @@ class Migration(migrations.Migration): 'abstract': False, }, ), + migrations.CreateModel( + name='SessionPurposeName', + fields=[ + ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), + ('name', models.CharField(max_length=255)), + ('desc', models.TextField(blank=True)), + ('used', models.BooleanField(default=True)), + ('order', models.IntegerField(default=0)), + ('timeslot_types', jsonfield.fields.JSONField(default=[], help_text='Allowed TimeSlotTypeNames', max_length=256, validators=[ietf.utils.validators.JSONForeignKeyListValidator('name.TimeSlotTypeName')])), + ('on_agenda', models.BooleanField(default=True, help_text='Are sessions of this purpose visible on the agenda by default?')), + ], + options={ + 'ordering': ['order', 'name'], + 'abstract': False, + }, + ), migrations.CreateModel( name='SessionStatusName', fields=[ @@ -489,6 +557,20 @@ class Migration(migrations.Migration): 'abstract': False, }, ), + migrations.CreateModel( + name='SlideSubmissionStatusName', + fields=[ + ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), + ('name', models.CharField(max_length=255)), + ('desc', models.TextField(blank=True)), + ('used', models.BooleanField(default=True)), + ('order', models.IntegerField(default=0)), + ], + options={ + 'ordering': ['order', 'name'], + 'abstract': False, + }, + ), migrations.CreateModel( name='StdLevelName', fields=[ @@ -517,6 +599,20 @@ class Migration(migrations.Migration): 'abstract': False, }, ), + migrations.CreateModel( + name='TimerangeName', + fields=[ + ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), + ('name', models.CharField(max_length=255)), + ('desc', models.TextField(blank=True)), + ('used', models.BooleanField(default=True)), + ('order', models.IntegerField(default=0)), + ], + options={ + 'ordering': ['order', 'name'], + 'abstract': False, + }, + ), migrations.CreateModel( name='TimeSlotTypeName', fields=[ @@ -545,4 +641,50 @@ class Migration(migrations.Migration): 'abstract': False, }, ), + migrations.CreateModel( + name='ExtResourceName', + fields=[ + ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), + ('name', models.CharField(max_length=255)), + ('desc', models.TextField(blank=True)), + ('used', models.BooleanField(default=True)), + ('order', models.IntegerField(default=0)), + ('type', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ExtResourceTypeName')), + ], + options={ + 'ordering': ['order', 'name'], + 'abstract': False, + }, + ), + migrations.CreateModel( + name='DraftSubmissionStateName', + fields=[ + ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), + ('name', models.CharField(max_length=255)), + ('desc', models.TextField(blank=True)), + ('used', models.BooleanField(default=True)), + ('order', models.IntegerField(default=0)), + ('next_states', models.ManyToManyField(blank=True, related_name='previous_states', to='name.DraftSubmissionStateName')), + ], + options={ + 'ordering': ['order', 'name'], + 'abstract': False, + }, + ), + migrations.CreateModel( + name='CountryName', + fields=[ + ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), + ('name', models.CharField(max_length=255)), + ('desc', models.TextField(blank=True)), + ('used', models.BooleanField(default=True)), + ('order', models.IntegerField(default=0)), + ('in_eu', models.BooleanField(default=False, verbose_name='In EU')), + ('continent', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ContinentName')), + ], + options={ + 'ordering': ['order', 'name'], + 'abstract': False, + }, + ), ] diff --git a/ietf/name/migrations/0002_agendatypename.py b/ietf/name/migrations/0002_agendatypename.py deleted file mode 100644 index f916f60da1..0000000000 --- a/ietf/name/migrations/0002_agendatypename.py +++ /dev/null @@ -1,29 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.13 on 2018-07-10 13:47 - - -from django.db import migrations, models - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0001_initial'), - ] - - operations = [ - migrations.CreateModel( - name='AgendaTypeName', - fields=[ - ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), - ('name', models.CharField(max_length=255)), - ('desc', models.TextField(blank=True)), - ('used', models.BooleanField(default=True)), - ('order', models.IntegerField(default=0)), - ], - options={ - 'ordering': ['order', 'name'], - 'abstract': False, - }, - ), - ] diff --git a/ietf/name/migrations/0002_telechatagendasectionname.py b/ietf/name/migrations/0002_telechatagendasectionname.py new file mode 100644 index 0000000000..d9fedeae75 --- /dev/null +++ b/ietf/name/migrations/0002_telechatagendasectionname.py @@ -0,0 +1,27 @@ +# Generated by Django 2.2.28 on 2023-03-10 16:12 + +from django.db import migrations, models + + +class Migration(migrations.Migration): + + dependencies = [ + ('name', '0001_initial'), + ] + + operations = [ + migrations.CreateModel( + name='TelechatAgendaSectionName', + fields=[ + ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), + ('name', models.CharField(max_length=255)), + ('desc', models.TextField(blank=True)), + ('used', models.BooleanField(default=True)), + ('order', models.IntegerField(default=0)), + ], + options={ + 'ordering': ['order', 'name'], + 'abstract': False, + }, + ), + ] diff --git a/ietf/name/migrations/0003_agendatypename_data.py b/ietf/name/migrations/0003_agendatypename_data.py deleted file mode 100644 index 4db2056d73..0000000000 --- a/ietf/name/migrations/0003_agendatypename_data.py +++ /dev/null @@ -1,56 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.13 on 2018-07-10 13:47 - - -from django.db import migrations - -agenda_type_names = ( - { - 'slug': 'ietf', - 'name': 'IETF Agenda', - 'desc': '', - 'used': True, - 'order': 0, - }, - { - 'slug': 'ad', - 'name': 'AD Office Hours', - 'desc': '', - 'used': True, - 'order': 0, - }, - { - 'slug': 'side', - 'name': 'Side Meetings', - 'desc': '', - 'used': True, - 'order': 0, - }, - { - 'slug': 'workshop', - 'name': 'Workshops', - 'desc': '', - 'used': True, - 'order': 0, - }, -) - -def forward(apps, schema_editor): - AgendaTypeName = apps.get_model('name', 'AgendaTypeName') - for entry in agenda_type_names: - AgendaTypeName.objects.create(**entry) - -def reverse(apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0002_agendatypename'), - ('group', '0002_groupfeatures_historicalgroupfeatures'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/name/migrations/0003_populate_telechatagendasectionname.py b/ietf/name/migrations/0003_populate_telechatagendasectionname.py new file mode 100644 index 0000000000..41d4f86727 --- /dev/null +++ b/ietf/name/migrations/0003_populate_telechatagendasectionname.py @@ -0,0 +1,28 @@ +# Generated by Django 2.2.28 on 2023-03-10 16:30 + +from django.db import migrations + + +def forward(apps, schema_editor): + TelechatAgendaSectionName = apps.get_model('name', 'TelechatAgendaSectionName') + for slug, name, desc, order in ( + ('roll_call', 'Roll Call', 'Roll call section', 1), + ('minutes', 'Minutes', 'Minutes section', 2), + ('action_items', 'Action Items', 'Action items section', 3), + ): + TelechatAgendaSectionName.objects.create(slug=slug, name=name, desc=desc, order=order) + + +def reverse(apps, schema_editor): + pass + + +class Migration(migrations.Migration): + + dependencies = [ + ('name', '0002_telechatagendasectionname'), + ] + + operations = [ + migrations.RunPython(forward, reverse), + ] diff --git a/ietf/name/migrations/0004_add_prefix_to_doctypenames.py b/ietf/name/migrations/0004_add_prefix_to_doctypenames.py deleted file mode 100644 index 05c103319f..0000000000 --- a/ietf/name/migrations/0004_add_prefix_to_doctypenames.py +++ /dev/null @@ -1,28 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.16 on 2018-10-19 11:34 - - -from django.db import migrations - -def forward(apps, schema_editor): - DocTypeName = apps.get_model('name','DocTypeName') - DocTypeName.objects.filter(slug='liaison').update(prefix='liaison') - DocTypeName.objects.filter(slug='review').update(prefix='review') - DocTypeName.objects.filter(slug='shepwrit').update(prefix='shepherd') - -def reverse(apps, schema_editor): - DocTypeName = apps.get_model('name','DocTypeName') - DocTypeName.objects.filter(slug='liaison').update(prefix='') - DocTypeName.objects.filter(slug='review').update(prefix='') - DocTypeName.objects.filter(slug='shepwrit').update(prefix='') - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0003_agendatypename_data'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/name/migrations/0004_statements.py b/ietf/name/migrations/0004_statements.py new file mode 100644 index 0000000000..ea37799746 --- /dev/null +++ b/ietf/name/migrations/0004_statements.py @@ -0,0 +1,21 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations + +def forward(apps, schema_editor): + DocTypeName = apps.get_model("name", "DocTypeName") + DocTypeName.objects.create(slug="statement", name="Statement", prefix="statement", desc="", used=True) + + +def reverse(apps, schema_editor): + DocTypeName = apps.get_model("name", "DocTypeName") + DocTypeName.objects.filter(slug="statement").delete() + +class Migration(migrations.Migration): + dependencies = [ + ("name", "0003_populate_telechatagendasectionname"), + ] + + operations = [ + migrations.RunPython(forward, reverse), + ] diff --git a/ietf/name/migrations/0005_feedbacktypename_schema.py b/ietf/name/migrations/0005_feedbacktypename_schema.py new file mode 100644 index 0000000000..cedb129be3 --- /dev/null +++ b/ietf/name/migrations/0005_feedbacktypename_schema.py @@ -0,0 +1,20 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations, models + +class Migration(migrations.Migration): + dependencies = [ + ("name", "0004_statements"), + ] + + operations = [ + migrations.AddField( + model_name="FeedbackTypeName", + name="legend", + field=models.CharField( + default="", + help_text="One-character legend for feedback classification form", + max_length=1, + ), + ), + ] diff --git a/ietf/name/migrations/0005_reviewassignmentstatename.py b/ietf/name/migrations/0005_reviewassignmentstatename.py deleted file mode 100644 index d75f9b7748..0000000000 --- a/ietf/name/migrations/0005_reviewassignmentstatename.py +++ /dev/null @@ -1,30 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.18 on 2019-01-04 13:59 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0004_add_prefix_to_doctypenames'), - ] - - operations = [ - migrations.CreateModel( - name='ReviewAssignmentStateName', - fields=[ - ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), - ('name', models.CharField(max_length=255)), - ('desc', models.TextField(blank=True)), - ('used', models.BooleanField(default=True)), - ('order', models.IntegerField(default=0)), - ], - options={ - 'ordering': ['order', 'name'], - 'abstract': False, - }, - ), - ] diff --git a/ietf/name/migrations/0006_adjust_statenames.py b/ietf/name/migrations/0006_adjust_statenames.py deleted file mode 100644 index 6c6092407f..0000000000 --- a/ietf/name/migrations/0006_adjust_statenames.py +++ /dev/null @@ -1,103 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.18 on 2019-01-04 14:02 - - -from django.db import migrations - -def forward(apps, schema_editor): - ReviewRequestStateName = apps.get_model('name','ReviewRequestStateName') - ReviewAssignmentStateName = apps.get_model('name','ReviewAssignmentStateName') - - # TODO: Remove these newly unused states in a future release - ReviewRequestStateName.objects.filter(slug__in=['accepted', 'rejected', 'no-response', 'part-completed', 'completed', 'unknown']).update(used=False) - name, created = ReviewRequestStateName.objects.get_or_create(slug = 'assigned') - if created: - name.name = 'Assigned' - name.desc = 'The ReviewRequest has been assigned to at least one reviewer' - name.used = True - name.order = 0 - name.save() - - assignment_states = [ - { 'slug': 'assigned', - 'name': 'Assigned', - 'desc': 'The review has been assigned to this reviewer', - 'used': True, - 'order': 0 - }, - { 'slug':'accepted', - 'name':'Accepted', - 'desc':'The reviewer has accepted the assignment', - 'used': True, - 'order':0 - }, - { 'slug':'rejected', - 'name':'Rejected', - 'desc':'The reviewer has rejected the assignment', - 'used': True, - 'order':0 - }, - { 'slug':'withdrawn', - 'name':'Withdrawn by Team', - 'desc':'The team secretary has withdrawn the assignment', - 'used': True, - 'order':0 - }, - { 'slug':'overtaken', - 'name':'Overtaken By Events', - 'desc':'The review was abandoned because of circumstances', - 'used': True, - 'order':0 - }, - { 'slug':'no-response', - 'name':'No Response', - 'desc':'The reviewer did not provide a review by the deadline', - 'used': True, - 'order':0 - }, - { 'slug':'part-completed', - 'name':'Partially Completed', - 'desc':'The reviewer partially completed the assignment', - 'used': True, - 'order':0 - }, - { 'slug':'completed', - 'name':'Completed', - 'desc':'The reviewer completed the assignment', - 'used': True, - 'order':0 - }, - { 'slug':'unknown', - 'name':'Unknown', - 'desc':'The assignment is was imported from an earlier database and its state could not be computed', - 'used':'True', - 'order':0 - } - ] - - for entry in assignment_states: - name, created = ReviewAssignmentStateName.objects.get_or_create(slug=entry['slug']) - if created: - for k, v in entry.items(): - setattr(name, k, v) - name.save() - -def reverse(apps, schema_editor): - ReviewRequestStateName = apps.get_model('name','ReviewRequestStateName') - ReviewAssignmentStateName = apps.get_model('name','ReviewAssignmentStateName') - ReviewRequestStateName.objects.filter(slug__in=['accepted', 'rejected', 'no-response', 'part-completed', 'completed', 'unknown']).update(used=True) - ReviewRequestStateName.objects.filter(slug='assigned').update(used=False) - ReviewAssignmentStateName.objects.update(used=False) - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0005_reviewassignmentstatename'), - ('doc','0011_reviewassignmentdocevent'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/name/migrations/0006_feedbacktypename_data.py b/ietf/name/migrations/0006_feedbacktypename_data.py new file mode 100644 index 0000000000..f11fca889b --- /dev/null +++ b/ietf/name/migrations/0006_feedbacktypename_data.py @@ -0,0 +1,36 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations + +def forward(apps, schema_editor): + FeedbackTypeName = apps.get_model("name", "FeedbackTypeName") + FeedbackTypeName.objects.create(slug="obe", name="Overcome by events") + for slug, legend, order in ( + ('comment', 'C', 1), + ('nomina', 'N', 2), + ('questio', 'Q', 3), + ('obe', 'O', 4), + ('junk', 'J', 5), + ('read', 'R', 6), + ): + ft = FeedbackTypeName.objects.get(slug=slug) + ft.legend = legend + ft.order = order + ft.save() + +def reverse(apps, schema_editor): + FeedbackTypeName = apps.get_model("name", "FeedbackTypeName") + FeedbackTypeName.objects.filter(slug="obe").delete() + for ft in FeedbackTypeName.objects.all(): + ft.legend = "" + ft.order = 0 + ft.save() + +class Migration(migrations.Migration): + dependencies = [ + ("name", "0005_feedbacktypename_schema"), + ] + + operations = [ + migrations.RunPython(forward, reverse), + ] diff --git a/ietf/name/migrations/0007_appeal_artifact_typename.py b/ietf/name/migrations/0007_appeal_artifact_typename.py new file mode 100644 index 0000000000..1a604dad84 --- /dev/null +++ b/ietf/name/migrations/0007_appeal_artifact_typename.py @@ -0,0 +1,59 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations, models + + +def forward(apps, schema_editor): + AppealArtifactTypeName = apps.get_model("name", "AppealArtifactTypeName") + for slug, name, desc, order in [ + ("appeal", "Appeal", "The content of an appeal", 1), + ( + "appeal_with_response", + "Response (with appeal included)", + "The content of an appeal combined with the content of a response", + 2, + ), + ("response", "Response", "The content of a response to an appeal", 3), + ( + "reply_to_response", + "Reply to response", + "The content of a reply to an appeal response", + 4, + ), + ("other_content", "Other content", "Other content related to an appeal", 5), + ]: + AppealArtifactTypeName.objects.create( + slug=slug, name=name, desc=desc, order=order + ) + + +def reverse(apps, schema_editor): + AppealArtifactTypeName = apps.get_model("name", "AppealArtifactTypeName") + AppealArtifactTypeName.objects.delete() + + +class Migration(migrations.Migration): + dependencies = [ + ("name", "0006_feedbacktypename_data"), + ] + + operations = [ + migrations.CreateModel( + name="AppealArtifactTypeName", + fields=[ + ( + "slug", + models.CharField(max_length=32, primary_key=True, serialize=False), + ), + ("name", models.CharField(max_length=255)), + ("desc", models.TextField(blank=True)), + ("used", models.BooleanField(default=True)), + ("order", models.IntegerField(default=0)), + ], + options={ + "ordering": ["order", "name"], + "abstract": False, + }, + ), + migrations.RunPython(forward, reverse), + ] diff --git a/ietf/name/migrations/0007_fix_m2m_slug_id_length.py b/ietf/name/migrations/0007_fix_m2m_slug_id_length.py deleted file mode 100644 index 578415bc58..0000000000 --- a/ietf/name/migrations/0007_fix_m2m_slug_id_length.py +++ /dev/null @@ -1,19 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0006_adjust_statenames'), - ] - - operations = [ - migrations.RunSQL("ALTER TABLE doc_ballottype_positions MODIFY ballotpositionname_id varchar(32);", ""), - migrations.RunSQL("ALTER TABLE doc_dochistory_tags MODIFY doctagname_id varchar(32);", ""), - migrations.RunSQL("ALTER TABLE group_group_unused_tags MODIFY doctagname_id varchar(32);", ""), - migrations.RunSQL("ALTER TABLE group_grouphistory_unused_tags MODIFY doctagname_id varchar(32);", ""), - ] diff --git a/ietf/name/migrations/0008_removed_objfalse.py b/ietf/name/migrations/0008_removed_objfalse.py new file mode 100644 index 0000000000..2c1f7e759d --- /dev/null +++ b/ietf/name/migrations/0008_removed_objfalse.py @@ -0,0 +1,24 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations + +def forward(apps, schema_editor): + IprDisclosureStateName = apps.get_model("name", "IprDisclosureStateName") + IprDisclosureStateName.objects.create(slug="removed_objfalse", name="Removed Objectively False", order=5) + IprEventTypeName = apps.get_model("name", "IprEventTypeName") + IprEventTypeName.objects.create(slug="removed_objfalse", name="Removed Objectively False") + +def reverse(apps, schema_editor): + IprDisclosureStateName = apps.get_model("name", "IprDisclosureStateName") + IprDisclosureStateName.objects.filter(slug="removed_objfalse").delete() + IprEventTypeName = apps.get_model("name", "IprEventTypeName") + IprEventTypeName.objects.filter(slug="removed_objfalse").delete() + +class Migration(migrations.Migration): + dependencies = [ + ("name", "0007_appeal_artifact_typename"), + ] + + operations = [ + migrations.RunPython(forward, reverse), + ] diff --git a/ietf/name/migrations/0008_reviewerqueuepolicyname.py b/ietf/name/migrations/0008_reviewerqueuepolicyname.py deleted file mode 100644 index caee6384b3..0000000000 --- a/ietf/name/migrations/0008_reviewerqueuepolicyname.py +++ /dev/null @@ -1,38 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.23 on 2019-11-18 08:35 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - def forward(apps, schema_editor): - ReviewerQueuePolicyName = apps.get_model('name', 'ReviewerQueuePolicyName') - ReviewerQueuePolicyName.objects.create(slug='RotateAlphabetically', name='Rotate alphabetically') - ReviewerQueuePolicyName.objects.create(slug='LeastRecentlyUsed', name='Least recently used') - - def reverse(self, apps): - pass - - dependencies = [ - ('name', '0007_fix_m2m_slug_id_length'), - ] - - operations = [ - migrations.CreateModel( - name='ReviewerQueuePolicyName', - fields=[ - ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), - ('name', models.CharField(max_length=255)), - ('desc', models.TextField(blank=True)), - ('used', models.BooleanField(default=True)), - ('order', models.IntegerField(default=0)), - ], - options={ - 'ordering': ['order', 'name'], - 'abstract': False, - }, - ), - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/name/migrations/0009_add_verified_errata_to_doctagname.py b/ietf/name/migrations/0009_add_verified_errata_to_doctagname.py deleted file mode 100644 index 94c1134d67..0000000000 --- a/ietf/name/migrations/0009_add_verified_errata_to_doctagname.py +++ /dev/null @@ -1,23 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -from django.db import migrations - -def forward(apps, schema_editor): - DocTagName = apps.get_model('name','DocTagName') - DocTagName.objects.get_or_create(slug='verified-errata', name='Has verified errata', desc='', used=True, order=0) - -def reverse(apps, schema_editor): - DocTagName = apps.get_model('name','DocTagName') - DocTagName.objects.filter(slug='verified-errata').delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0008_reviewerqueuepolicyname'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/name/migrations/0009_iabworkshops.py b/ietf/name/migrations/0009_iabworkshops.py new file mode 100644 index 0000000000..1819815860 --- /dev/null +++ b/ietf/name/migrations/0009_iabworkshops.py @@ -0,0 +1,29 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations + +def forward(apps, schema_editor): + GroupTypeName = apps.get_model("name", "GroupTypeName") + GroupTypeName.objects.create( + slug = "iabworkshop", + name = "IAB Workshop", + desc = "IAB Workshop", + used = True, + order = 0, + verbose_name = "IAB Workshop", + + ) + +def reverse(apps, schema_editor): + GroupTypeName = apps.get_model("name", "GroupTypeName") + GroupTypeName.objects.filter(slug="iabworkshop").delete() + + +class Migration(migrations.Migration): + dependencies = [ + ("name", "0008_removed_objfalse"), + ] + + operations = [ + migrations.RunPython(forward, reverse) + ] diff --git a/ietf/name/migrations/0010_rfc_doctype_names.py b/ietf/name/migrations/0010_rfc_doctype_names.py new file mode 100644 index 0000000000..8d7a565f23 --- /dev/null +++ b/ietf/name/migrations/0010_rfc_doctype_names.py @@ -0,0 +1,30 @@ +# Generated by Django 4.2.2 on 2023-06-14 20:39 + +from django.db import migrations + + +def forward(apps, schema_editor): + DocTypeName = apps.get_model("name", "DocTypeName") + DocTypeName.objects.get_or_create( + slug="rfc", + name="RFC", + used=True, + prefix="rfc", + ) + + DocRelationshipName = apps.get_model("name", "DocRelationshipName") + DocRelationshipName.objects.get_or_create( + slug="became_rfc", + name="became RFC", + used=True, + revname="came from draft", + ) + +class Migration(migrations.Migration): + dependencies = [ + ("name", "0009_iabworkshops"), + ] + + operations = [ + migrations.RunPython(forward), + ] diff --git a/ietf/name/migrations/0010_timerangename.py b/ietf/name/migrations/0010_timerangename.py deleted file mode 100644 index b6b17c9da4..0000000000 --- a/ietf/name/migrations/0010_timerangename.py +++ /dev/null @@ -1,58 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.27 on 2020-02-04 05:43 -from __future__ import unicode_literals - -from django.db import migrations, models - - -def forward(apps, schema_editor): - TimerangeName = apps.get_model('name', 'TimerangeName') - timeranges = [ - ('monday-morning', 'Monday morning'), - ('monday-afternoon-early', 'Monday early afternoon'), - ('monday-afternoon-late', 'Monday late afternoon'), - ('tuesday-morning', 'Tuesday morning'), - ('tuesday-afternoon-early', 'Tuesday early afternoon'), - ('tuesday-afternoon-late', 'Tuesday late afternoon'), - ('wednesday-morning', 'Wednesday morning'), - ('wednesday-afternoon-early', 'Wednesday early afternoon'), - ('wednesday-afternoon-late', 'Wednesday late afternoon'), - ('thursday-morning', 'Thursday morning'), - ('thursday-afternoon-early', 'Thursday early afternoon'), - ('thursday-afternoon-late', 'Thursday late afternoon'), - ('friday-morning', 'Friday morning'), - ('friday-afternoon-early', 'Friday early afternoon'), - ('friday-afternoon-late', 'Friday late afternoon'), - ] - for order, (slug, desc) in enumerate(timeranges): - TimerangeName.objects.create(slug=slug, name=slug, desc=desc, used=True, order=order) - - -def reverse(apps, schema_editor): - pass - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0009_add_verified_errata_to_doctagname'), - ] - - operations = [ - migrations.CreateModel( - name='TimerangeName', - fields=[ - ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), - ('name', models.CharField(max_length=255)), - ('desc', models.TextField(blank=True)), - ('used', models.BooleanField(default=True)), - ('order', models.IntegerField(default=0)), - ], - options={ - 'ordering': ['order', 'name'], - 'abstract': False, - }, - ), - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/name/migrations/0011_constraintname_editor_label.py b/ietf/name/migrations/0011_constraintname_editor_label.py deleted file mode 100644 index 0425785a48..0000000000 --- a/ietf/name/migrations/0011_constraintname_editor_label.py +++ /dev/null @@ -1,36 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- - -from __future__ import unicode_literals - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0010_timerangename'), - ] - - def fill_in_editor_labels(apps, schema_editor): - ConstraintName = apps.get_model('name', 'ConstraintName') - for cn in ConstraintName.objects.all(): - cn.editor_label = { - 'conflict': "(1)", - 'conflic2': "(2)", - 'conflic3': "(3)", - 'bethere': "(person)", - }.get(cn.slug, cn.slug) - cn.save() - - def noop(apps, schema_editor): - pass - - operations = [ - migrations.AddField( - model_name='constraintname', - name='editor_label', - field=models.CharField(blank=True, help_text='Very short label for producing warnings inline in the sessions in the schedule editor.', max_length=32), - ), - migrations.RunPython(fill_in_editor_labels, noop, elidable=True), - ] diff --git a/ietf/name/migrations/0011_subseries.py b/ietf/name/migrations/0011_subseries.py new file mode 100644 index 0000000000..b3fe107924 --- /dev/null +++ b/ietf/name/migrations/0011_subseries.py @@ -0,0 +1,38 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations + + +def forward(apps, schema_editor): + DocTypeName = apps.get_model("name", "DocTypeName") + DocRelationshipName = apps.get_model("name", "DocRelationshipName") + for slug, name, prefix in [ + ("std", "Standard", "std"), + ("bcp", "Best Current Practice", "bcp"), + ("fyi", "For Your Information", "fyi"), + ]: + DocTypeName.objects.create( + slug=slug, name=name, prefix=prefix, desc="", used=True + ) + DocRelationshipName.objects.create( + slug="contains", + name="Contains", + revname="Is part of", + desc="This document contains other documents (e.g., STDs contain RFCs)", + used=True, + ) + + +def reverse(apps, schema_editor): + DocTypeName = apps.get_model("name", "DocTypeName") + DocRelationshipName = apps.get_model("name", "DocRelationshipName") + DocTypeName.objects.filter(slug__in=["std", "bcp", "fyi"]).delete() + DocRelationshipName.objects.filter(slug="contains").delete() + + +class Migration(migrations.Migration): + dependencies = [ + ("name", "0010_rfc_doctype_names"), + ] + + operations = [migrations.RunPython(forward, reverse)] diff --git a/ietf/name/migrations/0012_adjust_important_dates.py b/ietf/name/migrations/0012_adjust_important_dates.py new file mode 100644 index 0000000000..7a3252bb5c --- /dev/null +++ b/ietf/name/migrations/0012_adjust_important_dates.py @@ -0,0 +1,29 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations + +def markdown_names(apps, schema_editor): + ImportantDateName = apps.get_model("name", "ImportantDateName") + changes = [ + ('bofproposals', "Preliminary BOF proposals requested. To request a __BoF__ session use the [IETF BoF Request Tool](/doc/bof-requests)."), + ('openreg', "IETF Online Registration Opens [Register Here](https://www.ietf.org/how/meetings/register/)."), + ('opensched', "Working Group and BOF scheduling begins. To request a Working Group session, use the [IETF Meeting Session Request Tool](/secr/sreq/). If you are working on a BOF request, it is highly recommended to tell the IESG now by sending an [email to iesg@ietf.org](mailtp:iesg@ietf.org) to get advance help with the request."), + ('cutoffwgreq', "Cut-off date for requests to schedule Working Group Meetings at UTC 23:59. To request a __Working Group__ session, use the [IETF Meeting Session Request Tool](/secr/sreq/)."), + ('idcutoff', "Internet-Draft submission cut-off (for all Internet-Drafts, including -00) by UTC 23:59. Upload using the [I-D Submission Tool](/submit/)."), + ('cutoffwgreq', "Cut-off date for requests to schedule Working Group Meetings at UTC 23:59. To request a __Working Group__ session, use the [IETF Meeting Session Request Tool](/secr/sreq/)."), + ('bofprelimcutoff', "Cut-off date for BOF proposal requests. To request a __BoF__ session use the [IETF BoF Request Tool](/doc/bof-requests)."), + ('cutoffbofreq', "Cut-off date for BOF proposal requests to Area Directors at UTC 23:59. To request a __BoF__ session use the [IETF BoF Request Tool](/doc/bof-requests)."), + ] + for slug, newDescription in changes: + datename = ImportantDateName.objects.get(pk=slug) # If the slug does not exist, then Django will throw an exception :-) + datename.desc = newDescription + datename.save() + +class Migration(migrations.Migration): + dependencies = [ + ("name", "0011_subseries"), + ] + + operations = [ + migrations.RunPython(markdown_names), + ] diff --git a/ietf/name/migrations/0012_role_name_robots.py b/ietf/name/migrations/0012_role_name_robots.py deleted file mode 100644 index bc5c37980e..0000000000 --- a/ietf/name/migrations/0012_role_name_robots.py +++ /dev/null @@ -1,23 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -from django.db import migrations - -def forward(apps, schema_editor): - RoleName = apps.get_model('name','RoleName') - RoleName.objects.get_or_create(slug='robot', name='Automation Robot', desc='A role for API access by external scripts or entities, such as the mail archive, registrations system, etc.', used=True, order=0) - -def reverse(apps, schema_editor): - RoleName = apps.get_model('name','RoleName') - RoleName.objects.filter(slug='robot').delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0011_constraintname_editor_label'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/name/migrations/0013_add_auth48_docurltagname.py b/ietf/name/migrations/0013_add_auth48_docurltagname.py deleted file mode 100644 index 6bede736be..0000000000 --- a/ietf/name/migrations/0013_add_auth48_docurltagname.py +++ /dev/null @@ -1,28 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 2.0.13 on 2020-06-02 10:13 - -from django.db import migrations - -def forward(apps, schema_editor): - DocUrlTagName = apps.get_model('name', 'DocUrlTagName') - DocUrlTagName.objects.create( - slug='auth48', - name='RFC Editor Auth48 status', - used=True, - ) - -def reverse(apps, schema_editor): - DocUrlTagName = apps.get_model('name', 'DocUrlTagName') - auth48_tag = DocUrlTagName.objects.get(slug='auth48') - auth48_tag.delete() - -class Migration(migrations.Migration): - """Add DocUrlTagName entry for RFC Ed Auth48 URL""" - dependencies = [ - ('name', '0012_role_name_robots'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/name/migrations/0013_narrativeminutes.py b/ietf/name/migrations/0013_narrativeminutes.py new file mode 100644 index 0000000000..89aa75a371 --- /dev/null +++ b/ietf/name/migrations/0013_narrativeminutes.py @@ -0,0 +1,35 @@ +# Copyright The IETF Trust 2023, All Rights Reserved + +from django.db import migrations, models + + +def forward(apps, schema_editor): + DocTypeName = apps.get_model("name", "DocTypeName") + DocTypeName.objects.create( + slug="narrativeminutes", + name="Narrative Minutes", + desc="", + used=True, + order=0, + prefix="narrative-minutes", + ) + + +def reverse(apps, schema_editor): + DocTypeName = apps.get_model("name", "DocTypeName") + DocTypeName.objects.filter(slug="narrativeminutes").delete() + + +class Migration(migrations.Migration): + dependencies = [ + ("name", "0012_adjust_important_dates"), + ] + + operations = [ + migrations.AlterField( + model_name="doctypename", + name="prefix", + field=models.CharField(default="", max_length=32), + ), + migrations.RunPython(forward, reverse), + ] diff --git a/ietf/name/migrations/0014_change_legacy_stream_desc.py b/ietf/name/migrations/0014_change_legacy_stream_desc.py new file mode 100644 index 0000000000..8297e86274 --- /dev/null +++ b/ietf/name/migrations/0014_change_legacy_stream_desc.py @@ -0,0 +1,21 @@ +# Copyright The IETF Trust 2024, All Rights Reserved + +from django.db import migrations + +def forward(apps, schema_editor): + StreamName = apps.get_model("name", "StreamName") + StreamName.objects.filter(pk="legacy").update(desc="Legacy") + +def reverse(apps, schema_editor): + StreamName = apps.get_model("name", "StreamName") + StreamName.objects.filter(pk="legacy").update(desc="Legacy stream") + +class Migration(migrations.Migration): + + dependencies = [ + ("name", "0013_narrativeminutes"), + ] + + operations = [ + migrations.RunPython(forward, reverse) + ] diff --git a/ietf/name/migrations/0014_extres.py b/ietf/name/migrations/0014_extres.py deleted file mode 100644 index 576a98ea36..0000000000 --- a/ietf/name/migrations/0014_extres.py +++ /dev/null @@ -1,51 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.29 on 2020-03-19 13:56 -from __future__ import unicode_literals - -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0013_add_auth48_docurltagname'), - ] - - operations = [ - migrations.CreateModel( - name='ExtResourceName', - fields=[ - ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), - ('name', models.CharField(max_length=255)), - ('desc', models.TextField(blank=True)), - ('used', models.BooleanField(default=True)), - ('order', models.IntegerField(default=0)), - ], - options={ - 'ordering': ['order', 'name'], - 'abstract': False, - }, - ), - migrations.CreateModel( - name='ExtResourceTypeName', - fields=[ - ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), - ('name', models.CharField(max_length=255)), - ('desc', models.TextField(blank=True)), - ('used', models.BooleanField(default=True)), - ('order', models.IntegerField(default=0)), - ], - options={ - 'ordering': ['order', 'name'], - 'abstract': False, - }, - ), - migrations.AddField( - model_name='extresourcename', - name='type', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ExtResourceTypeName'), - ), - ] diff --git a/ietf/name/migrations/0015_last_call_name.py b/ietf/name/migrations/0015_last_call_name.py new file mode 100644 index 0000000000..ac210a274f --- /dev/null +++ b/ietf/name/migrations/0015_last_call_name.py @@ -0,0 +1,22 @@ +# Copyright 2025, IETF Trust + +from django.db import migrations + + +def forward(apps, schema_editor): + ReviewTypeName = apps.get_model("name", "ReviewTypeName") + ReviewTypeName.objects.filter(slug="lc").update(name="IETF Last Call") + +def reverse(apps, schema_editor): + ReviewTypeName = apps.get_model("name", "ReviewTypeName") + ReviewTypeName.objects.filter(slug="lc").update(name="Last Call") + +class Migration(migrations.Migration): + + dependencies = [ + ("name", "0014_change_legacy_stream_desc"), + ] + + operations = [ + migrations.RunPython(forward, reverse) + ] diff --git a/ietf/name/migrations/0015_populate_extres.py b/ietf/name/migrations/0015_populate_extres.py deleted file mode 100644 index 64a6bf08ae..0000000000 --- a/ietf/name/migrations/0015_populate_extres.py +++ /dev/null @@ -1,63 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.29 on 2020-03-19 11:42 -from __future__ import unicode_literals - -from collections import namedtuple - -from django.db import migrations - - -def forward(apps, schema_editor): - ExtResourceName = apps.get_model('name','ExtResourceName') - ExtResourceTypeName = apps.get_model('name','ExtResourceTypeName') - - ExtResourceTypeName.objects.create(slug='email', name="Email address", desc="Email address", used=True, order=0) - ExtResourceTypeName.objects.create(slug='url', name="URL", desc="URL", used=True, order=0) - ExtResourceTypeName.objects.create(slug='string', name="string", desc="string", used=True, order=0) - - resourcename = namedtuple('resourcename', ['slug', 'name', 'type']) - resourcenames= [ - resourcename("webpage", "Additional Web Page", "url"), - resourcename("faq", "Frequently Asked Questions", "url"), - resourcename("github_username","GitHub Username", "string"), - resourcename("github_org","GitHub Organization", "url"), - resourcename("github_repo","GitHub Repository", "url"), - resourcename("gitlab_username","GitLab Username", "string"), - resourcename("tracker","Issuer Tracker", "url"), - resourcename("slack","Slack Channel", "url"), - resourcename("wiki","Wiki", "url"), - resourcename("yc_entry","Yang Catalog Entry", "url"), - resourcename("yc_impact","Yang Impact Analysis", "url"), - resourcename("jabber_room","Jabber Room", "url"), - resourcename("jabber_log","Jabber Log", "url"), - resourcename("mailing_list","Mailing List", "url"), - resourcename("mailing_list_archive","Mailing List Archive","url"), - resourcename("repo","Other Repository", "url") - ] - - for name in resourcenames: - ExtResourceName.objects.create(slug=name.slug, name=name.name, desc=name.name, used=True, order=0, type_id=name.type) - - - -def reverse(apps, schema_editor): - ExtResourceName = apps.get_model('name','ExtResourceName') - ExtResourceTypeName = apps.get_model('name','ExtResourceTypeName') - - ExtResourceName.objects.all().delete() - ExtResourceTypeName.objects.all().delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0014_extres'), - ('group', '0033_extres'), - ('doc', '0034_extres'), - ('person', '0015_extres'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/name/migrations/0016_add_research_area_groups.py b/ietf/name/migrations/0016_add_research_area_groups.py deleted file mode 100644 index 79ab86d40d..0000000000 --- a/ietf/name/migrations/0016_add_research_area_groups.py +++ /dev/null @@ -1,28 +0,0 @@ -# Generated by Django 2.2.14 on 2020-07-28 09:29 - -from django.db import migrations - -def forward(apps, schema_editor): - GroupTypeName = apps.get_model('name','GroupTypeName') - GroupTypeName.objects.create( - slug = 'rag', - name = 'RAG', - desc = 'Research Area Group', - used = True, - order = 0, - verbose_name='Research Area Group' - ) - -def reverse(apps, schema_editor): - GroupTypeName = apps.get_model('name','GroupTypeName') - GroupTypeName.objects.filter(slug='rag').delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0015_populate_extres'), - ] - - operations = [ - migrations.RunPython(forward,reverse), - ] diff --git a/ietf/name/migrations/0016_attendancetypename_registrationtickettypename.py b/ietf/name/migrations/0016_attendancetypename_registrationtickettypename.py new file mode 100644 index 0000000000..9376d3a4c6 --- /dev/null +++ b/ietf/name/migrations/0016_attendancetypename_registrationtickettypename.py @@ -0,0 +1,47 @@ +# Generated by Django 4.2.17 on 2025-01-02 18:21 + +from django.db import migrations, models + + +class Migration(migrations.Migration): + + dependencies = [ + ("name", "0015_last_call_name"), + ] + + operations = [ + migrations.CreateModel( + name="AttendanceTypeName", + fields=[ + ( + "slug", + models.CharField(max_length=32, primary_key=True, serialize=False), + ), + ("name", models.CharField(max_length=255)), + ("desc", models.TextField(blank=True)), + ("used", models.BooleanField(default=True)), + ("order", models.IntegerField(default=0)), + ], + options={ + "ordering": ["order", "name"], + "abstract": False, + }, + ), + migrations.CreateModel( + name="RegistrationTicketTypeName", + fields=[ + ( + "slug", + models.CharField(max_length=32, primary_key=True, serialize=False), + ), + ("name", models.CharField(max_length=255)), + ("desc", models.TextField(blank=True)), + ("used", models.BooleanField(default=True)), + ("order", models.IntegerField(default=0)), + ], + options={ + "ordering": ["order", "name"], + "abstract": False, + }, + ), + ] diff --git a/ietf/name/migrations/0017_populate_new_reg_names.py b/ietf/name/migrations/0017_populate_new_reg_names.py new file mode 100644 index 0000000000..51954885c0 --- /dev/null +++ b/ietf/name/migrations/0017_populate_new_reg_names.py @@ -0,0 +1,39 @@ +# Generated by Django 4.2.17 on 2025-01-02 18:26 + +from django.db import migrations + +def forward(apps, schema_editor): + AttendanceTypeName = apps.get_model('name', 'AttendanceTypeName') + RegistrationTicketTypeName = apps.get_model('name', 'RegistrationTicketTypeName') + AttendanceTypeName.objects.create(slug='onsite', name='Onsite') + AttendanceTypeName.objects.create(slug='remote', name='Remote') + AttendanceTypeName.objects.create(slug='hackathon_onsite', name='Hackathon Onsite') + AttendanceTypeName.objects.create(slug='hackathon_remote', name='Hackathon Remote') + AttendanceTypeName.objects.create(slug='anrw_onsite', name='ANRW Onsite') + AttendanceTypeName.objects.create(slug='unknown', name='Unknown') + RegistrationTicketTypeName.objects.create(slug='week_pass', name='Week Pass') + RegistrationTicketTypeName.objects.create(slug='one_day', name='One Day') + RegistrationTicketTypeName.objects.create(slug='student', name='Student') + RegistrationTicketTypeName.objects.create(slug='hackathon_only', name='Hackathon Only') + RegistrationTicketTypeName.objects.create(slug='hackathon_combo', name='Hackathon Combo') + RegistrationTicketTypeName.objects.create(slug='anrw_only', name='ANRW Only') + RegistrationTicketTypeName.objects.create(slug='anrw_combo', name='ANRW Combo') + RegistrationTicketTypeName.objects.create(slug='unknown', name='Unknown') + + +def reverse(apps, schema_editor): + AttendanceTypeName = apps.get_model('name', 'AttendanceTypeName') + RegistrationTicketTypeName = apps.get_model('name', 'RegistrationTicketTypeName') + AttendanceTypeName.objects.delete() + RegistrationTicketTypeName.objects.delete() + + +class Migration(migrations.Migration): + + dependencies = [ + ("name", "0016_attendancetypename_registrationtickettypename"), + ] + + operations = [ + migrations.RunPython(forward, reverse), + ] diff --git a/ietf/name/migrations/0017_update_constraintname_order_and_label.py b/ietf/name/migrations/0017_update_constraintname_order_and_label.py deleted file mode 100644 index 3b552baf61..0000000000 --- a/ietf/name/migrations/0017_update_constraintname_order_and_label.py +++ /dev/null @@ -1,46 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved - -from django.db import migrations, models - - -class Migration(migrations.Migration): - dependencies = [ - ('name', '0016_add_research_area_groups'), - ] - - def update_editor_labels(apps, schema_editor): - ConstraintName = apps.get_model('name', 'ConstraintName') - for cn in ConstraintName.objects.all(): - cn.editor_label = { - 'bethere': "{count}", - 'wg_adjacent': "", - 'conflict': "1", - 'conflic2': "2", - 'conflic3': "3", - 'time_relation': "Δ", - 'timerange': "", - }.get(cn.slug, cn.editor_label) - - cn.order = { - 'conflict': 1, - 'conflic2': 2, - 'conflic3': 3, - 'bethere': 4, - 'timerange': 5, - 'time_relation': 6, - 'wg_adjacent': 7, - }.get(cn.slug, cn.order) - - cn.save() - - def noop(apps, schema_editor): - pass - - operations = [ - migrations.AlterField( - model_name='constraintname', - name='editor_label', - field=models.CharField(blank=True, help_text='Very short label for producing warnings inline in the sessions in the schedule editor.', max_length=64), - ), - migrations.RunPython(update_editor_labels, noop, elidable=True), - ] diff --git a/ietf/name/migrations/0018_alter_rolenames.py b/ietf/name/migrations/0018_alter_rolenames.py new file mode 100644 index 0000000000..f931de2e97 --- /dev/null +++ b/ietf/name/migrations/0018_alter_rolenames.py @@ -0,0 +1,36 @@ +# Copyright The IETF Trust 2025, All Rights Reserved# Generated by Django 4.2.21 on 2025-05-30 16:35 + +from django.db import migrations + + +def forward(apps, schema_editor): + RoleName = apps.get_model("name", "RoleName") + RoleName.objects.filter(slug__in=["liaison_contact", "liaison_cc_contact"]).update( + used=False + ) + RoleName.objects.get_or_create( + slug="liaison_coordinator", + defaults={ + "name": "Liaison Coordinator", + "desc": "Coordinates liaison handling for the IAB", + "order": 14, + }, + ) + RoleName.objects.filter(slug__contains="trac-").update(used=False) + + +def reverse(apps, schema_editor): + RoleName = apps.get_model("name", "RoleName") + RoleName.objects.filter(slug__in=["liaison_contact", "liaison_cc_contact"]).update( + used=True + ) + RoleName.objects.filter(slug="liaison_coordinator").delete() + # Intentionally not restoring trac-* RoleNames to used=True + + +class Migration(migrations.Migration): + dependencies = [ + ("name", "0017_populate_new_reg_names"), + ] + + operations = [migrations.RunPython(forward, reverse)] diff --git a/ietf/name/migrations/0018_slidesubmissionstatusname.py b/ietf/name/migrations/0018_slidesubmissionstatusname.py deleted file mode 100644 index b0f1e18e4d..0000000000 --- a/ietf/name/migrations/0018_slidesubmissionstatusname.py +++ /dev/null @@ -1,42 +0,0 @@ -# Generated by Django 2.2.14 on 2020-08-03 11:53 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0017_update_constraintname_order_and_label'), - ] - - def forward(apps, schema_editor): - SlideSubmissionStatusName = apps.get_model('name', 'SlideSubmissionStatusName') - slide_submission_status_names = [ - ('pending', 'Pending'), - ('approved', 'Approved'), - ('rejected', 'Rejected'), - ] - for order, (slug, desc) in enumerate(slide_submission_status_names): - SlideSubmissionStatusName.objects.create(slug=slug, name=slug, desc=desc, used=True, order=order) - - - def reverse(apps, schema_editor): - pass - - operations = [ - migrations.CreateModel( - name='SlideSubmissionStatusName', - fields=[ - ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), - ('name', models.CharField(max_length=255)), - ('desc', models.TextField(blank=True)), - ('used', models.BooleanField(default=True)), - ('order', models.IntegerField(default=0)), - ], - options={ - 'ordering': ['order', 'name'], - 'abstract': False, - }, - ), - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/name/migrations/0019_add_timeslottypename_private.py b/ietf/name/migrations/0019_add_timeslottypename_private.py deleted file mode 100644 index f515769aa6..0000000000 --- a/ietf/name/migrations/0019_add_timeslottypename_private.py +++ /dev/null @@ -1,26 +0,0 @@ -# Generated by Django 2.2.15 on 2020-08-20 02:40 - -from django.db import migrations, models - -def set_private_bit_on_timeslottypename(apps, schema_editor): - TimeSlotTypeName = apps.get_model('name', 'TimeSlotTypeName') - - TimeSlotTypeName.objects.filter(slug__in=['lead', 'offagenda']).update(private=True) - -def noop(apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0018_slidesubmissionstatusname'), - ] - - operations = [ - migrations.AddField( - model_name='timeslottypename', - name='private', - field=models.BooleanField(default=False, help_text='Whether sessions of this type should be kept off the public agenda'), - ), - migrations.RunPython(set_private_bit_on_timeslottypename, noop), - ] diff --git a/ietf/name/migrations/0019_alter_sessionpurposename_timeslot_types.py b/ietf/name/migrations/0019_alter_sessionpurposename_timeslot_types.py new file mode 100644 index 0000000000..a0ca81836d --- /dev/null +++ b/ietf/name/migrations/0019_alter_sessionpurposename_timeslot_types.py @@ -0,0 +1,27 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +from django.db import migrations, models +import ietf.utils.validators + + +class Migration(migrations.Migration): + dependencies = [ + ("name", "0018_alter_rolenames"), + ] + + operations = [ + migrations.AlterField( + model_name="sessionpurposename", + name="timeslot_types", + field=models.JSONField( + default=list, + help_text="Allowed TimeSlotTypeNames", + max_length=256, + validators=[ + ietf.utils.validators.JSONForeignKeyListValidator( + "name.TimeSlotTypeName" + ) + ], + ), + ), + ] diff --git a/ietf/name/migrations/0020_add_rescheduled_session_name.py b/ietf/name/migrations/0020_add_rescheduled_session_name.py deleted file mode 100644 index d86c5237fc..0000000000 --- a/ietf/name/migrations/0020_add_rescheduled_session_name.py +++ /dev/null @@ -1,24 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0019_add_timeslottypename_private'), - ] - - def add_rescheduled_session_status_name(apps, schema_editor): - SessionStatusName = apps.get_model('name', 'SessionStatusName') - SessionStatusName.objects.get_or_create( - slug='resched', - name="Rescheduled", - ) - - def noop(apps, schema_editor): - pass - - operations = [ - migrations.RunPython(add_rescheduled_session_status_name, noop, elidable=True), - ] diff --git a/ietf/name/migrations/0021_add_ad_appr_draftsubmissionstatename.py b/ietf/name/migrations/0021_add_ad_appr_draftsubmissionstatename.py deleted file mode 100644 index 825c93e6a0..0000000000 --- a/ietf/name/migrations/0021_add_ad_appr_draftsubmissionstatename.py +++ /dev/null @@ -1,70 +0,0 @@ -# Generated by Django 2.2.17 on 2020-11-18 07:41 - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0020_add_rescheduled_session_name'), - ] - - def up(apps, schema_editor): - DraftSubmissionStateName = apps.get_model('name', 'DraftSubmissionStateName') - new_state_name = DraftSubmissionStateName.objects.create( - slug='ad-appr', - name='Awaiting AD Approval', - used=True, - ) - new_state_name.next_states.add(DraftSubmissionStateName.objects.get(slug='posted')) - new_state_name.next_states.add(DraftSubmissionStateName.objects.get(slug='cancel')) - DraftSubmissionStateName.objects.get(slug='uploaded').next_states.add(new_state_name) - - # Order so the '-appr' states are together - for slug, order in ( - ('confirmed', 0), - ('uploaded', 1), - ('auth', 2), - ('aut-appr', 3), - ('grp-appr', 4), - ('ad-appr', 5), - ('manual', 6), - ('cancel', 7), - ('posted', 8), - ('waiting-for-draft', 9), - ): - state_name = DraftSubmissionStateName.objects.get(slug=slug) - state_name.order = order - state_name.save() - - def down(apps, schema_editor): - DraftSubmissionStateName = apps.get_model('name', 'DraftSubmissionStateName') - Submission = apps.get_model('submit', 'Submission') - - name_to_delete = DraftSubmissionStateName.objects.get(slug='ad-appr') - - # Refuse to migrate if there are any Submissions using the state we're about to remove - assert(Submission.objects.filter(state=name_to_delete).count() == 0) - - DraftSubmissionStateName.objects.get(slug='uploaded').next_states.remove(name_to_delete) - name_to_delete.delete() - - # restore original order - for slug, order in ( - ('confirmed', 0), - ('uploaded', 1), - ('auth', 2), - ('aut-appr', 3), - ('grp-appr', 4), - ('manual', 5), - ('cancel', 6), - ('posted', 7), - ('waiting-for-draft', 8), - ): - state_name = DraftSubmissionStateName.objects.get(slug=slug) - state_name.order = order - state_name.save() - - operations = [ - migrations.RunPython(up, down), - ] diff --git a/ietf/name/migrations/0022_add_liaison_contact_rolenames.py b/ietf/name/migrations/0022_add_liaison_contact_rolenames.py deleted file mode 100644 index e5edb23998..0000000000 --- a/ietf/name/migrations/0022_add_liaison_contact_rolenames.py +++ /dev/null @@ -1,41 +0,0 @@ -# Generated by Django 2.2.17 on 2020-12-09 07:02 - -from django.db import migrations - - -def forward(apps, schema_editor): - """Perform forward migration - - Adds RoleNames for liaison contacts - """ - RoleName = apps.get_model('name', 'RoleName') - RoleName.objects.create( - slug='liaison_contact', - name='Liaison Contact', - ) - RoleName.objects.create( - slug='liaison_cc_contact', - name='Liaison CC Contact', - ) - - -def reverse(apps, schema_editor): - """Perform reverse migration - - Removes RoleNames for liaison contacts - """ - RoleName = apps.get_model('name', 'RoleName') - RoleName.objects.filter( - slug__in=['liaison_contact', 'liaison_cc_contact'] - ).delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0021_add_ad_appr_draftsubmissionstatename'), - ('group', '0040_lengthen_used_roles_fields'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/name/migrations/0023_change_stream_descriptions.py b/ietf/name/migrations/0023_change_stream_descriptions.py deleted file mode 100644 index e6e578e229..0000000000 --- a/ietf/name/migrations/0023_change_stream_descriptions.py +++ /dev/null @@ -1,45 +0,0 @@ -# Copyright The IETF Trust 2019-2021, All Rights Reserved -# -*- coding: utf-8 -*- - -from django.db import migrations - - -def forward(apps, schema_editor): - # Change the StreamName descriptions to match what appears in modern RFCs - StreamName = apps.get_model('name', 'StreamName') - for streamName in StreamName.objects.all(): - if streamName.name == "IETF": - streamName.desc = "Internent Engineering Task Force (IETF)" - elif streamName.name == "IRTF": - streamName.desc = "Internet Research Task Force (IRTF)" - elif streamName.name == "IAB": - streamName.desc = "Internet Architecture Board (IAB)" - elif streamName.name == "ISE": - streamName.desc = "Independent Submission" - streamName.save() - - -def reverse(apps, schema_editor): - StreamName = apps.get_model('name', 'StreamName') - for streamName in StreamName.objects.all(): - if streamName.name == "IETF": - streamName.desc = "IETF stream" - elif streamName.name == "IRTF": - streamName.desc = "IRTF Stream" - elif streamName.name == "IAB": - streamName.desc = "IAB stream" - elif streamName.name == "ISE": - streamName.desc = "Independent Submission Editor stream" - streamName.save() - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0022_add_liaison_contact_rolenames'), - ] - - operations = [ - migrations.RunPython(forward,reverse), - - ] - diff --git a/ietf/name/migrations/0024_constraintname_is_group_conflict.py b/ietf/name/migrations/0024_constraintname_is_group_conflict.py deleted file mode 100644 index 47810c6da0..0000000000 --- a/ietf/name/migrations/0024_constraintname_is_group_conflict.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.2.20 on 2021-05-19 09:45 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - dependencies = [ - ('name', '0023_change_stream_descriptions'), - ] - - operations = [ - migrations.AddField( - model_name='constraintname', - name='is_group_conflict', - field=models.BooleanField(default=False, - help_text='Does this constraint capture a conflict between groups?'), - ), - ] diff --git a/ietf/name/migrations/0025_set_constraintname_is_group_conflict.py b/ietf/name/migrations/0025_set_constraintname_is_group_conflict.py deleted file mode 100644 index 86deb9bd2d..0000000000 --- a/ietf/name/migrations/0025_set_constraintname_is_group_conflict.py +++ /dev/null @@ -1,25 +0,0 @@ -# Generated by Django 2.2.20 on 2021-05-19 09:55 - -from django.db import migrations - - -def forward(apps, schema_editor): - """Set is_group_conflict for ConstraintNames that need it to be True""" - ConstraintName = apps.get_model('name', 'ConstraintName') - ConstraintName.objects.filter( - slug__in=['conflict', 'conflic2', 'conflic3'] - ).update(is_group_conflict=True) - - -def reverse(apps, schema_editor): - pass # nothing to be done - - -class Migration(migrations.Migration): - dependencies = [ - ('name', '0024_constraintname_is_group_conflict'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/name/migrations/0026_add_conflict_constraintnames.py b/ietf/name/migrations/0026_add_conflict_constraintnames.py deleted file mode 100644 index d404e2cf30..0000000000 --- a/ietf/name/migrations/0026_add_conflict_constraintnames.py +++ /dev/null @@ -1,75 +0,0 @@ -# Generated by Django 2.2.20 on 2021-05-05 10:05 - -from collections import namedtuple -from django.db import migrations -from django.db.models import Max - - -# Simple type for representing constraint name data that will be -# modified. -ConstraintInfo = namedtuple( - 'ConstraintInfo', - ['replaces', 'slug', 'name', 'desc', 'editor_label'], -) - -constraint_names_to_add = [ - ConstraintInfo( - replaces='conflict', - slug='chair_conflict', - name='Chair conflict', - desc='Indicates other WGs the chairs also lead or will be active participants in', - editor_label='', - ), - ConstraintInfo( - replaces='conflic2', - slug='tech_overlap', - name='Technology overlap', - desc='Indicates WGs with a related technology or a closely related charter', - editor_label='', - ), - ConstraintInfo( - replaces='conflic3', - slug='key_participant', - name='Key participant conflict', - desc='Indicates WGs with which key participants (presenter, secretary, etc.) may overlap', - editor_label='', - ) -] - - -def forward(apps, schema_editor): - ConstraintName = apps.get_model('name', 'ConstraintName') - max_order = ConstraintName.objects.all().aggregate(Max('order'))['order__max'] - - for index, new_constraint in enumerate(constraint_names_to_add): - # hack_constraint is the constraint type relabeled by the hack fix in #2754 - hack_constraint = ConstraintName.objects.get(slug=new_constraint.replaces) - ConstraintName.objects.create( - slug=new_constraint.slug, - name=new_constraint.name, - desc=new_constraint.desc, - used=hack_constraint.used, - order=max_order + index + 1, - penalty=hack_constraint.penalty, - editor_label=new_constraint.editor_label, - is_group_conflict=True, - ) - - -def reverse(apps, schema_editor): - ConstraintName = apps.get_model('name', 'ConstraintName') - for new_constraint in constraint_names_to_add: - ConstraintName.objects.filter(slug=new_constraint.slug).delete() - -class Migration(migrations.Migration): - dependencies = [ - ('name', '0025_set_constraintname_is_group_conflict'), - # Reversing this migration requires that the 'day' field be removed from - # the Constraint model, so we indirectly depend on the migration that - # removed it. - ('meeting', '0027_add_constraint_options_and_joint_groups'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] \ No newline at end of file diff --git a/ietf/name/migrations/0027_add_bofrequest.py b/ietf/name/migrations/0027_add_bofrequest.py deleted file mode 100644 index c4e31bb2db..0000000000 --- a/ietf/name/migrations/0027_add_bofrequest.py +++ /dev/null @@ -1,23 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -# Generated by Django 2.2.23 on 2021-05-21 12:48 - -from django.db import migrations - -def forward(apps,schema_editor): - DocTypeName = apps.get_model('name','DocTypeName') - DocTypeName.objects.create(prefix='bofreq', slug='bofreq', name="BOF Request", desc="", used=True, order=0) - -def reverse(apps,schema_editor): - DocTypeName = apps.get_model('name','DocTypeName') - DocTypeName.objects.filter(slug='bofreq').delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0026_add_conflict_constraintnames'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/name/migrations/0028_iabasg.py b/ietf/name/migrations/0028_iabasg.py deleted file mode 100644 index f339341ea9..0000000000 --- a/ietf/name/migrations/0028_iabasg.py +++ /dev/null @@ -1,22 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -from django.db import migrations - -def forward(apps, schema_editor): - GroupTypeName = apps.get_model('name', 'GroupTypeName') - GroupTypeName.objects.create(slug='iabasg', name='IAB ASG', verbose_name='IAB Administrative Support Group', desc='') - -def reverse(apps, schema_editor): - GroupTypeName = apps.get_model('name', 'GroupTypeName') - GroupTypeName.objects.filter(slug='iabasg').delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0027_add_bofrequest'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/name/migrations/0029_proceedingsmaterialtypename.py b/ietf/name/migrations/0029_proceedingsmaterialtypename.py deleted file mode 100644 index 97606c17b3..0000000000 --- a/ietf/name/migrations/0029_proceedingsmaterialtypename.py +++ /dev/null @@ -1,29 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -# Generated by Django 2.2.24 on 2021-07-26 16:42 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0028_iabasg'), - ] - - operations = [ - migrations.CreateModel( - name='ProceedingsMaterialTypeName', - fields=[ - ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), - ('name', models.CharField(max_length=255)), - ('desc', models.TextField(blank=True)), - ('used', models.BooleanField(default=True)), - ('order', models.IntegerField(default=0)), - ], - options={ - 'ordering': ['order', 'name'], - 'abstract': False, - }, - ), - ] diff --git a/ietf/name/migrations/0030_populate_proceedingsmaterialtypename.py b/ietf/name/migrations/0030_populate_proceedingsmaterialtypename.py deleted file mode 100644 index 3e3936a50d..0000000000 --- a/ietf/name/migrations/0030_populate_proceedingsmaterialtypename.py +++ /dev/null @@ -1,34 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -# Generated by Django 2.2.24 on 2021-07-26 16:55 - -from django.db import migrations - - -def forward(apps, schema_editor): - ProceedingsMaterialTypeName = apps.get_model('name', 'ProceedingsMaterialTypeName') - names = [ - {'slug': 'supporters', 'name': 'Sponsors and Supporters', 'desc': 'Sponsors and supporters', 'order': 0}, - {'slug': 'host_speaker_series', 'name': 'Host Speaker Series', 'desc': 'Host speaker series', 'order': 1}, - {'slug': 'social_event', 'name': 'Social Event', 'desc': 'Social event', 'order': 2}, - {'slug': 'wiki', 'name': 'Meeting Wiki', 'desc': 'Meeting wiki', 'order': 3}, - {'slug': 'additional_information', 'name': 'Additional Information', 'desc': 'Any other materials', 'order': 4}, - ] - for name in names: - ProceedingsMaterialTypeName.objects.create(used=True, **name) - - -def reverse(apps, schema_editor): - pass - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0029_proceedingsmaterialtypename'), - ('meeting', '0046_meetinghost'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/name/migrations/0031_add_procmaterials.py b/ietf/name/migrations/0031_add_procmaterials.py deleted file mode 100644 index 3ceb575a99..0000000000 --- a/ietf/name/migrations/0031_add_procmaterials.py +++ /dev/null @@ -1,42 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -# Generated by Django 2.2.24 on 2021-07-27 08:04 - -from django.db import migrations - - -def forward(apps, schema_editor): - DocTypeName = apps.get_model('name', 'DocTypeName') - DocTypeName.objects.create( - prefix='proc-materials', - slug='procmaterials', - name="Proceedings Materials", - desc="", - used=True, - order=0 - ) - - -def reverse(apps, schema_editor): - DocTypeName = apps.get_model('name', 'DocTypeName') - DocTypeName.objects.filter(slug='procmaterials').delete() - - -class Migration(migrations.Migration): - # Most of these dependencies are needed to permit the reverse migration - # to work. Without them Django does not replay enough migrations and during - # migration believes that there are foreign key references to the old - # PK (name) on the Document model. - dependencies = [ - ('doc', '0043_bofreq_docevents'), - ('group', '0044_populate_groupfeatures_parent_type_fields'), - ('liaisons', '0006_document_primary_key_cleanup'), - ('meeting', '0018_document_primary_key_cleanup'), - ('name', '0030_populate_proceedingsmaterialtypename'), - ('review', '0014_document_primary_key_cleanup'), - ('submit', '0008_submissionextresource'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/name/migrations/0032_agendafiltertypename.py b/ietf/name/migrations/0032_agendafiltertypename.py deleted file mode 100644 index 6c6fa4eab2..0000000000 --- a/ietf/name/migrations/0032_agendafiltertypename.py +++ /dev/null @@ -1,27 +0,0 @@ -# Generated by Django 2.2.19 on 2021-04-02 12:03 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0031_add_procmaterials'), - ] - - operations = [ - migrations.CreateModel( - name='AgendaFilterTypeName', - fields=[ - ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), - ('name', models.CharField(max_length=255)), - ('desc', models.TextField(blank=True)), - ('used', models.BooleanField(default=True)), - ('order', models.IntegerField(default=0)), - ], - options={ - 'ordering': ['order', 'name'], - 'abstract': False, - }, - ), - ] diff --git a/ietf/name/migrations/0033_populate_agendafiltertypename.py b/ietf/name/migrations/0033_populate_agendafiltertypename.py deleted file mode 100644 index 9f450ca24d..0000000000 --- a/ietf/name/migrations/0033_populate_agendafiltertypename.py +++ /dev/null @@ -1,33 +0,0 @@ -# Generated by Django 2.2.20 on 2021-04-20 13:56 - -from django.db import migrations - - -def forward(apps, schema_editor): - AgendaFilterTypeName = apps.get_model('name', 'AgendaFilterTypeName') - names = ( - ('none', 'None', 'Not used except for a timeslot-type column (e.g., officehours)'), - ('normal', 'Normal', 'Non-heading filter button'), - ('heading', 'Heading', 'Column heading button'), - ('special', 'Special', 'Button in the catch-all column'), - ) - for order, (slug, name, desc) in enumerate(names): - AgendaFilterTypeName.objects.get_or_create( - slug=slug, - defaults=dict(name=name, desc=desc, order=order, used=True) - ) - - -def reverse(apps, schema_editor): - pass # nothing to do, model about to be destroyed anyway - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0032_agendafiltertypename'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/name/migrations/0034_sessionpurposename.py b/ietf/name/migrations/0034_sessionpurposename.py deleted file mode 100644 index ca22e1a155..0000000000 --- a/ietf/name/migrations/0034_sessionpurposename.py +++ /dev/null @@ -1,33 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -# Generated by Django 2.2.24 on 2021-09-16 09:42 - -from django.db import migrations, models -import ietf.name.models -import jsonfield - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0033_populate_agendafiltertypename'), - ] - - operations = [ - migrations.CreateModel( - name='SessionPurposeName', - fields=[ - ('slug', models.CharField(max_length=32, primary_key=True, serialize=False)), - ('name', models.CharField(max_length=255)), - ('desc', models.TextField(blank=True)), - ('used', models.BooleanField(default=True)), - ('order', models.IntegerField(default=0)), - ('timeslot_types', jsonfield.fields.JSONField(default=[], help_text='Allowed TimeSlotTypeNames', max_length=256, validators=[ietf.name.models.JSONForeignKeyListValidator('name.TimeSlotTypeName')])), - ('on_agenda', models.BooleanField(default=True, help_text='Are sessions of this purpose visible on the agenda by default?')), - ], - options={ - 'ordering': ['order', 'name'], - 'abstract': False, - }, - ), - ] diff --git a/ietf/name/migrations/0035_populate_sessionpurposename.py b/ietf/name/migrations/0035_populate_sessionpurposename.py deleted file mode 100644 index 8af7d60d1e..0000000000 --- a/ietf/name/migrations/0035_populate_sessionpurposename.py +++ /dev/null @@ -1,53 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved - -# Generated by Django 2.2.24 on 2021-09-16 09:42 - -from django.db import migrations - - -def forward(apps, schema_editor): - SessionPurposeName = apps.get_model('name', 'SessionPurposeName') - TimeSlotTypeName = apps.get_model('name', 'TimeSlotTypeName') - - for order, (slug, name, desc, tstypes, on_agenda, used) in enumerate(( - ('none', 'None', 'Value not set (do not use for new sessions)', [], True, False), - ('regular', 'Regular', 'Regular group session', ['regular'], True, True), - ('tutorial', 'Tutorial', 'Tutorial or training session', ['other'], True, True), - ('officehours', 'Office hours', 'Office hours session', ['other'], True, True), - ('coding', 'Coding', 'Coding session', ['other'], True, True), - ('admin', 'Administrative', 'Meeting administration', ['other', 'reg'], True, True), - ('social', 'Social', 'Social event or activity', ['break', 'other'], True, True), - ('plenary', 'Plenary', 'Plenary session', ['plenary'], True, True), - ('presentation', 'Presentation', 'Presentation session', ['other', 'regular'], True, True), - ('open_meeting', 'Open meeting', 'Open meeting', ['other'], True, True), - ('closed_meeting', 'Closed meeting', 'Closed meeting', ['other', 'regular'], False, True), - )): - # verify that we're not about to use an invalid type - for ts_type in tstypes: - TimeSlotTypeName.objects.get(pk=ts_type) # throws an exception unless exists - - SessionPurposeName.objects.create( - slug=slug, - name=name, - desc=desc, - used=used, - order=order, - timeslot_types = tstypes, - on_agenda=on_agenda, - ) - - -def reverse(apps, schema_editor): - SessionPurposeName = apps.get_model('name', 'SessionPurposeName') - SessionPurposeName.objects.all().delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0034_sessionpurposename'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/name/migrations/0036_depopulate_timeslottypename_private.py b/ietf/name/migrations/0036_depopulate_timeslottypename_private.py deleted file mode 100644 index 352ab8d582..0000000000 --- a/ietf/name/migrations/0036_depopulate_timeslottypename_private.py +++ /dev/null @@ -1,31 +0,0 @@ -# Generated by Django 2.2.24 on 2021-10-25 16:58 - -from django.db import migrations - - -PRIVATE_TIMESLOT_SLUGS = {'lead', 'offagenda'} # from DB 2021 Oct - - -def forward(apps, schema_editor): - TimeSlotTypeName = apps.get_model('name', 'TimeSlotTypeName') - slugs = TimeSlotTypeName.objects.filter(private=True).values_list('slug', flat=True) - if set(slugs) != PRIVATE_TIMESLOT_SLUGS: - # the reverse migration will not restore the database, refuse to migrate - raise ValueError('Disagreement between migration data and database') - - -def reverse(apps, schema_editor): - TimeSlotTypeName = apps.get_model('name', 'TimeSlotTypeName') - TimeSlotTypeName.objects.filter(slug__in=PRIVATE_TIMESLOT_SLUGS).update(private=True) - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0035_populate_sessionpurposename'), - ('meeting', '0051_populate_session_on_agenda'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/name/migrations/0037_remove_timeslottypename_private.py b/ietf/name/migrations/0037_remove_timeslottypename_private.py deleted file mode 100644 index 2a86780564..0000000000 --- a/ietf/name/migrations/0037_remove_timeslottypename_private.py +++ /dev/null @@ -1,17 +0,0 @@ -# Generated by Django 2.2.24 on 2021-10-25 17:23 - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0036_depopulate_timeslottypename_private'), - ] - - operations = [ - migrations.RemoveField( - model_name='timeslottypename', - name='private', - ), - ] diff --git a/ietf/name/migrations/0038_disuse_offagenda_and_reserved.py b/ietf/name/migrations/0038_disuse_offagenda_and_reserved.py deleted file mode 100644 index be0b507bde..0000000000 --- a/ietf/name/migrations/0038_disuse_offagenda_and_reserved.py +++ /dev/null @@ -1,24 +0,0 @@ -# Generated by Django 2.2.24 on 2021-10-29 06:44 - -from django.db import migrations - - -def forward(apps, schema_editor): - TimeSlotTypeName = apps.get_model('name', 'TimeSlotTypeName') - TimeSlotTypeName.objects.filter(slug__in=('offagenda', 'reserved')).update(used=False) - - -def reverse(apps, schema_editor): - TimeSlotTypeName = apps.get_model('name', 'TimeSlotTypeName') - TimeSlotTypeName.objects.filter(slug__in=('offagenda', 'reserved')).update(used=True) - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0037_remove_timeslottypename_private'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/name/migrations/0039_depopulate_constraintname_editor_label.py b/ietf/name/migrations/0039_depopulate_constraintname_editor_label.py deleted file mode 100644 index 9949002bec..0000000000 --- a/ietf/name/migrations/0039_depopulate_constraintname_editor_label.py +++ /dev/null @@ -1,39 +0,0 @@ -# Generated by Django 2.2.27 on 2022-03-11 07:55 - -from django.db import migrations - - -def forward(apps, schema_editor): - pass # nothing to do, row will be dropped - - -def reverse(apps, schema_editor): - # Restore data from when this migration was written. Dumped with: - # >>> from pprint import pp - # >>> from ietf.name.models import ConstraintName - # >>> pp(list(ConstraintName.objects.values_list('slug', 'editor_label'))) - ConstraintName = apps.get_model('name', 'ConstraintName') - for slug, editor_label in [ - ('conflict', '1'), - ('conflic2', '2'), - ('conflic3', '3'), - ('bethere', '{count}'), - ('timerange', ''), - ('time_relation', 'Δ'), - ('wg_adjacent', ''), - ('chair_conflict', ''), - ('tech_overlap', ''), - ('key_participant', ''), - ]: - ConstraintName.objects.filter(slug=slug).update(editor_label=editor_label) - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0038_disuse_offagenda_and_reserved'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/name/migrations/0040_remove_constraintname_editor_label.py b/ietf/name/migrations/0040_remove_constraintname_editor_label.py deleted file mode 100644 index 85390c77a6..0000000000 --- a/ietf/name/migrations/0040_remove_constraintname_editor_label.py +++ /dev/null @@ -1,17 +0,0 @@ -# Generated by Django 2.2.27 on 2022-03-11 10:05 - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0039_depopulate_constraintname_editor_label'), - ] - - operations = [ - migrations.RemoveField( - model_name='constraintname', - name='editor_label', - ), - ] diff --git a/ietf/name/migrations/0041_update_rfcedtyp.py b/ietf/name/migrations/0041_update_rfcedtyp.py deleted file mode 100644 index 8ebae6aa11..0000000000 --- a/ietf/name/migrations/0041_update_rfcedtyp.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2022 All Rights Reserved - -from django.db import migrations - -def forward(apps, schema_editor): - GroupTypeName = apps.get_model('name', 'GroupTypeName') - GroupTypeName.objects.filter(slug='rfcedtyp').update(order=2, verbose_name='RFC Editor Group') - -def reverse(apps, schema_editor): - GroupTypeName = apps.get_model('name', 'GroupTypeName') - GroupTypeName.objects.filter(slug='rfcedtyp').update(order=0, verbose_name='The RFC Editor') - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0040_remove_constraintname_editor_label'), - ] - - operations = [ - migrations.RunPython(forward,reverse), - ] diff --git a/ietf/name/migrations/0042_editorial_stream.py b/ietf/name/migrations/0042_editorial_stream.py deleted file mode 100644 index 1b9aaaf076..0000000000 --- a/ietf/name/migrations/0042_editorial_stream.py +++ /dev/null @@ -1,30 +0,0 @@ -# Copyright The IETF Trust 2022 All Rights Reserved - -from django.db import migrations - -def forward(apps, schema_editor): - StreamName = apps.get_model('name', 'StreamName') - StreamName.objects.create( - slug = 'editorial', - name = 'Editorial', - desc = 'Editorial', - used = True, - order = 5, - ) - StreamName.objects.filter(slug='legacy').update(order=6) - - -def reverse(apps, schema_editor): - StreamName = apps.get_model('name', 'StreamName') - StreamName.objects.filter(slug='editorial').delete() - StreamName.objects.filter(slug='legacy').update(order=5) - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0041_update_rfcedtyp'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/name/migrations/0043_editorial_stream_grouptype.py b/ietf/name/migrations/0043_editorial_stream_grouptype.py deleted file mode 100644 index 5fe839174f..0000000000 --- a/ietf/name/migrations/0043_editorial_stream_grouptype.py +++ /dev/null @@ -1,26 +0,0 @@ -# Copyright The IETF Trust 2022 All Rights Reserved - -from django.db import migrations - -def forward(apps, schema_editor): - GroupTypeName = apps.get_model('name', 'GroupTypeName') - GroupTypeName.objects.create( - slug = 'editorial', - name = 'Editorial', - desc = 'Editorial Stream Group', - used = True, - ) - -def reverse(apps, schema_editor): - GroupTypeName = apps.get_model('name', 'GroupTypeName') - GroupTypeName.objects.filter(slug='editorial').delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0042_editorial_stream'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/name/migrations/0044_validating_draftsubmissionstatename.py b/ietf/name/migrations/0044_validating_draftsubmissionstatename.py deleted file mode 100644 index de82bbeef9..0000000000 --- a/ietf/name/migrations/0044_validating_draftsubmissionstatename.py +++ /dev/null @@ -1,40 +0,0 @@ -# Generated by Django 2.2.28 on 2022-05-17 11:35 - -from django.db import migrations - - -def forward(apps, schema_editor): - DraftSubmissionStateName = apps.get_model('name', 'DraftSubmissionStateName') - new_state = DraftSubmissionStateName.objects.create( - slug='validating', - name='Validating Submitted Draft', - desc='Running validation checks on received submission', - used=True, - order=1 + DraftSubmissionStateName.objects.order_by('-order').first().order, - ) - new_state.next_states.set( - DraftSubmissionStateName.objects.filter( - slug__in=['cancel', 'uploaded'], - ) - ) - - -def reverse(apps, schema_editor): - Submission = apps.get_model('submit', 'Submission') - # Any submissions in the state we are about to delete would be deleted. - # Remove these manually if you really mean to do this. - assert Submission.objects.filter(state__slug='validating').count() == 0 - DraftSubmissionStateName = apps.get_model('name', 'DraftSubmissionStateName') - DraftSubmissionStateName.objects.filter(slug='validating').delete() - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0043_editorial_stream_grouptype'), - ('submit', '0001_initial'), # ensure Submission model exists - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/name/migrations/0045_polls_and_chatlogs.py b/ietf/name/migrations/0045_polls_and_chatlogs.py deleted file mode 100644 index 1014a9dcef..0000000000 --- a/ietf/name/migrations/0045_polls_and_chatlogs.py +++ /dev/null @@ -1,35 +0,0 @@ -# Copyright The IETF Trust 2022, All Rights Reserved -from django.db import migrations - -def forward(apps, schema_editor): - DocTypeName = apps.get_model("name", "DocTypeName") - DocTypeName.objects.create( - slug = "chatlog", - name = "Chat Log", - prefix = "chatlog", - desc = "", - order = 0, - used = True, - ) - DocTypeName.objects.create( - slug = "polls", - name = "Polls", - prefix = "polls", - desc = "", - order = 0, - used = True, - ) - -def reverse(apps, schema_editor): - DocTypeName = apps.get_model("name", "DocTypeName") - DocTypeName.objects.filter(slug__in=("chatlog", "polls")).delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0044_validating_draftsubmissionstatename'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/name/models.py b/ietf/name/models.py index f0d56523d8..24104c5f45 100644 --- a/ietf/name/models.py +++ b/ietf/name/models.py @@ -1,8 +1,6 @@ # Copyright The IETF Trust 2010-2020, All Rights Reserved # -*- coding: utf-8 -*- -import jsonfield - from django.db import models from ietf.utils.models import ForeignKey @@ -42,8 +40,8 @@ class DocRelationshipName(NameModel): class DocTypeName(NameModel): """Draft, Agenda, Minutes, Charter, Discuss, Guideline, Email, - Review, Issue, Wiki""" - prefix = models.CharField(max_length=16, default="") + Review, Issue, Wiki, RFC""" + prefix = models.CharField(max_length=32, default="") class DocTagName(NameModel): """Waiting for Reference, IANA Coordination, Revised ID Needed, External Party, AD Followup, Point Raised - Writeup Needed, ...""" @@ -73,8 +71,8 @@ class SessionStatusName(NameModel): """Waiting for Approval, Approved, Waiting for Scheduling, Scheduled, Cancelled, Disapproved""" class SessionPurposeName(NameModel): """Regular, Tutorial, Office Hours, Coding, Social, Admin""" - timeslot_types = jsonfield.JSONField( - max_length=256, blank=False, default=[], + timeslot_types = models.JSONField( + max_length=256, blank=False, default=list, help_text='Allowed TimeSlotTypeNames', validators=[JSONForeignKeyListValidator('name.TimeSlotTypeName')], ) @@ -94,13 +92,14 @@ class NomineePositionStateName(NameModel): """Status of a candidate for a position: None, Accepted, Declined""" class FeedbackTypeName(NameModel): """Type of feedback: questionnaires, nominations, comments""" + legend = models.CharField(max_length=1, default="", help_text="One-character legend for feedback classification form") class DBTemplateTypeName(NameModel): """reStructuredText, Plain, Django""" class DraftSubmissionStateName(NameModel): """Uploaded, Awaiting Submitter Authentication, Awaiting Approval from Previous Version Authors, Awaiting Initial Version Approval, Awaiting Manual Post, Cancelled, Posted""" - next_states = models.ManyToManyField('DraftSubmissionStateName', related_name="previous_states", blank=True) + next_states = models.ManyToManyField('name.DraftSubmissionStateName', related_name="previous_states", blank=True) class RoomResourceName(NameModel): "Room resources: Audio Stream, Meetecho, . . ." class IprDisclosureStateName(NameModel): @@ -148,3 +147,11 @@ class ExtResourceName(NameModel): type = ForeignKey(ExtResourceTypeName) class SlideSubmissionStatusName(NameModel): "Pending, Accepted, Rejected" +class TelechatAgendaSectionName(NameModel): + """roll_call, minutes, action_items""" +class AppealArtifactTypeName(NameModel): + pass +class AttendanceTypeName(NameModel): + """onsite, remote, hackathon_onsite, hackathon_remote""" +class RegistrationTicketTypeName(NameModel): + """week, one_day, student""" diff --git a/ietf/name/resources.py b/ietf/name/resources.py index 764c5e1089..0cb0e41e0b 100644 --- a/ietf/name/resources.py +++ b/ietf/name/resources.py @@ -18,7 +18,8 @@ ReviewAssignmentStateName, ReviewRequestStateName, ReviewResultName, ReviewTypeName, RoleName, RoomResourceName, SessionStatusName, StdLevelName, StreamName, TimeSlotTypeName, TopicAudienceName, ReviewerQueuePolicyName, TimerangeName, ExtResourceTypeName, ExtResourceName, - SlideSubmissionStatusName, ProceedingsMaterialTypeName, SessionPurposeName ) + SlideSubmissionStatusName, ProceedingsMaterialTypeName, SessionPurposeName, TelechatAgendaSectionName, + AppealArtifactTypeName, AttendanceTypeName, RegistrationTicketTypeName ) class TimeSlotTypeNameResource(ModelResource): class Meta: @@ -720,3 +721,64 @@ class Meta: "on_agenda": ALL, } api.name.register(SessionPurposeNameResource()) + + +class TelechatAgendaSectionNameResource(ModelResource): + class Meta: + queryset = TelechatAgendaSectionName.objects.all() + serializer = api.Serializer() + cache = SimpleCache() + #resource_name = 'telechatagendasectionname' + ordering = ['slug', ] + filtering = { + "slug": ALL, + "name": ALL, + "desc": ALL, + "used": ALL, + "order": ALL, + } +api.name.register(TelechatAgendaSectionNameResource()) + +class AppealArtifactTypeNameResource(ModelResource): + class Meta: + cache = SimpleCache() + queryset = AppealArtifactTypeName.objects.all() + serializer = api.Serializer() + filtering = { + "slug": ALL, + "name": ALL, + "desc": ALL, + "used": ALL, + "order": ALL, + } +api.name.register(AppealArtifactTypeNameResource()) + + +class AttendanceTypeNameResource(ModelResource): + class Meta: + cache = SimpleCache() + queryset = AttendanceTypeName.objects.all() + serializer = api.Serializer() + filtering = { + "slug": ALL, + "name": ALL, + "desc": ALL, + "used": ALL, + "order": ALL, + } +api.name.register(AttendanceTypeNameResource()) + + +class RegistrationTicketTypeNameResource(ModelResource): + class Meta: + cache = SimpleCache() + queryset = RegistrationTicketTypeName.objects.all() + serializer = api.Serializer() + filtering = { + "slug": ALL, + "name": ALL, + "desc": ALL, + "used": ALL, + "order": ALL, + } +api.name.register(RegistrationTicketTypeNameResource()) diff --git a/ietf/name/serializers.py b/ietf/name/serializers.py new file mode 100644 index 0000000000..a764f56051 --- /dev/null +++ b/ietf/name/serializers.py @@ -0,0 +1,11 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +"""django-rest-framework serializers""" +from rest_framework import serializers + +from .models import StreamName + + +class StreamNameSerializer(serializers.ModelSerializer): + class Meta: + model = StreamName + fields = ["slug", "name", "desc"] diff --git a/ietf/nomcom/admin.py b/ietf/nomcom/admin.py index 4b18cc005c..1675d731a5 100644 --- a/ietf/nomcom/admin.py +++ b/ietf/nomcom/admin.py @@ -21,9 +21,9 @@ class NomComAdmin(admin.ModelAdmin): admin.site.register(NomCom, NomComAdmin) class NominationAdmin(admin.ModelAdmin): - list_display = ['id', 'position', 'candidate_name', 'candidate_email', 'candidate_phone', 'nominee', 'comments', 'nominator_email', 'user', 'time', 'share_nominator'] + list_display = ['id', 'position', 'candidate_name', 'candidate_email', 'candidate_phone', 'nominee', 'comments', 'nominator_email', 'person', 'time', 'share_nominator'] list_filter = ['time', 'share_nominator'] - raw_id_fields = ['nominee', 'comments', 'user'] + raw_id_fields = ['nominee', 'comments', 'person'] admin.site.register(Nomination, NominationAdmin) class NomineeAdmin(admin.ModelAdmin): @@ -51,9 +51,9 @@ def nominee(self, obj): return ", ".join(n.person.ascii for n in obj.nominees.all()) nominee.admin_order_field = 'nominees__person__ascii' # type: ignore # https://github.com/python/mypy/issues/2087 - list_display = ['id', 'nomcom', 'author', 'nominee', 'subject', 'type', 'user', 'time'] + list_display = ['id', 'nomcom', 'author', 'nominee', 'subject', 'type', 'person', 'time'] list_filter = ['nomcom', 'type', 'time', ] - raw_id_fields = ['positions', 'topics', 'user'] + raw_id_fields = ['positions', 'topics', 'person'] admin.site.register(Feedback, FeedbackAdmin) diff --git a/ietf/nomcom/decorators.py b/ietf/nomcom/decorators.py index a002f7c7e0..43250bd306 100644 --- a/ietf/nomcom/decorators.py +++ b/ietf/nomcom/decorators.py @@ -3,10 +3,11 @@ import functools +from urllib.parse import quote as urlquote from django.urls import reverse from django.http import HttpResponseRedirect -from django.utils.http import urlquote + def nomcom_private_key_required(view_func): diff --git a/ietf/nomcom/factories.py b/ietf/nomcom/factories.py index 8ef4e07faa..286e0229ab 100644 --- a/ietf/nomcom/factories.py +++ b/ietf/nomcom/factories.py @@ -9,7 +9,7 @@ from ietf.nomcom.models import NomCom, Position, Feedback, Nominee, NomineePosition, Nomination, Topic from ietf.group.factories import GroupFactory -from ietf.person.factories import PersonFactory, UserFactory +from ietf.person.factories import PersonFactory import debug # pyflakes:ignore @@ -66,7 +66,7 @@ def provide_private_key_to_test_client(testcase): session = testcase.client.session - session['NOMCOM_PRIVATE_KEY_%s'%testcase.nc.year()] = key + session['NOMCOM_PRIVATE_KEY_%s'%testcase.nc.year()] = key.decode("utf8") session.save() def nomcom_kwargs_for_year(year=None, *args, **kwargs): @@ -84,6 +84,7 @@ def nomcom_kwargs_for_year(year=None, *args, **kwargs): class NomComFactory(factory.django.DjangoModelFactory): class Meta: model = NomCom + skip_postgeneration_save = True group = factory.SubFactory(GroupFactory,type_id='nomcom') @@ -167,6 +168,7 @@ class Meta: class FeedbackFactory(factory.django.DjangoModelFactory): class Meta: model = Feedback + skip_postgeneration_save = True nomcom = factory.SubFactory(NomComFactory) subject = factory.Faker('sentence') @@ -176,6 +178,7 @@ class Meta: def comments(obj, create, extracted, **kwargs): comment_text = Faker().paragraph() obj.comments = obj.nomcom.encrypt(comment_text) + obj.save() class TopicFactory(factory.django.DjangoModelFactory): class Meta: @@ -196,7 +199,7 @@ class Meta: candidate_email = factory.LazyAttribute(lambda obj: obj.nominee.person.email()) candidate_phone = factory.Faker('phone_number') comments = factory.SubFactory(FeedbackFactory) - nominator_email = factory.LazyAttribute(lambda obj: obj.user.email) - user = factory.SubFactory(UserFactory) + nominator_email = factory.LazyAttribute(lambda obj: obj.person.user.email) + person = factory.SubFactory(PersonFactory) share_nominator = False diff --git a/ietf/nomcom/forms.py b/ietf/nomcom/forms.py index 20bf508e8e..5987b22637 100644 --- a/ietf/nomcom/forms.py +++ b/ietf/nomcom/forms.py @@ -15,12 +15,13 @@ from ietf.nomcom.models import ( NomCom, Nomination, Nominee, NomineePosition, Position, Feedback, ReminderDates, Topic, Volunteer ) from ietf.nomcom.utils import (NOMINATION_RECEIPT_TEMPLATE, FEEDBACK_RECEIPT_TEMPLATE, - get_user_email, validate_private_key, validate_public_key, + get_person_email, validate_private_key, validate_public_key, make_nomineeposition, make_nomineeposition_for_newperson, create_feedback_email) from ietf.person.models import Email from ietf.person.fields import (SearchableEmailField, SearchableEmailsField, SearchablePersonField, SearchablePersonsField ) +from ietf.utils.fields import ModelMultipleChoiceField from ietf.utils.mail import send_mail from ietf.mailtrigger.utils import gather_address_lists @@ -256,7 +257,7 @@ class NominateForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.nomcom = kwargs.pop('nomcom', None) - self.user = kwargs.pop('user', None) + self.person = kwargs.pop('person', None) self.public = kwargs.pop('public', None) super(NominateForm, self).__init__(*args, **kwargs) @@ -273,7 +274,7 @@ def __init__(self, *args, **kwargs): if not self.public: self.fields.pop('confirmation') - author = get_user_email(self.user) + author = get_person_email(self.person) if author: self.fields['nominator_email'].initial = author.address help_text = """(Nomcom Chair/Member: please fill this in. Use your own email address if the person making the @@ -303,7 +304,7 @@ def save(self, commit=True): author = None if self.public: - author = get_user_email(self.user) + author = get_person_email(self.person) else: if nominator_email: emails = Email.objects.filter(address=nominator_email) @@ -314,7 +315,7 @@ def save(self, commit=True): feedback = Feedback.objects.create(nomcom=self.nomcom, comments=self.nomcom.encrypt(qualifications), type=FeedbackTypeName.objects.get(slug='nomina'), - user=self.user) + person=self.person) feedback.positions.add(position) feedback.nominees.add(nominee) @@ -326,7 +327,7 @@ def save(self, commit=True): nomination.nominee = nominee nomination.comments = feedback nomination.share_nominator = share_nominator - nomination.user = self.user + nomination.person = self.person if commit: nomination.save() @@ -343,7 +344,7 @@ def save(self, commit=True): 'year': self.nomcom.year(), } path = nomcom_template_path + NOMINATION_RECEIPT_TEMPLATE - send_mail(None, to_email, from_email, subject, path, context, cc=cc) + send_mail(None, to_email, from_email, subject, path, context, cc=cc, copy=False, save=False) return nomination @@ -361,7 +362,7 @@ class NominateNewPersonForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.nomcom = kwargs.pop('nomcom', None) - self.user = kwargs.pop('user', None) + self.person = kwargs.pop('person', None) self.public = kwargs.pop('public', None) super(NominateNewPersonForm, self).__init__(*args, **kwargs) @@ -375,7 +376,7 @@ def __init__(self, *args, **kwargs): if not self.public: self.fields.pop('confirmation') - author = get_user_email(self.user) + author = get_person_email(self.person) if author: self.fields['nominator_email'].initial = author.address help_text = """(Nomcom Chair/Member: please fill this in. Use your own email address if the person making the @@ -416,7 +417,7 @@ def save(self, commit=True): author = None if self.public: - author = get_user_email(self.user) + author = get_person_email(self.person) else: if nominator_email: emails = Email.objects.filter(address=nominator_email) @@ -429,7 +430,7 @@ def save(self, commit=True): feedback = Feedback.objects.create(nomcom=self.nomcom, comments=self.nomcom.encrypt(qualifications), type=FeedbackTypeName.objects.get(slug='nomina'), - user=self.user) + person=self.person) feedback.positions.add(position) feedback.nominees.add(nominee) @@ -441,7 +442,7 @@ def save(self, commit=True): nomination.nominee = nominee nomination.comments = feedback nomination.share_nominator = share_nominator - nomination.user = self.user + nomination.person = self.person if commit: nomination.save() @@ -458,7 +459,7 @@ def save(self, commit=True): 'year': self.nomcom.year(), } path = nomcom_template_path + NOMINATION_RECEIPT_TEMPLATE - send_mail(None, to_email, from_email, subject, path, context, cc=cc) + send_mail(None, to_email, from_email, subject, path, context, cc=cc, copy=False, save=False) return nomination @@ -476,7 +477,7 @@ class FeedbackForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.nomcom = kwargs.pop('nomcom', None) - self.user = kwargs.pop('user', None) + self.person = kwargs.pop('person', None) self.public = kwargs.pop('public', None) self.position = kwargs.pop('position', None) self.nominee = kwargs.pop('nominee', None) @@ -484,7 +485,7 @@ def __init__(self, *args, **kwargs): super(FeedbackForm, self).__init__(*args, **kwargs) - author = get_user_email(self.user) + author = get_person_email(self.person) if self.public: self.fields.pop('nominator_email') @@ -514,7 +515,7 @@ def save(self, commit=True): author = None if self.public: - author = get_user_email(self.user) + author = get_person_email(self.person) else: nominator_email = self.cleaned_data['nominator_email'] if nominator_email: @@ -525,7 +526,7 @@ def save(self, commit=True): feedback.author = author.address feedback.nomcom = self.nomcom - feedback.user = self.user + feedback.person = self.person feedback.type = FeedbackTypeName.objects.get(slug='comment') feedback.comments = self.nomcom.encrypt(comment_text) feedback.save() @@ -551,7 +552,7 @@ def save(self, commit=True): } path = nomcom_template_path + FEEDBACK_RECEIPT_TEMPLATE # TODO - make the thing above more generic - send_mail(None, to_email, from_email, subject, path, context, cc=cc, copy=False) + send_mail(None, to_email, from_email, subject, path, context, cc=cc, copy=False, save=False) class Meta: model = Feedback @@ -578,7 +579,7 @@ class QuestionnaireForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.nomcom = kwargs.pop('nomcom', None) - self.user = kwargs.pop('user', None) + self.person = kwargs.pop('person', None) super(QuestionnaireForm, self).__init__(*args, **kwargs) self.fields['nominee'] = PositionNomineeField(nomcom=self.nomcom, required=True) @@ -588,13 +589,13 @@ def save(self, commit=True): comment_text = self.cleaned_data['comment_text'] (position, nominee) = self.cleaned_data['nominee'] - author = get_user_email(self.user) + author = get_person_email(self.person) if author: feedback.author = author feedback.nomcom = self.nomcom - feedback.user = self.user + feedback.person = self.person feedback.type = FeedbackTypeName.objects.get(slug='questio') feedback.comments = self.nomcom.encrypt(comment_text) feedback.save() @@ -653,15 +654,15 @@ def clean_key(self): class PendingFeedbackForm(forms.ModelForm): - type = forms.ModelChoiceField(queryset=FeedbackTypeName.objects.all().order_by('pk'), widget=forms.RadioSelect, empty_label='Unclassified', required=False) + type = forms.ModelChoiceField(queryset=FeedbackTypeName.objects.all(), widget=forms.RadioSelect, empty_label='Unclassified', required=False) class Meta: model = Feedback fields = ('type', ) - def set_nomcom(self, nomcom, user): + def set_nomcom(self, nomcom, person): self.nomcom = nomcom - self.user = user + self.person = person #self.fields['nominee'] = MultiplePositionNomineeField(nomcom=self.nomcom, #required=True, #widget=forms.SelectMultiple, @@ -670,7 +671,7 @@ def set_nomcom(self, nomcom, user): def save(self, commit=True): feedback = super(PendingFeedbackForm, self).save(commit=False) feedback.nomcom = self.nomcom - feedback.user = self.user + feedback.person = self.person feedback.save() return feedback @@ -700,9 +701,9 @@ class Meta: model = Feedback fields = ('type', ) - def set_nomcom(self, nomcom, user, instances=None): + def set_nomcom(self, nomcom, person, instances=None): self.nomcom = nomcom - self.user = user + self.person = person instances = instances or [] self.feedback_type = None for i in instances: @@ -719,9 +720,9 @@ def set_nomcom(self, nomcom, user, instances=None): required= self.feedback_type.slug != 'comment', help_text='Hold down "Control", or "Command" on a Mac, to select more than one.') if self.feedback_type.slug == 'comment': - self.fields['topic'] = forms.ModelMultipleChoiceField(queryset=self.nomcom.topic_set.all(), - help_text='Hold down "Control" or "Command" on a Mac, to select more than one.', - required=False,) + self.fields['topic'] = ModelMultipleChoiceField(queryset=self.nomcom.topic_set.all(), + help_text='Hold down "Control" or "Command" on a Mac, to select more than one.', + required=False,) else: self.fields['position'] = forms.ModelChoiceField(queryset=Position.objects.get_by_nomcom(self.nomcom).filter(is_open=True), label="Position") self.fields['searched_email'] = SearchableEmailField(only_users=False,help_text="Try to find the candidate you are classifying with this field first. Only use the name and email fields below if this search does not find the candidate.",label="Candidate",required=False) @@ -782,7 +783,7 @@ def save(self, commit=True): nominee=nominee, comments=feedback, nominator_email=nominator_email, - user=self.user) + person=self.person) return feedback else: feedback.save() @@ -847,7 +848,7 @@ class Meta: class NominationResponseCommentForm(forms.Form): comments = forms.CharField(widget=forms.Textarea,required=False,help_text="Any comments provided will be encrypted and will only be visible to the NomCom.", strip=False) -class NomcomVolunteerMultipleChoiceField(forms.ModelMultipleChoiceField): +class NomcomVolunteerMultipleChoiceField(ModelMultipleChoiceField): def label_from_instance(self, obj): year = obj.year() return f'Volunteer for the {year}/{year+1} Nominating Committee' diff --git a/ietf/nomcom/management/commands/create_test_nomcom.py b/ietf/nomcom/management/commands/create_test_nomcom.py index bfca892b88..3365ea8f40 100644 --- a/ietf/nomcom/management/commands/create_test_nomcom.py +++ b/ietf/nomcom/management/commands/create_test_nomcom.py @@ -11,7 +11,7 @@ from ietf.nomcom.factories import nomcom_kwargs_for_year, NomComFactory, NomineePositionFactory, key from ietf.person.factories import EmailFactory from ietf.group.models import Group -from ietf.person.models import User +from ietf.person.models import Person, User class Command(BaseCommand): help = ("Create (or delete) a nomcom for test and development purposes.") @@ -27,7 +27,9 @@ def handle(self, *args, **options): if opt_delete: if Group.objects.filter(acronym='nomcom7437').exists(): Group.objects.filter(acronym='nomcom7437').delete() - User.objects.filter(username__in=['testchair','testmember','testcandidate']).delete() + users_to_delete = ['testchair','testmember','testcandidate'] + Person.objects.filter(user__username__in=users_to_delete).delete() + User.objects.filter(username__in=users_to_delete).delete() self.stdout.write("Deleted test group 'nomcom7437' and its related objects.") else: self.stderr.write("test nomcom 'nomcom7437' does not exist; nothing to do.\n") @@ -57,6 +59,6 @@ def handle(self, *args, **options): position__nomcom=nc, position__name='Test Area Director', position__is_iesg_position=True, ) - self.stdout.write("%s\n" % key) + self.stdout.write("%s\n" % key.decode()) self.stdout.write("Nomcom 7437 created. The private key can also be found at any time\nin ietf/nomcom/factories.py. Note that it is NOT a secure key.\n") diff --git a/ietf/nomcom/management/commands/feedback_email.py b/ietf/nomcom/management/commands/feedback_email.py index 32e9c9aa28..3846208d58 100644 --- a/ietf/nomcom/management/commands/feedback_email.py +++ b/ietf/nomcom/management/commands/feedback_email.py @@ -42,8 +42,11 @@ def handle(self, *args, **options): except NomCom.DoesNotExist: raise CommandError("NomCom %s does not exist or it isn't active" % year) - binary_input = io.open(email, 'rb') if email else sys.stdin.buffer - self.msg = binary_input.read() + if email: + with io.open(email, 'rb') as binary_input: + self.msg = binary_input.read() + else: + self.msg = sys.stdin.buffer.read() try: feedback = create_feedback_email(self.nomcom, self.msg) diff --git a/ietf/nomcom/management/commands/send_reminders.py b/ietf/nomcom/management/commands/send_reminders.py deleted file mode 100644 index bc10425430..0000000000 --- a/ietf/nomcom/management/commands/send_reminders.py +++ /dev/null @@ -1,37 +0,0 @@ -# Copyright The IETF Trust 2013-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -import syslog - -from django.core.management.base import BaseCommand - -from ietf.nomcom.models import NomCom, NomineePosition -from ietf.nomcom.utils import send_accept_reminder_to_nominee,send_questionnaire_reminder_to_nominee -from ietf.utils.timezone import date_today - - -def log(message): - syslog.syslog(message) - -def is_time_to_send(nomcom,send_date,nomination_date): - if nomcom.reminder_interval: - days_passed = (send_date - nomination_date).days - return days_passed > 0 and days_passed % nomcom.reminder_interval == 0 - else: - return bool(nomcom.reminderdates_set.filter(date=send_date)) - -class Command(BaseCommand): - help = ("Send acceptance and questionnaire reminders to nominees") - - def handle(self, *args, **options): - for nomcom in NomCom.objects.filter(group__state__slug='active'): - nps = NomineePosition.objects.filter(nominee__nomcom=nomcom,nominee__duplicated__isnull=True) - for nominee_position in nps.pending(): - if is_time_to_send(nomcom, date_today(), nominee_position.time.date()): - send_accept_reminder_to_nominee(nominee_position) - log('Sent accept reminder to %s' % nominee_position.nominee.email.address) - for nominee_position in nps.accepted().without_questionnaire_response(): - if is_time_to_send(nomcom, date_today(), nominee_position.time.date()): - send_questionnaire_reminder_to_nominee(nominee_position) - log('Sent questionnaire reminder to %s' % nominee_position.nominee.email.address) diff --git a/ietf/nomcom/management/tests.py b/ietf/nomcom/management/tests.py index 7bda2b5aa5..08c0e1fe32 100644 --- a/ietf/nomcom/management/tests.py +++ b/ietf/nomcom/management/tests.py @@ -1,7 +1,7 @@ # Copyright The IETF Trust 2021, All Rights Reserved # -*- coding: utf-8 -*- """Tests of nomcom management commands""" -import mock +from unittest import mock import sys from collections import namedtuple diff --git a/ietf/nomcom/migrations/0001_initial.py b/ietf/nomcom/migrations/0001_initial.py index 9ef8e54767..81eaabc604 100644 --- a/ietf/nomcom/migrations/0001_initial.py +++ b/ietf/nomcom/migrations/0001_initial.py @@ -1,12 +1,8 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-20 10:52 - +# Generated by Django 2.2.28 on 2023-03-20 19:22 from django.conf import settings from django.db import migrations, models import django.db.models.deletion -import ietf.nomcom.fields import ietf.nomcom.models import ietf.utils.models import ietf.utils.storage @@ -17,11 +13,11 @@ class Migration(migrations.Migration): initial = True dependencies = [ - ('group', '0001_initial'), - migrations.swappable_dependency(settings.AUTH_USER_MODEL), ('dbtemplate', '0001_initial'), ('name', '0001_initial'), ('person', '0001_initial'), + migrations.swappable_dependency(settings.AUTH_USER_MODEL), + ('group', '0001_initial'), ] operations = [ @@ -31,20 +27,13 @@ class Migration(migrations.Migration): ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('author', models.EmailField(blank=True, max_length=254, verbose_name='Author')), ('subject', models.TextField(blank=True, verbose_name='Subject')), - ('comments', ietf.nomcom.fields.EncryptedTextField(verbose_name='Comments')), + ('comments', models.BinaryField(verbose_name='Comments')), ('time', models.DateTimeField(auto_now_add=True)), ], options={ 'ordering': ['time'], }, ), - migrations.CreateModel( - name='FeedbackLastSeen', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('time', models.DateTimeField(auto_now=True)), - ], - ), migrations.CreateModel( name='NomCom', fields=[ @@ -54,6 +43,9 @@ class Migration(migrations.Migration): ('reminder_interval', models.PositiveIntegerField(blank=True, help_text='If the nomcom user sets the interval field then a cron command will send reminders to the nominees who have not responded using the following formula: (today - nomination_date) % interval == 0.', null=True)), ('initial_text', models.TextField(blank=True, verbose_name='Help text for nomination form')), ('show_nominee_pictures', models.BooleanField(default=True, help_text='Display pictures of each nominee (if available) on the feedback pages', verbose_name='Show nominee pictures')), + ('show_accepted_nominees', models.BooleanField(default=True, help_text='Show accepted nominees on the public nomination page', verbose_name='Show accepted nominees')), + ('is_accepting_volunteers', models.BooleanField(default=False, help_text='Is this nomcom currently accepting volunteers?', verbose_name='Accepting volunteers')), + ('first_call_for_volunteers', models.DateField(blank=True, null=True, verbose_name='Date of the first call for volunteers')), ('group', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='group.Group')), ], options={ @@ -61,22 +53,6 @@ class Migration(migrations.Migration): 'verbose_name_plural': 'NomComs', }, ), - migrations.CreateModel( - name='Nomination', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('candidate_name', models.CharField(max_length=255, verbose_name='Candidate name')), - ('candidate_email', models.EmailField(max_length=255, verbose_name='Candidate email')), - ('candidate_phone', models.CharField(blank=True, max_length=255, verbose_name='Candidate phone')), - ('nominator_email', models.EmailField(blank=True, max_length=254, verbose_name='Nominator Email')), - ('time', models.DateTimeField(auto_now_add=True)), - ('share_nominator', models.BooleanField(default=False, help_text='Check this box to allow the NomCom to let the person you are nominating know that you were one of the people who nominated them. If you do not check this box, your name will be confidential and known only within NomCom.', verbose_name='Share nominator name with candidate')), - ('comments', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.Feedback')), - ], - options={ - 'verbose_name_plural': 'Nominations', - }, - ), migrations.CreateModel( name='Nominee', fields=[ @@ -86,38 +62,41 @@ class Migration(migrations.Migration): ('nomcom', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.NomCom')), ], options={ - 'ordering': ['-nomcom__group__acronym', 'email__address'], 'verbose_name_plural': 'Nominees', + 'ordering': ['-nomcom__group__acronym', 'person__name'], }, ), migrations.CreateModel( - name='NomineePosition', + name='Topic', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('time', models.DateTimeField(auto_now_add=True)), - ('nominee', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.Nominee')), + ('subject', models.CharField(help_text='This short description will appear on the Feedback pages.', max_length=255, verbose_name='Name')), + ('accepting_feedback', models.BooleanField(default=False, verbose_name='Is accepting feedback')), + ('audience', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.TopicAudienceName', verbose_name='Who can provide feedback (intended audience)')), + ('description', ietf.utils.models.ForeignKey(editable=False, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='description', to='dbtemplate.DBTemplate')), + ('nomcom', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.NomCom')), ], options={ - 'ordering': ['nominee'], - 'verbose_name': 'Nominee position', - 'verbose_name_plural': 'Nominee positions', + 'verbose_name_plural': 'Topics', }, ), migrations.CreateModel( - name='Position', + name='Volunteer', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('name', models.CharField(help_text='This short description will appear on the Nomination and Feedback pages. Be as descriptive as necessary. Past examples: "Transport AD", "IAB Member"', max_length=255, verbose_name='Name')), - ('is_open', models.BooleanField(default=False, help_text='Set is_open when the nomcom is working on a position. Clear it when an appointment is confirmed.', verbose_name='Is open')), - ('accepting_nominations', models.BooleanField(default=False, verbose_name='Is accepting nominations')), - ('accepting_feedback', models.BooleanField(default=False, verbose_name='Is accepting feedback')), + ('affiliation', models.CharField(blank=True, max_length=255)), ('nomcom', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.NomCom')), - ('questionnaire', ietf.utils.models.ForeignKey(editable=False, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='questionnaire', to='dbtemplate.DBTemplate')), - ('requirement', ietf.utils.models.ForeignKey(editable=False, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='requirement', to='dbtemplate.DBTemplate')), + ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), + ], + ), + migrations.CreateModel( + name='TopicFeedbackLastSeen', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('time', models.DateTimeField(auto_now=True)), + ('reviewer', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), + ('topic', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.Topic')), ], - options={ - 'verbose_name_plural': 'Positions', - }, ), migrations.CreateModel( name='ReminderDates', @@ -128,37 +107,37 @@ class Migration(migrations.Migration): ], ), migrations.CreateModel( - name='Topic', + name='Position', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('subject', models.CharField(help_text='This short description will appear on the Feedback pages.', max_length=255, verbose_name='Name')), + ('name', models.CharField(help_text='This short description will appear on the Nomination and Feedback pages. Be as descriptive as necessary. Past examples: "Transport AD", "IAB Member"', max_length=255, verbose_name='Name')), + ('is_open', models.BooleanField(default=False, help_text='Set is_open when the nomcom is working on a position. Clear it when an appointment is confirmed.', verbose_name='Is open')), + ('accepting_nominations', models.BooleanField(default=False, verbose_name='Is accepting nominations')), ('accepting_feedback', models.BooleanField(default=False, verbose_name='Is accepting feedback')), - ('audience', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.TopicAudienceName')), - ('description', ietf.utils.models.ForeignKey(editable=False, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='description', to='dbtemplate.DBTemplate')), + ('is_iesg_position', models.BooleanField(default=False, verbose_name='Is IESG Position')), ('nomcom', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.NomCom')), + ('questionnaire', ietf.utils.models.ForeignKey(editable=False, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='questionnaire', to='dbtemplate.DBTemplate')), + ('requirement', ietf.utils.models.ForeignKey(editable=False, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='requirement', to='dbtemplate.DBTemplate')), ], options={ - 'verbose_name_plural': 'Topics', + 'verbose_name_plural': 'Positions', }, ), migrations.CreateModel( - name='TopicFeedbackLastSeen', + name='NomineePosition', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('time', models.DateTimeField(auto_now=True)), - ('reviewer', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), - ('topic', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.Topic')), + ('time', models.DateTimeField(auto_now_add=True)), + ('nominee', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.Nominee')), + ('position', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.Position')), + ('state', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.NomineePositionStateName')), ], - ), - migrations.AddField( - model_name='nomineeposition', - name='position', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.Position'), - ), - migrations.AddField( - model_name='nomineeposition', - name='state', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.NomineePositionStateName'), + options={ + 'verbose_name': 'Nominee position', + 'verbose_name_plural': 'Nominee positions', + 'ordering': ['nominee'], + 'unique_together': {('position', 'nominee')}, + }, ), migrations.AddField( model_name='nominee', @@ -170,30 +149,33 @@ class Migration(migrations.Migration): name='person', field=ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='person.Person'), ), - migrations.AddField( - model_name='nomination', - name='nominee', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.Nominee'), - ), - migrations.AddField( - model_name='nomination', - name='position', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.Position'), - ), - migrations.AddField( - model_name='nomination', - name='user', - field=ietf.utils.models.ForeignKey(editable=False, on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL), - ), - migrations.AddField( - model_name='feedbacklastseen', - name='nominee', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.Nominee'), + migrations.CreateModel( + name='Nomination', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('candidate_name', models.CharField(max_length=255, verbose_name='Candidate name')), + ('candidate_email', models.EmailField(max_length=255, verbose_name='Candidate email')), + ('candidate_phone', models.CharField(blank=True, max_length=255, verbose_name='Candidate phone')), + ('nominator_email', models.EmailField(blank=True, max_length=254, verbose_name='Nominator Email')), + ('time', models.DateTimeField(auto_now_add=True)), + ('share_nominator', models.BooleanField(default=False, help_text='Check this box to allow the NomCom to let the person you are nominating know that you were one of the people who nominated them. If you do not check this box, your name will be confidential and known only within NomCom.', verbose_name='Share nominator name with candidate')), + ('comments', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.Feedback')), + ('nominee', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.Nominee')), + ('position', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.Position')), + ('user', ietf.utils.models.ForeignKey(editable=False, null=True, on_delete=django.db.models.deletion.SET_NULL, to=settings.AUTH_USER_MODEL)), + ], + options={ + 'verbose_name_plural': 'Nominations', + }, ), - migrations.AddField( - model_name='feedbacklastseen', - name='reviewer', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person'), + migrations.CreateModel( + name='FeedbackLastSeen', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('time', models.DateTimeField(auto_now=True)), + ('nominee', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.Nominee')), + ('reviewer', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), + ], ), migrations.AddField( model_name='feedback', @@ -223,14 +205,14 @@ class Migration(migrations.Migration): migrations.AddField( model_name='feedback', name='user', - field=ietf.utils.models.ForeignKey(blank=True, editable=False, null=True, on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL), - ), - migrations.AlterUniqueTogether( - name='nomineeposition', - unique_together=set([('position', 'nominee')]), + field=ietf.utils.models.ForeignKey(blank=True, editable=False, null=True, on_delete=django.db.models.deletion.SET_NULL, to=settings.AUTH_USER_MODEL), ), migrations.AlterUniqueTogether( name='nominee', - unique_together=set([('email', 'nomcom')]), + unique_together={('email', 'nomcom')}, + ), + migrations.AddIndex( + model_name='feedback', + index=models.Index(fields=['time'], name='nomcom_feed_time_35cf38_idx'), ), ] diff --git a/ietf/nomcom/migrations/0002_add_uncheck_reminder.py b/ietf/nomcom/migrations/0002_add_uncheck_reminder.py new file mode 100644 index 0000000000..c7a57601f2 --- /dev/null +++ b/ietf/nomcom/migrations/0002_add_uncheck_reminder.py @@ -0,0 +1,18 @@ +# Generated by Django 2.2.28 on 2023-03-25 02:52 + +from django.db import migrations, models + + +class Migration(migrations.Migration): + + dependencies = [ + ('nomcom', '0001_initial'), + ] + + operations = [ + migrations.AlterField( + model_name='nomcom', + name='send_questionnaire', + field=models.BooleanField(default=False, help_text='If you check this box, questionnaires are sent automatically after nominations. DO NOT CHECK if they are not ready yet.', verbose_name='Send questionnaires automatically'), + ), + ] diff --git a/ietf/nomcom/migrations/0002_auto_20180918_0550.py b/ietf/nomcom/migrations/0002_auto_20180918_0550.py deleted file mode 100644 index 44fd8d4df8..0000000000 --- a/ietf/nomcom/migrations/0002_auto_20180918_0550.py +++ /dev/null @@ -1,29 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.15 on 2018-09-18 05:50 - - -from django.conf import settings -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('nomcom', '0001_initial'), - ] - - operations = [ - migrations.AlterField( - model_name='feedback', - name='user', - field=ietf.utils.models.ForeignKey(blank=True, editable=False, null=True, on_delete=django.db.models.deletion.SET_NULL, to=settings.AUTH_USER_MODEL), - ), - migrations.AlterField( - model_name='nomination', - name='user', - field=ietf.utils.models.ForeignKey(editable=False, null=True, on_delete=django.db.models.deletion.SET_NULL, to=settings.AUTH_USER_MODEL), - ), - ] diff --git a/ietf/nomcom/migrations/0003_alter_nomination_share_nominator.py b/ietf/nomcom/migrations/0003_alter_nomination_share_nominator.py new file mode 100644 index 0000000000..d5108c3fdc --- /dev/null +++ b/ietf/nomcom/migrations/0003_alter_nomination_share_nominator.py @@ -0,0 +1,21 @@ +# Generated by Django 4.2.2 on 2023-06-08 18:06 + +from django.db import migrations, models + + +class Migration(migrations.Migration): + dependencies = [ + ("nomcom", "0002_add_uncheck_reminder"), + ] + + operations = [ + migrations.AlterField( + model_name="nomination", + name="share_nominator", + field=models.BooleanField( + default=False, + help_text="Check this box to allow the NomCom to let the person you are nominating know that you were one of the people who nominated them. If you do not check this box, your name will be confidential and known only within NomCom.", + verbose_name="OK to share nominator's name with candidate", + ), + ), + ] diff --git a/ietf/nomcom/migrations/0003_nomcom_show_accepted_nominees.py b/ietf/nomcom/migrations/0003_nomcom_show_accepted_nominees.py deleted file mode 100644 index cc7d732834..0000000000 --- a/ietf/nomcom/migrations/0003_nomcom_show_accepted_nominees.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.15 on 2018-09-26 11:10 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('nomcom', '0002_auto_20180918_0550'), - ] - - operations = [ - migrations.AddField( - model_name='nomcom', - name='show_accepted_nominees', - field=models.BooleanField(default=True, help_text='Show accepted nominees on the public nomination page', verbose_name='Show accepted nominees'), - ), - ] diff --git a/ietf/nomcom/migrations/0004_set_show_accepted_nominees_false_on_existing_nomcoms.py b/ietf/nomcom/migrations/0004_set_show_accepted_nominees_false_on_existing_nomcoms.py deleted file mode 100644 index 7f4746a274..0000000000 --- a/ietf/nomcom/migrations/0004_set_show_accepted_nominees_false_on_existing_nomcoms.py +++ /dev/null @@ -1,24 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.15 on 2018-09-26 11:12 - - -from django.db import migrations - -def forward(apps, schema_editor): - NomCom = apps.get_model('nomcom','NomCom') - NomCom.objects.update(show_accepted_nominees=False) - -def reverse(apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('nomcom', '0003_nomcom_show_accepted_nominees'), - ] - - operations = [ - migrations.RunPython(forward,reverse) - ] - diff --git a/ietf/nomcom/migrations/0004_volunteer_origin_volunteer_time_volunteer_withdrawn.py b/ietf/nomcom/migrations/0004_volunteer_origin_volunteer_time_volunteer_withdrawn.py new file mode 100644 index 0000000000..9eaebf2069 --- /dev/null +++ b/ietf/nomcom/migrations/0004_volunteer_origin_volunteer_time_volunteer_withdrawn.py @@ -0,0 +1,27 @@ +# Generated by Django 4.2.7 on 2023-11-05 09:45 + +from django.db import migrations, models + + +class Migration(migrations.Migration): + dependencies = [ + ("nomcom", "0003_alter_nomination_share_nominator"), + ] + + operations = [ + migrations.AddField( + model_name="volunteer", + name="origin", + field=models.CharField(default="datatracker", max_length=32), + ), + migrations.AddField( + model_name="volunteer", + name="time", + field=models.DateTimeField(auto_now_add=True, null=True, blank=True), + ), + migrations.AddField( + model_name="volunteer", + name="withdrawn", + field=models.DateTimeField(blank=True, null=True), + ), + ] diff --git a/ietf/nomcom/migrations/0005_auto_20181008_0602.py b/ietf/nomcom/migrations/0005_auto_20181008_0602.py deleted file mode 100644 index cd56b09938..0000000000 --- a/ietf/nomcom/migrations/0005_auto_20181008_0602.py +++ /dev/null @@ -1,20 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.16 on 2018-10-08 06:02 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('nomcom', '0004_set_show_accepted_nominees_false_on_existing_nomcoms'), - ] - - operations = [ - migrations.AlterModelOptions( - name='nominee', - options={'ordering': ['-nomcom__group__acronym', 'person__name'], 'verbose_name_plural': 'Nominees'}, - ), - ] diff --git a/ietf/nomcom/migrations/0005_user_to_person.py b/ietf/nomcom/migrations/0005_user_to_person.py new file mode 100644 index 0000000000..66a6e99642 --- /dev/null +++ b/ietf/nomcom/migrations/0005_user_to_person.py @@ -0,0 +1,100 @@ +# Generated by Django 4.2.2 on 2023-06-14 19:47 + +from django.db import migrations +from django.db.models import OuterRef, Subquery +import django.db.models.deletion +import ietf.utils.models + + +def forward(apps, schema_editor): + Nomination = apps.get_model('nomcom', 'Nomination') + Person = apps.get_model("person", "Person") + Nomination.objects.exclude( + user__isnull=True + ).update( + person=Subquery( + Person.objects.filter(user_id=OuterRef("user_id")).values("pk")[:1] + ) + ) + + Feedback = apps.get_model('nomcom', 'Feedback') + Feedback.objects.exclude( + user__isnull=True + ).update( + person=Subquery( + Person.objects.filter(user_id=OuterRef("user_id")).values("pk")[:1] + ) + ) + +def reverse(apps, schema_editor): + Nomination = apps.get_model('nomcom', 'Nomination') + Person = apps.get_model("person", "Person") + Nomination.objects.exclude( + person__isnull=True + ).update( + user_id=Subquery( + Person.objects.filter(pk=OuterRef("person_id")).values("user_id")[:1] + ) + ) + + Feedback = apps.get_model('nomcom', 'Feedback') + Feedback.objects.exclude( + person__isnull=True + ).update( + user_id=Subquery( + Person.objects.filter(pk=OuterRef("person_id")).values("user_id")[:1] + ) + ) + +class Migration(migrations.Migration): + dependencies = [ + ("person", "0001_initial"), + ("nomcom", "0004_volunteer_origin_volunteer_time_volunteer_withdrawn"), + ] + + operations = [ + migrations.AddField( + model_name="feedback", + name="person", + field=ietf.utils.models.ForeignKey( + blank=True, + editable=False, + null=True, + on_delete=django.db.models.deletion.SET_NULL, + to="person.person", + ), + ), + migrations.AddField( + model_name="nomination", + name="person", + field=ietf.utils.models.ForeignKey( + editable=False, + null=True, + on_delete=django.db.models.deletion.SET_NULL, + to="person.person", + ), + ), + migrations.RunPython(forward, reverse), + migrations.RemoveField( + model_name="feedback", + name="user", + field=ietf.utils.models.ForeignKey( + blank=True, + editable=False, + null=True, + on_delete=django.db.models.deletion.SET_NULL, + to="user.user", + ), + ), + migrations.RemoveField( + model_name="nomination", + name="user", + field=ietf.utils.models.ForeignKey( + blank=True, + editable=False, + null=True, + on_delete=django.db.models.deletion.SET_NULL, + to="user.user", + ), + ), + ] diff --git a/ietf/nomcom/migrations/0006_auto_20190716_1216.py b/ietf/nomcom/migrations/0006_auto_20190716_1216.py deleted file mode 100644 index 7c3a374f4b..0000000000 --- a/ietf/nomcom/migrations/0006_auto_20190716_1216.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.22 on 2019-07-16 12:16 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('nomcom', '0005_auto_20181008_0602'), - ] - - operations = [ - migrations.AlterField( - model_name='feedback', - name='comments', - field=models.BinaryField(verbose_name='Comments'), - ), - ] diff --git a/ietf/nomcom/migrations/0007_position_is_iesg_position.py b/ietf/nomcom/migrations/0007_position_is_iesg_position.py deleted file mode 100644 index edd4cf0f08..0000000000 --- a/ietf/nomcom/migrations/0007_position_is_iesg_position.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.27 on 2020-01-07 14:41 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('nomcom', '0006_auto_20190716_1216'), - ] - - operations = [ - migrations.AddField( - model_name='position', - name='is_iesg_position', - field=models.BooleanField(default=False, verbose_name='Is IESG Position'), - ), - ] diff --git a/ietf/nomcom/migrations/0008_auto_20201008_0506.py b/ietf/nomcom/migrations/0008_auto_20201008_0506.py deleted file mode 100644 index b14e8da276..0000000000 --- a/ietf/nomcom/migrations/0008_auto_20201008_0506.py +++ /dev/null @@ -1,20 +0,0 @@ -# Generated by Django 2.2.16 on 2020-10-08 05:06 - -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('nomcom', '0007_position_is_iesg_position'), - ] - - operations = [ - migrations.AlterField( - model_name='topic', - name='audience', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.TopicAudienceName', verbose_name='Who can provide feedback (intended audience)'), - ), - ] diff --git a/ietf/nomcom/migrations/0009_auto_20201109_0439.py b/ietf/nomcom/migrations/0009_auto_20201109_0439.py deleted file mode 100644 index 8b6efe2a86..0000000000 --- a/ietf/nomcom/migrations/0009_auto_20201109_0439.py +++ /dev/null @@ -1,17 +0,0 @@ -# Generated by Django 2.2.17 on 2020-11-09 04:39 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('nomcom', '0008_auto_20201008_0506'), - ] - - operations = [ - migrations.AddIndex( - model_name='feedback', - index=models.Index(fields=['time'], name='nomcom_feed_time_35cf38_idx'), - ), - ] diff --git a/ietf/nomcom/migrations/0010_nomcom_first_call_for_volunteers.py b/ietf/nomcom/migrations/0010_nomcom_first_call_for_volunteers.py deleted file mode 100644 index 01f2f3335c..0000000000 --- a/ietf/nomcom/migrations/0010_nomcom_first_call_for_volunteers.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.2.20 on 2021-04-22 14:32 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('nomcom', '0009_auto_20201109_0439'), - ] - - operations = [ - migrations.AddField( - model_name='nomcom', - name='first_call_for_volunteers', - field=models.DateField(blank=True, null=True, verbose_name='Date of the first call for volunteers'), - ), - ] diff --git a/ietf/nomcom/migrations/0011_volunteers.py b/ietf/nomcom/migrations/0011_volunteers.py deleted file mode 100644 index 651e1dc54a..0000000000 --- a/ietf/nomcom/migrations/0011_volunteers.py +++ /dev/null @@ -1,30 +0,0 @@ -# Generated by Django 2.2.24 on 2021-06-09 12:15 - -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0019_auto_20210604_1443'), - ('nomcom', '0010_nomcom_first_call_for_volunteers'), - ] - - operations = [ - migrations.AddField( - model_name='nomcom', - name='is_accepting_volunteers', - field=models.BooleanField(default=False, help_text='Is this nomcom is currently accepting volunteers?', verbose_name='Accepting volunteers'), - ), - migrations.CreateModel( - name='Volunteer', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('affiliation', models.CharField(blank=True, max_length=255)), - ('nomcom', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='nomcom.NomCom')), - ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), - ], - ), - ] diff --git a/ietf/nomcom/migrations/0012_populate_volunteers.py b/ietf/nomcom/migrations/0012_populate_volunteers.py deleted file mode 100644 index 4da1bbaf09..0000000000 --- a/ietf/nomcom/migrations/0012_populate_volunteers.py +++ /dev/null @@ -1,103 +0,0 @@ -# Copyright The IETF Trust 2021 All Rights Reserved -# Generated by Django 2.2.24 on 2021-06-10 12:50 - -from django.db import migrations - -def forward(apps, schema_editor): - NomCom = apps.get_model('nomcom','NomCom') - nc = NomCom.objects.filter(group__acronym='nomcom2021').first() - if nc is None: - return # nothing to do if the NomCom in question does not exist - - nc.is_accepting_volunteers = True - nc.save() - - volunteers = [ - (21684, 'Futurewei Technologies'), # Barry Leiba - (117988, 'Google LLC'), # Benjamin M. Schwartz - (122671, 'Fastmail Pty Ltd'), # Bron Gondwana - (124329, 'Huawei'), # Cheng Li - (113580, 'China Mobile'), # Weiqiang Cheng - (22933, 'LabN Consulting'), # Christian Hopps - (102391, 'Futurewei Technologies, Inc'), # Donald E. Eastlake 3rd - (111477, 'Huawei'), # Dhruv Dhody - (12695, 'Mozilla'), # Eric Rescorla - (17141, 'APNIC'), # George G. Michaelson - (108833, 'Huawei'), # Luigi Iannone - (118908, 'Huawei Technologies'), # Giuseppe Fioccola - (5376, 'Vigil Security, LLC'), # Russ Housley - (118100, 'Equinix'), # Ignas Bagdonas - (107287, 'rtfm llp'), # Jim Reid - (123344, 'Netflix'), # Theresa Enghardt - (109226, 'Huawei'), # Italo Busi - (113152, 'Ericsson'), # Jaime Jimenez - (111354, 'Juniper Networks'), # John Drake - (112342, 'Cisco Systems, Inc.'), # Jakob Heitz - (109207, 'Huawei Technologies Co., Ltd.'), # 江元龙 - (110737, 'Huawei Technologies'), # Jie Dong - (109330, 'Ericsson'), # John Preuß Mattsson - (123589, 'Nokia'), # Julien Maisonneuve - (124655, 'UK National Cyber Security Centre (NCSC)'), # Kirsty Paine - (119463, 'Akamai Technologies, Inc.'), # Kyle Rose - (109983, 'Huawei Technologies Co.,Ltd.'), # Bo Wu - (2097, 'Cisco Systems'), # Eliot Lear - (567, 'UCLA'), # Lixia Zhang - (107762, 'Huawei'), # Mach Chen - (125198, 'Juniper Networks'), # Melchior Aelmans - (104294, 'Ericsson'), # Magnus Westerlund - (104495, 'Impedance Mismatch LLC'), # Marc Petit-Huguenin - (119947, 'Painless Security, LLC'), # Margaret Cullen - (102830, 'Independent'), # Mary Barnes - (116593, 'Fastly'), # Patrick McManus - (102254, 'Sandelman Software Works'), # Michael Richardson - (20356, 'Apple'), # Ted Lemon - (103881, 'Fastly'), # Mark Nottingham - (106741, 'NthPermutation Security'), # Michael StJohns - (116323, 'Moulay Ismail University of Meknes, Morocco'), # Nabil Benamar - (20106, 'W3C/MIT'), # Samuel Weiler - (105691, 'cisco'), # Ole Trøan - (121160, 'Independent, Amazon'), # Padma Pillay-Esnault - (115824, 'Cisco Systems'), # Pascal Thubert - (122637, 'Huawei Technologies Co.,Ltd.'), # Shuping Peng - (112952, 'Huawei'), # Haibo Wang - (5234, 'IIJ Research Lab & Arrcus Inc'), # Randy Bush - (101568, 'Juniper Networks'), # Ron Bonica - (123443, 'Huawei Technologies Co.,Ltd.'), # Bing Liu (Remy) - (18321, 'Episteme Technology Consulting LLC'), # Pete Resnick - (18427, 'Akamai'), # Rich Salz - (126259, 'Huawei Technologies Co.,Ltd.'), # Fan Yang - (115724, 'Juniper Networks'), # Shraddha Hegde - (125509, 'Tencent'), # Shuai Zhao - (110614, 'Huawei'), # Tal Mizrahi - (123395, 'APNIC'), # Tom Harrison - (116516, 'Juniper Networks'), # Tarek Saad - (11834, 'Futurewei USA'), # Toerless Eckert - (123962, 'Open-Xchange'), # Vittorio Bertola - (126530, 'Huawei'), # Yali Wang - (106199, 'Ericsson'), # Wassim Haddad - (125173, 'Huawei'), # Wei Pan - (111299, 'ZTE Corporation'), # Xiao Min - (113285, 'Huawei Technologies'), # XiPeng Xiao - (116337, 'Futurewei Technologies Inc.'), # Yingzhen Qu - (123974, 'ZTE'), # Zheng Zhang - (117500, 'Huawei'), # Guangying Zheng - (115934, 'Huawei Technologies'), # Haomian Zheng - (110966, 'Juniper'), # Zhaohui (Jeffrey) Zhang - ] - - for pk, affiliation in volunteers: - nc.volunteer_set.create(person_id=pk, affiliation=affiliation) - -def reverse(apps, schema_editor): - Volunteer = apps.get_model('nomcom','Volunteer') - Volunteer.objects.all().delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('nomcom', '0011_volunteers'), - ] - - operations = [ - migrations.RunPython(forward,reverse) - ] diff --git a/ietf/nomcom/migrations/0013_update_accepting_volunteers_helptext.py b/ietf/nomcom/migrations/0013_update_accepting_volunteers_helptext.py deleted file mode 100644 index 718e3b49f6..0000000000 --- a/ietf/nomcom/migrations/0013_update_accepting_volunteers_helptext.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.2.28 on 2022-07-23 13:14 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('nomcom', '0012_populate_volunteers'), - ] - - operations = [ - migrations.AlterField( - model_name='nomcom', - name='is_accepting_volunteers', - field=models.BooleanField(default=False, help_text='Is this nomcom currently accepting volunteers?', verbose_name='Accepting volunteers'), - ), - ] diff --git a/ietf/nomcom/models.py b/ietf/nomcom/models.py index 868787429f..c206e467bd 100644 --- a/ietf/nomcom/models.py +++ b/ietf/nomcom/models.py @@ -7,7 +7,6 @@ from django.db import models from django.db.models.signals import post_delete from django.conf import settings -from django.contrib.auth.models import User from django.template.loader import render_to_string from django.template.defaultfilters import linebreaks # type: ignore @@ -43,12 +42,13 @@ class ReminderDates(models.Model): class NomCom(models.Model): + # TODO-BLOBSTORE: migrate this to a database field instead of a FileField and update code accordingly public_key = models.FileField(storage=NoLocationMigrationFileSystemStorage(location=settings.NOMCOM_PUBLIC_KEYS_DIR), upload_to=upload_path_handler, blank=True, null=True) group = ForeignKey(Group) send_questionnaire = models.BooleanField(verbose_name='Send questionnaires automatically', default=False, - help_text='If you check this box, questionnaires are sent automatically after nominations.') + help_text='If you check this box, questionnaires are sent automatically after nominations. DO NOT CHECK if they are not ready yet.') reminder_interval = models.PositiveIntegerField(help_text='If the nomcom user sets the interval field then a cron command will ' 'send reminders to the nominees who have not responded using ' 'the following formula: (today - nomination_date) % interval == 0.', @@ -128,9 +128,9 @@ class Nomination(models.Model): nominee = ForeignKey('Nominee') comments = ForeignKey('Feedback') nominator_email = models.EmailField(verbose_name='Nominator Email', blank=True) - user = ForeignKey(User, editable=False, null=True, on_delete=models.SET_NULL) + person = ForeignKey(Person, editable=False, null=True, on_delete=models.SET_NULL) time = models.DateTimeField(auto_now_add=True) - share_nominator = models.BooleanField(verbose_name='Share nominator name with candidate', default=False, + share_nominator = models.BooleanField(verbose_name='OK to share nominator\'s name with candidate', default=False, help_text='Check this box to allow the NomCom to let the ' 'person you are nominating know that you were ' 'one of the people who nominated them. If you ' @@ -148,7 +148,7 @@ class Nominee(models.Model): email = ForeignKey(Email) person = ForeignKey(Person, blank=True, null=True) - nominee_position = models.ManyToManyField('Position', through='NomineePosition') + nominee_position = models.ManyToManyField('nomcom.Position', through='nomcom.NomineePosition') duplicated = ForeignKey('Nominee', blank=True, null=True) nomcom = ForeignKey('NomCom') @@ -188,8 +188,9 @@ class Meta: def save(self, **kwargs): if not self.pk and not self.state_id: + # Don't need to set update_fields because the self.pk test means this is a new instance self.state = NomineePositionStateName.objects.get(slug='pending') - super(NomineePosition, self).save(**kwargs) + super().save(**kwargs) def __str__(self): return "%s - %s - %s" % (self.nominee, self.state, self.position) @@ -292,13 +293,13 @@ def get_description(self): class Feedback(models.Model): nomcom = ForeignKey('NomCom') author = models.EmailField(verbose_name='Author', blank=True) - positions = models.ManyToManyField('Position', blank=True) - nominees = models.ManyToManyField('Nominee', blank=True) - topics = models.ManyToManyField('Topic', blank=True) + positions = models.ManyToManyField('nomcom.Position', blank=True) + nominees = models.ManyToManyField('nomcom.Nominee', blank=True) + topics = models.ManyToManyField('nomcom.Topic', blank=True) subject = models.TextField(verbose_name='Subject', blank=True) comments = models.BinaryField(verbose_name='Comments') type = ForeignKey(FeedbackTypeName, blank=True, null=True) - user = ForeignKey(User, editable=False, blank=True, null=True, on_delete=models.SET_NULL) + person = ForeignKey(Person, editable=False, blank=True, null=True, on_delete=models.SET_NULL) time = models.DateTimeField(auto_now_add=True) objects = FeedbackManager() @@ -326,7 +327,10 @@ class Volunteer(models.Model): nomcom = ForeignKey('NomCom') person = ForeignKey(Person) affiliation = models.CharField(blank=True, max_length=255) - + time = models.DateTimeField(auto_now_add=True, null=True, blank=True) + origin = models.CharField(max_length=32, default='datatracker') + withdrawn = models.DateTimeField(blank=True, null=True) + def __str__(self): return f'{self.person} for {self.nomcom}' diff --git a/ietf/nomcom/resources.py b/ietf/nomcom/resources.py index c87e72eae6..109a136419 100644 --- a/ietf/nomcom/resources.py +++ b/ietf/nomcom/resources.py @@ -115,11 +115,11 @@ class Meta: api.nomcom.register(NomineePositionResource()) from ietf.name.resources import FeedbackTypeNameResource -from ietf.utils.resources import UserResource +from ietf.person.resources import PersonResource class FeedbackResource(ModelResource): nomcom = ToOneField(NomComResource, 'nomcom') type = ToOneField(FeedbackTypeNameResource, 'type', null=True) - user = ToOneField(UserResource, 'user', null=True) + person = ToOneField(PersonResource, 'person', null=True) positions = ToManyField(PositionResource, 'positions', null=True) nominees = ToManyField(NomineeResource, 'nominees', null=True) class Meta: @@ -136,18 +136,18 @@ class Meta: "time": ALL, "nomcom": ALL_WITH_RELATIONS, "type": ALL_WITH_RELATIONS, - "user": ALL_WITH_RELATIONS, + "person": ALL_WITH_RELATIONS, "positions": ALL_WITH_RELATIONS, "nominees": ALL_WITH_RELATIONS, } api.nomcom.register(FeedbackResource()) -from ietf.utils.resources import UserResource +from ietf.person.resources import PersonResource class NominationResource(ModelResource): position = ToOneField(PositionResource, 'position') nominee = ToOneField(NomineeResource, 'nominee') comments = ToOneField(FeedbackResource, 'comments') - user = ToOneField(UserResource, 'user', null=True) + person = ToOneField(PersonResource, 'person', null=True) class Meta: cache = SimpleCache() queryset = Nomination.objects.all() @@ -164,7 +164,7 @@ class Meta: "position": ALL_WITH_RELATIONS, "nominee": ALL_WITH_RELATIONS, "comments": ALL_WITH_RELATIONS, - "user": ALL_WITH_RELATIONS, + "person": ALL_WITH_RELATIONS, } api.nomcom.register(NominationResource()) diff --git a/ietf/nomcom/tasks.py b/ietf/nomcom/tasks.py new file mode 100644 index 0000000000..3d063a6b26 --- /dev/null +++ b/ietf/nomcom/tasks.py @@ -0,0 +1,10 @@ +# Copyright The IETF Trust 2024, All Rights Reserved + +from celery import shared_task + +from .utils import send_reminders + + +@shared_task +def send_nomcom_reminders_task(): + send_reminders() diff --git a/ietf/nomcom/templatetags/nomcom_tags.py b/ietf/nomcom/templatetags/nomcom_tags.py index 05a2c2e8b8..8f795be80e 100644 --- a/ietf/nomcom/templatetags/nomcom_tags.py +++ b/ietf/nomcom/templatetags/nomcom_tags.py @@ -1,12 +1,14 @@ -# Copyright The IETF Trust 2013-2019, All Rights Reserved +# Copyright The IETF Trust 2013-2023, All Rights Reserved import os import tempfile import re +from collections import defaultdict + from django import template from django.conf import settings from django.template.defaultfilters import linebreaksbr, force_escape -from django.utils.encoding import force_text, DjangoUnicodeDecodeError +from django.utils.encoding import force_str, DjangoUnicodeDecodeError from django.utils.safestring import mark_safe import debug # pyflakes:ignore @@ -55,8 +57,10 @@ def formatted_email(address): @register.simple_tag def decrypt(string, request, year, plain=False): - key = retrieve_nomcom_private_key(request, year) - + try: + key = retrieve_nomcom_private_key(request, year) + except UnicodeError: + return f"-*- Encrypted text [Error retrieving private key, contact the secretariat ({settings.SECRETARIAT_SUPPORT_EMAIL})]" if not key: return '-*- Encrypted text [No private key provided] -*-' @@ -68,7 +72,7 @@ def decrypt(string, request, year, plain=False): code, out, error = pipe(command % (settings.OPENSSL_COMMAND, encrypted_file.name), key) try: - out = force_text(out) + out = force_str(out) except DjangoUnicodeDecodeError: pass if code != 0: @@ -82,3 +86,11 @@ def decrypt(string, request, year, plain=False): if not plain: return force_escape(linebreaksbr(out)) return mark_safe(force_escape(out)) + +@register.filter +def feedback_totals(staterank_list): + totals = defaultdict(lambda: 0) + for fb_dict in staterank_list: + for fbtype_name, fbtype_count, _ in fb_dict['feedback']: + totals[fbtype_name] += fbtype_count + return totals.values() diff --git a/ietf/nomcom/test_data.py b/ietf/nomcom/test_data.py index fc4c4fded0..c969022ec6 100644 --- a/ietf/nomcom/test_data.py +++ b/ietf/nomcom/test_data.py @@ -94,7 +94,8 @@ def check_comments(encryped, plain, privatekey_file): decrypted_file.close() encrypted_file.close() - decrypted_comments = io.open(decrypted_file.name, 'rb').read().decode('utf-8') + with io.open(decrypted_file.name, 'rb') as fd: + decrypted_comments = fd.read().decode('utf-8') os.unlink(encrypted_file.name) os.unlink(decrypted_file.name) @@ -116,7 +117,8 @@ def nomcom_test_data(): nomcom_test_cert_file, privatekey_file = generate_cert() nomcom.public_key.storage = FileSystemStorage(location=settings.NOMCOM_PUBLIC_KEYS_DIR) - nomcom.public_key.save('cert', File(io.open(nomcom_test_cert_file.name, 'r'))) + with io.open(nomcom_test_cert_file.name, 'r') as fd: + nomcom.public_key.save('cert', File(fd)) # chair and member create_person(group, "chair", username=CHAIR_USER, email_address='%s%s'%(CHAIR_USER,EMAIL_DOMAIN)) diff --git a/ietf/nomcom/tests.py b/ietf/nomcom/tests.py index 2cc9ac328d..210788ce07 100644 --- a/ietf/nomcom/tests.py +++ b/ietf/nomcom/tests.py @@ -1,9 +1,9 @@ -# Copyright The IETF Trust 2012-2022, All Rights Reserved -# -*- coding: utf-8 -*- +# Copyright The IETF Trust 2012-2025, All Rights Reserved import datetime import io +from unittest import mock import random import shutil @@ -23,32 +23,49 @@ import debug # pyflakes:ignore +from ietf.api.views import EmailIngestionError from ietf.dbtemplate.factories import DBTemplateFactory from ietf.dbtemplate.models import DBTemplate -from ietf.doc.factories import DocEventFactory, WgDocumentAuthorFactory, \ - NewRevisionDocEventFactory, DocumentAuthorFactory +from ietf.doc.factories import ( + DocEventFactory, + WgDocumentAuthorFactory, + NewRevisionDocEventFactory, + DocumentAuthorFactory, + RfcAuthorFactory, + WgDraftFactory, WgRfcFactory, +) from ietf.group.factories import GroupFactory, GroupHistoryFactory, RoleFactory, RoleHistoryFactory from ietf.group.models import Group, Role -from ietf.meeting.factories import MeetingFactory, AttendedFactory +from ietf.meeting.factories import MeetingFactory, AttendedFactory, RegistrationFactory +from ietf.meeting.models import Registration from ietf.message.models import Message from ietf.nomcom.test_data import nomcom_test_data, generate_cert, check_comments, \ COMMUNITY_USER, CHAIR_USER, \ MEMBER_USER, SECRETARIAT_USER, EMAIL_DOMAIN, NOMCOM_YEAR from ietf.nomcom.models import NomineePosition, Position, Nominee, \ NomineePositionStateName, Feedback, FeedbackTypeName, \ - Nomination, FeedbackLastSeen, TopicFeedbackLastSeen, ReminderDates -from ietf.nomcom.management.commands.send_reminders import Command, is_time_to_send + Nomination, FeedbackLastSeen, TopicFeedbackLastSeen, ReminderDates, \ + NomCom from ietf.nomcom.factories import NomComFactory, FeedbackFactory, TopicFactory, \ nomcom_kwargs_for_year, provide_private_key_to_test_client, \ key -from ietf.nomcom.utils import get_nomcom_by_year, make_nomineeposition, \ - get_hash_nominee_position, is_eligible, list_eligible, \ - get_eligibility_date, suggest_affiliation, \ - decorate_volunteers_with_qualifications +from ietf.nomcom.tasks import send_nomcom_reminders_task +from ietf.nomcom.utils import ( + get_nomcom_by_year, + make_nomineeposition, + get_hash_nominee_position, + is_eligible, + list_eligible, + get_eligibility_date, + suggest_affiliation, + ingest_feedback_email, + decorate_volunteers_with_qualifications, + send_reminders, + _is_time_to_send_reminder, + get_qualified_author_queryset, +) from ietf.person.factories import PersonFactory, EmailFactory from ietf.person.models import Email, Person -from ietf.stats.models import MeetingRegistration -from ietf.stats.factories import MeetingRegistrationFactory from ietf.utils.mail import outbox, empty_outbox, get_payload_text from ietf.utils.test_utils import login_testing_unauthorized, TestCase, unicontent from ietf.utils.timezone import date_today, datetime_today, datetime_from_date, DEADLINE_TZINFO @@ -98,6 +115,7 @@ def setUp(self): self.private_nominate_newperson_url = reverse('ietf.nomcom.views.private_nominate_newperson', kwargs={'year': self.year}) self.add_questionnaire_url = reverse('ietf.nomcom.views.private_questionnaire', kwargs={'year': self.year}) self.private_feedback_url = reverse('ietf.nomcom.views.private_feedback', kwargs={'year': self.year}) + self.private_feedback_email_url = reverse('ietf.nomcom.views.private_feedback_email', kwargs={'year': self.year}) self.positions_url = reverse('ietf.nomcom.views.list_positions', kwargs={'year': self.year}) self.edit_position_url = reverse('ietf.nomcom.views.edit_position', kwargs={'year': self.year}) @@ -120,7 +138,7 @@ def access_member_url(self, url): self.check_url_status(url, 200) self.client.logout() login_testing_unauthorized(self, MEMBER_USER, url) - return self.check_url_status(url, 200) + self.check_url_status(url, 200) def access_chair_url(self, url): login_testing_unauthorized(self, COMMUNITY_USER, url) @@ -132,7 +150,7 @@ def access_secretariat_url(self, url): login_testing_unauthorized(self, COMMUNITY_USER, url) login_testing_unauthorized(self, CHAIR_USER, url) login_testing_unauthorized(self, SECRETARIAT_USER, url) - return self.check_url_status(url, 200) + self.check_url_status(url, 200) def test_private_index_view(self): """Verify private home view""" @@ -597,6 +615,8 @@ def test_public_nominate(self): self.nominate_view(public=True,confirmation=True) self.assertEqual(len(outbox), messages_before + 3) + self.assertEqual(Message.objects.count(), 2) + self.assertFalse(Message.objects.filter(subject="Nomination receipt").exists()) self.assertEqual('IETF Nomination Information', outbox[-3]['Subject']) self.assertEqual(self.email_from, outbox[-3]['From']) @@ -623,8 +643,7 @@ def test_public_nominate(self): def test_private_nominate(self): self.access_member_url(self.private_nominate_url) - return self.nominate_view(public=False) - self.client.logout() + self.nominate_view(public=False) def test_public_nominate_newperson(self): login_testing_unauthorized(self, COMMUNITY_USER, self.public_nominate_url) @@ -664,13 +683,13 @@ def test_public_nominate_newperson(self): def test_private_nominate_newperson(self): self.access_member_url(self.private_nominate_url) - return self.nominate_newperson_view(public=False) - self.client.logout() + self.nominate_newperson_view(public=False, confirmation=True) + self.assertFalse(Message.objects.filter(subject="Nomination receipt").exists()) def test_private_nominate_newperson_who_already_exists(self): EmailFactory(address='nominee@example.com') self.access_member_url(self.private_nominate_newperson_url) - return self.nominate_newperson_view(public=False) + self.nominate_newperson_view(public=False) def test_public_nominate_with_automatic_questionnaire(self): nomcom = get_nomcom_by_year(self.year) @@ -686,20 +705,16 @@ def test_public_nominate_with_automatic_questionnaire(self): self.assertIn('nominee@', outbox[1]['To']) - def nominate_view(self, *args, **kwargs): - public = kwargs.pop('public', True) - searched_email = kwargs.pop('searched_email', None) - nominee_email = kwargs.pop('nominee_email', 'nominee@example.com') + def nominate_view(self, public=True, searched_email=None, + nominee_email='nominee@example.com', + nominator_email=COMMUNITY_USER+EMAIL_DOMAIN, + position='IAOC', confirmation=False): + if not searched_email: - searched_email = Email.objects.filter(address=nominee_email).first() - if not searched_email: - searched_email = EmailFactory(address=nominee_email, primary=True, origin='test') + searched_email = Email.objects.filter(address=nominee_email).first() or EmailFactory(address=nominee_email, primary=True, origin='test') if not searched_email.person: searched_email.person = PersonFactory() searched_email.save() - nominator_email = kwargs.pop('nominator_email', "%s%s" % (COMMUNITY_USER, EMAIL_DOMAIN)) - position_name = kwargs.pop('position', 'IAOC') - confirmation = kwargs.pop('confirmation', False) if public: nominate_url = self.public_nominate_url @@ -715,14 +730,15 @@ def nominate_view(self, *args, **kwargs): # save the cert file in tmp #nomcom.public_key.storage.location = tempfile.gettempdir() - nomcom.public_key.save('cert', File(io.open(self.cert_file.name, 'r'))) + with io.open(self.cert_file.name, 'r') as fd: + nomcom.public_key.save('cert', File(fd)) response = self.client.get(nominate_url) self.assertEqual(response.status_code, 200) q = PyQuery(response.content) self.assertEqual(len(q("#nominate-form")), 1) - position = Position.objects.get(name=position_name) + position = Position.objects.get(name=position) comment_text = 'Test nominate view. Comments with accents äöåÄÖÅ éáíóú âêîôû ü àèìòù.' candidate_phone = '123456' @@ -760,12 +776,9 @@ def nominate_view(self, *args, **kwargs): comments=feedback, nominator_email="%s%s" % (COMMUNITY_USER, EMAIL_DOMAIN)) - def nominate_newperson_view(self, *args, **kwargs): - public = kwargs.pop('public', True) - nominee_email = kwargs.pop('nominee_email', 'nominee@example.com') - nominator_email = kwargs.pop('nominator_email', "%s%s" % (COMMUNITY_USER, EMAIL_DOMAIN)) - position_name = kwargs.pop('position', 'IAOC') - confirmation = kwargs.pop('confirmation', False) + def nominate_newperson_view(self, public=True, nominee_email='nominee@example.com', + nominator_email=COMMUNITY_USER+EMAIL_DOMAIN, + position='IAOC', confirmation=False): if public: nominate_url = self.public_nominate_newperson_url @@ -781,14 +794,15 @@ def nominate_newperson_view(self, *args, **kwargs): # save the cert file in tmp #nomcom.public_key.storage.location = tempfile.gettempdir() - nomcom.public_key.save('cert', File(io.open(self.cert_file.name, 'r'))) + with io.open(self.cert_file.name, 'r') as fd: + nomcom.public_key.save('cert', File(fd)) response = self.client.get(nominate_url) self.assertEqual(response.status_code, 200) q = PyQuery(response.content) self.assertEqual(len(q("#nominate-form")), 1) - position = Position.objects.get(name=position_name) + position = Position.objects.get(name=position) candidate_email = nominee_email candidate_name = 'nominee' comment_text = 'Test nominate view. Comments with accents äöåÄÖÅ éáíóú âêîôû ü àèìòù.' @@ -840,18 +854,15 @@ def nominate_newperson_view(self, *args, **kwargs): def test_add_questionnaire(self): self.access_chair_url(self.add_questionnaire_url) - return self.add_questionnaire() - self.client.logout() + self.add_questionnaire() - def add_questionnaire(self, *args, **kwargs): - public = kwargs.pop('public', False) - nominee_email = kwargs.pop('nominee_email', 'nominee@example.com') - nominator_email = kwargs.pop('nominator_email', "%s%s" % (COMMUNITY_USER, EMAIL_DOMAIN)) - position_name = kwargs.pop('position', 'IAOC') + def add_questionnaire(self, public=False, nominee_email='nominee@example.com', + nominator_email=COMMUNITY_USER+EMAIL_DOMAIN, + position='IAOC'): self.nominate_view(public=public, nominee_email=nominee_email, - position=position_name, + position=position, nominator_email=nominator_email) response = self.client.get(self.add_questionnaire_url) @@ -863,13 +874,14 @@ def add_questionnaire(self, *args, **kwargs): # save the cert file in tmp #nomcom.public_key.storage.location = tempfile.gettempdir() - nomcom.public_key.save('cert', File(io.open(self.cert_file.name, 'r'))) + with io.open(self.cert_file.name, 'r') as fd: + nomcom.public_key.save('cert', File(fd)) response = self.client.get(self.add_questionnaire_url) self.assertEqual(response.status_code, 200) self.assertContains(response, "questionnnaireform") - position = Position.objects.get(name=position_name) + position = Position.objects.get(name=position) nominee = Nominee.objects.get(email__address=nominee_email) comment_text = 'Test add questionnaire view. Comments with accents äöåÄÖÅ éáíóú âêîôû ü àèìòù.' @@ -901,6 +913,8 @@ def test_public_feedback(self): # We're interested in the confirmation receipt here self.assertEqual(len(outbox),3) self.assertEqual('NomCom comment confirmation', outbox[2]['Subject']) + self.assertEqual(Message.objects.count(), 2) + self.assertFalse(Message.objects.filter(subject="NomCom comment confirmation").exists()) email_body = get_payload_text(outbox[2]) self.assertIn(position, email_body) self.assertNotIn('$', email_body) @@ -915,18 +929,15 @@ def test_public_feedback(self): def test_private_feedback(self): self.access_member_url(self.private_feedback_url) - return self.feedback_view(public=False) + self.feedback_view(public=False) - def feedback_view(self, *args, **kwargs): - public = kwargs.pop('public', True) - nominee_email = kwargs.pop('nominee_email', 'nominee@example.com') - nominator_email = kwargs.pop('nominator_email', "%s%s" % (COMMUNITY_USER, EMAIL_DOMAIN)) - position_name = kwargs.pop('position', 'IAOC') - confirmation = kwargs.pop('confirmation', False) + def feedback_view(self, public=True, nominee_email='nominee@example.com', + nominator_email=COMMUNITY_USER+EMAIL_DOMAIN, + position='IAOC', confirmation=False): self.nominate_view(public=public, nominee_email=nominee_email, - position=position_name, + position=position, nominator_email=nominator_email) feedback_url = self.public_feedback_url @@ -942,13 +953,14 @@ def feedback_view(self, *args, **kwargs): # save the cert file in tmp #nomcom.public_key.storage.location = tempfile.gettempdir() - nomcom.public_key.save('cert', File(io.open(self.cert_file.name, 'r'))) + with io.open(self.cert_file.name, 'r') as fd: + nomcom.public_key.save('cert', File(fd)) response = self.client.get(feedback_url) self.assertEqual(response.status_code, 200) self.assertNotContains(response, "feedbackform") - position = Position.objects.get(name=position_name) + position = Position.objects.get(name=position) nominee = Nominee.objects.get(email__address=nominee_email) feedback_url += "?nominee=%d&position=%d" % (nominee.id, position.id) @@ -964,7 +976,7 @@ def feedback_view(self, *args, **kwargs): comments = 'Test feedback view. Comments with accents äöåÄÖÅ éáíóú âêîôû ü àèìòù.' test_data = {'comment_text': comments, - 'position_name': position.name, + 'position': position.name, 'nominee_name': nominee.email.person.name, 'nominee_email': nominee.email.address, 'confirmation': confirmation} @@ -1007,6 +1019,43 @@ def feedback_view(self, *args, **kwargs): nominee_position.save() + def test_private_feedback_email(self): + self.access_chair_url(self.private_feedback_email_url) + + feedback_url = self.private_feedback_email_url + response = self.client.get(feedback_url) + self.assertEqual(response.status_code, 200) + + nomcom = get_nomcom_by_year(self.year) + if not nomcom.public_key: + self.assertNotContains(response, "paste-email-feedback-form") + + # save the cert file in tmp + with io.open(self.cert_file.name, 'r') as fd: + nomcom.public_key.save('cert', File(fd)) + + response = self.client.get(feedback_url) + self.assertEqual(response.status_code, 200) + self.assertContains(response, "paste-email-feedback-form") + + headers = \ + "From: Zaphod Beeblebrox \n" \ + "Subject: Ford Prefect\n\n" + body = \ + "Hey, you sass that hoopy Ford Prefect?\n" \ + "There's a frood who really knows where his towel is.\n" + + test_data = {'email_text': body} + response = self.client.post(feedback_url, test_data) + self.assertEqual(response.status_code, 200) + self.assertContains(response, 'Missing email headers') + + test_data = {'email_text': headers + body} + response = self.client.post(feedback_url, test_data) + self.assertEqual(response.status_code, 200) + self.assertContains(response, 'The feedback email has been registered.') + + class NomineePositionStateSaveTest(TestCase): """Tests for the NomineePosition save override method""" @@ -1035,8 +1084,8 @@ def test_state_specified(self): state=NomineePositionStateName.objects.get(slug='accepted')) self.assertEqual(nominee_position.state.slug, 'accepted') - def test_nomine_position_unique(self): - """Verify nomine and position are unique together""" + def test_nominee_position_unique(self): + """Verify nominee and position are unique together""" position = Position.objects.get(name='OAM') NomineePosition.objects.create(position=position, nominee=self.nominee) @@ -1066,7 +1115,8 @@ def test_encrypted_comments(self): # save the cert file in tmp #nomcom.public_key.storage.location = tempfile.gettempdir() - nomcom.public_key.save('cert', File(io.open(self.cert_file.name, 'r'))) + with io.open(self.cert_file.name, 'r') as fd: + nomcom.public_key.save('cert', File(fd)) comment_text = 'Plain text. Comments with accents äöåÄÖÅ éáíóú âêîôû ü àèìòù.' comments = nomcom.encrypt(comment_text) @@ -1080,6 +1130,47 @@ def test_encrypted_comments(self): self.assertNotEqual(feedback.comments, comment_text) self.assertEqual(check_comments(feedback.comments, comment_text, self.privatekey_file), True) + @mock.patch("ietf.nomcom.utils.create_feedback_email") + def test_ingest_feedback_email(self, mock_create_feedback_email): + message = b"This is nomcom feedback" + no_nomcom_year = date_today().year + 10 # a guess at a year with no nomcoms + while NomCom.objects.filter(group__acronym__icontains=no_nomcom_year).exists(): + no_nomcom_year += 1 + inactive_nomcom = NomComFactory(group__state_id="conclude", group__acronym=f"nomcom{no_nomcom_year + 1}") + + # cases where the nomcom does not exist, so admins are notified + for bad_year in (no_nomcom_year, inactive_nomcom.year()): + with self.assertRaises(EmailIngestionError) as context: + ingest_feedback_email(message, bad_year) + self.assertIn("does not exist", context.exception.msg) + self.assertIsNotNone(context.exception.email_body) # error message to be sent + self.assertIsNone(context.exception.email_recipients) # default recipients (i.e., admin) + self.assertIsNone(context.exception.email_original_message) # no original message + self.assertFalse(context.exception.email_attach_traceback) # no traceback + self.assertFalse(mock_create_feedback_email.called) + + # nomcom exists but an error occurs, so feedback goes to the nomcom chair + active_nomcom = NomComFactory(group__acronym=f"nomcom{no_nomcom_year + 2}") + mock_create_feedback_email.side_effect = ValueError("ouch!") + with self.assertRaises(EmailIngestionError) as context: + ingest_feedback_email(message, active_nomcom.year()) + self.assertIn(f"Error ingesting nomcom {active_nomcom.year()}", context.exception.msg) + self.assertIsNotNone(context.exception.email_body) # error message to be sent + self.assertEqual(context.exception.email_recipients, active_nomcom.chair_emails()) + self.assertEqual(context.exception.email_original_message, message) + self.assertFalse(context.exception.email_attach_traceback) # no traceback + self.assertTrue(mock_create_feedback_email.called) + self.assertEqual(mock_create_feedback_email.call_args, mock.call(active_nomcom, message)) + mock_create_feedback_email.reset_mock() + + # and, finally, success + mock_create_feedback_email.side_effect = None + mock_create_feedback_email.return_value = FeedbackFactory(author="someone@example.com") + ingest_feedback_email(message, active_nomcom.year()) + self.assertTrue(mock_create_feedback_email.called) + self.assertEqual(mock_create_feedback_email.call_args, mock.call(active_nomcom, message)) + + class ReminderTest(TestCase): def setUp(self): @@ -1089,7 +1180,8 @@ def setUp(self): self.nomcom = get_nomcom_by_year(NOMCOM_YEAR) self.cert_file, self.privatekey_file = get_cert_files() #self.nomcom.public_key.storage.location = tempfile.gettempdir() - self.nomcom.public_key.save('cert', File(io.open(self.cert_file.name, 'r'))) + with io.open(self.cert_file.name, 'r') as fd: + self.nomcom.public_key.save('cert', File(fd)) gen = Position.objects.get(nomcom=self.nomcom,name='GEN') rai = Position.objects.get(nomcom=self.nomcom,name='RAI') @@ -1121,7 +1213,7 @@ def setUp(self): feedback = Feedback.objects.create(nomcom=self.nomcom, comments=self.nomcom.encrypt('some non-empty comments'), type=FeedbackTypeName.objects.get(slug='questio'), - user=User.objects.get(username=CHAIR_USER)) + person=User.objects.get(username=CHAIR_USER).person) feedback.positions.add(gen) feedback.nominees.add(n) @@ -1129,36 +1221,41 @@ def tearDown(self): teardown_test_public_keys_dir(self) super().tearDown() - def test_is_time_to_send(self): + def test_is_time_to_send_reminder(self): self.nomcom.reminder_interval = 4 today = date_today() - self.assertTrue(is_time_to_send(self.nomcom,today+datetime.timedelta(days=4),today)) + self.assertTrue( + _is_time_to_send_reminder(self.nomcom, today + datetime.timedelta(days=4), today) + ) for delta in range(4): - self.assertFalse(is_time_to_send(self.nomcom,today+datetime.timedelta(days=delta),today)) + self.assertFalse( + _is_time_to_send_reminder( + self.nomcom, today + datetime.timedelta(days=delta), today + ) + ) self.nomcom.reminder_interval = None - self.assertFalse(is_time_to_send(self.nomcom,today,today)) + self.assertFalse(_is_time_to_send_reminder(self.nomcom, today, today)) self.nomcom.reminderdates_set.create(date=today) - self.assertTrue(is_time_to_send(self.nomcom,today,today)) + self.assertTrue(_is_time_to_send_reminder(self.nomcom, today, today)) - def test_command(self): - c = Command() - messages_before=len(outbox) + def test_send_reminders(self): + messages_before = len(outbox) self.nomcom.reminder_interval = 3 self.nomcom.save() - c.handle(None,None) + send_reminders() self.assertEqual(len(outbox), messages_before + 2) self.assertIn('nominee1@example.org', outbox[-1]['To']) self.assertIn('please complete', outbox[-1]['Subject']) self.assertIn('nominee1@example.org', outbox[-2]['To']) self.assertIn('please accept', outbox[-2]['Subject']) - messages_before=len(outbox) + messages_before = len(outbox) self.nomcom.reminder_interval = 4 self.nomcom.save() - c.handle(None,None) + send_reminders() self.assertEqual(len(outbox), messages_before + 1) self.assertIn('nominee2@example.org', outbox[-1]['To']) self.assertIn('please accept', outbox[-1]['Subject']) - + def test_remind_accept_view(self): url = reverse('ietf.nomcom.views.send_reminder_mail', kwargs={'year': NOMCOM_YEAR,'type':'accept'}) login_testing_unauthorized(self, CHAIR_USER, url) @@ -1280,6 +1377,36 @@ def test_cannot_modify_nominees(self): q = PyQuery(response.content) self.assertIn('not active', q('.alert-warning').text() ) + def test_filter_nominees(self): + url = reverse( + "ietf.nomcom.views.private_index", kwargs={"year": self.nc.year()} + ) + login_testing_unauthorized(self, self.chair.user.username, url) + response = self.client.get(url) + self.assertEqual(response.status_code, 200) + + states = list(NomineePositionStateName.objects.values_list("slug", flat=True)) + states += ["not-declined", "questionnaire"] + for state in states: + response = self.client.get(url, {"state": state}) + self.assertEqual(response.status_code, 200) + q = PyQuery(response.content) + nps = [] + if state == "not-declined": + nps = NomineePosition.objects.exclude(state__slug="declined") + elif state == "questionnaire": + nps = [ + np + for np in NomineePosition.objects.not_duplicated() + if np.questionnaires + ] + else: + nps = NomineePosition.objects.filter(state__slug=state) + # nomination state is in third table column + self.assertEqual( + len(nps), len(q("#nominee-position-table td:nth-child(3)")) + ) + def test_email_pasting_closed(self): url = reverse('ietf.nomcom.views.private_feedback_email', kwargs={'year':self.nc.year()}) login_testing_unauthorized(self, self.chair.user.username, url) @@ -1376,6 +1503,35 @@ def test_can_view_but_not_edit_templates(self): q = PyQuery(response.content) self.assertFalse( q('#templateform') ) +class FeedbackIndexTests(TestCase): + + def setUp(self): + super().setUp() + setup_test_public_keys_dir(self) + self.nc = NomComFactory.create(**nomcom_kwargs_for_year()) + self.author = PersonFactory.create().email_set.first().address + self.member = self.nc.group.role_set.filter(name='member').first().person + self.nominee = self.nc.nominee_set.order_by('pk').first() + self.position = self.nc.position_set.first() + for type_id in ['comment','nomina','questio']: + f = FeedbackFactory.create(author=self.author,nomcom=self.nc,type_id=type_id) + f.positions.add(self.position) + f.nominees.add(self.nominee) + + def tearDown(self): + teardown_test_public_keys_dir(self) + super().tearDown() + + def test_feedback_index_totals(self): + url = reverse('ietf.nomcom.views.view_feedback',kwargs={'year':self.nc.year()}) + login_testing_unauthorized(self, self.member.user.username, url) + provide_private_key_to_test_client(self) + response = self.client.get(url) + self.assertEqual(response.status_code,200) + q = PyQuery(response.content) + r = q('tfoot').eq(0).find('td').contents() + self.assertEqual([a.strip() for a in r], ['1', '1', '1', '0']) + class FeedbackLastSeenTests(TestCase): def setUp(self): @@ -1409,7 +1565,7 @@ def test_feedback_index_badges(self): response = self.client.get(url) self.assertEqual(response.status_code,200) q = PyQuery(response.content) - self.assertEqual( len(q('.bg-success')), 4 ) + self.assertEqual( len(q('.text-bg-success')), 4 ) f = self.nc.feedback_set.first() f.time = self.hour_ago @@ -1419,20 +1575,20 @@ def test_feedback_index_badges(self): response = self.client.get(url) self.assertEqual(response.status_code,200) q = PyQuery(response.content) - self.assertEqual( len(q('.bg-success')), 3 ) + self.assertEqual( len(q('.text-bg-success')), 3 ) FeedbackLastSeen.objects.update(time=self.second_from_now) response = self.client.get(url) self.assertEqual(response.status_code,200) q = PyQuery(response.content) - self.assertEqual( len(q('.bg-success')), 1 ) + self.assertEqual( len(q('.text-bg-success')), 1 ) TopicFeedbackLastSeen.objects.create(reviewer=self.member,topic=self.topic) TopicFeedbackLastSeen.objects.update(time=self.second_from_now) response = self.client.get(url) self.assertEqual(response.status_code,200) q = PyQuery(response.content) - self.assertEqual( len(q('.bg-success')), 0 ) + self.assertEqual( len(q('.text-bg-success')), 0 ) def test_feedback_nominee_badges(self): url = reverse('ietf.nomcom.views.view_feedback_nominee', kwargs={'year':self.nc.year(), 'nominee_id':self.nominee.id}) @@ -1441,7 +1597,7 @@ def test_feedback_nominee_badges(self): response = self.client.get(url) self.assertEqual(response.status_code,200) q = PyQuery(response.content) - self.assertEqual( len(q('.bg-success')), 3 ) + self.assertEqual( len(q('.text-bg-success')), 3 ) f = self.nc.feedback_set.first() f.time = self.hour_ago @@ -1451,13 +1607,13 @@ def test_feedback_nominee_badges(self): response = self.client.get(url) self.assertEqual(response.status_code,200) q = PyQuery(response.content) - self.assertEqual( len(q('.bg-success')), 2 ) + self.assertEqual( len(q('.text-bg-success')), 2 ) FeedbackLastSeen.objects.update(time=self.second_from_now) response = self.client.get(url) self.assertEqual(response.status_code,200) q = PyQuery(response.content) - self.assertEqual( len(q('.bg-success')), 0 ) + self.assertEqual( len(q('.text-bg-success')), 0 ) def test_feedback_topic_badges(self): url = reverse('ietf.nomcom.views.view_feedback_topic', kwargs={'year':self.nc.year(), 'topic_id':self.topic.id}) @@ -1466,7 +1622,7 @@ def test_feedback_topic_badges(self): response = self.client.get(url) self.assertEqual(response.status_code,200) q = PyQuery(response.content) - self.assertEqual( len(q('.bg-success')), 1 ) + self.assertEqual( len(q('.text-bg-success')), 1 ) f = self.topic.feedback_set.first() f.time = self.hour_ago @@ -1476,20 +1632,23 @@ def test_feedback_topic_badges(self): response = self.client.get(url) self.assertEqual(response.status_code,200) q = PyQuery(response.content) - self.assertEqual( len(q('.bg-success')), 0 ) + self.assertEqual( len(q('.text-bg-success')), 0 ) TopicFeedbackLastSeen.objects.update(time=self.second_from_now) response = self.client.get(url) self.assertEqual(response.status_code,200) q = PyQuery(response.content) - self.assertEqual( len(q('.bg-success')), 0 ) + self.assertEqual( len(q('.text-bg-success')), 0 ) class NewActiveNomComTests(TestCase): def setUp(self): super().setUp() setup_test_public_keys_dir(self) - self.nc = NomComFactory.create(**nomcom_kwargs_for_year(year=random.randint(1992,2100))) + # Pin nomcom years to be after 2008 or later so that ietf.nomcom.utils.list_eligible can + # return something other than empty. Note that anything after 2022 is suspect, and that + # we should revisit this when implementing RFC 9389. + self.nc = NomComFactory.create(**nomcom_kwargs_for_year(year=random.randint(2008,2100))) self.chair = self.nc.group.role_set.filter(name='chair').first().person self.saved_days_to_expire_nomination_link = settings.DAYS_TO_EXPIRE_NOMINATION_LINK @@ -1568,6 +1727,16 @@ def test_provide_private_key(self): login_testing_unauthorized(self,self.chair.user.username,url) response = self.client.get(url) self.assertEqual(response.status_code,200) + # Check that we get an error if there's an encoding problem talking to openssl + # "\xc3\x28" is an invalid utf8 string + with mock.patch("ietf.nomcom.utils.pipe", return_value=(0, b"\xc3\x28", None)): + response = self.client.post(url, {'key': force_str(key)}) + self.assertFormError( + response.context["form"], + None, + "An internal error occurred while adding your private key to your session." + f"Please contact the secretariat for assistance ({settings.SECRETARIAT_SUPPORT_EMAIL})", + ) response = self.client.post(url,{'key': force_str(key)}) self.assertEqual(response.status_code,302) @@ -1906,7 +2075,15 @@ def first_meeting_of_year(year): if not ' ' in ascii: continue first_name, last_name = ascii.rsplit(None, 1) - MeetingRegistration.objects.create(meeting=meeting, first_name=first_name, last_name=last_name, person=person, country_code='WO', email=email, attended=True) + RegistrationFactory( + meeting=meeting, + first_name=first_name, + last_name=last_name, + person=person, + country_code='WO', + email=email, + attended=True + ) for view in ('public_eligible','private_eligible'): url = reverse(f'ietf.nomcom.views.{view}',kwargs={'year':self.nc.year()}) for username in (self.chair.user.username,'secretary'): @@ -1929,7 +2106,7 @@ def first_meeting_of_year(year): for number in range(meeting_start, meeting_start+8): m = MeetingFactory.create(type_id='ietf', number=number) for p in people: - m.meetingregistration_set.create(person=p) + RegistrationFactory(meeting=m, person=p, checkedin=True, attended=True) for p in people: self.nc.volunteer_set.create(person=p,affiliation='something') for view in ('public_volunteers','private_volunteers'): @@ -1947,9 +2124,13 @@ def first_meeting_of_year(year): login_testing_unauthorized(self,self.chair.user.username,url) response = self.client.get(url) self.assertContains(response,people[-1].email(),status_code=200) - - - + unqualified_person = PersonFactory() + url = reverse('ietf.nomcom.views.qualified_volunteer_list_for_announcement',kwargs={'year':year}) + self.client.logout() + login_testing_unauthorized(self,self.chair.user.username,url) + response = self.client.get(url) + self.assertContains(response, people[-1].plain_name(), status_code=200) + self.assertNotContains(response, unqualified_person.plain_name()) class NomComIndexTests(TestCase): def setUp(self): @@ -1973,10 +2154,10 @@ def do_common_work(self,url,expected_form): response = self.client.get(url) self.assertEqual(response.status_code,200) q=PyQuery(response.content) - text_bits = [x.xpath('./text()') for x in q('.alert-warning')] + text_bits = [x.xpath('.//text()') for x in q('.alert-warning')] flat_text_bits = [item for sublist in text_bits for item in sublist] self.assertTrue(any(['not yet' in y for y in flat_text_bits])) - self.assertEqual(bool(q('form:not(.navbar-form)')),expected_form) + self.assertEqual(bool(q('#content form:not(.navbar-form)')),expected_form) self.client.logout() def test_not_yet(self): @@ -2065,7 +2246,7 @@ def test_public_accepting_feedback(self): self.assertIn('not currently accepting feedback', unicontent(response)) test_data = {'comment_text': 'junk', - 'position_name': pos.name, + 'position': pos.name, 'nominee_name': pos.nominee_set.first().email.person.name, 'nominee_email': pos.nominee_set.first().email.address, 'confirmation': False, @@ -2274,6 +2455,85 @@ def test_get_eligibility_date(self): NomComFactory(group__acronym=f'nomcom{this_year}', first_call_for_volunteers=datetime.date(this_year,5,6)) self.assertEqual(get_eligibility_date(),datetime.date(this_year,5,6)) + def test_get_qualified_author_queryset(self): + """get_qualified_author_queryset implements the eligiblity rules correctly + + This is not an exhaustive test of corner cases. Overlaps considerably with + rfc8989EligibilityTests.test_elig_by_author(). + """ + people = PersonFactory.create_batch(2) + extra_person = PersonFactory() + base_qs = Person.objects.filter(pk__in=[person.pk for person in people]) + now = datetime.datetime.now(tz=datetime.UTC) + one_year = datetime.timedelta(days=365) + + # Authors with no qualifying drafts + self.assertCountEqual( + get_qualified_author_queryset(base_qs, now - 5 * one_year, now), [] + ) + + # Authors with one qualifying draft + approved_draft = WgDraftFactory(authors=people, states=[("draft", "active")]) + DocEventFactory( + type="iesg_approved", + doc=approved_draft, + time=now - 4 * one_year, + ) + self.assertCountEqual( + get_qualified_author_queryset(base_qs, now - 5 * one_year, now), [] + ) + + # Create a draft that was published into an RFC. Give it an extra author who + # should not be eligible. + published_draft = WgDraftFactory(authors=people, states=[("draft", "rfc")]) + DocEventFactory( + type="iesg_approved", + doc=published_draft, + time=now - 5.5 * one_year, # < 6 years ago + ) + rfc = WgRfcFactory( + authors=people + [extra_person], + group=published_draft.group, + ) + DocEventFactory( + type="published_rfc", + doc=rfc, + time=now - 0.5 * one_year, # < 1 year ago + ) + # Period 6 years ago to 1 year ago - authors are eligible due to the + # iesg-approved draft in this window + self.assertCountEqual( + get_qualified_author_queryset(base_qs, now - 6 * one_year, now - one_year), + people, + ) + + # Period 5 years ago to now - authors are eligible due to the RFC publication + self.assertCountEqual( + get_qualified_author_queryset(base_qs, now - 5 * one_year, now), + people, + ) + + # Use the extra_person to check that a single doc can't count both as an + # RFC _and_ an approved draft. Use an eligibility interval that includes both + # the approval and the RFC publication + self.assertCountEqual( + get_qualified_author_queryset(base_qs, now - 6 * one_year, now), + people, # does not include extra_person! + ) + + # Now add an RfcAuthor for only one of the two authors to the RFC. This should + # remove the other author from the eligibility list because the DocumentAuthor + # records are no longer used. + RfcAuthorFactory( + document=rfc, + person=people[0], + titlepage_name="P. Zero", + ) + self.assertCountEqual( + get_qualified_author_queryset(base_qs, now - 5 * one_year, now), + [people[0]], + ) + class rfc8713EligibilityTests(TestCase): @@ -2292,16 +2552,29 @@ def setUp(self): self.eligible_people = list() self.ineligible_people = list() + # Section 4.14 qualification criteria for combo_len in range(0,6): for combo in combinations(meetings,combo_len): p = PersonFactory() for m in combo: - MeetingRegistrationFactory(person=p, meeting=m, attended=True) + RegistrationFactory(person=p, meeting=m, attended=True) if combo_len<3: self.ineligible_people.append(p) else: self.eligible_people.append(p) + # Section 4.15 disqualification criteria + def ineligible_person_with_role(**kwargs): + p = RoleFactory(**kwargs).person + for m in meetings: + RegistrationFactory(person=p, meeting=m, attended=True) + self.ineligible_people.append(p) + for group in ['isocbot', 'ietf-trust', 'llc-board', 'iab']: + for role in ['member', 'chair']: + ineligible_person_with_role(group__acronym=group, name_id=role) + ineligible_person_with_role(group__type_id='area', group__state_id='active',name_id='ad') + ineligible_person_with_role(group=self.nomcom.group, name_id='chair') + # No-one is eligible for the other_nomcom self.other_nomcom = NomComFactory(group__acronym='nomcom2018',first_call_for_volunteers=datetime.date(2018,5,1)) @@ -2309,8 +2582,7 @@ def setUp(self): self.other_date = datetime.date(2009,5,1) self.other_people = PersonFactory.create_batch(1) for date in (datetime.date(2009,3,1), datetime.date(2008,11,1), datetime.date(2008,7,1)): - MeetingRegistrationFactory(person=self.other_people[0],meeting__date=date, meeting__type_id='ietf', attended=True) - + RegistrationFactory(person=self.other_people[0], meeting__date=date, meeting__type_id='ietf', attended=True) def test_is_person_eligible(self): for person in self.eligible_people: @@ -2354,7 +2626,7 @@ def setUp(self): for combo in combinations(meetings,combo_len): p = PersonFactory() for m in combo: - MeetingRegistrationFactory(person=p, meeting=m, attended=True) + RegistrationFactory(person=p, meeting=m, attended=True) if combo_len<3: self.ineligible_people.append(p) else: @@ -2402,7 +2674,7 @@ def test_elig_by_meetings(self): for combo in combinations(prev_five,combo_len): p = PersonFactory() for m in combo: - MeetingRegistrationFactory(person=p, meeting=m, attended=True) # not checkedin because this forces looking at older meetings + RegistrationFactory(person=p, meeting=m, attended=True) # not checkedin because this forces looking at older meetings AttendedFactory(session__meeting=m, session__type_id='plenary',person=p) if combo_len<3: ineligible_people.append(p) @@ -2417,8 +2689,9 @@ def test_elig_by_meetings(self): for person in ineligible_people: self.assertFalse(is_eligible(person,nomcom)) - Person.objects.filter(pk__in=[p.pk for p in eligible_people+ineligible_people]).delete() - + people = Person.objects.filter(pk__in=[p.pk for p in eligible_people + ineligible_people]) + Registration.objects.filter(person__in=people).delete() + people.delete() def test_elig_by_office_active_groups(self): @@ -2545,33 +2818,41 @@ def test_elig_by_author(self): ineligible = set() p = PersonFactory() - ineligible.add(p) - + ineligible.add(p) # no RFCs or iesg-approved drafts p = PersonFactory() - da = WgDocumentAuthorFactory(person=p) - DocEventFactory(type='published_rfc',doc=da.document,time=middle_date) - ineligible.add(p) + doc = WgRfcFactory(authors=[p]) + DocEventFactory(type='published_rfc', doc=doc, time=middle_date) + ineligible.add(p) # only one RFC p = PersonFactory() - da = WgDocumentAuthorFactory(person=p) + da = WgDocumentAuthorFactory( + person=p, + document__states=[("draft", "active"), ("draft-rfceditor", "ref")], + ) DocEventFactory(type='iesg_approved',doc=da.document,time=last_date) - da = WgDocumentAuthorFactory(person=p) - DocEventFactory(type='published_rfc',doc=da.document,time=first_date) - eligible.add(p) + doc = WgRfcFactory(authors=[p]) + DocEventFactory(type='published_rfc', doc=doc, time=first_date) + eligible.add(p) # one RFC and one iesg-approved draft p = PersonFactory() - da = WgDocumentAuthorFactory(person=p) + da = WgDocumentAuthorFactory( + person=p, + document__states=[("draft", "active"), ("draft-rfceditor", "ref")], + ) DocEventFactory(type='iesg_approved',doc=da.document,time=middle_date) - da = WgDocumentAuthorFactory(person=p) - DocEventFactory(type='published_rfc',doc=da.document,time=day_before_first_date) - ineligible.add(p) + doc = WgRfcFactory(authors=[p]) + DocEventFactory(type='published_rfc', doc=doc, time=day_before_first_date) + ineligible.add(p) # RFC is out of the eligibility window p = PersonFactory() - da = WgDocumentAuthorFactory(person=p) + da = WgDocumentAuthorFactory( + person=p, + document__states=[("draft", "active"), ("draft-rfceditor", "ref")], + ) DocEventFactory(type='iesg_approved',doc=da.document,time=day_after_last_date) - da = WgDocumentAuthorFactory(person=p) - DocEventFactory(type='published_rfc',doc=da.document,time=middle_date) - ineligible.add(p) + doc = WgRfcFactory(authors=[p]) + DocEventFactory(type='published_rfc', doc=doc, time=middle_date) + ineligible.add(p) # iesg approval is outside the eligibility window for person in eligible: self.assertTrue(is_eligible(person,nomcom)) @@ -2582,7 +2863,7 @@ def test_elig_by_author(self): self.assertEqual(set(list_eligible(nomcom=nomcom)),set(eligible)) Person.objects.filter(pk__in=[p.pk for p in eligible.union(ineligible)]).delete() -class rfc8989bisEligibilityTests(TestCase): +class rfc9389EligibilityTests(TestCase): def setUp(self): super().setUp() @@ -2602,7 +2883,7 @@ def setUp(self): def test_registration_is_not_enough(self): p = PersonFactory() for meeting in self.meetings: - MeetingRegistrationFactory(person=p, meeting=meeting, checkedin=False) + RegistrationFactory(person=p, meeting=meeting, checkedin=False) self.assertFalse(is_eligible(p, self.nomcom)) def test_elig_by_meetings(self): @@ -2619,7 +2900,7 @@ def test_elig_by_meetings(self): for method in attendance_methods: p = PersonFactory() for meeting in combo: - MeetingRegistrationFactory(person=p, meeting=meeting, reg_type='onsite', checkedin=(method in ('checkedin', 'both'))) + RegistrationFactory(person=p, meeting=meeting, checkedin=(method in ('checkedin', 'both'))) if method in ('session', 'both'): AttendedFactory(session__meeting=meeting, session__type_id='plenary',person=p) if combo_len<3: @@ -2635,6 +2916,7 @@ def test_elig_by_meetings(self): for person in ineligible_people: self.assertFalse(is_eligible(person,self.nomcom)) + class VolunteerTests(TestCase): def test_volunteer(self): @@ -2651,7 +2933,7 @@ def test_volunteer(self): self.assertContains(r, 'NomCom is not accepting volunteers at this time', status_code=200) nomcom.is_accepting_volunteers = True nomcom.save() - MeetingRegistrationFactory(person=person, affiliation='mtg_affiliation', checkedin=True) + RegistrationFactory(person=person, affiliation='mtg_affiliation', checkedin=True) r = self.client.get(url) self.assertContains(r, 'Volunteer for NomCom', status_code=200) self.assertContains(r, 'mtg_affiliation') @@ -2698,15 +2980,38 @@ def test_volunteer(self): def test_suggest_affiliation(self): person = PersonFactory() - self.assertEqual(suggest_affiliation(person), '') - da = DocumentAuthorFactory(person=person,affiliation='auth_affil') + self.assertEqual(suggest_affiliation(person), "") + rfc_da = DocumentAuthorFactory( + person=person, + document__type_id="rfc", + affiliation="", + ) + rfc = rfc_da.document + DocEventFactory(doc=rfc, type="published_rfc") + self.assertEqual(suggest_affiliation(person), "") + + rfc_da.affiliation = "rfc_da_affil" + rfc_da.save() + self.assertEqual(suggest_affiliation(person), "rfc_da_affil") + + rfc_ra = RfcAuthorFactory(person=person, document=rfc, affiliation="") + self.assertEqual(suggest_affiliation(person), "") + + rfc_ra.affiliation = "rfc_ra_affil" + rfc_ra.save() + self.assertEqual(suggest_affiliation(person), "rfc_ra_affil") + + da = DocumentAuthorFactory(person=person, affiliation="auth_affil") NewRevisionDocEventFactory(doc=da.document) - self.assertEqual(suggest_affiliation(person), 'auth_affil') + self.assertEqual(suggest_affiliation(person), "auth_affil") + nc = NomComFactory() - nc.volunteer_set.create(person=person,affiliation='volunteer_affil') - self.assertEqual(suggest_affiliation(person), 'volunteer_affil') - MeetingRegistrationFactory(person=person, affiliation='meeting_affil') - self.assertEqual(suggest_affiliation(person), 'meeting_affil') + nc.volunteer_set.create(person=person, affiliation="volunteer_affil") + self.assertEqual(suggest_affiliation(person), "volunteer_affil") + + RegistrationFactory(person=person, affiliation="meeting_affil") + self.assertEqual(suggest_affiliation(person), "meeting_affil") + class VolunteerDecoratorUnitTests(TestCase): def test_decorate_volunteers_with_qualifications(self): @@ -2723,7 +3028,7 @@ def test_decorate_volunteers_with_qualifications(self): ('106', datetime.date(2019, 11, 16)), ]] for m in meetings: - MeetingRegistrationFactory(meeting=m, person=meeting_person, attended=True) + RegistrationFactory(meeting=m, person=meeting_person, attended=True) AttendedFactory(session__meeting=m, session__type_id='plenary', person=meeting_person) nomcom.volunteer_set.create(person=meeting_person) @@ -2742,15 +3047,15 @@ def test_decorate_volunteers_with_qualifications(self): author_person = PersonFactory() for i in range(2): - da = WgDocumentAuthorFactory(person=author_person) + doc = WgRfcFactory(authors=[author_person]) DocEventFactory( type='published_rfc', - doc=da.document, + doc=doc, time=datetime.datetime( elig_date.year - 3, elig_date.month, 28 if elig_date.month == 2 and elig_date.day == 29 else elig_date.day, - tzinfo=datetime.timezone.utc, + tzinfo=datetime.UTC, ) ) nomcom.volunteer_set.create(person=author_person) @@ -2766,3 +3071,120 @@ def test_decorate_volunteers_with_qualifications(self): self.assertEqual(v.qualifications,'path_2') if v.person == author_person: self.assertEqual(v.qualifications,'path_3') + +class ReclassifyFeedbackTests(TestCase): + """Tests for feedback reclassification""" + + def setUp(self): + super().setUp() + setup_test_public_keys_dir(self) + self.nc = NomComFactory.create(**nomcom_kwargs_for_year()) + self.chair = self.nc.group.role_set.filter(name='chair').first().person + self.member = self.nc.group.role_set.filter(name='member').first().person + self.nominee = self.nc.nominee_set.order_by('pk').first() + self.position = self.nc.position_set.first() + self.topic = self.nc.topic_set.first() + + def tearDown(self): + teardown_test_public_keys_dir(self) + super().tearDown() + + def test_download_feedback_nominee(self): + # not really a reclassification test, but in closely adjacent code + fb = FeedbackFactory.create(nomcom=self.nc,type_id='questio') + fb.positions.add(self.position) + fb.nominees.add(self.nominee) + fb.save() + self.assertEqual(Feedback.objects.questionnaires().count(), 1) + + url = reverse('ietf.nomcom.views.view_feedback_nominee', kwargs={'year':self.nc.year(), 'nominee_id':self.nominee.id}) + login_testing_unauthorized(self,self.member.user.username,url) + provide_private_key_to_test_client(self) + response = self.client.post(url, {'feedback_id': fb.id, 'submit': 'download'}) + self.assertEqual(response.status_code, 403) + + self.client.logout() + self.client.login(username=self.chair.user.username, password=self.chair.user.username + "+password") + provide_private_key_to_test_client(self) + + response = self.client.post(url, {'feedback_id': fb.id, 'submit': 'download'}) + self.assertEqual(response.status_code, 200) + self.assertIn('questionnaire-', response['Content-Disposition']) + + def test_reclassify_feedback_nominee(self): + fb = FeedbackFactory.create(nomcom=self.nc,type_id='comment') + fb.positions.add(self.position) + fb.nominees.add(self.nominee) + fb.save() + self.assertEqual(Feedback.objects.comments().count(), 1) + + url = reverse('ietf.nomcom.views.view_feedback_nominee', kwargs={'year':self.nc.year(), 'nominee_id':self.nominee.id}) + login_testing_unauthorized(self,self.member.user.username,url) + provide_private_key_to_test_client(self) + response = self.client.post(url, {'feedback_id': fb.id, 'type': 'obe', 'submit': 'reclassify'}) + self.assertEqual(response.status_code, 403) + + self.client.logout() + self.client.login(username=self.chair.user.username, password=self.chair.user.username + "+password") + provide_private_key_to_test_client(self) + + response = self.client.post(url, {'feedback_id': fb.id, 'type': 'obe', 'submit': 'reclassify'}) + self.assertEqual(response.status_code, 200) + + fb = Feedback.objects.get(id=fb.id) + self.assertEqual(fb.type_id,'obe') + self.assertEqual(Feedback.objects.comments().count(), 0) + self.assertEqual(Feedback.objects.filter(type='obe').count(), 1) + + def test_reclassify_feedback_topic(self): + fb = FeedbackFactory.create(nomcom=self.nc,type_id='comment') + fb.topics.add(self.topic) + fb.save() + self.assertEqual(Feedback.objects.comments().count(), 1) + + url = reverse('ietf.nomcom.views.view_feedback_topic', kwargs={'year':self.nc.year(), 'topic_id':self.topic.id}) + login_testing_unauthorized(self,self.member.user.username,url) + provide_private_key_to_test_client(self) + response = self.client.post(url, {'feedback_id': fb.id, 'type': 'unclassified'}) + self.assertEqual(response.status_code, 403) + + self.client.logout() + self.client.login(username=self.chair.user.username, password=self.chair.user.username + "+password") + provide_private_key_to_test_client(self) + + response = self.client.post(url, {'feedback_id': fb.id, 'type': 'unclassified'}) + self.assertEqual(response.status_code, 200) + + fb = Feedback.objects.get(id=fb.id) + self.assertEqual(fb.type_id,None) + self.assertEqual(Feedback.objects.comments().count(), 0) + self.assertEqual(Feedback.objects.filter(type=None).count(), 1) + + def test_reclassify_feedback_unrelated(self): + fb = FeedbackFactory(nomcom=self.nc, type_id='read') + self.assertEqual(Feedback.objects.filter(type='read').count(), 1) + + url = reverse('ietf.nomcom.views.view_feedback_unrelated', kwargs={'year':self.nc.year()}) + login_testing_unauthorized(self,self.member.user.username,url) + provide_private_key_to_test_client(self) + response = self.client.post(url, {'feedback_id': fb.id, 'type': 'junk'}) + self.assertEqual(response.status_code, 403) + + self.client.logout() + self.client.login(username=self.chair.user.username, password=self.chair.user.username + "+password") + provide_private_key_to_test_client(self) + + response = self.client.post(url, {'feedback_id': fb.id, 'type': 'junk'}) + self.assertEqual(response.status_code, 200) + + fb = Feedback.objects.get(id=fb.id) + self.assertEqual(fb.type_id, 'junk') + self.assertEqual(Feedback.objects.filter(type='read').count(), 0) + self.assertEqual(Feedback.objects.filter(type='junk').count(), 1) + + +class TaskTests(TestCase): + @mock.patch("ietf.nomcom.tasks.send_reminders") + def test_send_nomcom_reminders_task(self, mock_send): + send_nomcom_reminders_task() + self.assertEqual(mock_send.call_count, 1) diff --git a/ietf/nomcom/urls.py b/ietf/nomcom/urls.py index f7a19e2226..a3b0c42d3c 100644 --- a/ietf/nomcom/urls.py +++ b/ietf/nomcom/urls.py @@ -22,8 +22,8 @@ url(r'^(?P\d{4})/private/view-feedback/nominee/(?P\d+)$', views.view_feedback_nominee), url(r'^(?P\d{4})/private/view-feedback/topic/(?P\d+)$', views.view_feedback_topic), url(r'^(?P\d{4})/private/edit/nominee/(?P\d+)$', views.edit_nominee), - url(r'^(?P\d{4})/private/merge-nominee/?$', views.private_merge_nominee), - url(r'^(?P\d{4})/private/merge-person/?$', views.private_merge_person), + url(r'^(?P\d{4})/private/merge-nominee/$', views.private_merge_nominee), + url(r'^(?P\d{4})/private/merge-person/$', views.private_merge_person), url(r'^(?P\d{4})/private/send-reminder-mail/(?P\w+)/$', views.send_reminder_mail), url(r'^(?P\d{4})/private/extract-email-lists/$', views.extract_email_lists), url(r'^(?P\d{4})/private/edit-members/$', views.edit_members), @@ -41,6 +41,7 @@ url(r'^(?P\d{4})/private/chair/eligible/$', views.private_eligible), url(r'^(?P\d{4})/private/chair/volunteers/$', views.private_volunteers), url(r'^(?P\d{4})/private/chair/volunteers/csv/$', views.private_volunteers_csv), + url(r'^(?P\d{4})/private/chair/volunteers/announce-list/$', views.qualified_volunteer_list_for_announcement), url(r'^(?P\d{4})/$', views.year_index), url(r'^(?P\d{4})/requirements/$', views.requirements), diff --git a/ietf/nomcom/utils.py b/ietf/nomcom/utils.py index 3227771af5..a2ab680df6 100644 --- a/ietf/nomcom/utils.py +++ b/ietf/nomcom/utils.py @@ -1,4 +1,4 @@ -# Copyright The IETF Trust 2012-2022, All Rights Reserved +# Copyright The IETF Trust 2012-2023, All Rights Reserved # -*- coding: utf-8 -*- @@ -12,11 +12,13 @@ from collections import defaultdict from email import message_from_string, message_from_bytes +from email.errors import HeaderParseError from email.header import decode_header from email.iterators import typed_subpart_iterator from email.utils import parseaddr +from textwrap import dedent -from django.db.models import Q, Count +from django.db.models import Q, Count, F, QuerySet from django.conf import settings from django.contrib.sites.models import Site from django.core.exceptions import ObjectDoesNotExist @@ -25,11 +27,12 @@ from django.shortcuts import get_object_or_404 from ietf.dbtemplate.models import DBTemplate -from ietf.doc.models import DocEvent, NewRevisionDocEvent +from ietf.doc.models import DocEvent, NewRevisionDocEvent, Document from ietf.group.models import Group, Role from ietf.person.models import Email, Person from ietf.mailtrigger.utils import gather_address_lists -from ietf.meeting.models import Meeting, Attended +from ietf.meeting.models import Meeting +from ietf.meeting.utils import participants_for_meeting from ietf.utils.pipe import pipe from ietf.utils.mail import send_mail_text, send_mail, get_payload_text from ietf.utils.log import log @@ -65,8 +68,11 @@ ] # See RFC8713 section 4.15 +# This potentially over-disqualifies past nomcom chairs if some +# nomcom 2+ nomcoms ago is still in the active state DISQUALIFYING_ROLE_QUERY_EXPRESSION = ( Q(group__acronym__in=['isocbot', 'ietf-trust', 'llc-board', 'iab'], name_id__in=['member', 'chair']) | Q(group__type_id='area', group__state='active',name_id='ad') + | Q(group__type_id='nomcom', group__state='active', name_id='chair') ) @@ -83,26 +89,21 @@ def get_year_by_nomcom(nomcom): return m.group(0) -def get_user_email(user): - # a user object already has an email field, but we don't want to - # overwrite anything that might be there, and we don't know that - # what's there is the right thing, so we cache the lookup results in a - # separate attribute - if not hasattr(user, "_email_cache"): - user._email_cache = None - if hasattr(user, "person"): - emails = user.person.email_set.filter(active=True).order_by('-time') - if emails: - user._email_cache = emails[0] - for email in emails: - if email.address == user.username: - user._email_cache = email +def get_person_email(person): + if not hasattr(person, "_email_cache"): + person._email_cache = None + emails = person.email_set.filter(active=True).order_by('-time') + if emails: + person._email_cache = emails[0] + for email in emails: + if email.address.lower() == person.user.username.lower(): + person._email_cache = email else: try: - user._email_cache = Email.objects.get(address=user.username) + person._email_cache = Email.objects.get(address=person.user.username) except ObjectDoesNotExist: pass - return user._email_cache + return person._email_cache def get_hash_nominee_position(date, nominee_position_id): return hmac.new(settings.NOMCOM_APP_SECRET, f"{date}{nominee_position_id}".encode('utf-8'), hashlib.sha256).hexdigest() @@ -172,18 +173,25 @@ def command_line_safe_secret(secret): return base64.encodebytes(secret).decode('utf-8').rstrip() def retrieve_nomcom_private_key(request, year): + """Retrieve decrypted nomcom private key from the session store + + Retrieves encrypted, ascii-armored private key from the session store, encodes + as utf8 bytes, then decrypts. Raises UnicodeError if the value in the session + store cannot be encoded as utf8. + """ private_key = request.session.get('NOMCOM_PRIVATE_KEY_%s' % year, None) if not private_key: return private_key - command = "%s bf -d -in /dev/stdin -k \"%s\" -a" + command = "%s aes-128-ecb -d -in /dev/stdin -k \"%s\" -a -iter 1000" code, out, error = pipe( command % ( settings.OPENSSL_COMMAND, command_line_safe_secret(settings.NOMCOM_APP_SECRET) ), - private_key + # The openssl command expects ascii-armored input, so utf8 encoding should be valid + private_key.encode("utf8") ) if code != 0: log("openssl error: %s:\n Error %s: %s" %(command, code, error)) @@ -191,10 +199,16 @@ def retrieve_nomcom_private_key(request, year): def store_nomcom_private_key(request, year, private_key): + """Put encrypted nomcom private key in the session store + + Encrypts the private key using openssl, then decodes the ascii-armored output + as utf8 and adds to the session store. Raises UnicodeError if the openssl's + output cannot be decoded as utf8. + """ if not private_key: request.session['NOMCOM_PRIVATE_KEY_%s' % year] = '' else: - command = "%s bf -e -in /dev/stdin -k \"%s\" -a" + command = "%s aes-128-ecb -e -in /dev/stdin -k \"%s\" -a -iter 1000" code, out, error = pipe( command % ( settings.OPENSSL_COMMAND, @@ -205,8 +219,9 @@ def store_nomcom_private_key(request, year, private_key): if code != 0: log("openssl error: %s:\n Error %s: %s" %(command, code, error)) if error and error!=b"*** WARNING : deprecated key derivation used.\nUsing -iter or -pbkdf2 would be better.\n": - out = '' - request.session['NOMCOM_PRIVATE_KEY_%s' % year] = out + out = b'' + # The openssl command output in 'out' is an ascii-armored value, so should be utf8-decodable + request.session['NOMCOM_PRIVATE_KEY_%s' % year] = out.decode("utf8") def validate_private_key(key): @@ -428,7 +443,11 @@ def make_nomineeposition_for_newperson(nomcom, candidate_name, candidate_email, def getheader(header_text, default="utf-8"): """Decode the specified header""" - tuples = decode_header(header_text) + try: + tuples = decode_header(header_text) + except TypeError: + return "" + header_sections = [ text.decode(charset or default) if isinstance(text, bytes) else text for text, charset in tuples] return "".join(header_sections) @@ -477,6 +496,9 @@ def parse_email(text): body = get_body(msg) subject = getheader(msg['Subject']) __, addr = parseaddr(msg['From']) + if not addr: + raise HeaderParseError + return addr.lower(), subject, body @@ -513,7 +535,7 @@ def list_eligible(nomcom=None, date=None, base_qs=None): elif eligibility_date.year in (2021,2022): return list_eligible_8989(date=eligibility_date, base_qs=base_qs) elif eligibility_date.year > 2022: - return list_eligible_8989bis(date=eligibility_date, base_qs=base_qs) + return list_eligible_9389(date=eligibility_date, base_qs=base_qs) else: return Person.objects.none() @@ -531,7 +553,7 @@ def decorate_volunteers_with_qualifications(volunteers, nomcom=None, date=None, qualifications.append('path_2') if v.person in author_qs: qualifications.append('path_3') - v.qualifications = ", ".join(qualifications) + v.qualifications = "+".join(qualifications) else: for v in volunteers: v.qualifications = '' @@ -551,8 +573,72 @@ def list_eligible_8788(date, base_qs=None): def get_8989_eligibility_querysets(date, base_qs): return get_threerule_eligibility_querysets(date, base_qs, three_of_five_callable=three_of_five_eligible_8713) -def get_8989bis_eligibility_querysets(date, base_qs): - return get_threerule_eligibility_querysets(date, base_qs, three_of_five_callable=three_of_five_eligible_8989bis) +def get_9389_eligibility_querysets(date, base_qs): + return get_threerule_eligibility_querysets(date, base_qs, three_of_five_callable=three_of_five_eligible_9389) + + +def get_qualified_author_queryset( + base_qs: QuerySet[Person], + eligibility_period_start: datetime.datetime, + eligibility_period_end: datetime.datetime, +): + """Filter a Person queryset, keeping those qualified by RFC 8989's author path + + The author path is defined by "path 3" in section 4 of RFC 8989. It qualifies + a person who has been a front-page listed author or editor of at least two IETF- + stream RFCs within the last five years. An I-D in the RFC Editor queue that was + approved by the IESG is treated as an RFC, using the date of entry to the RFC + Editor queue as the date for qualification. + + This method does not strictly enforce "in the RFC Editor queue" for IESG-approved + drafts when computing eligibility. In the overwhelming majority of cases, an IESG- + approved draft immediately enters the queue and goes on to be published, so this + simplification makes the calculation much easier and virtually never affects + eligibility. + + Arguments eligibility_period_start and eligibility_period_end are datetimes that + mark the start and end of the eligibility period. These should be five years apart. + """ + # First, get the RFCs using publication date + qualifying_rfc_pub_events = DocEvent.objects.filter( + type='published_rfc', + time__gte=eligibility_period_start, + time__lte=eligibility_period_end, + ) + qualifying_rfcs = Document.objects.filter( + type_id="rfc", + docevent__in=qualifying_rfc_pub_events + ).annotate( + rfcauthor_count=Count("rfcauthor") + ) + rfcs_with_rfcauthors = qualifying_rfcs.filter(rfcauthor_count__gt=0).distinct() + rfcs_without_rfcauthors = qualifying_rfcs.filter(rfcauthor_count=0).distinct() + + # Second, get the IESG-approved I-Ds excluding any we're already counting as rfcs + qualifying_approval_events = DocEvent.objects.filter( + type='iesg_approved', + time__gte=eligibility_period_start, + time__lte=eligibility_period_end, + ) + qualifying_drafts = Document.objects.filter( + type_id="draft", + docevent__in=qualifying_approval_events, + ).exclude( + relateddocument__relationship_id="became_rfc", + relateddocument__target__in=qualifying_rfcs, + ).distinct() + + return base_qs.filter( + Q(documentauthor__document__in=qualifying_drafts) + | Q(rfcauthor__document__in=rfcs_with_rfcauthors) + | Q(documentauthor__document__in=rfcs_without_rfcauthors) + ).annotate( + document_author_count=Count('documentauthor'), + rfc_author_count=Count("rfcauthor") + ).annotate( + authorship_count=F("document_author_count") + F("rfc_author_count") + ).filter(authorship_count__gte=2) + def get_threerule_eligibility_querysets(date, base_qs, three_of_five_callable): if not base_qs: @@ -586,14 +672,7 @@ def get_threerule_eligibility_querysets(date, base_qs, three_of_five_callable): ) ).distinct() - rfc_pks = set(DocEvent.objects.filter(type='published_rfc', time__gte=five_years_ago, time__lte=date_as_dt).values_list('doc__pk', flat=True)) - iesgappr_pks = set(DocEvent.objects.filter(type='iesg_approved', time__gte=five_years_ago, time__lte=date_as_dt).values_list('doc__pk',flat=True)) - qualifying_pks = rfc_pks.union(iesgappr_pks.difference(rfc_pks)) - author_qs = base_qs.filter( - documentauthor__document__pk__in=qualifying_pks - ).annotate( - document_author_count = Count('documentauthor') - ).filter(document_author_count__gte=2) + author_qs = get_qualified_author_queryset(base_qs, five_years_ago, date_as_dt) return three_of_five_qs, officer_qs, author_qs def list_eligible_8989(date, base_qs=None): @@ -605,10 +684,10 @@ def list_eligible_8989(date, base_qs=None): author_pks = author_qs.values_list('pk',flat=True) return remove_disqualified(Person.objects.filter(pk__in=set(three_of_five_pks).union(set(officer_pks)).union(set(author_pks)))) -def list_eligible_8989bis(date, base_qs=None): +def list_eligible_9389(date, base_qs=None): if not base_qs: base_qs = Person.objects.all() - three_of_five_qs, officer_qs, author_qs = get_8989bis_eligibility_querysets(date, base_qs) + three_of_five_qs, officer_qs, author_qs = get_9389_eligibility_querysets(date, base_qs) three_of_five_pks = three_of_five_qs.values_list('pk',flat=True) officer_pks = officer_qs.values_list('pk',flat=True) author_pks = author_qs.values_list('pk',flat=True) @@ -644,46 +723,67 @@ def previous_five_meetings(date = None): return Meeting.objects.filter(type='ietf',date__lte=date).order_by('-date')[:5] def three_of_five_eligible_8713(previous_five, queryset=None): - """ Return a list of Person records who attended at least + """ Return a list of Person records who attended at least 3 of the 5 type_id='ietf' meetings before the given date. Does not disqualify anyone based on held roles. This variant bases the calculation on MeetingRegistration.attended """ if queryset is None: queryset = Person.objects.all() - return queryset.filter(meetingregistration__meeting__in=list(previous_five),meetingregistration__attended=True).annotate(mtg_count=Count('meetingregistration')).filter(mtg_count__gte=3) + return queryset.filter(registration__meeting__in=list(previous_five), registration__attended=True).annotate(mtg_count=Count('registration')).filter(mtg_count__gte=3) -def three_of_five_eligible_8989bis(previous_five, queryset=None): +def three_of_five_eligible_9389(previous_five, queryset=None): """ Return a list of Person records who attended at least 3 of the 5 type_id='ietf' meetings before the given date. Does not disqualify anyone based on held roles. This variant bases the calculation on Meeting.Session and MeetingRegistration.checked_in - Leadership will have to create a new RFC specifying eligibility (RFC8989 is timing out) before it can be used. """ if queryset is None: queryset = Person.objects.all() counts = defaultdict(lambda: 0) for meeting in previous_five: - checked_in = meeting.meetingregistration_set.filter(reg_type='onsite', checkedin=True).values_list('person', flat=True) - sessions = meeting.session_set.filter(Q(type='plenary') | Q(group__type__in=['wg', 'rg'])) - attended = Attended.objects.filter(session__in=sessions).values_list('person', flat=True) + checked_in, attended = participants_for_meeting(meeting) for id in set(checked_in) | set(attended): counts[id] += 1 return queryset.filter(pk__in=[id for id, count in counts.items() if count >= 3]) -def suggest_affiliation(person): - recent_meeting = person.meetingregistration_set.order_by('-meeting__date').first() - affiliation = recent_meeting.affiliation if recent_meeting else '' - if not affiliation: - recent_volunteer = person.volunteer_set.order_by('-nomcom__group__acronym').first() - if recent_volunteer: - affiliation = recent_volunteer.affiliation - if not affiliation: - recent_draft_revision = NewRevisionDocEvent.objects.filter(doc__type_id='draft',doc__documentauthor__person=person).order_by('-time').first() - if recent_draft_revision: - affiliation = recent_draft_revision.doc.documentauthor_set.filter(person=person).first().affiliation - return affiliation +def suggest_affiliation(person) -> str: + """Heuristically suggest a current affiliation for a Person""" + recent_meeting = person.registration_set.order_by('-meeting__date').first() + if recent_meeting and recent_meeting.affiliation: + return recent_meeting.affiliation + + recent_volunteer = person.volunteer_set.order_by('-nomcom__group__acronym').first() + if recent_volunteer and recent_volunteer.affiliation: + return recent_volunteer.affiliation + + recent_draft_revision = NewRevisionDocEvent.objects.filter( + doc__type_id="draft", + doc__documentauthor__person=person, + ).order_by("-time").first() + if recent_draft_revision: + draft_author = recent_draft_revision.doc.documentauthor_set.filter( + person=person + ).first() + if draft_author and draft_author.affiliation: + return draft_author.affiliation + + recent_rfc_publication = DocEvent.objects.filter( + Q(doc__documentauthor__person=person) | Q(doc__rfcauthor__person=person), + doc__type_id="rfc", + type="published_rfc", + ).order_by("-time").first() + if recent_rfc_publication: + rfc = recent_rfc_publication.doc + if rfc.rfcauthor_set.exists(): + rfc_author = rfc.rfcauthor_set.filter(person=person).first() + else: + rfc_author = rfc.documentauthor_set.filter(person=person).first() + if rfc_author and rfc_author.affiliation: + return rfc_author.affiliation + return "" + def extract_volunteers(year): nomcom = get_nomcom_by_year(year) @@ -697,3 +797,58 @@ def extract_volunteers(year): decorate_volunteers_with_qualifications(volunteers,nomcom=nomcom) volunteers = sorted(volunteers,key=lambda v:(not v.eligible,v.person.last_name())) return nomcom, volunteers + + +def ingest_feedback_email(message: bytes, year: int): + from ietf.api.views import EmailIngestionError # avoid circular import + from .models import NomCom + try: + nomcom = NomCom.objects.get(group__acronym__icontains=str(year), + group__state__slug='active') + except NomCom.DoesNotExist: + raise EmailIngestionError( + f"Error ingesting nomcom email: nomcom {year} does not exist or is not active", + email_body=dedent(f"""\ + An email for nomcom {year} was posted to ingest_feedback_email, but no + active nomcom exists for that year. + """), + ) + + try: + feedback = create_feedback_email(nomcom, message) + except Exception as err: + raise EmailIngestionError( + f"Error ingesting nomcom {year} feedback email", + email_recipients=nomcom.chair_emails(), + email_body=dedent(f"""\ + An error occurred while ingesting feedback email for nomcom {year}. + + {{error_summary}} + """), + email_original_message=message, + ) from err + log("Received nomcom email from %s" % feedback.author) + + +def _is_time_to_send_reminder(nomcom, send_date, nomination_date): + if nomcom.reminder_interval: + days_passed = (send_date - nomination_date).days + return days_passed > 0 and days_passed % nomcom.reminder_interval == 0 + else: + return bool(nomcom.reminderdates_set.filter(date=send_date)) + + +def send_reminders(): + from .models import NomCom, NomineePosition + for nomcom in NomCom.objects.filter(group__state__slug="active"): + nps = NomineePosition.objects.filter( + nominee__nomcom=nomcom, nominee__duplicated__isnull=True + ) + for nominee_position in nps.pending(): + if _is_time_to_send_reminder(nomcom, date_today(), nominee_position.time.date()): + send_accept_reminder_to_nominee(nominee_position) + log(f"Sent accept reminder to {nominee_position.nominee.email.address}") + for nominee_position in nps.accepted().without_questionnaire_response(): + if _is_time_to_send_reminder(nomcom, date_today(), nominee_position.time.date()): + send_questionnaire_reminder_to_nominee(nominee_position) + log(f"Sent questionnaire reminder to {nominee_position.nominee.email.address}") diff --git a/ietf/nomcom/views.py b/ietf/nomcom/views.py index 6f02b16e29..3f90be5253 100644 --- a/ietf/nomcom/views.py +++ b/ietf/nomcom/views.py @@ -1,10 +1,10 @@ -# Copyright The IETF Trust 2012-2020, All Rights Reserved +# Copyright The IETF Trust 2012-2023, All Rights Reserved # -*- coding: utf-8 -*- import datetime import re -from collections import OrderedDict, Counter +from collections import Counter import csv import hmac @@ -14,12 +14,14 @@ from django.contrib.auth.models import AnonymousUser from django.core.paginator import Paginator, EmptyPage, PageNotAnInteger from django.forms.models import modelformset_factory, inlineformset_factory -from django.http import Http404, HttpResponseRedirect, HttpResponse +from django.http import Http404, HttpResponseRedirect, HttpResponse, HttpResponseForbidden from django.shortcuts import render, get_object_or_404, redirect from django.template.loader import render_to_string from django.urls import reverse -from django.utils.encoding import force_bytes, force_text +from django.utils.encoding import force_bytes, force_str +from django.utils.text import slugify +from email.errors import HeaderParseError from ietf.dbtemplate.models import DBTemplate from ietf.dbtemplate.views import group_template_edit, group_template_show @@ -55,7 +57,7 @@ def index(request): for nomcom in nomcom_list: year = int(nomcom.acronym[6:]) nomcom.year = year - nomcom.label = "%s/%s" % (year, year+1) + nomcom.label = str(year) if year > 2012: nomcom.url = "/nomcom/%04d" % year else: @@ -75,7 +77,6 @@ def year_index(request, year): return render(request, 'nomcom/year_index.html', {'nomcom': nomcom, 'year': year, - 'selected': 'index', 'template': template}) def announcements(request): @@ -158,8 +159,16 @@ def private_key(request, year): if request.method == 'POST': form = PrivateKeyForm(data=request.POST) if form.is_valid(): - store_nomcom_private_key(request, year, force_bytes(form.cleaned_data.get('key', ''))) - return HttpResponseRedirect(back_url) + try: + store_nomcom_private_key(request, year, force_bytes(form.cleaned_data.get('key', ''))) + except UnicodeError: + form.add_error( + None, + "An internal error occurred while adding your private key to your session." + f"Please contact the secretariat for assistance ({settings.SECRETARIAT_SUPPORT_EMAIL})" + ) + else: + return HttpResponseRedirect(back_url) else: form = PrivateKeyForm() @@ -172,8 +181,7 @@ def private_key(request, year): {'nomcom': nomcom, 'year': year, 'back_url': back_url, - 'form': form, - 'selected': 'private_key'}) + 'form': form}) @role_required("Nomcom") @@ -181,6 +189,7 @@ def private_index(request, year): nomcom = get_nomcom_by_year(year) all_nominee_positions = NomineePosition.objects.get_by_nomcom(nomcom).not_duplicated() is_chair = nomcom.group.has_role(request.user, "chair") + mailto = None if is_chair and request.method == 'POST': if nomcom.group.state_id != 'active': messages.warning(request, "This nomcom is not active. Request administrative assistance if Nominee state needs to change.") @@ -198,15 +207,18 @@ def private_index(request, year): elif action == "set_as_pending": nominations.update(state='pending') messages.success(request,'The selected nominations have been set as pending') + elif action == 'email': + mailto = ','.join([np.nominee.email.email_address() for np in nominations]) else: messages.warning(request, "Please, select some nominations to work with") filters = {} questionnaire_state = "questionnaire" + not_declined_state = "not-declined" selected_state = request.GET.get('state') selected_position = request.GET.get('position') - if selected_state and not selected_state == questionnaire_state: + if selected_state and selected_state not in [questionnaire_state, not_declined_state]: filters['state__slug'] = selected_state if selected_position: @@ -218,13 +230,15 @@ def private_index(request, year): if selected_state == questionnaire_state: nominee_positions = [np for np in nominee_positions if np.questionnaires] + elif selected_state == not_declined_state: + nominee_positions = nominee_positions.exclude(state__slug='declined') positions = Position.objects.get_by_nomcom(nomcom=nomcom) stats = [ { 'position__name':p.name, 'position__id':p.pk, 'position': p, } for p in positions] - states = list(NomineePositionStateName.objects.values('slug', 'name')) + [{'slug': questionnaire_state, 'name': 'Questionnaire'}] + states = [{'slug': questionnaire_state, 'name': 'Accepted and sent Questionnaire'}, {'slug': not_declined_state, 'name': 'Not declined'}] + list(NomineePositionStateName.objects.values('slug', 'name')) positions = set([ n.position for n in all_nominee_positions.order_by('position__name') ]) for s in stats: for state in states: @@ -267,8 +281,8 @@ def private_index(request, year): 'positions': positions, 'selected_state': selected_state, 'selected_position': selected_position and int(selected_position) or None, - 'selected': 'index', 'is_chair': is_chair, + 'mailto': mailto, }) @@ -291,13 +305,11 @@ def send_reminder_mail(request, year, type): interesting_state = 'pending' mail_path = nomcom_template_path + NOMINEE_ACCEPT_REMINDER_TEMPLATE reminder_description = 'accept (or decline) a nomination' - selected_tab = 'send_accept_reminder' state_description = NomineePositionStateName.objects.get(slug=interesting_state).name elif type=='questionnaire': interesting_state = 'accepted' mail_path = nomcom_template_path + NOMINEE_QUESTIONNAIRE_REMINDER_TEMPLATE reminder_description = 'complete the questionnaire for a nominated position' - selected_tab = 'send_questionnaire_reminder' state_description = NomineePositionStateName.objects.get(slug=interesting_state).name+' but no questionnaire has been received' else: raise Http404 @@ -332,7 +344,6 @@ def send_reminder_mail(request, year, type): 'year': year, 'nominees': annotated_nominees, 'mail_template': mail_template, - 'selected': selected_tab, 'reminder_description': reminder_description, 'state_description': state_description, 'is_chair_task' : True, @@ -359,7 +370,6 @@ def private_merge_person(request, year): {'nomcom': nomcom, 'year': year, 'form': form, - 'selected': 'merge_person', 'is_chair_task' : True, }) @@ -384,7 +394,6 @@ def private_merge_nominee(request, year): {'nomcom': nomcom, 'year': year, 'form': form, - 'selected': 'merge_nominee', 'is_chair_task' : True, }) @@ -394,8 +403,7 @@ def requirements(request, year): return render(request, 'nomcom/requirements.html', {'nomcom': nomcom, 'positions': positions, - 'year': year, - 'selected': 'requirements'}) + 'year': year}) def questionnaires(request, year): @@ -404,8 +412,7 @@ def questionnaires(request, year): return render(request, 'nomcom/questionnaires.html', {'nomcom': nomcom, 'positions': positions, - 'year': year, - 'selected': 'questionnaires'}) + 'year': year}) @login_required @@ -439,40 +446,38 @@ def nominate(request, year, public, newperson): messages.warning(request, "This Nomcom is not yet accepting nominations") return render(request, template, {'nomcom': nomcom, - 'year': year, - 'selected': 'nominate'}) + 'year': year}) if nomcom.group.state_id == 'conclude': messages.warning(request, "Nominations to this Nomcom are closed.") return render(request, template, {'nomcom': nomcom, - 'year': year, - 'selected': 'nominate'}) + 'year': year}) + person = request.user.person if request.method == 'POST': if newperson: - form = NominateNewPersonForm(data=request.POST, nomcom=nomcom, user=request.user, public=public) + form = NominateNewPersonForm(data=request.POST, nomcom=nomcom, person=person, public=public) else: - form = NominateForm(data=request.POST, nomcom=nomcom, user=request.user, public=public) + form = NominateForm(data=request.POST, nomcom=nomcom, person=person, public=public) if form.is_valid(): form.save() messages.success(request, 'Your nomination has been registered. Thank you for the nomination.') if newperson: return redirect('ietf.nomcom.views.%s_nominate' % ('public' if public else 'private'), year=year) else: - form = NominateForm(nomcom=nomcom, user=request.user, public=public) + form = NominateForm(nomcom=nomcom, person=person, public=public) else: if newperson: - form = NominateNewPersonForm(nomcom=nomcom, user=request.user, public=public) + form = NominateNewPersonForm(nomcom=nomcom, person=person, public=public) else: - form = NominateForm(nomcom=nomcom, user=request.user, public=public) + form = NominateForm(nomcom=nomcom, person=person, public=public) return render(request, template, {'form': form, 'nomcom': nomcom, 'year': year, - 'positions': nomcom.position_set.filter(is_open=True), - 'selected': 'nominate'}) + 'positions': nomcom.position_set.filter(is_open=True)}) @login_required def public_feedback(request, year): @@ -490,6 +495,7 @@ def feedback(request, year, public): nominee = None position = None topic = None + person = request.user.person if nomcom.group.state_id != 'conclude': selected_nominee = request.GET.get('nominee') selected_position = request.GET.get('position') @@ -501,7 +507,7 @@ def feedback(request, year, public): topic = get_object_or_404(Topic,id=selected_topic) if topic.audience_id == 'nomcom' and not nomcom.group.has_role(request.user, ['chair','advisor','liaison','member']): raise Http404() - if topic.audience_id == 'nominees' and not nomcom.nominee_set.filter(person=request.user.person).exists(): + if topic.audience_id == 'nominees' and not nomcom.nominee_set.filter(person=person).exists(): raise Http404() if public: @@ -513,12 +519,12 @@ def feedback(request, year, public): if not nomcom.group.has_role(request.user, ['chair','advisor','liaison','member']): topics = topics.exclude(audience_id='nomcom') - if not nomcom.nominee_set.filter(person=request.user.person).exists(): + if not nomcom.nominee_set.filter(person=person).exists(): topics = topics.exclude(audience_id='nominees') user_comments = Feedback.objects.filter(nomcom=nomcom, type='comment', - author__in=request.user.person.email_set.filter(active='True')) + author__in=person.email_set.filter(active='True')) counter = Counter(user_comments.values_list('positions','nominees')) counts = dict() for pos,nom in counter: @@ -536,7 +542,6 @@ def feedback(request, year, public): return render(request, 'nomcom/feedback.html', { 'nomcom': nomcom, 'year': year, - 'selected': 'feedback', 'counts' : counts, 'base_template': base_template }) @@ -547,7 +552,6 @@ def feedback(request, year, public): 'form': None, 'nomcom': nomcom, 'year': year, - 'selected': 'feedback', 'positions': positions, 'topics': topics, 'counts' : counts, @@ -561,7 +565,6 @@ def feedback(request, year, public): 'form': None, 'nomcom': nomcom, 'year': year, - 'selected': 'feedback', 'positions': positions, 'topics': topics, 'counts' : counts, @@ -571,11 +574,11 @@ def feedback(request, year, public): if request.method == 'POST': if nominee and position: form = FeedbackForm(data=request.POST, - nomcom=nomcom, user=request.user, + nomcom=nomcom, person=person, public=public, position=position, nominee=nominee) elif topic: form = FeedbackForm(data=request.POST, - nomcom=nomcom, user=request.user, + nomcom=nomcom, person=person, public=public, topic=topic) else: form = None @@ -594,10 +597,10 @@ def feedback(request, year, public): pass else: if nominee and position: - form = FeedbackForm(nomcom=nomcom, user=request.user, public=public, + form = FeedbackForm(nomcom=nomcom, person=person, public=public, position=position, nominee=nominee) elif topic: - form = FeedbackForm(nomcom=nomcom, user=request.user, public=public, + form = FeedbackForm(nomcom=nomcom, person=person, public=public, topic=topic) else: form = None @@ -608,7 +611,6 @@ def feedback(request, year, public): 'year': year, 'positions': positions, 'topics': topics, - 'selected': 'feedback', 'counts': counts, 'topic_counts': topic_counts, 'base_template': base_template @@ -634,7 +636,6 @@ def private_feedback_email(request, year): return render(request, template, {'nomcom': nomcom, 'year': year, - 'selected': 'feedback_email', 'is_chair_task' : True, }) @@ -644,15 +645,17 @@ def private_feedback_email(request, year): form = FeedbackEmailForm(data=request.POST, nomcom=nomcom) if form.is_valid(): - form.save() - form = FeedbackEmailForm(nomcom=nomcom) - messages.success(request, 'The feedback email has been registered.') + try: + form.save() + form = FeedbackEmailForm(nomcom=nomcom) + messages.success(request, 'The feedback email has been registered.') + except HeaderParseError: + messages.error(request, 'Missing email headers') return render(request, template, {'form': form, 'nomcom': nomcom, - 'year': year, - 'selected': 'feedback_email'}) + 'year': year}) @role_required("Nomcom Chair", "Nomcom Advisor") def private_questionnaire(request, year): @@ -660,6 +663,7 @@ def private_questionnaire(request, year): has_publickey = nomcom.public_key and True or False questionnaire_response = None template = 'nomcom/private_questionnaire.html' + person = request.user.person if not has_publickey: messages.warning(request, "This Nomcom is not yet accepting questionnaires.") @@ -674,27 +678,25 @@ def private_questionnaire(request, year): return render(request, template, {'nomcom': nomcom, 'year': year, - 'selected': 'questionnaire', 'is_chair_task' : True, }) if request.method == 'POST': form = QuestionnaireForm(data=request.POST, - nomcom=nomcom, user=request.user) + nomcom=nomcom, person=person) if form.is_valid(): form.save() messages.success(request, 'The questionnaire response has been registered.') - questionnaire_response = force_text(form.cleaned_data['comment_text']) - form = QuestionnaireForm(nomcom=nomcom, user=request.user) + questionnaire_response = force_str(form.cleaned_data['comment_text']) + form = QuestionnaireForm(nomcom=nomcom, person=person) else: - form = QuestionnaireForm(nomcom=nomcom, user=request.user) + form = QuestionnaireForm(nomcom=nomcom, person=person) return render(request, template, {'form': form, 'questionnaire_response': questionnaire_response, 'nomcom': nomcom, - 'year': year, - 'selected': 'questionnaire'}) + 'year': year}) def process_nomination_status(request, year, nominee_position_id, state, date, hash): @@ -726,15 +728,13 @@ def process_nomination_status(request, year, nominee_position_id, state, date, h if form.cleaned_data['comments']: # This Feedback object is of type comment instead of nomina in order to not # make answering "who nominated themselves" harder. - who = request.user - if isinstance(who,AnonymousUser): - who = None + who = None if isinstance(request.user, AnonymousUser) else request.user.person f = Feedback.objects.create(nomcom = nomcom, author = nominee_position.nominee.email, subject = '%s nomination %s'%(nominee_position.nominee.name(),state), comments = nomcom.encrypt(form.cleaned_data['comments']), type_id = 'comment', - user = who, + person = who, ) f.positions.add(nominee_position.position) f.nominees.add(nominee_position.nominee) @@ -748,10 +748,8 @@ def process_nomination_status(request, year, nominee_position_id, state, date, h 'nominee_position': nominee_position, 'state': state, 'need_confirmation': need_confirmation, - 'selected': 'feedback', 'form': form }) - @role_required("Nomcom") @nomcom_private_key_required def view_feedback(request, year): @@ -759,7 +757,7 @@ def view_feedback(request, year): nominees = Nominee.objects.get_by_nomcom(nomcom).not_duplicated().distinct() independent_feedback_types = [] nominee_feedback_types = [] - for ft in FeedbackTypeName.objects.all(): + for ft in FeedbackTypeName.objects.filter(used=True): if ft.slug in settings.NOMINEE_FEEDBACK_TYPES: nominee_feedback_types.append(ft) else: @@ -782,8 +780,9 @@ def nominee_staterank(nominee): sorted_nominees = sorted(nominees,key=lambda x:x.staterank) + reviewer = request.user.person for nominee in sorted_nominees: - last_seen = FeedbackLastSeen.objects.filter(reviewer=request.user.person,nominee=nominee).first() + last_seen = FeedbackLastSeen.objects.filter(reviewer=reviewer,nominee=nominee).first() nominee_feedback = [] for ft in nominee_feedback_types: qs = nominee.feedback_set.by_type(ft.slug) @@ -798,7 +797,7 @@ def nominee_staterank(nominee): nominees_feedback.append( {'nominee':nominee, 'feedback':nominee_feedback} ) independent_feedback = [ft.feedback_set.get_by_nomcom(nomcom).count() for ft in independent_feedback_types] for topic in nomcom.topic_set.all(): - last_seen = TopicFeedbackLastSeen.objects.filter(reviewer=request.user.person,topic=topic).first() + last_seen = TopicFeedbackLastSeen.objects.filter(reviewer=reviewer,topic=topic).first() topic_feedback = [] for ft in topic_feedback_types: qs = topic.feedback_set.by_type(ft.slug) @@ -814,7 +813,6 @@ def nominee_staterank(nominee): return render(request, 'nomcom/view_feedback.html', {'year': year, - 'selected': 'view_feedback', 'nominees': nominees, 'nominee_feedback_types': nominee_feedback_types, 'independent_feedback_types': independent_feedback_types, @@ -822,7 +820,8 @@ def nominee_staterank(nominee): 'topics_feedback': topics_feedback, 'independent_feedback': independent_feedback, 'nominees_feedback': nominees_feedback, - 'nomcom': nomcom}) + 'nomcom': nomcom, + }) @role_required("Nomcom Chair", "Nomcom Advisor") @@ -845,6 +844,7 @@ def view_feedback_pending(request, year): except EmptyPage: feedback_page = paginator.page(paginator.num_pages) extra_step = False + person = request.user.person if request.method == 'POST' and request.POST.get('end'): extra_ids = request.POST.get('extra_ids', None) extra_step = True @@ -853,7 +853,7 @@ def view_feedback_pending(request, year): formset.absolute_max = 2000 formset.validate_max = False for form in formset.forms: - form.set_nomcom(nomcom, request.user) + form.set_nomcom(nomcom, person) if formset.is_valid(): formset.save() if extra_ids: @@ -865,7 +865,7 @@ def view_feedback_pending(request, year): extra.append(feedback) formset = FullFeedbackFormSet(queryset=Feedback.objects.filter(id__in=[i.id for i in extra])) for form in formset.forms: - form.set_nomcom(nomcom, request.user, extra) + form.set_nomcom(nomcom, person, extra) extra_ids = None else: messages.success(request, 'Feedback saved') @@ -873,7 +873,7 @@ def view_feedback_pending(request, year): elif request.method == 'POST': formset = FeedbackFormSet(request.POST) for form in formset.forms: - form.set_nomcom(nomcom, request.user) + form.set_nomcom(nomcom, person) if formset.is_valid(): extra = [] nominations = [] @@ -893,12 +893,12 @@ def view_feedback_pending(request, year): if nominations: formset = FullFeedbackFormSet(queryset=Feedback.objects.filter(id__in=[i.id for i in nominations])) for form in formset.forms: - form.set_nomcom(nomcom, request.user, nominations) + form.set_nomcom(nomcom, person, nominations) extra_ids = ','.join(['%s:%s' % (i.id, i.type.pk) for i in extra]) else: formset = FullFeedbackFormSet(queryset=Feedback.objects.filter(id__in=[i.id for i in extra])) for form in formset.forms: - form.set_nomcom(nomcom, request.user, extra) + form.set_nomcom(nomcom, person, extra) if moved: messages.success(request, '%s messages classified. You must enter more information for the following feedback.' % moved) else: @@ -907,24 +907,13 @@ def view_feedback_pending(request, year): else: formset = FeedbackFormSet(queryset=feedback_page.object_list) for form in formset.forms: - form.set_nomcom(nomcom, request.user) - type_dict = OrderedDict() - for t in FeedbackTypeName.objects.all().order_by('pk'): - rest = t.name - slug = rest[0] - rest = rest[1:] - while slug in type_dict and rest: - slug = rest[0] - rest = rest[1] - type_dict[slug] = t + form.set_nomcom(nomcom, person) return render(request, 'nomcom/view_feedback_pending.html', {'year': year, - 'selected': 'feedback_pending', 'formset': formset, 'extra_step': extra_step, - 'type_dict': type_dict, 'extra_ids': extra_ids, - 'types': FeedbackTypeName.objects.all().order_by('pk'), + 'types': FeedbackTypeName.objects.filter(used=True), 'nomcom': nomcom, 'is_chair_task' : True, 'page': feedback_page, @@ -935,60 +924,139 @@ def view_feedback_pending(request, year): @nomcom_private_key_required def view_feedback_unrelated(request, year): nomcom = get_nomcom_by_year(year) + + if request.method == 'POST': + if not nomcom.group.has_role(request.user, ['chair','advisor']): + return HttpResponseForbidden('Restricted to roles: Nomcom Chair, Nomcom Advisor') + feedback_id = request.POST.get('feedback_id', None) + feedback = get_object_or_404(Feedback, id=feedback_id) + type = request.POST.get('type', None) + if type: + if type == 'unclassified': + feedback.type = None + messages.success(request, 'The selected feedback has been de-classified. Please reclassify it in the Pending emails tab.') + else: + feedback.type = FeedbackTypeName.objects.get(slug=type) + messages.success(request, f'The selected feedback has been reclassified as {feedback.type.name}.') + feedback.save() + else: + return render(request, 'nomcom/view_feedback_unrelated.html', + {'year': year, + 'nomcom': nomcom, + 'feedback_types': FeedbackTypeName.objects.filter(used=True).exclude(slug__in=settings.NOMINEE_FEEDBACK_TYPES), + 'reclassify_feedback': feedback, + 'is_chair_task' : True, + }) + feedback_types = [] - for ft in FeedbackTypeName.objects.exclude(slug__in=settings.NOMINEE_FEEDBACK_TYPES): + for ft in FeedbackTypeName.objects.filter(used=True).exclude(slug__in=settings.NOMINEE_FEEDBACK_TYPES): feedback_types.append({'ft': ft, 'feedback': ft.feedback_set.get_by_nomcom(nomcom)}) - return render(request, 'nomcom/view_feedback_unrelated.html', {'year': year, - 'selected': 'view_feedback', 'feedback_types': feedback_types, - 'nomcom': nomcom}) + 'nomcom': nomcom, + }) @role_required("Nomcom") @nomcom_private_key_required def view_feedback_topic(request, year, topic_id): - nomcom = get_nomcom_by_year(year) + # At present, the only feedback type for topics is 'comment'. + # Reclassifying from 'comment' to 'comment' is a no-op, + # so the only meaningful action is to de-classify it. + if request.method == 'POST': + nomcom = get_nomcom_by_year(year) + if not nomcom.group.has_role(request.user, ['chair','advisor']): + return HttpResponseForbidden('Restricted to roles: Nomcom Chair, Nomcom Advisor') + feedback_id = request.POST.get('feedback_id', None) + feedback = get_object_or_404(Feedback, id=feedback_id) + feedback.type = None + feedback.topics.clear() + feedback.save() + messages.success(request, 'The selected feedback has been de-classified. Please reclassify it in the Pending emails tab.') + topic = get_object_or_404(Topic, id=topic_id) + nomcom = get_nomcom_by_year(year) feedback_types = FeedbackTypeName.objects.filter(slug__in=['comment',]) + reviewer = request.user.person - last_seen = TopicFeedbackLastSeen.objects.filter(reviewer=request.user.person,topic=topic).first() - last_seen_time = (last_seen and last_seen.time) or datetime.datetime(year=1, month=1, day=1, tzinfo=datetime.timezone.utc) + last_seen = TopicFeedbackLastSeen.objects.filter(reviewer=reviewer,topic=topic).first() + last_seen_time = (last_seen and last_seen.time) or datetime.datetime(year=1, month=1, day=1, tzinfo=datetime.UTC) if last_seen: last_seen.save() else: - TopicFeedbackLastSeen.objects.create(reviewer=request.user.person,topic=topic) + TopicFeedbackLastSeen.objects.create(reviewer=reviewer,topic=topic) return render(request, 'nomcom/view_feedback_topic.html', {'year': year, - 'selected': 'view_feedback', 'topic': topic, 'feedback_types': feedback_types, 'last_seen_time' : last_seen_time, - 'nomcom': nomcom}) + 'nomcom': nomcom, + }) @role_required("Nomcom") @nomcom_private_key_required def view_feedback_nominee(request, year, nominee_id): nomcom = get_nomcom_by_year(year) nominee = get_object_or_404(Nominee, id=nominee_id) - feedback_types = FeedbackTypeName.objects.filter(slug__in=settings.NOMINEE_FEEDBACK_TYPES) - - last_seen = FeedbackLastSeen.objects.filter(reviewer=request.user.person,nominee=nominee).first() - last_seen_time = (last_seen and last_seen.time) or datetime.datetime(year=1, month=1, day=1, tzinfo=datetime.timezone.utc) + feedback_types = FeedbackTypeName.objects.filter(used=True, slug__in=settings.NOMINEE_FEEDBACK_TYPES) + reviewer = request.user.person + if request.method == 'POST': + if not nomcom.group.has_role(request.user, ['chair','advisor']): + return HttpResponseForbidden('Restricted to roles: Nomcom Chair, Nomcom Advisor') + feedback_id = request.POST.get('feedback_id', None) + feedback = get_object_or_404(Feedback, id=feedback_id) + submit = request.POST.get('submit', None) + if submit == 'download': + fn = f'questionnaire-{slugify(nominee.name())}-{feedback.time.date()}.txt' + response = render_to_string('nomcom/download_questionnaire.txt', + {'year': year, + 'nominee': nominee, + 'feedback': feedback, + 'positions': ','.join([str(p) for p in feedback.positions.all()]), + }, + request=request) + response = HttpResponse( + response, + content_type=f"text/plain; charset={settings.DEFAULT_CHARSET}", + ) + response['Content-Disposition'] = f'attachment; filename="{fn}"' + return response + elif submit == 'reclassify': + type = request.POST.get('type', None) + if type: + if type == 'unclassified': + feedback.type = None + feedback.nominees.clear() + messages.success(request, 'The selected feedback has been de-classified. Please reclassify it in the Pending emails tab.') + else: + feedback.type = FeedbackTypeName.objects.get(slug=type) + messages.success(request, f'The selected feedback has been reclassified as {feedback.type.name}.') + feedback.save() + else: + return render(request, 'nomcom/view_feedback_nominee.html', + {'year': year, + 'nomcom': nomcom, + 'feedback_types': feedback_types, + 'reclassify_feedback': feedback, + 'is_chair_task': True, + }) + + last_seen = FeedbackLastSeen.objects.filter(reviewer=reviewer,nominee=nominee).first() + last_seen_time = (last_seen and last_seen.time) or datetime.datetime(year=1, month=1, day=1, tzinfo=datetime.UTC) if last_seen: last_seen.save() else: - FeedbackLastSeen.objects.create(reviewer=request.user.person,nominee=nominee) + FeedbackLastSeen.objects.create(reviewer=reviewer,nominee=nominee) return render(request, 'nomcom/view_feedback_nominee.html', {'year': year, - 'selected': 'view_feedback', 'nominee': nominee, 'feedback_types': feedback_types, 'last_seen_time' : last_seen_time, - 'nomcom': nomcom}) + 'nomcom': nomcom, + }) @role_required("Nomcom Chair", "Nomcom Advisor") @@ -1008,7 +1076,6 @@ def edit_nominee(request, year, nominee_id): return render(request, 'nomcom/edit_nominee.html', {'year': year, - 'selected': 'index', 'nominee': nominee, 'form': form, 'nomcom': nomcom, @@ -1046,7 +1113,6 @@ def edit_nomcom(request, year): 'formset': formset, 'nomcom': nomcom, 'year': year, - 'selected': 'edit_nomcom', 'is_chair_task' : True, }) @@ -1060,7 +1126,6 @@ def list_templates(request, year): return render(request, 'nomcom/list_templates.html', {'template_list': template_list, 'year': year, - 'selected': 'edit_templates', 'nomcom': nomcom, 'is_chair_task' : True, }) @@ -1094,11 +1159,45 @@ def edit_template(request, year, template_id): def list_positions(request, year): nomcom = get_nomcom_by_year(year) positions = nomcom.position_set.order_by('-is_open') + if request.method == 'POST': + if nomcom.group.state_id != 'active': + messages.warning(request, "This nomcom is not active. Request administrative assistance if Position state needs to change.") + else: + action = request.POST.get('action') + positions_to_modify = request.POST.getlist('selected') + if positions_to_modify: + positions = positions.filter(id__in=positions_to_modify) + if action == "set_iesg": + positions.update(is_iesg_position=True) + messages.success(request,'The selected positions have been set as IESG Positions') + elif action == "unset_iesg": + positions.update(is_iesg_position=False) + messages.success(request,'The selected positions have been set as NOT IESG Positions') + elif action == "set_open": + positions.update(is_open=True) + messages.success(request,'The selected positions have been set as Open') + elif action == "unset_open": + positions.update(is_open=False) + messages.success(request,'The selected positions have been set as NOT Open') + elif action == "set_accept_nom": + positions.update(accepting_nominations=True) + messages.success(request,'The selected positions have been set as Accepting Nominations') + elif action == "unset_accept_nom": + positions.update(accepting_nominations=False) + messages.success(request,'The selected positions have been set as NOT Accepting Nominations') + elif action == "set_accept_fb": + positions.update(accepting_feedback=True) + messages.success(request,'The selected positions have been set as Accepting Feedback') + elif action == "unset_accept_fb": + positions.update(accepting_feedback=False) + messages.success(request,'The selected positions have been set as NOT Accepting Feedback') + positions = nomcom.position_set.order_by('-is_open') + else: + messages.warning(request, "Please select some positions to work with") return render(request, 'nomcom/list_positions.html', {'positions': positions, 'year': year, - 'selected': 'edit_positions', 'nomcom': nomcom, 'is_chair_task' : True, }) @@ -1165,7 +1264,6 @@ def list_topics(request, year): return render(request, 'nomcom/list_topics.html', {'topics': topics, 'year': year, - 'selected': 'edit_topics', 'nomcom': nomcom, 'is_chair_task' : True, }) @@ -1231,15 +1329,15 @@ def configuration_help(request, year): @role_required("Nomcom Chair", "Nomcom Advisor") def edit_members(request, year): nomcom = get_nomcom_by_year(year) - if nomcom.group.state_id=='conclude': permission_denied(request, 'This nomcom is closed.') + person = request.user.person if request.method=='POST': form = EditMembersForm(nomcom, data=request.POST) if form.is_valid(): - update_role_set(nomcom.group, 'member', form.cleaned_data['members'], request.user.person) - update_role_set(nomcom.group, 'liaison', form.cleaned_data['liaisons'], request.user.person) + update_role_set(nomcom.group, 'member', form.cleaned_data['members'], person) + update_role_set(nomcom.group, 'liaison', form.cleaned_data['liaisons'], person) return HttpResponseRedirect(reverse('ietf.nomcom.views.private_index',kwargs={'year':year})) else: form = EditMembersForm(nomcom) @@ -1247,8 +1345,7 @@ def edit_members(request, year): return render(request, 'nomcom/new_edit_members.html', {'nomcom' : nomcom, 'year' : year, - 'form': form, - }) + 'form': form}) @role_required("Nomcom Chair", "Nomcom Advisor") def extract_email_lists(request, year): @@ -1268,8 +1365,7 @@ def extract_email_lists(request, year): 'pending': pending, 'accepted': accepted, 'noresp': noresp, - 'bypos': bypos, - }) + 'bypos': bypos}) @login_required def volunteer(request): @@ -1284,7 +1380,7 @@ def volunteer(request): form = VolunteerForm(person=person, data=request.POST) if form.is_valid(): for nc in form.cleaned_data['nomcoms']: - nc.volunteer_set.create(person=person, affiliation=form.cleaned_data['affiliation']) + nc.volunteer_set.get_or_create(person=person, defaults={"affiliation": form.cleaned_data["affiliation"], "origin":"datatracker"}) return redirect('ietf.ietfauth.views.profile') else: form = VolunteerForm(person=person,initial=dict(nomcoms=can_volunteer, affiliation=suggest_affiliation(person))) @@ -1333,3 +1429,12 @@ def private_volunteers_csv(request, year, public=False): writer.writerow([v.person.last_name(), v.person.first_name(), v.person.ascii_name(), v.affiliation, v.person.email(), v.qualifications, v.eligible]) return response +@role_required("Nomcom Chair", "Nomcom Advisor", "Secretariat") +def qualified_volunteer_list_for_announcement(request, year, public=False): + _, volunteers = extract_volunteers(year) + qualified_volunteers = [v for v in volunteers if v.eligible] + return render(request, 'nomcom/qualified_volunteer_list_for_announcement.txt', + dict(volunteers=qualified_volunteers), + content_type="text/plain; charset=%s"%settings.DEFAULT_CHARSET) + + diff --git a/ietf/person/admin.py b/ietf/person/admin.py index cd8ca2abf1..f46edcf8ae 100644 --- a/ietf/person/admin.py +++ b/ietf/person/admin.py @@ -7,6 +7,7 @@ from ietf.person.models import Email, Alias, Person, PersonalApiKey, PersonEvent, PersonApiKeyEvent, PersonExtResource from ietf.person.name import name_parts +from ietf.utils.admin import SaferStackedInline, SaferTabularInline from ietf.utils.validators import validate_external_resource_value @@ -16,7 +17,7 @@ class EmailAdmin(simple_history.admin.SimpleHistoryAdmin): search_fields = ["address", "person__name", ] admin.site.register(Email, EmailAdmin) -class EmailInline(admin.TabularInline): +class EmailInline(SaferTabularInline): model = Email class AliasAdmin(admin.ModelAdmin): @@ -25,7 +26,7 @@ class AliasAdmin(admin.ModelAdmin): raw_id_fields = ["person"] admin.site.register(Alias, AliasAdmin) -class AliasInline(admin.StackedInline): +class AliasInline(SaferStackedInline): model = Alias class PersonAdmin(simple_history.admin.SimpleHistoryAdmin): diff --git a/ietf/person/api.py b/ietf/person/api.py new file mode 100644 index 0000000000..960785a3d4 --- /dev/null +++ b/ietf/person/api.py @@ -0,0 +1,45 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +"""DRF API Views""" +from rest_framework import mixins, viewsets +from rest_framework.decorators import action +from rest_framework.permissions import IsAuthenticated +from rest_framework.response import Response + +from ietf.api.permissions import BelongsToOwnPerson, IsOwnPerson +from ietf.ietfauth.utils import send_new_email_confirmation_request + +from .models import Email, Person +from .serializers import NewEmailSerializer, EmailSerializer, PersonSerializer + + +class EmailViewSet(mixins.UpdateModelMixin, viewsets.GenericViewSet): + """Email viewset + + Only allows updating an existing email for now. + """ + permission_classes = [IsAuthenticated & BelongsToOwnPerson] + queryset = Email.objects.all() + serializer_class = EmailSerializer + lookup_value_regex = '.+@.+' # allow @-sign in the pk + + +class PersonViewSet(mixins.RetrieveModelMixin, viewsets.GenericViewSet): + """Person viewset""" + permission_classes = [IsAuthenticated & IsOwnPerson] + queryset = Person.objects.all() + serializer_class = PersonSerializer + + @action(detail=True, methods=["post"], serializer_class=NewEmailSerializer) + def email(self, request, pk=None): + """Add an email address for this Person + + Always succeeds if the email address is valid. Causes a confirmation email to be sent to the + requested address and completion of that handshake will actually add the email address. If the + address already exists, an alert will be sent instead of the confirmation email. + """ + person = self.get_object() + serializer = self.get_serializer(data=request.data) + serializer.is_valid(raise_exception=True) + # This may or may not actually send a confirmation, but doesn't reveal that to the user. + send_new_email_confirmation_request(person, serializer.validated_data["address"]) + return Response(serializer.data) diff --git a/ietf/person/factories.py b/ietf/person/factories.py index 8e80932c91..98756f26c8 100644 --- a/ietf/person/factories.py +++ b/ietf/person/factories.py @@ -8,7 +8,7 @@ import faker.config import os import random -import shutil +from PIL import Image from unidecode import unidecode from unicodedata import normalize @@ -16,7 +16,7 @@ from django.conf import settings from django.contrib.auth.models import User from django.utils.text import slugify -from django.utils.encoding import force_text +from django.utils.encoding import force_str import debug # pyflakes:ignore @@ -26,26 +26,29 @@ fake = faker.Factory.create() -def setup(): - global acceptable_fakers - # The transliteration of some Arabic and Devanagari names introduces - # non-alphabetic characters that don't work with the draft author - # extraction code, and also don't seem to match the way people with Arabic - # names romanize Arabic names. Exclude those locales from name generation - # in order to avoid test failures. - locales = set( [ l for l in faker.config.AVAILABLE_LOCALES if not (l.startswith('ar_') or l.startswith('sg_') or l=='fr_QC') ] ) - acceptable_fakers = [faker.Faker(locale) for locale in locales] -setup() +# The transliteration of some Arabic and Devanagari names introduces +# non-alphabetic characters that don't work with the draft author +# extraction code, and also don't seem to match the way people with Arabic +# names romanize Arabic names. Exclude those locales from name generation +# in order to avoid test failures. +_acceptable_fakers = [ + faker.Faker(locale) + for locale in set(faker.config.AVAILABLE_LOCALES) + if not (locale.startswith('ar_') or locale.startswith('sg_') or locale == 'fr_QC') +] + def random_faker(): - global acceptable_fakers - return random.sample(acceptable_fakers, 1)[0] + """Helper to get a random faker acceptable for User names""" + return random.sample(_acceptable_fakers, 1)[0] + class UserFactory(factory.django.DjangoModelFactory): class Meta: model = User django_get_or_create = ('username',) exclude = ['faker', ] + skip_postgeneration_save = True faker = factory.LazyFunction(random_faker) # normalize these i18n Unicode strings in the same way the database does @@ -55,20 +58,23 @@ class Meta: slugify(unidecode(u.last_name)), n, fake.domain_name())) # type: ignore username = factory.LazyAttribute(lambda u: u.email) + # Consider using PostGenerationMethodCall instead @factory.post_generation def set_password(obj, create, extracted, **kwargs): # pylint: disable=no-self-argument obj.set_password( '%s+password' % obj.username ) # pylint: disable=no-value-for-parameter + obj.save() class PersonFactory(factory.django.DjangoModelFactory): class Meta: model = Person + skip_postgeneration_save = True user = factory.SubFactory(UserFactory) name = factory.LazyAttribute(lambda p: normalize_name('%s %s'%(p.user.first_name, p.user.last_name))) # Some i18n names, e.g., "शिला के.सी." have a dot at the end that is also part of the ASCII, e.g., "Shilaa Kesii." # That trailing dot breaks extract_authors(). Avoid this issue by stripping the dot from the ASCII. # Some others have a trailing semicolon (e.g., "உயிரோவியம் தங்கராஐ;") - strip those, too. - ascii = factory.LazyAttribute(lambda p: force_text(unidecode_name(p.name)).rstrip(".;")) + ascii = factory.LazyAttribute(lambda p: force_str(unidecode_name(p.name)).rstrip(".;")) class Params: with_bio = factory.Trait(biography = "\n\n".join(fake.paragraphs())) # type: ignore @@ -99,13 +105,13 @@ def default_photo(obj, create, extracted, **kwargs): # pylint: disable=no-self-a media_name = "%s/%s.jpg" % (settings.PHOTOS_DIRNAME, photo_name) obj.photo = media_name obj.photo_thumb = media_name - photosrc = os.path.join(settings.TEST_DATA_DIR, "profile-default.jpg") photodst = os.path.join(settings.PHOTOS_DIR, photo_name + '.jpg') - if not os.path.exists(photodst): - shutil.copy(photosrc, photodst) + img = Image.new('RGB', (200, 200)) + img.save(photodst) def delete_file(file): os.unlink(file) atexit.register(delete_file, photodst) + obj.save() class AliasFactory(factory.django.DjangoModelFactory): class Meta: @@ -154,10 +160,22 @@ class Meta: class PersonalApiKeyFactory(factory.django.DjangoModelFactory): person = factory.SubFactory(PersonFactory) - endpoint = FuzzyChoice(PERSON_API_KEY_ENDPOINTS) - + endpoint = FuzzyChoice(v for v, n in PERSON_API_KEY_ENDPOINTS) + class Meta: model = PersonalApiKey + skip_postgeneration_save = True + + @factory.post_generation + def validate_model(obj, create, extracted, **kwargs): + """Validate the model after creation + + Passing validate_model=False will disable the validation. + """ + do_clean = True if extracted is None else extracted + if do_clean: + obj.full_clean() + class PersonApiKeyEventFactory(factory.django.DjangoModelFactory): key = factory.SubFactory(PersonalApiKeyFactory) diff --git a/ietf/person/forms.py b/ietf/person/forms.py index 81ee362561..7eef8aa17b 100644 --- a/ietf/person/forms.py +++ b/ietf/person/forms.py @@ -1,15 +1,26 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved +# Copyright The IETF Trust 2018-2025, All Rights Reserved # -*- coding: utf-8 -*- from django import forms + from ietf.person.models import Person +from ietf.utils.fields import MultiEmailField, NameAddrEmailField class MergeForm(forms.Form): source = forms.IntegerField(label='Source Person ID') target = forms.IntegerField(label='Target Person ID') + def __init__(self, *args, **kwargs): + self.readonly = False + if 'readonly' in kwargs: + self.readonly = kwargs.pop('readonly') + super().__init__(*args, **kwargs) + if self.readonly: + self.fields['source'].widget.attrs['readonly'] = True + self.fields['target'].widget.attrs['readonly'] = True + def clean_source(self): return self.get_person(self.cleaned_data['source']) @@ -21,3 +32,11 @@ def get_person(self, pk): return Person.objects.get(pk=pk) except Person.DoesNotExist: raise forms.ValidationError("ID does not exist") + + +class MergeRequestForm(forms.Form): + to = MultiEmailField() + frm = NameAddrEmailField() + reply_to = MultiEmailField() + subject = forms.CharField() + body = forms.CharField(widget=forms.Textarea) diff --git a/ietf/person/management/commands/purge_old_personal_api_key_events.py b/ietf/person/management/commands/purge_old_personal_api_key_events.py deleted file mode 100644 index a32edf866c..0000000000 --- a/ietf/person/management/commands/purge_old_personal_api_key_events.py +++ /dev/null @@ -1,59 +0,0 @@ -# Copyright The IETF Trust 2021, All Rights Reserved -# -*- coding: utf-8 -*- - -from datetime import timedelta -from django.core.management.base import BaseCommand, CommandError -from django.db.models import Max, Min -from django.utils import timezone - -from ietf.person.models import PersonApiKeyEvent - - -class Command(BaseCommand): - help = 'Purge PersonApiKeyEvent instances older than KEEP_DAYS days' - - def add_arguments(self, parser): - parser.add_argument('keep_days', type=int, - help='Delete events older than this many days') - parser.add_argument('-n', '--dry-run', action='store_true', default=False, - help="Don't delete events, just show what would be done") - - def handle(self, *args, **options): - keep_days = options['keep_days'] - dry_run = options['dry_run'] - - def _format_count(count, unit='day'): - return '{} {}{}'.format(count, unit, ('' if count == 1 else 's')) - - if keep_days < 0: - raise CommandError('Negative keep_days not allowed ({} was specified)'.format(keep_days)) - - self.stdout.write('purge_old_personal_api_key_events: Finding events older than {}\n'.format(_format_count(keep_days))) - if dry_run: - self.stdout.write('Dry run requested, records will not be deleted\n') - self.stdout.flush() - - now = timezone.now() - old_events = PersonApiKeyEvent.objects.filter( - time__lt=now - timedelta(days=keep_days) - ) - - stats = old_events.aggregate(Min('time'), Max('time')) - old_count = old_events.count() - if old_count == 0: - self.stdout.write('No events older than {} found\n'.format(_format_count(keep_days))) - return - - oldest_date = stats['time__min'] - oldest_ago = now - oldest_date - newest_date = stats['time__max'] - newest_ago = now - newest_date - - action_fmt = 'Would delete {}\n' if dry_run else 'Deleting {}\n' - self.stdout.write(action_fmt.format(_format_count(old_count, 'event'))) - self.stdout.write(' Oldest at {} ({} ago)\n'.format(oldest_date, _format_count(oldest_ago.days))) - self.stdout.write(' Most recent at {} ({} ago)\n'.format(newest_date, _format_count(newest_ago.days))) - self.stdout.flush() - - if not dry_run: - old_events.delete() diff --git a/ietf/person/management/commands/tests.py b/ietf/person/management/commands/tests.py deleted file mode 100644 index 291a6ace5f..0000000000 --- a/ietf/person/management/commands/tests.py +++ /dev/null @@ -1,122 +0,0 @@ -# Copyright The IETF Trust 2021, All Rights Reserved -# -*- coding: utf-8 -*- - -import datetime -from io import StringIO - -from django.core.management import call_command, CommandError -from django.utils import timezone - -from ietf.person.factories import PersonApiKeyEventFactory -from ietf.person.models import PersonApiKeyEvent, PersonEvent -from ietf.utils.test_utils import TestCase - - -class CommandTests(TestCase): - @staticmethod - def _call_command(command_name, *args, **options): - out = StringIO() - options['stdout'] = out - call_command(command_name, *args, **options) - return out.getvalue() - - def _assert_purge_results(self, cmd_output, expected_delete_count, expected_kept_events): - self.assertNotIn('Dry run requested', cmd_output) - if expected_delete_count == 0: - delete_text = 'No events older than' - else: - delete_text = 'Deleting {} event'.format(expected_delete_count) - self.assertIn(delete_text, cmd_output) - self.assertCountEqual( - PersonApiKeyEvent.objects.all(), - expected_kept_events, - 'Wrong events were deleted' - ) - - def _assert_purge_dry_run_results(self, cmd_output, expected_delete_count, expected_kept_events): - self.assertIn('Dry run requested', cmd_output) - if expected_delete_count == 0: - delete_text = 'No events older than' - else: - delete_text = 'Would delete {} event'.format(expected_delete_count) - self.assertIn(delete_text, cmd_output) - self.assertCountEqual( - PersonApiKeyEvent.objects.all(), - expected_kept_events, - 'Events were deleted when dry-run option was used' - ) - - def test_purge_old_personal_api_key_events(self): - keep_days = 10 - - # Remember how many PersonEvents were present so we can verify they're cleaned up properly. - personevents_before = PersonEvent.objects.count() - - now = timezone.now() - # The first of these events will be timestamped a fraction of a second more than keep_days - # days ago by the time we call the management command, so will just barely chosen for purge. - old_events = [ - PersonApiKeyEventFactory(time=now - datetime.timedelta(days=n)) - for n in range(keep_days, 2 * keep_days + 1) - ] - num_old_events = len(old_events) - - recent_events = [ - PersonApiKeyEventFactory(time=now - datetime.timedelta(days=n)) - for n in range(0, keep_days) - ] - # We did not create recent_event timestamped exactly keep_days ago because it would - # be treated as an old_event by the management command. Create an event a few seconds - # on the "recent" side of keep_days old to test the threshold. - recent_events.append( - PersonApiKeyEventFactory( - time=now + datetime.timedelta(seconds=3) - datetime.timedelta(days=keep_days) - ) - ) - num_recent_events = len(recent_events) - - # call with dry run - output = self._call_command('purge_old_personal_api_key_events', str(keep_days), '--dry-run') - self._assert_purge_dry_run_results(output, num_old_events, old_events + recent_events) - - # call for real - output = self._call_command('purge_old_personal_api_key_events', str(keep_days)) - self._assert_purge_results(output, num_old_events, recent_events) - self.assertEqual(PersonEvent.objects.count(), personevents_before + num_recent_events, - 'PersonEvents were not cleaned up properly') - - # repeat - there should be nothing left to delete - output = self._call_command('purge_old_personal_api_key_events', '--dry-run', str(keep_days)) - self._assert_purge_dry_run_results(output, 0, recent_events) - - output = self._call_command('purge_old_personal_api_key_events', str(keep_days)) - self._assert_purge_results(output, 0, recent_events) - self.assertEqual(PersonEvent.objects.count(), personevents_before + num_recent_events, - 'PersonEvents were not cleaned up properly') - - # and now delete the remaining events - output = self._call_command('purge_old_personal_api_key_events', '0') - self._assert_purge_results(output, num_recent_events, []) - self.assertEqual(PersonEvent.objects.count(), personevents_before, - 'PersonEvents were not cleaned up properly') - - def test_purge_old_personal_api_key_events_rejects_invalid_arguments(self): - """The purge_old_personal_api_key_events command should reject invalid arguments""" - event = PersonApiKeyEventFactory(time=timezone.now() - datetime.timedelta(days=30)) - - with self.assertRaises(CommandError): - self._call_command('purge_old_personal_api_key_events') - - with self.assertRaises(CommandError): - self._call_command('purge_old_personal_api_key_events', '-15') - - with self.assertRaises(CommandError): - self._call_command('purge_old_personal_api_key_events', '15.3') - - with self.assertRaises(CommandError): - self._call_command('purge_old_personal_api_key_events', '15', '15') - - with self.assertRaises(CommandError): - self._call_command('purge_old_personal_api_key_events', 'abc', '15') - - self.assertCountEqual(PersonApiKeyEvent.objects.all(), [event]) diff --git a/ietf/person/migrations/0001_initial.py b/ietf/person/migrations/0001_initial.py index 72f6c9105b..d0f6f239a3 100644 --- a/ietf/person/migrations/0001_initial.py +++ b/ietf/person/migrations/0001_initial.py @@ -1,16 +1,16 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-20 10:52 +# Generated by Django 2.2.28 on 2023-03-20 19:22 - -import datetime from django.conf import settings +import django.contrib.postgres.fields.citext import django.core.validators from django.db import migrations, models import django.db.models.deletion +import django.utils.timezone import ietf.person.models import ietf.utils.models import ietf.utils.storage +import jsonfield.fields +import simple_history.models class Migration(migrations.Migration): @@ -19,53 +19,64 @@ class Migration(migrations.Migration): dependencies = [ migrations.swappable_dependency(settings.AUTH_USER_MODEL), + ('name', '0001_initial'), ] operations = [ migrations.CreateModel( - name='Alias', + name='Person', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('name', models.CharField(db_index=True, max_length=255)), + ('time', models.DateTimeField(default=django.utils.timezone.now)), + ('name', models.CharField(db_index=True, help_text='Preferred long form of name.', max_length=255, validators=[ietf.person.models.name_character_validator], verbose_name='Full Name (Unicode)')), + ('plain', models.CharField(blank=True, default='', help_text="Use this if you have a Spanish double surname. Don't use this for nicknames, and don't use it unless you've actually observed that the datatracker shows your name incorrectly.", max_length=64, verbose_name='Plain Name correction (Unicode)')), + ('ascii', models.CharField(help_text='Name as rendered in ASCII (Latin, unaccented) characters.', max_length=255, verbose_name='Full Name (ASCII)')), + ('ascii_short', models.CharField(blank=True, help_text='Example: A. Nonymous. Fill in this with initials and surname only if taking the initials and surname of the ASCII name above produces an incorrect initials-only form. (Blank is OK).', max_length=32, null=True, verbose_name='Abbreviated Name (ASCII)')), + ('pronouns_selectable', jsonfield.fields.JSONCharField(blank=True, default=list, max_length=120, null=True, verbose_name='Pronouns')), + ('pronouns_freetext', models.CharField(blank=True, help_text='Optionally provide your personal pronouns. These will be displayed on your public profile page and alongside your name in Meetecho and, in future, other systems. Select any number of the checkboxes OR provide a custom string up to 30 characters.', max_length=30, null=True, verbose_name=' ')), + ('biography', models.TextField(blank=True, help_text='Short biography for use on leadership pages. Use plain text or reStructuredText markup.')), + ('photo', models.ImageField(blank=True, default=None, storage=ietf.utils.storage.NoLocationMigrationFileSystemStorage(location=None), upload_to='photo')), + ('photo_thumb', models.ImageField(blank=True, default=None, storage=ietf.utils.storage.NoLocationMigrationFileSystemStorage(location=None), upload_to='photo')), + ('name_from_draft', models.CharField(editable=False, help_text='Name as found in an Internet-Draft submission.', max_length=255, null=True, verbose_name='Full Name (from submission)')), + ('user', ietf.utils.models.OneToOneField(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, to=settings.AUTH_USER_MODEL)), + ], + ), + migrations.CreateModel( + name='PersonEvent', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('time', models.DateTimeField(default=django.utils.timezone.now, help_text='When the event happened')), + ('type', models.CharField(choices=[('apikey_login', 'API key login'), ('email_address_deactivated', 'Email address deactivated')], max_length=50)), + ('desc', models.TextField()), + ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), ], options={ - 'verbose_name_plural': 'Aliases', + 'ordering': ['-time', '-id'], }, ), migrations.CreateModel( - name='Email', + name='PersonApiKeyEvent', fields=[ - ('address', models.CharField(max_length=64, primary_key=True, serialize=False, validators=[django.core.validators.EmailValidator()])), - ('time', models.DateTimeField(auto_now_add=True)), - ('primary', models.BooleanField(default=False)), - ('active', models.BooleanField(default=True)), + ('personevent_ptr', models.OneToOneField(auto_created=True, on_delete=django.db.models.deletion.CASCADE, parent_link=True, primary_key=True, serialize=False, to='person.PersonEvent')), ], + bases=('person.personevent',), ), migrations.CreateModel( - name='Person', + name='PersonExtResource', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('time', models.DateTimeField(default=datetime.datetime.now)), - ('name', models.CharField(db_index=True, help_text='Preferred form of name.', max_length=255, verbose_name='Full Name (Unicode)')), - ('ascii', models.CharField(help_text='Name as rendered in ASCII (Latin, unaccented) characters.', max_length=255, verbose_name='Full Name (ASCII)')), - ('ascii_short', models.CharField(blank=True, help_text='Example: A. Nonymous. Fill in this with initials and surname only if taking the initials and surname of the ASCII name above produces an incorrect initials-only form. (Blank is OK).', max_length=32, null=True, verbose_name='Abbreviated Name (ASCII)')), - ('affiliation', models.CharField(blank=True, help_text='Employer, university, sponsor, etc.', max_length=255)), - ('address', models.TextField(blank=True, help_text='Postal mailing address.', max_length=255)), - ('biography', models.TextField(blank=True, help_text='Short biography for use on leadership pages. Use plain text or reStructuredText markup.')), - ('photo', models.ImageField(blank=True, default=None, storage=ietf.utils.storage.NoLocationMigrationFileSystemStorage(location=None), upload_to='photo')), - ('photo_thumb', models.ImageField(blank=True, default=None, storage=ietf.utils.storage.NoLocationMigrationFileSystemStorage(location=None), upload_to='photo')), - ('user', ietf.utils.models.OneToOneField(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL)), + ('display_name', models.CharField(blank=True, default='', max_length=255)), + ('value', models.CharField(max_length=2083)), + ('name', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ExtResourceName')), + ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), ], - options={ - 'abstract': False, - }, ), migrations.CreateModel( name='PersonalApiKey', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('endpoint', models.CharField(choices=[('/api/iesg/position', '/api/iesg/position')], max_length=128)), - ('created', models.DateTimeField(default=datetime.datetime.now)), + ('endpoint', models.CharField(choices=[('/api/appauth/authortools', '/api/appauth/authortools'), ('/api/appauth/bibxml', '/api/appauth/bibxml'), ('/api/iesg/position', '/api/iesg/position'), ('/api/meeting/session/video/url', '/api/meeting/session/video/url'), ('/api/notify/meeting/bluesheet', '/api/notify/meeting/bluesheet'), ('/api/notify/meeting/registration', '/api/notify/meeting/registration'), ('/api/notify/session/attendees', '/api/notify/session/attendees'), ('/api/notify/session/chatlog', '/api/notify/session/chatlog'), ('/api/notify/session/polls', '/api/notify/session/polls'), ('/api/v2/person/person', '/api/v2/person/person')], max_length=128)), + ('created', models.DateTimeField(default=django.utils.timezone.now)), ('valid', models.BooleanField(default=True)), ('salt', models.BinaryField(default=ietf.person.models.salt, max_length=12)), ('count', models.IntegerField(default=0)), @@ -74,58 +85,87 @@ class Migration(migrations.Migration): ], ), migrations.CreateModel( - name='PersonEvent', + name='HistoricalPerson', fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('time', models.DateTimeField(default=datetime.datetime.now, help_text='When the event happened')), - ('type', models.CharField(choices=[('apikey_login', 'API key login')], max_length=50)), - ('desc', models.TextField()), + ('id', models.IntegerField(auto_created=True, blank=True, db_index=True, verbose_name='ID')), + ('time', models.DateTimeField(default=django.utils.timezone.now)), + ('name', models.CharField(db_index=True, help_text='Preferred long form of name.', max_length=255, validators=[ietf.person.models.name_character_validator], verbose_name='Full Name (Unicode)')), + ('plain', models.CharField(blank=True, default='', help_text="Use this if you have a Spanish double surname. Don't use this for nicknames, and don't use it unless you've actually observed that the datatracker shows your name incorrectly.", max_length=64, verbose_name='Plain Name correction (Unicode)')), + ('ascii', models.CharField(help_text='Name as rendered in ASCII (Latin, unaccented) characters.', max_length=255, verbose_name='Full Name (ASCII)')), + ('ascii_short', models.CharField(blank=True, help_text='Example: A. Nonymous. Fill in this with initials and surname only if taking the initials and surname of the ASCII name above produces an incorrect initials-only form. (Blank is OK).', max_length=32, null=True, verbose_name='Abbreviated Name (ASCII)')), + ('pronouns_selectable', jsonfield.fields.JSONCharField(blank=True, default=list, max_length=120, null=True, verbose_name='Pronouns')), + ('pronouns_freetext', models.CharField(blank=True, help_text='Optionally provide your personal pronouns. These will be displayed on your public profile page and alongside your name in Meetecho and, in future, other systems. Select any number of the checkboxes OR provide a custom string up to 30 characters.', max_length=30, null=True, verbose_name=' ')), + ('biography', models.TextField(blank=True, help_text='Short biography for use on leadership pages. Use plain text or reStructuredText markup.')), + ('photo', models.TextField(blank=True, default=None, max_length=100)), + ('photo_thumb', models.TextField(blank=True, default=None, max_length=100)), + ('name_from_draft', models.CharField(editable=False, help_text='Name as found in an Internet-Draft submission.', max_length=255, null=True, verbose_name='Full Name (from submission)')), + ('history_id', models.AutoField(primary_key=True, serialize=False)), + ('history_date', models.DateTimeField(db_index=True)), + ('history_change_reason', models.CharField(max_length=100, null=True)), + ('history_type', models.CharField(choices=[('+', 'Created'), ('~', 'Changed'), ('-', 'Deleted')], max_length=1)), + ('history_user', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='+', to=settings.AUTH_USER_MODEL)), + ('user', models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to=settings.AUTH_USER_MODEL)), ], options={ - 'ordering': ['-time', '-id'], + 'verbose_name': 'historical person', + 'verbose_name_plural': 'historical persons', + 'ordering': ('-history_date', '-history_id'), + 'get_latest_by': ('history_date', 'history_id'), }, + bases=(simple_history.models.HistoricalChanges, models.Model), ), migrations.CreateModel( - name='PersonHistory', + name='HistoricalEmail', fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('time', models.DateTimeField(default=datetime.datetime.now)), - ('name', models.CharField(db_index=True, help_text='Preferred form of name.', max_length=255, verbose_name='Full Name (Unicode)')), - ('ascii', models.CharField(help_text='Name as rendered in ASCII (Latin, unaccented) characters.', max_length=255, verbose_name='Full Name (ASCII)')), - ('ascii_short', models.CharField(blank=True, help_text='Example: A. Nonymous. Fill in this with initials and surname only if taking the initials and surname of the ASCII name above produces an incorrect initials-only form. (Blank is OK).', max_length=32, null=True, verbose_name='Abbreviated Name (ASCII)')), - ('affiliation', models.CharField(blank=True, help_text='Employer, university, sponsor, etc.', max_length=255)), - ('address', models.TextField(blank=True, help_text='Postal mailing address.', max_length=255)), - ('biography', models.TextField(blank=True, help_text='Short biography for use on leadership pages. Use plain text or reStructuredText markup.')), - ('photo', models.ImageField(blank=True, default=None, storage=ietf.utils.storage.NoLocationMigrationFileSystemStorage(location=None), upload_to='photo')), - ('photo_thumb', models.ImageField(blank=True, default=None, storage=ietf.utils.storage.NoLocationMigrationFileSystemStorage(location=None), upload_to='photo')), - ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='history_set', to='person.Person')), - ('user', ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL)), + ('address', django.contrib.postgres.fields.citext.CICharField(db_index=True, max_length=64, validators=[django.core.validators.EmailValidator()])), + ('time', models.DateTimeField(blank=True, editable=False)), + ('primary', models.BooleanField(default=False)), + ('origin', models.CharField(help_text="The origin of the address: the user's email address, or 'author: DRAFTNAME' if an Internet-Draft, or 'role: GROUP/ROLE' if a role.", max_length=150)), + ('active', models.BooleanField(default=True)), + ('history_id', models.AutoField(primary_key=True, serialize=False)), + ('history_date', models.DateTimeField(db_index=True)), + ('history_change_reason', models.CharField(max_length=100, null=True)), + ('history_type', models.CharField(choices=[('+', 'Created'), ('~', 'Changed'), ('-', 'Deleted')], max_length=1)), + ('history_user', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='+', to=settings.AUTH_USER_MODEL)), + ('person', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='person.Person')), ], options={ - 'abstract': False, + 'verbose_name': 'historical email', + 'verbose_name_plural': 'historical emails', + 'ordering': ('-history_date', '-history_id'), + 'get_latest_by': ('history_date', 'history_id'), }, + bases=(simple_history.models.HistoricalChanges, models.Model), ), migrations.CreateModel( - name='PersonApiKeyEvent', + name='Email', fields=[ - ('personevent_ptr', models.OneToOneField(auto_created=True, on_delete=django.db.models.deletion.CASCADE, parent_link=True, primary_key=True, serialize=False, to='person.PersonEvent')), - ('key', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.PersonalApiKey')), + ('address', django.contrib.postgres.fields.citext.CICharField(max_length=64, primary_key=True, serialize=False, validators=[django.core.validators.EmailValidator()])), + ('time', models.DateTimeField(auto_now_add=True)), + ('primary', models.BooleanField(default=False)), + ('origin', models.CharField(help_text="The origin of the address: the user's email address, or 'author: DRAFTNAME' if an Internet-Draft, or 'role: GROUP/ROLE' if a role.", max_length=150)), + ('active', models.BooleanField(default=True)), + ('person', ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='person.Person')), ], - bases=('person.personevent',), ), - migrations.AddField( - model_name='personevent', - name='person', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person'), + migrations.CreateModel( + name='Alias', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('name', models.CharField(db_index=True, max_length=255)), + ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), + ], + options={ + 'verbose_name_plural': 'Aliases', + }, ), - migrations.AddField( - model_name='email', - name='person', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='person.Person'), + migrations.AddIndex( + model_name='personevent', + index=models.Index(fields=['-time', '-id'], name='person_pers_time_8dfbc5_idx'), ), migrations.AddField( - model_name='alias', - name='person', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person'), + model_name='personapikeyevent', + name='key', + field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.PersonalApiKey'), ), ] diff --git a/ietf/person/migrations/0002_alter_historicalperson_ascii_and_more.py b/ietf/person/migrations/0002_alter_historicalperson_ascii_and_more.py new file mode 100644 index 0000000000..98d5da75d6 --- /dev/null +++ b/ietf/person/migrations/0002_alter_historicalperson_ascii_and_more.py @@ -0,0 +1,82 @@ +# Generated by Django 4.2.13 on 2024-05-22 18:50 + +from django.db import migrations, models +import ietf.person.models + + +class Migration(migrations.Migration): + + dependencies = [ + ("person", "0001_initial"), + ] + + operations = [ + migrations.AlterField( + model_name="historicalperson", + name="ascii", + field=models.CharField( + help_text="Name as rendered in ASCII (Latin, unaccented) characters.", + max_length=255, + validators=[ietf.person.models.name_character_validator], + verbose_name="Full Name (ASCII)", + ), + ), + migrations.AlterField( + model_name="historicalperson", + name="ascii_short", + field=models.CharField( + blank=True, + help_text="Example: A. Nonymous. Fill in this with initials and surname only if taking the initials and surname of the ASCII name above produces an incorrect initials-only form. (Blank is OK).", + max_length=32, + null=True, + validators=[ietf.person.models.name_character_validator], + verbose_name="Abbreviated Name (ASCII)", + ), + ), + migrations.AlterField( + model_name="historicalperson", + name="plain", + field=models.CharField( + blank=True, + default="", + help_text="Use this if you have a Spanish double surname. Don't use this for nicknames, and don't use it unless you've actually observed that the datatracker shows your name incorrectly.", + max_length=64, + validators=[ietf.person.models.name_character_validator], + verbose_name="Plain Name correction (Unicode)", + ), + ), + migrations.AlterField( + model_name="person", + name="ascii", + field=models.CharField( + help_text="Name as rendered in ASCII (Latin, unaccented) characters.", + max_length=255, + validators=[ietf.person.models.name_character_validator], + verbose_name="Full Name (ASCII)", + ), + ), + migrations.AlterField( + model_name="person", + name="ascii_short", + field=models.CharField( + blank=True, + help_text="Example: A. Nonymous. Fill in this with initials and surname only if taking the initials and surname of the ASCII name above produces an incorrect initials-only form. (Blank is OK).", + max_length=32, + null=True, + validators=[ietf.person.models.name_character_validator], + verbose_name="Abbreviated Name (ASCII)", + ), + ), + migrations.AlterField( + model_name="person", + name="plain", + field=models.CharField( + blank=True, + default="", + help_text="Use this if you have a Spanish double surname. Don't use this for nicknames, and don't use it unless you've actually observed that the datatracker shows your name incorrectly.", + max_length=64, + validators=[ietf.person.models.name_character_validator], + verbose_name="Plain Name correction (Unicode)", + ), + ), + ] diff --git a/ietf/person/migrations/0002_auto_20180330_0808.py b/ietf/person/migrations/0002_auto_20180330_0808.py deleted file mode 100644 index ec2d2d8938..0000000000 --- a/ietf/person/migrations/0002_auto_20180330_0808.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.11 on 2018-03-30 08:08 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0001_initial'), - ] - - operations = [ - migrations.AlterField( - model_name='personalapikey', - name='endpoint', - field=models.CharField(choices=[('/api/iesg/position', '/api/iesg/position'), ('/api/meeting/session/video/url', '/api/meeting/session/video/url')], max_length=128), - ), - ] diff --git a/ietf/person/migrations/0003_alter_personalapikey_endpoint.py b/ietf/person/migrations/0003_alter_personalapikey_endpoint.py new file mode 100644 index 0000000000..202af4b101 --- /dev/null +++ b/ietf/person/migrations/0003_alter_personalapikey_endpoint.py @@ -0,0 +1,42 @@ +# Generated by Django 4.2.16 on 2024-10-24 21:39 + +from django.db import migrations, models + + +class Migration(migrations.Migration): + + dependencies = [ + ("person", "0002_alter_historicalperson_ascii_and_more"), + ] + + operations = [ + migrations.AlterField( + model_name="personalapikey", + name="endpoint", + field=models.CharField( + choices=[ + ("/api/appauth/authortools", "/api/appauth/authortools"), + ("/api/appauth/bibxml", "/api/appauth/bibxml"), + ("/api/iesg/position", "/api/iesg/position"), + ( + "/api/meeting/session/recording-name", + "/api/meeting/session/recording-name", + ), + ( + "/api/meeting/session/video/url", + "/api/meeting/session/video/url", + ), + ("/api/notify/meeting/bluesheet", "/api/notify/meeting/bluesheet"), + ( + "/api/notify/meeting/registration", + "/api/notify/meeting/registration", + ), + ("/api/notify/session/attendees", "/api/notify/session/attendees"), + ("/api/notify/session/chatlog", "/api/notify/session/chatlog"), + ("/api/notify/session/polls", "/api/notify/session/polls"), + ("/api/v2/person/person", "/api/v2/person/person"), + ], + max_length=128, + ), + ), + ] diff --git a/ietf/person/migrations/0003_auto_20180504_1519.py b/ietf/person/migrations/0003_auto_20180504_1519.py deleted file mode 100644 index 928c637c53..0000000000 --- a/ietf/person/migrations/0003_auto_20180504_1519.py +++ /dev/null @@ -1,111 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.12 on 2018-05-04 15:19 - - -import datetime -from django.conf import settings -import django.core.validators -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - migrations.swappable_dependency(settings.AUTH_USER_MODEL), - ('person', '0002_auto_20180330_0808'), - ] - - operations = [ - migrations.CreateModel( - name='HistoricalEmail', - fields=[ - ('address', models.CharField(db_index=True, max_length=64, validators=[django.core.validators.EmailValidator()])), - ('time', models.DateTimeField(blank=True, editable=False)), - ('primary', models.BooleanField(default=False)), - ('origin', models.CharField(default='', editable=False, max_length=150)), - ('active', models.BooleanField(default=True)), - ('history_id', models.AutoField(primary_key=True, serialize=False)), - ('history_date', models.DateTimeField()), - ('history_change_reason', models.CharField(max_length=100, null=True)), - ('history_type', models.CharField(choices=[('+', 'Created'), ('~', 'Changed'), ('-', 'Deleted')], max_length=1)), - ('history_user', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='+', to=settings.AUTH_USER_MODEL)), - ], - options={ - 'ordering': ('-history_date', '-history_id'), - 'get_latest_by': 'history_date', - 'verbose_name': 'historical email', - }, - ), - migrations.CreateModel( - name='HistoricalPerson', - fields=[ - ('id', models.IntegerField(auto_created=True, blank=True, db_index=True, verbose_name='ID')), - ('time', models.DateTimeField(default=datetime.datetime.now)), - ('name', models.CharField(db_index=True, help_text='Preferred form of name.', max_length=255, verbose_name='Full Name (Unicode)')), - ('ascii', models.CharField(help_text='Name as rendered in ASCII (Latin, unaccented) characters.', max_length=255, verbose_name='Full Name (ASCII)')), - ('ascii_short', models.CharField(blank=True, help_text='Example: A. Nonymous. Fill in this with initials and surname only if taking the initials and surname of the ASCII name above produces an incorrect initials-only form. (Blank is OK).', max_length=32, null=True, verbose_name='Abbreviated Name (ASCII)')), - ('biography', models.TextField(blank=True, help_text='Short biography for use on leadership pages. Use plain text or reStructuredText markup.')), - ('photo', models.TextField(blank=True, default=None, max_length=100)), - ('photo_thumb', models.TextField(blank=True, default=None, max_length=100)), - ('name_from_draft', models.CharField(editable=False, help_text='Name as found in a draft submission.', max_length=255, null=True, verbose_name='Full Name (from submission)')), - ('history_id', models.AutoField(primary_key=True, serialize=False)), - ('history_date', models.DateTimeField()), - ('history_change_reason', models.CharField(max_length=100, null=True)), - ('history_type', models.CharField(choices=[('+', 'Created'), ('~', 'Changed'), ('-', 'Deleted')], max_length=1)), - ('history_user', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='+', to=settings.AUTH_USER_MODEL)), - ('user', models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to=settings.AUTH_USER_MODEL)), - ], - options={ - 'ordering': ('-history_date', '-history_id'), - 'get_latest_by': 'history_date', - 'verbose_name': 'historical person', - }, - ), - migrations.RemoveField( - model_name='personhistory', - name='person', - ), - migrations.RemoveField( - model_name='personhistory', - name='user', - ), - migrations.RemoveField( - model_name='person', - name='address', - ), - migrations.RemoveField( - model_name='person', - name='affiliation', - ), - migrations.AddField( - model_name='email', - name='origin', - field=models.CharField(default='', editable=False, max_length=150), - ), - migrations.AddField( - model_name='person', - name='name_from_draft', - field=models.CharField(editable=False, help_text='Name as found in a draft submission.', max_length=255, null=True, verbose_name='Full Name (from submission)'), - ), - migrations.DeleteModel( - name='PersonHistory', - ), - migrations.AddField( - model_name='historicalemail', - name='person', - field=ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='person.Person'), - ), - migrations.AddField( - model_name='historicalperson', - name='consent', - field=models.NullBooleanField(default=None, verbose_name='I hereby give my consent to the use of the personal details I have provided (photo, bio, name, email) within the IETF Datatracker'), - ), - migrations.AddField( - model_name='person', - name='consent', - field=models.NullBooleanField(default=None, verbose_name='I hereby give my consent to the use of the personal details I have provided (photo, bio, name, email) within the IETF Datatracker'), - ), - ] diff --git a/ietf/person/migrations/0004_alter_person_photo_alter_person_photo_thumb.py b/ietf/person/migrations/0004_alter_person_photo_alter_person_photo_thumb.py new file mode 100644 index 0000000000..f34382fa70 --- /dev/null +++ b/ietf/person/migrations/0004_alter_person_photo_alter_person_photo_thumb.py @@ -0,0 +1,38 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +from django.db import migrations, models +import ietf.utils.storage + + +class Migration(migrations.Migration): + + dependencies = [ + ("person", "0003_alter_personalapikey_endpoint"), + ] + + operations = [ + migrations.AlterField( + model_name="person", + name="photo", + field=models.ImageField( + blank=True, + default=None, + storage=ietf.utils.storage.BlobShadowFileSystemStorage( + kind="", location=None + ), + upload_to="photo", + ), + ), + migrations.AlterField( + model_name="person", + name="photo_thumb", + field=models.ImageField( + blank=True, + default=None, + storage=ietf.utils.storage.BlobShadowFileSystemStorage( + kind="", location=None + ), + upload_to="photo", + ), + ), + ] diff --git a/ietf/person/migrations/0004_populate_email_origin.py b/ietf/person/migrations/0004_populate_email_origin.py deleted file mode 100644 index 147b7b70c1..0000000000 --- a/ietf/person/migrations/0004_populate_email_origin.py +++ /dev/null @@ -1,110 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.12 on 2018-05-10 05:28 - - -import sys - -from django.db import migrations - -import debug # pyflakes:ignore - -def populate_email_origin(apps, schema_editor): - Submission = apps.get_model('submit', 'Submission') - Document = apps.get_model('doc', 'Document') - DocHistory = apps.get_model('doc', 'DocHistory') - DocumentAuthor = apps.get_model('doc', 'DocumentAuthor') - DocHistoryAuthor= apps.get_model('doc', 'DocHistoryAuthor') - Role = apps.get_model('group', 'Role') - RoleHistory = apps.get_model('group', 'RoleHistory') - ReviewRequest = apps.get_model('review', 'ReviewRequest') - LiaisonStatement= apps.get_model('liaisons', 'LiaisonStatement') - # - Email = apps.get_model('person', 'Email') - # - sys.stdout.write("\n") - # - sys.stdout.write("\n ** This migration may take some time. Expect at least a few minutes **.\n\n") - sys.stdout.write(" Initializing data structures...\n") - emails = dict([ (e.address, e) for e in Email.objects.filter(origin='') ]) - - count = 0 - sys.stdout.write(" Assigning email origins from Submission records...\n") - for o in Submission.objects.all().order_by('-submission_date'): - for a in o.authors: - addr = a['email'] - if addr in emails: - e = emails[addr] - if e.origin != o.name: - e.origin = "author: %s" % o.name - count += 1 - e.save() - del emails[addr] - sys.stdout.write(" Submission email origins assigned: %d\n" % count) - - for model in (DocumentAuthor, DocHistoryAuthor, ): - count = 0 - sys.stdout.write(" Assigning email origins from %s records...\n" % model.__name__) - for o in model.objects.filter(email__origin=''): - if not o.email.origin: - o.email.origin = "author: %s" % o.document.name - o.email.save() - count += 1 - sys.stdout.write(" %s email origins assigned: %d\n" % (model.__name__, count)) - - for model in (Role, RoleHistory, ): - count = 0 - sys.stdout.write(" Assigning email origins from %s records...\n" % model.__name__) - for o in model.objects.filter(email__origin=''): - if not o.email.origin: - o.email.origin = "role: %s %s" % (o.group.acronym, o.name.slug) - o.email.save() - count += 1 - sys.stdout.write(" %s email origins assigned: %d\n" % (model.__name__, count)) - - for model in (ReviewRequest, ): - count = 0 - sys.stdout.write(" Assigning email origins from %s records...\n" % model.__name__) - for o in model.objects.filter(reviewer__origin=''): - if not o.reviewer.origin: - o.reviewer.origin = "reviewer: %s" % (o.doc.name) - o.reviewer.save() - count += 1 - sys.stdout.write(" %s email origins assigned: %d\n" % (model.__name__, count)) - - for model in (LiaisonStatement, ): - count = 0 - sys.stdout.write(" Assigning email origins from %s records...\n" % model.__name__) - for o in model.objects.filter(from_contact__origin=''): - if not o.from_contact.origin: - o.from_contact.origin = "liaison: %s" % (','.join([ g.acronym for g in o.from_groups.all() ])) - o.from_contact.save() - count += 1 - sys.stdout.write(" %s email origins assigned: %d\n" % (model.__name__, count)) - - for model in (Document, DocHistory, ): - count = 0 - sys.stdout.write(" Assigning email origins from %s records...\n" % model.__name__) - for o in model.objects.filter(shepherd__origin=''): - if not o.shepherd.origin: - o.shepherd.origin = "shepherd: %s" % o.name - o.shepherd.save() - count += 1 - sys.stdout.write(" %s email origins assigned: %d\n" % (model.__name__, count)) - - sys.stdout.write("\n") - sys.stdout.write(" Email records with origin indication: %d\n" % Email.objects.exclude(origin='').count()) - sys.stdout.write(" Email records without origin indication: %d\n" % Email.objects.filter(origin='').count()) - -def reverse(apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0003_auto_20180504_1519'), - ] - - operations = [ - migrations.RunPython(populate_email_origin, reverse) - ] diff --git a/ietf/person/migrations/0005_alter_historicalperson_pronouns_selectable_and_more.py b/ietf/person/migrations/0005_alter_historicalperson_pronouns_selectable_and_more.py new file mode 100644 index 0000000000..2af874b1fa --- /dev/null +++ b/ietf/person/migrations/0005_alter_historicalperson_pronouns_selectable_and_more.py @@ -0,0 +1,34 @@ +# Copyright The IETF Trust 2025, All Rights Reserved + +from django.db import migrations, models + + +class Migration(migrations.Migration): + dependencies = [ + ("person", "0004_alter_person_photo_alter_person_photo_thumb"), + ] + + operations = [ + migrations.AlterField( + model_name="historicalperson", + name="pronouns_selectable", + field=models.JSONField( + blank=True, + default=list, + max_length=120, + null=True, + verbose_name="Pronouns", + ), + ), + migrations.AlterField( + model_name="person", + name="pronouns_selectable", + field=models.JSONField( + blank=True, + default=list, + max_length=120, + null=True, + verbose_name="Pronouns", + ), + ), + ] diff --git a/ietf/person/migrations/0005_populate_person_name_from_draft.py b/ietf/person/migrations/0005_populate_person_name_from_draft.py deleted file mode 100644 index d83dd22844..0000000000 --- a/ietf/person/migrations/0005_populate_person_name_from_draft.py +++ /dev/null @@ -1,48 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.12 on 2018-05-10 05:28 - - -import sys - -from django.db import migrations - -import debug # pyflakes:ignore - -def populate_person_name_from_draft(apps, schema_editor): - Submission = apps.get_model('submit', 'Submission') - Email = apps.get_model('person', 'Email') - # - sys.stdout.write("\n") - # - sys.stdout.write("\n ** This migration may take some time. Expect at least a few minutes **.\n\n") - sys.stdout.write(" Initializing data structures...\n") - persons = dict([ (e.address, e.person) for e in Email.objects.all() ]) - - count = 0 - sys.stdout.write(" Assigning Person.name_from_draft from Submission records...\n") - for o in Submission.objects.all().order_by('-submission_date'): - for a in o.authors: - name = a['name'] - email = a['email'] - if email in persons: - p = persons[email] - if not p.name_from_draft: - p.name_from_draft = name - count += 1 - p.save() - del persons[email] - sys.stdout.write(" Submission author names assigned: %d\n" % count) - -def reverse(apps, schema_editor): - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0004_populate_email_origin'), - ] - - operations = [ - migrations.RunPython(populate_person_name_from_draft, reverse) - ] diff --git a/ietf/person/migrations/0006_auto_20180910_0719.py b/ietf/person/migrations/0006_auto_20180910_0719.py deleted file mode 100644 index 9e17137a86..0000000000 --- a/ietf/person/migrations/0006_auto_20180910_0719.py +++ /dev/null @@ -1,24 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.13 on 2018-09-10 07:19 - - -from django.conf import settings -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0005_populate_person_name_from_draft'), - ] - - operations = [ - migrations.AlterField( - model_name='person', - name='user', - field=ietf.utils.models.OneToOneField(blank=True, null=True, on_delete=django.db.models.deletion.SET_NULL, to=settings.AUTH_USER_MODEL), - ), - ] diff --git a/ietf/person/migrations/0007_auto_20180929_1303.py b/ietf/person/migrations/0007_auto_20180929_1303.py deleted file mode 100644 index b4e41d7fa2..0000000000 --- a/ietf/person/migrations/0007_auto_20180929_1303.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.15 on 2018-09-29 13:03 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0006_auto_20180910_0719'), - ] - - operations = [ - migrations.AlterField( - model_name='personevent', - name='type', - field=models.CharField(choices=[('apikey_login', 'API key login'), ('gdpr_notice_email', 'GDPR consent request email sent'), ('email_address_deactivated', 'Email address deactivated')], max_length=50), - ), - ] diff --git a/ietf/person/migrations/0008_auto_20181014_1448.py b/ietf/person/migrations/0008_auto_20181014_1448.py deleted file mode 100644 index f6d765b02f..0000000000 --- a/ietf/person/migrations/0008_auto_20181014_1448.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.16 on 2018-10-14 14:48 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0007_auto_20180929_1303'), - ] - - operations = [ - migrations.AlterField( - model_name='personalapikey', - name='endpoint', - field=models.CharField(choices=[('/api/iesg/position', '/api/iesg/position'), ('/api/v2/person/person', '/api/v2/person/person'), ('/api/meeting/session/video/url', '/api/meeting/session/video/url')], max_length=128), - ), - ] diff --git a/ietf/person/migrations/0009_auto_20190118_0725.py b/ietf/person/migrations/0009_auto_20190118_0725.py deleted file mode 100644 index be609f109f..0000000000 --- a/ietf/person/migrations/0009_auto_20190118_0725.py +++ /dev/null @@ -1,26 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.18 on 2019-01-18 07:25 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0008_auto_20181014_1448'), - ] - - operations = [ - migrations.AlterField( - model_name='email', - name='origin', - field=models.CharField(help_text="The origin of the address: the user's email address, or 'author: DRAFTNAME' if a draft, or 'role: GROUP/ROLE' if a role.", max_length=150), - ), - migrations.AlterField( - model_name='historicalemail', - name='origin', - field=models.CharField(help_text="The origin of the address: the user's email address, or 'author: DRAFTNAME' if a draft, or 'role: GROUP/ROLE' if a role.", max_length=150), - ), - ] diff --git a/ietf/person/migrations/0010_auto_20200415_1133.py b/ietf/person/migrations/0010_auto_20200415_1133.py deleted file mode 100644 index 80aea5efa9..0000000000 --- a/ietf/person/migrations/0010_auto_20200415_1133.py +++ /dev/null @@ -1,20 +0,0 @@ -# -*- coding: utf-8 -*- -# Generated by Django 1.11.29 on 2020-04-15 11:33 -from __future__ import unicode_literals - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0009_auto_20190118_0725'), - ] - - operations = [ - migrations.AlterField( - model_name='personalapikey', - name='endpoint', - field=models.CharField(choices=[('/api/iesg/position', '/api/iesg/position'), ('/api/v2/person/person', '/api/v2/person/person'), ('/api/meeting/session/video/url', '/api/meeting/session/video/url'), ('/api/person/access/meetecho', '/api/person/access/meetecho')], max_length=128), - ), - ] diff --git a/ietf/person/migrations/0011_auto_20200608_1212.py b/ietf/person/migrations/0011_auto_20200608_1212.py deleted file mode 100644 index ea64411aa7..0000000000 --- a/ietf/person/migrations/0011_auto_20200608_1212.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.0.13 on 2020-06-08 12:12 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0010_auto_20200415_1133'), - ] - - operations = [ - migrations.AlterField( - model_name='personalapikey', - name='endpoint', - field=models.CharField(choices=[('/api/iesg/position', '/api/iesg/position'), ('/api/v2/person/person', '/api/v2/person/person'), ('/api/meeting/session/video/url', '/api/meeting/session/video/url'), ('/api/person/access/meetecho', '/api/person/access/meetecho'), ('/api/notify/meeting/registration', '/api/notify/meeting/registration')], max_length=128), - ), - ] diff --git a/ietf/person/migrations/0012_auto_20200624_1332.py b/ietf/person/migrations/0012_auto_20200624_1332.py deleted file mode 100644 index bba2ff621e..0000000000 --- a/ietf/person/migrations/0012_auto_20200624_1332.py +++ /dev/null @@ -1,23 +0,0 @@ -# Generated by Django 2.1.15 on 2020-06-24 13:32 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0011_auto_20200608_1212'), - ] - - operations = [ - migrations.AlterField( - model_name='historicalperson', - name='consent', - field=models.BooleanField(default=None, null=True, verbose_name='I hereby give my consent to the use of the personal details I have provided (photo, bio, name, email) within the IETF Datatracker'), - ), - migrations.AlterField( - model_name='person', - name='consent', - field=models.BooleanField(default=None, null=True, verbose_name='I hereby give my consent to the use of the personal details I have provided (photo, bio, name, email) within the IETF Datatracker'), - ), - ] diff --git a/ietf/person/migrations/0013_auto_20200711_1036.py b/ietf/person/migrations/0013_auto_20200711_1036.py deleted file mode 100644 index df031f6daf..0000000000 --- a/ietf/person/migrations/0013_auto_20200711_1036.py +++ /dev/null @@ -1,33 +0,0 @@ -# Generated by Django 2.2.14 on 2020-07-11 13:28 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0012_auto_20200624_1332'), - ] - - operations = [ - migrations.AddField( - model_name='historicalperson', - name='plain', - field=models.CharField(blank=True, default='', help_text='Preferred plain form of name, if different from the automatic plain form.', max_length=64, verbose_name='Plain Name (Unicode)'), - ), - migrations.AddField( - model_name='person', - name='plain', - field=models.CharField(blank=True, default='', help_text='Preferred plain form of name, if different from the automatic plain form.', max_length=64, verbose_name='Plain Name (Unicode)'), - ), - migrations.AlterField( - model_name='historicalperson', - name='name', - field=models.CharField(db_index=True, help_text='Preferred long form of name.', max_length=255, verbose_name='Full Name (Unicode)'), - ), - migrations.AlterField( - model_name='person', - name='name', - field=models.CharField(db_index=True, help_text='Preferred long form of name.', max_length=255, verbose_name='Full Name (Unicode)'), - ), - ] diff --git a/ietf/person/migrations/0014_auto_20200717_0743.py b/ietf/person/migrations/0014_auto_20200717_0743.py deleted file mode 100644 index 4c8742685c..0000000000 --- a/ietf/person/migrations/0014_auto_20200717_0743.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.2.14 on 2020-07-17 07:43 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0013_auto_20200711_1036'), - ] - - operations = [ - migrations.AlterField( - model_name='personalapikey', - name='endpoint', - field=models.CharField(choices=[('/api/iesg/position', '/api/iesg/position'), ('/api/v2/person/person', '/api/v2/person/person'), ('/api/meeting/session/video/url', '/api/meeting/session/video/url'), ('/api/notify/meeting/registration', '/api/notify/meeting/registration')], max_length=128), - ), - ] diff --git a/ietf/person/migrations/0015_extres.py b/ietf/person/migrations/0015_extres.py deleted file mode 100644 index 1b98285857..0000000000 --- a/ietf/person/migrations/0015_extres.py +++ /dev/null @@ -1,28 +0,0 @@ -# -*- coding: utf-8 -*- -# Generated by Django 1.11.29 on 2020-04-15 10:20 -from __future__ import unicode_literals - -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0014_extres'), - ('person', '0014_auto_20200717_0743'), - ] - - operations = [ - migrations.CreateModel( - name='PersonExtResource', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('display_name', models.CharField(blank=True, default='', max_length=255)), - ('value', models.CharField(max_length=2083)), - ('name', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ExtResourceName')), - ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), - ], - ), - ] diff --git a/ietf/person/migrations/0016_auto_20200807_0750.py b/ietf/person/migrations/0016_auto_20200807_0750.py deleted file mode 100644 index a956d608ba..0000000000 --- a/ietf/person/migrations/0016_auto_20200807_0750.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.2.15 on 2020-08-07 07:50 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0015_extres'), - ] - - operations = [ - migrations.AlterField( - model_name='personalapikey', - name='endpoint', - field=models.CharField(choices=[('/api/iesg/position', '/api/iesg/position'), ('/api/v2/person/person', '/api/v2/person/person'), ('/api/meeting/session/video/url', '/api/meeting/session/video/url'), ('/api/notify/meeting/registration', '/api/notify/meeting/registration'), ('/api/notify/meeting/bluesheet', '/api/notify/meeting/bluesheet')], max_length=128), - ), - ] diff --git a/ietf/person/migrations/0017_auto_20201028_0902.py b/ietf/person/migrations/0017_auto_20201028_0902.py deleted file mode 100644 index 275de3e1bf..0000000000 --- a/ietf/person/migrations/0017_auto_20201028_0902.py +++ /dev/null @@ -1,23 +0,0 @@ -# Generated by Django 2.2.16 on 2020-10-28 09:02 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0016_auto_20200807_0750'), - ] - - operations = [ - migrations.AlterField( - model_name='historicalperson', - name='plain', - field=models.CharField(blank=True, default='', help_text="Use this if you have a Spanish double surname. Don't use this for nicknames, and don't use it unless you've actually observed that the datatracker shows your name incorrectly.", max_length=64, verbose_name='Plain Name correction (Unicode)'), - ), - migrations.AlterField( - model_name='person', - name='plain', - field=models.CharField(blank=True, default='', help_text="Use this if you have a Spanish double surname. Don't use this for nicknames, and don't use it unless you've actually observed that the datatracker shows your name incorrectly.", max_length=64, verbose_name='Plain Name correction (Unicode)'), - ), - ] diff --git a/ietf/person/migrations/0018_auto_20201109_0439.py b/ietf/person/migrations/0018_auto_20201109_0439.py deleted file mode 100644 index 51f1d2d917..0000000000 --- a/ietf/person/migrations/0018_auto_20201109_0439.py +++ /dev/null @@ -1,17 +0,0 @@ -# Generated by Django 2.2.17 on 2020-11-09 04:39 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0017_auto_20201028_0902'), - ] - - operations = [ - migrations.AddIndex( - model_name='personevent', - index=models.Index(fields=['-time', '-id'], name='person_pers_time_8dfbc5_idx'), - ), - ] diff --git a/ietf/person/migrations/0019_auto_20210604_1443.py b/ietf/person/migrations/0019_auto_20210604_1443.py deleted file mode 100644 index 432f7541d5..0000000000 --- a/ietf/person/migrations/0019_auto_20210604_1443.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.2.20 on 2021-06-04 14:43 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0018_auto_20201109_0439'), - ] - - operations = [ - migrations.AlterField( - model_name='personalapikey', - name='endpoint', - field=models.CharField(choices=[('/api/iesg/position', '/api/iesg/position'), ('/api/meeting/session/video/url', '/api/meeting/session/video/url'), ('/api/notify/meeting/bluesheet', '/api/notify/meeting/bluesheet'), ('/api/notify/meeting/registration', '/api/notify/meeting/registration'), ('/api/v2/person/person', '/api/v2/person/person')], max_length=128), - ), - ] diff --git a/ietf/person/migrations/0020_auto_20210920_0924.py b/ietf/person/migrations/0020_auto_20210920_0924.py deleted file mode 100644 index bcb9dc1c3e..0000000000 --- a/ietf/person/migrations/0020_auto_20210920_0924.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.2.24 on 2021-09-20 09:24 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0019_auto_20210604_1443'), - ] - - operations = [ - migrations.AlterField( - model_name='personalapikey', - name='endpoint', - field=models.CharField(choices=[('/api/appauth/authortools', '/api/appauth/authortools'), ('/api/iesg/position', '/api/iesg/position'), ('/api/meeting/session/video/url', '/api/meeting/session/video/url'), ('/api/notify/meeting/bluesheet', '/api/notify/meeting/bluesheet'), ('/api/notify/meeting/registration', '/api/notify/meeting/registration'), ('/api/v2/person/person', '/api/v2/person/person')], max_length=128), - ), - ] diff --git a/ietf/person/migrations/0021_auto_20211210_0805.py b/ietf/person/migrations/0021_auto_20211210_0805.py deleted file mode 100644 index d65b1a99bd..0000000000 --- a/ietf/person/migrations/0021_auto_20211210_0805.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.2.25 on 2021-12-10 08:05 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0020_auto_20210920_0924'), - ] - - operations = [ - migrations.AlterField( - model_name='personalapikey', - name='endpoint', - field=models.CharField(choices=[('/api/appauth/authortools', '/api/appauth/authortools'), ('/api/appauth/bibxml', '/api/appauth/bibxml'), ('/api/iesg/position', '/api/iesg/position'), ('/api/meeting/session/video/url', '/api/meeting/session/video/url'), ('/api/notify/meeting/bluesheet', '/api/notify/meeting/bluesheet'), ('/api/notify/meeting/registration', '/api/notify/meeting/registration'), ('/api/v2/person/person', '/api/v2/person/person')], max_length=128), - ), - ] diff --git a/ietf/person/migrations/0022_auto_20220513_1456.py b/ietf/person/migrations/0022_auto_20220513_1456.py deleted file mode 100644 index 96f942496d..0000000000 --- a/ietf/person/migrations/0022_auto_20220513_1456.py +++ /dev/null @@ -1,31 +0,0 @@ -# Generated by Django 2.2.28 on 2022-05-13 14:56 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0021_auto_20211210_0805'), - ] - - operations = [ - migrations.AlterModelOptions( - name='historicalemail', - options={'get_latest_by': ('history_date', 'history_id'), 'ordering': ('-history_date', '-history_id'), 'verbose_name': 'historical email', 'verbose_name_plural': 'historical emails'}, - ), - migrations.AlterModelOptions( - name='historicalperson', - options={'get_latest_by': ('history_date', 'history_id'), 'ordering': ('-history_date', '-history_id'), 'verbose_name': 'historical person', 'verbose_name_plural': 'historical persons'}, - ), - migrations.AlterField( - model_name='historicalemail', - name='history_date', - field=models.DateTimeField(db_index=True), - ), - migrations.AlterField( - model_name='historicalperson', - name='history_date', - field=models.DateTimeField(db_index=True), - ), - ] diff --git a/ietf/person/migrations/0023_auto_20220615_1006.py b/ietf/person/migrations/0023_auto_20220615_1006.py deleted file mode 100644 index 22b50a25a3..0000000000 --- a/ietf/person/migrations/0023_auto_20220615_1006.py +++ /dev/null @@ -1,18 +0,0 @@ -# Generated by Django 2.2.28 on 2022-06-15 10:06 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0022_auto_20220513_1456'), - ] - - operations = [ - migrations.AlterField( - model_name='personalapikey', - name='endpoint', - field=models.CharField(choices=[('/api/appauth/authortools', '/api/appauth/authortools'), ('/api/appauth/bibxml', '/api/appauth/bibxml'), ('/api/iesg/position', '/api/iesg/position'), ('/api/meeting/session/video/url', '/api/meeting/session/video/url'), ('/api/notify/meeting/bluesheet', '/api/notify/meeting/bluesheet'), ('/api/notify/meeting/registration', '/api/notify/meeting/registration'), ('/api/notify/session/attendees', '/api/notify/session/attendees'), ('/api/v2/person/person', '/api/v2/person/person')], max_length=128), - ), - ] diff --git a/ietf/person/migrations/0024_pronouns.py b/ietf/person/migrations/0024_pronouns.py deleted file mode 100644 index 1ef9514b03..0000000000 --- a/ietf/person/migrations/0024_pronouns.py +++ /dev/null @@ -1,45 +0,0 @@ -# Copyright The IETF Trust 2022, All Rights Reserved -# Generated by Django 2.2.28 on 2022-06-17 15:09 - -from django.db import migrations, models -import jsonfield.fields - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0023_auto_20220615_1006'), - ] - - operations = [ - migrations.AddField( - model_name='historicalperson', - name='pronouns_freetext', - field=models.CharField(blank=True, help_text='Optionally provide your personal pronouns. These will be displayed on your public profile page and alongside your name in Meetecho and, in future, other systems. Select any number of the checkboxes OR provide a custom string up to 30 characters.', max_length=30, null=True, verbose_name=' '), - ), - migrations.AddField( - model_name='historicalperson', - name='pronouns_selectable', - field=jsonfield.fields.JSONCharField(blank=True, default=list, max_length=120, null=True, verbose_name='Pronouns'), - ), - migrations.AddField( - model_name='person', - name='pronouns_freetext', - field=models.CharField(blank=True, help_text='Optionally provide your personal pronouns. These will be displayed on your public profile page and alongside your name in Meetecho and, in future, other systems. Select any number of the checkboxes OR provide a custom string up to 30 characters.', max_length=30, null=True, verbose_name=' '), - ), - migrations.AddField( - model_name='person', - name='pronouns_selectable', - field=jsonfield.fields.JSONCharField(blank=True, default=list, max_length=120, null=True, verbose_name='Pronouns'), - ), - migrations.AlterField( - model_name='historicalperson', - name='consent', - field=models.BooleanField(default=None, null=True, verbose_name='I hereby give my consent to the use of the personal details I have provided (photo, bio, name, pronouns, email) within the IETF Datatracker'), - ), - migrations.AlterField( - model_name='person', - name='consent', - field=models.BooleanField(default=None, null=True, verbose_name='I hereby give my consent to the use of the personal details I have provided (photo, bio, name, pronouns, email) within the IETF Datatracker'), - ), - ] diff --git a/ietf/person/migrations/0025_chat_and_polls_apikey.py b/ietf/person/migrations/0025_chat_and_polls_apikey.py deleted file mode 100644 index 03afdc5999..0000000000 --- a/ietf/person/migrations/0025_chat_and_polls_apikey.py +++ /dev/null @@ -1,17 +0,0 @@ -# Copyright The IETF Trust 2022, All Rights Reserved - -from django.db import migrations, models - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0024_pronouns'), - ] - - operations = [ - migrations.AlterField( - model_name='personalapikey', - name='endpoint', - field=models.CharField(choices=[('/api/appauth/authortools', '/api/appauth/authortools'), ('/api/appauth/bibxml', '/api/appauth/bibxml'), ('/api/iesg/position', '/api/iesg/position'), ('/api/meeting/session/video/url', '/api/meeting/session/video/url'), ('/api/notify/meeting/bluesheet', '/api/notify/meeting/bluesheet'), ('/api/notify/meeting/registration', '/api/notify/meeting/registration'), ('/api/notify/session/attendees', '/api/notify/session/attendees'), ('/api/notify/session/chatlog', '/api/notify/session/chatlog'), ('/api/notify/session/polls', '/api/notify/session/polls'), ('/api/v2/person/person', '/api/v2/person/person')], max_length=128), - ), - ] diff --git a/ietf/person/migrations/0026_drop_consent.py b/ietf/person/migrations/0026_drop_consent.py deleted file mode 100644 index 2acaad6786..0000000000 --- a/ietf/person/migrations/0026_drop_consent.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2022, All Rights Reserved - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0025_chat_and_polls_apikey'), - ] - - operations = [ - migrations.RemoveField( - model_name='historicalperson', - name='consent', - ), - migrations.RemoveField( - model_name='person', - name='consent', - ), - ] diff --git a/ietf/person/migrations/0027_personevent_drop_consent.py b/ietf/person/migrations/0027_personevent_drop_consent.py deleted file mode 100644 index e2443596fb..0000000000 --- a/ietf/person/migrations/0027_personevent_drop_consent.py +++ /dev/null @@ -1,18 +0,0 @@ -# Copyright The IETF Trust 2022, All Rights Reserved - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0026_drop_consent'), - ] - - operations = [ - migrations.AlterField( - model_name='personevent', - name='type', - field=models.CharField(choices=[('apikey_login', 'API key login'), ('email_address_deactivated', 'Email address deactivated')], max_length=50), - ), - ] diff --git a/ietf/person/migrations/0028_name_character_validator.py b/ietf/person/migrations/0028_name_character_validator.py deleted file mode 100644 index c7e4a93051..0000000000 --- a/ietf/person/migrations/0028_name_character_validator.py +++ /dev/null @@ -1,24 +0,0 @@ -# Copyright The IETF Trust 2022, All Rights Reserved - -from django.db import migrations, models -import ietf.person.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0027_personevent_drop_consent'), - ] - - operations = [ - migrations.AlterField( - model_name='historicalperson', - name='name', - field=models.CharField(db_index=True, help_text='Preferred long form of name.', max_length=255, validators=[ietf.person.models.name_character_validator], verbose_name='Full Name (Unicode)'), - ), - migrations.AlterField( - model_name='person', - name='name', - field=models.CharField(db_index=True, help_text='Preferred long form of name.', max_length=255, validators=[ietf.person.models.name_character_validator], verbose_name='Full Name (Unicode)'), - ), - ] diff --git a/ietf/person/migrations/0029_use_timezone_now_for_person_models.py b/ietf/person/migrations/0029_use_timezone_now_for_person_models.py deleted file mode 100644 index 4848573ee5..0000000000 --- a/ietf/person/migrations/0029_use_timezone_now_for_person_models.py +++ /dev/null @@ -1,34 +0,0 @@ -# Generated by Django 2.2.28 on 2022-07-12 11:24 - -from django.db import migrations, models -import django.utils.timezone - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0028_name_character_validator'), - ] - - operations = [ - migrations.AlterField( - model_name='historicalperson', - name='time', - field=models.DateTimeField(default=django.utils.timezone.now), - ), - migrations.AlterField( - model_name='person', - name='time', - field=models.DateTimeField(default=django.utils.timezone.now), - ), - migrations.AlterField( - model_name='personalapikey', - name='created', - field=models.DateTimeField(default=django.utils.timezone.now), - ), - migrations.AlterField( - model_name='personevent', - name='time', - field=models.DateTimeField(default=django.utils.timezone.now, help_text='When the event happened'), - ), - ] diff --git a/ietf/person/models.py b/ietf/person/models.py index a7e30b2de9..3ab89289a6 100644 --- a/ietf/person/models.py +++ b/ietf/person/models.py @@ -4,7 +4,6 @@ import email.utils import email.header -import jsonfield import uuid from hashids import Hashids @@ -12,6 +11,7 @@ from django.conf import settings from django.contrib.auth.models import User +from django.contrib.postgres.fields import CICharField from django.core.exceptions import ValidationError from django.core.validators import validate_email from django.db import models @@ -28,7 +28,7 @@ from ietf.name.models import ExtResourceName from ietf.person.name import name_parts, initials, plain_name from ietf.utils.mail import send_mail_preformatted -from ietf.utils.storage import NoLocationMigrationFileSystemStorage +from ietf.utils.storage import BlobShadowFileSystemStorage from ietf.utils.mail import formataddr from ietf.person.name import unidecode_name from ietf.utils import log @@ -36,8 +36,12 @@ def name_character_validator(value): - if '/' in value: - raise ValidationError('Name cannot contain "/" character.') + disallowed = "@:/" + found = set(disallowed).intersection(value) + if len(found) > 0: + raise ValidationError( + f"This name cannot contain the characters {', '.join(disallowed)}" + ) class Person(models.Model): @@ -47,17 +51,27 @@ class Person(models.Model): # The normal unicode form of the name. This must be # set to the same value as the ascii-form if equal. name = models.CharField("Full Name (Unicode)", max_length=255, db_index=True, help_text="Preferred long form of name.", validators=[name_character_validator]) - plain = models.CharField("Plain Name correction (Unicode)", max_length=64, default='', blank=True, help_text="Use this if you have a Spanish double surname. Don't use this for nicknames, and don't use it unless you've actually observed that the datatracker shows your name incorrectly.") + plain = models.CharField("Plain Name correction (Unicode)", max_length=64, default='', blank=True, help_text="Use this if you have a Spanish double surname. Don't use this for nicknames, and don't use it unless you've actually observed that the datatracker shows your name incorrectly.", validators=[name_character_validator]) # The normal ascii-form of the name. - ascii = models.CharField("Full Name (ASCII)", max_length=255, help_text="Name as rendered in ASCII (Latin, unaccented) characters.") + ascii = models.CharField("Full Name (ASCII)", max_length=255, help_text="Name as rendered in ASCII (Latin, unaccented) characters.", validators=[name_character_validator]) # The short ascii-form of the name. Also in alias table if non-null - ascii_short = models.CharField("Abbreviated Name (ASCII)", max_length=32, null=True, blank=True, help_text="Example: A. Nonymous. Fill in this with initials and surname only if taking the initials and surname of the ASCII name above produces an incorrect initials-only form. (Blank is OK).") - pronouns_selectable = jsonfield.JSONCharField("Pronouns", max_length=120, blank=True, null=True, default=list ) + ascii_short = models.CharField("Abbreviated Name (ASCII)", max_length=32, null=True, blank=True, help_text="Example: A. Nonymous. Fill in this with initials and surname only if taking the initials and surname of the ASCII name above produces an incorrect initials-only form. (Blank is OK).", validators=[name_character_validator]) + pronouns_selectable = models.JSONField("Pronouns", max_length=120, blank=True, null=True, default=list ) pronouns_freetext = models.CharField(" ", max_length=30, null=True, blank=True, help_text="Optionally provide your personal pronouns. These will be displayed on your public profile page and alongside your name in Meetecho and, in future, other systems. Select any number of the checkboxes OR provide a custom string up to 30 characters.") biography = models.TextField(blank=True, help_text="Short biography for use on leadership pages. Use plain text or reStructuredText markup.") - photo = models.ImageField(storage=NoLocationMigrationFileSystemStorage(), upload_to=settings.PHOTOS_DIRNAME, blank=True, default=None) - photo_thumb = models.ImageField(storage=NoLocationMigrationFileSystemStorage(), upload_to=settings.PHOTOS_DIRNAME, blank=True, default=None) - name_from_draft = models.CharField("Full Name (from submission)", null=True, max_length=255, editable=False, help_text="Name as found in a draft submission.") + photo = models.ImageField( + storage=BlobShadowFileSystemStorage(kind="photo"), + upload_to=settings.PHOTOS_DIRNAME, + blank=True, + default=None, + ) + photo_thumb = models.ImageField( + storage=BlobShadowFileSystemStorage(kind="photo"), + upload_to=settings.PHOTOS_DIRNAME, + blank=True, + default=None, + ) + name_from_draft = models.CharField("Full Name (from submission)", null=True, max_length=255, editable=False, help_text="Name as found in an Internet-Draft submission.") def __str__(self): return self.plain_name() @@ -73,7 +87,7 @@ def short(self): else: prefix, first, middle, last, suffix = self.ascii_parts() return (first and first[0]+"." or "")+(middle or "")+" "+last+(suffix and " "+suffix or "") - def plain_name(self): + def plain_name(self) -> str: if not hasattr(self, '_cached_plain_name'): if self.plain: self._cached_plain_name = self.plain @@ -144,6 +158,14 @@ def email(self): e = self.email_set.filter(active=True).order_by("-time").first() self._cached_email = e return self._cached_email + def email_allowing_inactive(self): + if not hasattr(self, "_cached_email_allowing_inactive"): + e = self.email() + if not e: + e = self.email_set.order_by("-time").first() + log.assertion(statement="e is not None", note=f"Person {self.pk} has no Email objects") + self._cached_email_allowing_inactive = e + return self._cached_email_allowing_inactive def email_address(self): e = self.email() if e: @@ -181,18 +203,40 @@ def has_drafts(self): def rfcs(self): from ietf.doc.models import Document - rfcs = list(Document.objects.filter(documentauthor__person=self, type='draft', states__slug='rfc')) - rfcs.sort(key=lambda d: d.canonical_name() ) + # When RfcAuthors are populated, this may over-return if an author is dropped + # from the author list between the final draft and the published RFC. Should + # ignore DocumentAuthors when an RfcAuthor exists for a draft. + rfcs = list(Document.objects.filter(type="rfc").filter(models.Q(documentauthor__person=self)|models.Q(rfcauthor__person=self)).distinct()) + rfcs.sort(key=lambda d: d.name ) return rfcs def active_drafts(self): from ietf.doc.models import Document - return Document.objects.filter(documentauthor__person=self, type='draft', states__slug='active').distinct().order_by('-time') + + return ( + Document.objects.filter( + documentauthor__person=self, + type="draft", + states__type="draft", + states__slug="active", + ) + .distinct() + .order_by("-time") + ) def expired_drafts(self): from ietf.doc.models import Document - return Document.objects.filter(documentauthor__person=self, type='draft', states__slug__in=['repl', 'expired', 'auth-rm', 'ietf-rm']).distinct().order_by('-time') + return ( + Document.objects.filter( + documentauthor__person=self, + type="draft", + states__type="draft", + states__slug__in=["repl", "expired", "auth-rm", "ietf-rm"], + ) + .distinct() + .order_by("-time") + ) def save(self, *args, **kwargs): created = not self.pk @@ -225,11 +269,16 @@ def available_api_endpoints(self): def cdn_photo_url(self, size=80): if self.photo: if settings.SERVE_CDN_PHOTOS: + if settings.SERVER_MODE != "production": + original_media_dir = settings.MEDIA_URL + settings.MEDIA_URL = "https://www.ietf.org/lib/dt/media/" source_url = self.photo.url if source_url.startswith(settings.IETF_HOST_URL): source_url = source_url[len(settings.IETF_HOST_URL):] elif source_url.startswith('/'): source_url = source_url[1:] + if settings.SERVER_MODE != "production": + settings.MEDIA_URL = original_media_dir return f'{settings.IETF_HOST_URL}cdn-cgi/image/fit=scale-down,width={size},height={size}/{source_url}' else: datatracker_photo_path = urlreverse('ietf.person.views.photo', kwargs={'email_or_name': self.email()}) @@ -277,11 +326,11 @@ class Meta: class Email(models.Model): history = HistoricalRecords() - address = models.CharField(max_length=64, primary_key=True, validators=[validate_email]) + address = CICharField(max_length=64, primary_key=True, validators=[validate_email]) person = ForeignKey(Person, null=True) time = models.DateTimeField(auto_now_add=True) primary = models.BooleanField(default=False) - origin = models.CharField(max_length=150, blank=False, help_text="The origin of the address: the user's email address, or 'author: DRAFTNAME' if a draft, or 'role: GROUP/ROLE' if a role.") # User.username or Document.name + origin = models.CharField(max_length=150, blank=False, help_text="The origin of the address: the user's email address, or 'author: DRAFTNAME' if an Internet-Draft, or 'role: GROUP/ROLE' if a role.") # User.username or Document.name active = models.BooleanField(default=True) # Old email addresses are *not* purged, as history # information points to persons through these @@ -344,6 +393,7 @@ def salt(): ("/api/iesg/position", "/api/iesg/position", "Area Director"), ("/api/v2/person/person", "/api/v2/person/person", "Robot"), ("/api/meeting/session/video/url", "/api/meeting/session/video/url", "Recording Manager"), + ("/api/meeting/session/recording-name", "/api/meeting/session/recording-name", "Recording Manager"), ("/api/notify/meeting/registration", "/api/notify/meeting/registration", "Robot"), ("/api/notify/meeting/bluesheet", "/api/notify/meeting/bluesheet", "Recording Manager"), ("/api/notify/session/attendees", "/api/notify/session/attendees", "Recording Manager"), diff --git a/ietf/person/name.py b/ietf/person/name.py index dc57f58f4b..0dbeaa9b99 100644 --- a/ietf/person/name.py +++ b/ietf/person/name.py @@ -59,7 +59,7 @@ def name_parts(name): last = parts[0] if len(parts) >= 2: # Handle reverse-order names with uppercase surname correctly - if len(first)>1 and re.search("^[A-Z-]+$", first): + if len(first)>1 and re.search("^[A-Z-]+$", first) and first != "JP": first, last = last, first.capitalize() # Handle exception for RFC Editor if (prefix, first, middle, last, suffix) == ('', 'Editor', '', 'Rfc', ''): diff --git a/ietf/person/serializers.py b/ietf/person/serializers.py new file mode 100644 index 0000000000..023d77d4bc --- /dev/null +++ b/ietf/person/serializers.py @@ -0,0 +1,39 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +"""DRF Serializers""" + +from rest_framework import serializers + +from ietf.ietfauth.validators import is_allowed_address + +from .models import Email, Person + + +class EmailSerializer(serializers.ModelSerializer): + """Email serializer for read/update""" + + address = serializers.EmailField(read_only=True) + + class Meta: + model = Email + fields = [ + "person", + "address", + "primary", + "active", + "origin", + ] + read_only_fields = ["person", "address", "origin"] + + +class NewEmailSerializer(serializers.Serializer): + """Serialize a new email address request""" + address = serializers.EmailField(validators=[is_allowed_address]) + + +class PersonSerializer(serializers.ModelSerializer): + """Person serializer""" + emails = EmailSerializer(many=True, source="email_set") + + class Meta: + model = Person + fields = ["id", "name", "emails"] diff --git a/ietf/person/tasks.py b/ietf/person/tasks.py new file mode 100644 index 0000000000..f0c979fa26 --- /dev/null +++ b/ietf/person/tasks.py @@ -0,0 +1,59 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +# +# Celery task definitions +# +import datetime + +from celery import shared_task + +from django.conf import settings +from django.utils import timezone + +from ietf.utils import log +from ietf.utils.mail import send_mail +from .models import PersonalApiKey, PersonApiKeyEvent + + +@shared_task +def send_apikey_usage_emails_task(days): + """Send usage emails to Persons who have API keys""" + earliest = timezone.now() - datetime.timedelta(days=days) + keys = PersonalApiKey.objects.filter( + valid=True, + personapikeyevent__time__gt=earliest, + ).distinct() + for key in keys: + events = PersonApiKeyEvent.objects.filter(key=key, time__gt=earliest) + count = events.count() + events = events[:32] + if count: + key_name = key.hash()[:8] + subject = "API key usage for key '%s' for the last %s days" % ( + key_name, + days, + ) + to = key.person.email_address() + frm = settings.DEFAULT_FROM_EMAIL + send_mail( + None, + to, + frm, + subject, + "utils/apikey_usage_report.txt", + { + "person": key.person, + "days": days, + "key": key, + "key_name": key_name, + "count": count, + "events": events, + }, + ) + +@shared_task +def purge_personal_api_key_events_task(keep_days): + keep_since = timezone.now() - datetime.timedelta(days=keep_days) + old_events = PersonApiKeyEvent.objects.filter(time__lt=keep_since) + count = len(old_events) + old_events.delete() + log.log(f"Deleted {count} PersonApiKeyEvents older than {keep_since}") diff --git a/ietf/person/tests.py b/ietf/person/tests.py index 11e1a5b663..f55d8b8a34 100644 --- a/ietf/person/tests.py +++ b/ietf/person/tests.py @@ -1,15 +1,15 @@ -# Copyright The IETF Trust 2014-2022, All Rights Reserved +# Copyright The IETF Trust 2014-2025, All Rights Reserved # -*- coding: utf-8 -*- import datetime import json +from unittest import mock from io import StringIO, BytesIO from PIL import Image from pyquery import PyQuery - from django.core.exceptions import ValidationError from django.http import HttpRequest from django.test import override_settings @@ -22,14 +22,16 @@ from ietf.community.models import CommunityList from ietf.group.factories import RoleFactory from ietf.group.models import Group +from ietf.message.models import Message from ietf.nomcom.models import NomCom from ietf.nomcom.test_data import nomcom_test_data from ietf.nomcom.factories import NomComFactory, NomineeFactory, NominationFactory, FeedbackFactory, PositionFactory -from ietf.person.factories import EmailFactory, PersonFactory, UserFactory -from ietf.person.models import Person, Alias +from ietf.person.factories import EmailFactory, PersonFactory, PersonApiKeyEventFactory +from ietf.person.models import Person, Alias, PersonApiKeyEvent +from ietf.person.tasks import purge_personal_api_key_events_task from ietf.person.utils import (merge_persons, determine_merge_order, send_merge_notification, handle_users, get_extra_primary, dedupe_aliases, move_related_objects, merge_nominees, - handle_reviewer_settings, merge_users, get_dots) + handle_reviewer_settings, get_dots) from ietf.review.models import ReviewerSettings from ietf.utils.test_utils import TestCase, login_testing_unauthorized from ietf.utils.mail import outbox, empty_outbox @@ -112,6 +114,14 @@ def test_person_profile_without_email(self): r = self.client.get(url) self.assertContains(r, person.name, status_code=200) + def test_case_insensitive(self): + # Case insensitive seach + person = PersonFactory(name="Test Person") + url = urlreverse("ietf.person.views.profile", kwargs={ "email_or_name": "test person"}) + r = self.client.get(url) + self.assertContains(r, person.name, status_code=200) + self.assertNotIn('More than one person', r.content.decode()) + def test_person_profile_duplicates(self): # same Person name and email - should not show on the profile as multiple Person records person = PersonFactory(name="bazquux@example.com", user__email="bazquux@example.com") @@ -157,6 +167,14 @@ def test_person_photo(self): img = Image.open(BytesIO(r.content)) self.assertEqual(img.width, 200) + def test_person_photo_duplicates(self): + person = PersonFactory(name="bazquux@example.com", user__username="bazquux@example.com", with_bio=True) + PersonFactory(name="bazquux@example.com", user__username="foobar@example.com", with_bio=True) + + url = urlreverse("ietf.person.views.photo", kwargs={ "email_or_name": person.plain_name()}) + r = self.client.get(url) + self.assertEqual(r.status_code, 404) + def test_name_methods(self): person = PersonFactory(name="Dr. Jens F. Möller", ) @@ -190,13 +208,13 @@ def test_merge(self): def test_merge_with_params(self): p1 = get_person_no_user() p2 = PersonFactory() - url = urlreverse("ietf.person.views.merge") + "?source={}&target={}".format(p1.pk, p2.pk) + url = urlreverse("ietf.person.views.merge_submit") + "?source={}&target={}".format(p1.pk, p2.pk) login_testing_unauthorized(self, "secretary", url) r = self.client.get(url) self.assertContains(r, 'retaining login', status_code=200) def test_merge_with_params_bad_id(self): - url = urlreverse("ietf.person.views.merge") + "?source=1000&target=2000" + url = urlreverse("ietf.person.views.merge_submit") + "?source=1000&target=2000" login_testing_unauthorized(self, "secretary", url) r = self.client.get(url) self.assertContains(r, 'ID does not exist', status_code=200) @@ -204,7 +222,7 @@ def test_merge_with_params_bad_id(self): def test_merge_post(self): p1 = get_person_no_user() p2 = PersonFactory() - url = urlreverse("ietf.person.views.merge") + url = urlreverse("ietf.person.views.merge_submit") expected_url = urlreverse("ietf.secr.rolodex.views.view", kwargs={'id': p2.pk}) login_testing_unauthorized(self, "secretary", url) data = {'source': p1.pk, 'target': p2.pk} @@ -228,9 +246,11 @@ def test_cdn_photo_url_cdn_off(self): self.assertNotIn('cdn-cgi/photo',p.cdn_photo_url()) def test_invalid_name_characters_rejected(self): - slash_person = PersonFactory.build(name='I have a /', user=None) # build() does not save the new object - with self.assertRaises(ValidationError): - slash_person.full_clean() # calls validators (save() does *not*) + for disallowed in "/:@": + # build() does not save the new object + person_with_bad_name = PersonFactory.build(name=f"I have a {disallowed}", user=None) + with self.assertRaises(ValidationError, msg=f"Name with a {disallowed} char should be rejected"): + person_with_bad_name.full_clean() # calls validators (save() does *not*) class PersonUtilsTests(TestCase): @@ -373,13 +393,24 @@ def test_merge_persons(self): request.user = user source = PersonFactory() target = PersonFactory() + mars = RoleFactory(name_id='chair',group__acronym='mars').group source_id = source.pk source_email = source.email_set.first() source_alias = source.alias_set.first() source_user = source.user + communitylist = CommunityList.objects.create(person=source, group=mars) + nomcom = NomComFactory() + position = PositionFactory(nomcom=nomcom) + nominee = NomineeFactory(nomcom=nomcom, person=mars.get_chair().person) + feedback = FeedbackFactory(person=source, author=source.email().address, nomcom=nomcom) + feedback.nominees.add(nominee) + nomination = NominationFactory(nominee=nominee, person=source, position=position, comments=feedback) merge_persons(request, source, target, file=StringIO()) self.assertTrue(source_email in target.email_set.all()) self.assertTrue(source_alias in target.alias_set.all()) + self.assertIn(communitylist, target.communitylist_set.all()) + self.assertIn(feedback, target.feedback_set.all()) + self.assertIn(nomination, target.nomination_set.all()) self.assertFalse(Person.objects.filter(id=source_id)) self.assertFalse(source_user.is_active) @@ -399,24 +430,6 @@ def test_merge_persons_reviewer_settings(self): rs = target.reviewersettings_set.first() self.assertEqual(rs.min_interval, 7) - def test_merge_users(self): - person = PersonFactory() - source = person.user - target = UserFactory() - mars = RoleFactory(name_id='chair',group__acronym='mars').group - communitylist = CommunityList.objects.create(user=source, group=mars) - nomcom = NomComFactory() - position = PositionFactory(nomcom=nomcom) - nominee = NomineeFactory(nomcom=nomcom, person=mars.get_chair().person) - feedback = FeedbackFactory(user=source, author=person.email().address, nomcom=nomcom) - feedback.nominees.add(nominee) - nomination = NominationFactory(nominee=nominee, user=source, position=position, comments=feedback) - - merge_users(source, target) - self.assertIn(communitylist, target.communitylist_set.all()) - self.assertIn(feedback, target.feedback_set.all()) - self.assertIn(nomination, target.nomination_set.all()) - def test_dots(self): noroles = PersonFactory() self.assertEqual(get_dots(noroles),[]) @@ -437,3 +450,40 @@ def test_dots(self): self.assertEqual(get_dots(ncmember),['nomcom']) ncchair = RoleFactory(group__acronym='nomcom2020',group__type_id='nomcom',name_id='chair').person self.assertEqual(get_dots(ncchair),['nomcom']) + + def test_send_merge_request(self): + empty_outbox() + message_count_before = Message.objects.count() + source = PersonFactory() + target = PersonFactory() + url = urlreverse('ietf.person.views.send_merge_request') + url = url + f'?source={source.pk}&target={target.pk}' + login_testing_unauthorized(self, 'secretary', url) + r = self.client.get(url) + initial = r.context['form'].initial + subject = 'Action requested: Merging possible duplicate IETF Datatracker accounts' + self.assertEqual(initial['to'], ', '.join([source.user.username, target.user.username])) + self.assertEqual(initial['subject'], subject) + self.assertEqual(initial['reply_to'], 'support@ietf.org') + self.assertEqual(r.status_code, 200) + r = self.client.post(url, data=initial) + self.assertEqual(r.status_code, 302) + self.assertEqual(len(outbox), 1) + self.assertIn(source.user.username, outbox[0]['To']) + message_count_after = Message.objects.count() + message = Message.objects.last() + self.assertEqual(message_count_after, message_count_before + 1) + self.assertIn(source.user.username, message.to) + + +class TaskTests(TestCase): + @mock.patch("ietf.person.tasks.log.log") + def test_purge_personal_api_key_events_task(self, mock_log): + now = timezone.now() + old_event = PersonApiKeyEventFactory(time=now - datetime.timedelta(days=1, minutes=1)) + young_event = PersonApiKeyEventFactory(time=now - datetime.timedelta(days=1, minutes=-1)) + purge_personal_api_key_events_task(keep_days=1) + self.assertFalse(PersonApiKeyEvent.objects.filter(pk=old_event.pk).exists()) + self.assertTrue(PersonApiKeyEvent.objects.filter(pk=young_event.pk).exists()) + self.assertTrue(mock_log.called) + self.assertIn("Deleted 1", mock_log.call_args[0][0]) diff --git a/ietf/person/urls.py b/ietf/person/urls.py index f37d8b46cf..f3eccd04b7 100644 --- a/ietf/person/urls.py +++ b/ietf/person/urls.py @@ -1,8 +1,12 @@ +# Copyright The IETF Trust 2009-2025, All Rights Reserved +# -*- coding: utf-8 -*- from ietf.person import views, ajax from ietf.utils.urls import url urlpatterns = [ - url(r'^merge/$', views.merge), + url(r'^merge/?$', views.merge), + url(r'^merge/submit/?$', views.merge_submit), + url(r'^merge/send_request/?$', views.send_merge_request), url(r'^search/(?P(person|email))/$', views.ajax_select2_search), url(r'^(?P[0-9]+)/email.json$', ajax.person_email_json), url(r'^(?P[^/]+)$', views.profile), diff --git a/ietf/person/utils.py b/ietf/person/utils.py index a7e3227349..5ed90591f9 100755 --- a/ietf/person/utils.py +++ b/ietf/person/utils.py @@ -3,27 +3,26 @@ import datetime -import os import pprint import sys -import syslog from django.contrib import admin from django.core.cache import cache from django.core.exceptions import ObjectDoesNotExist from django.db.models import Q +from django.http import Http404 import debug # pyflakes:ignore -from ietf.person.models import Person +from ietf.person.models import Person, Alias, Email +from ietf.utils import log from ietf.utils.mail import send_mail def merge_persons(request, source, target, file=sys.stdout, verbose=False): changes = [] # write log - syslog.openlog(str(os.path.basename(__file__)), syslog.LOG_PID, syslog.LOG_USER) - syslog.syslog("Merging person records {} => {}".format(source.pk,target.pk)) + log.log(f"Merging person records {source.pk} => {target.pk}") # handle primary emails for email in get_extra_primary(source,target): @@ -31,6 +30,20 @@ def merge_persons(request, source, target, file=sys.stdout, verbose=False): email.save() changes.append('EMAIL ACTION: {} no longer marked as primary'.format(email.address)) + # handle community list + for communitylist in source.communitylist_set.all(): + source.communitylist_set.remove(communitylist) + target.communitylist_set.add(communitylist) + + # handle feedback + for feedback in source.feedback_set.all(): + feedback.person = target + feedback.save() + # handle nominations + for nomination in source.nomination_set.all(): + nomination.person = target + nomination.save() + changes.append(handle_users(source, target)) reviewer_changes = handle_reviewer_settings(source, target) if reviewer_changes: @@ -47,7 +60,6 @@ def merge_persons(request, source, target, file=sys.stdout, verbose=False): # check for any remaining relationships and exit if more found objs = [source] -# request.user = User.objects.filter(is_superuser=True).first() deletable_objects = admin.utils.get_deleted_objects(objs, request, admin.site) deletable_objects_summary = deletable_objects[1] if len(deletable_objects_summary) > 1: # should only include one object (Person) @@ -104,8 +116,7 @@ def handle_users(source,target,check_only=False): if source.user and target.user: message = "DATATRACKER LOGIN ACTION: retaining login: {}, removing login: {}".format(target.user,source.user) if not check_only: - merge_users(source.user, target.user) - syslog.syslog('merge-person-records: deactivating user {}'.format(source.user.username)) + log.log(f"merge-person-records: deactivating user {source.user.username}") user = source.user source.user = None source.save() @@ -127,21 +138,6 @@ def move_related_objects(source, target, file, verbose=False): kwargs = { field_name:target } queryset.update(**kwargs) -def merge_users(source, target): - '''Move related objects from source user to target user''' - # handle community list - for communitylist in source.communitylist_set.all(): - source.communitylist_set.remove(communitylist) - target.communitylist_set.add(communitylist) - # handle feedback - for feedback in source.feedback_set.all(): - feedback.user = target - feedback.save() - # handle nominations - for nomination in source.nomination_set.all(): - nomination.user = target - nomination.save() - def dedupe_aliases(person): '''Check person for duplicate aliases and purge''' seen = [] @@ -199,11 +195,13 @@ def determine_merge_order(source,target): return source,target def get_active_balloters(ballot_type): - if (ballot_type.slug != "irsg-approve"): - active_balloters = get_active_ads() + if ballot_type.slug == 'irsg-approve': + return get_active_irsg() + elif ballot_type.slug == 'rsab-approve': + return get_active_rsab() else: - active_balloters = get_active_irsg() - return active_balloters + return get_active_ads() + def get_active_ads(): cache_key = "doc:active_ads" @@ -219,7 +217,15 @@ def get_active_irsg(): if not active_irsg_balloters: active_irsg_balloters = list(Person.objects.filter(role__group__acronym='irsg',role__name__in=['chair','member','atlarge']).distinct()) cache.set(cache_key, active_irsg_balloters) - return active_irsg_balloters + return active_irsg_balloters + +def get_active_rsab(): + cache_key = "doc:active_rsab_balloters" + active_rsab_balloters = cache.get(cache_key) + if not active_rsab_balloters: + active_rsab_balloters = list(Person.objects.filter(role__group__acronym='rsab', role__name="member").distinct()) + cache.set(cache_key, active_rsab_balloters) + return active_rsab_balloters def get_dots(person): roles = person.role_set.filter(group__state_id__in=('active','bof','proposed')) @@ -239,3 +245,17 @@ def get_dots(person): if roles.filter(group__acronym__startswith='nomcom', name_id__in=('chair','member')).exists(): dots.append('nomcom') return dots + +def lookup_persons(email_or_name): + aliases = Alias.objects.filter(name__iexact=email_or_name) + persons = set(a.person for a in aliases) + + if '@' in email_or_name: + emails = Email.objects.filter(address__iexact=email_or_name) + persons.update(e.person for e in emails) + + persons = [p for p in persons if p and p.id] + if not persons: + raise Http404 + persons.sort(key=lambda p: p.id) + return persons diff --git a/ietf/person/views.py b/ietf/person/views.py index 31bf43d82a..d0b5912431 100644 --- a/ietf/person/views.py +++ b/ietf/person/views.py @@ -1,23 +1,26 @@ -# Copyright The IETF Trust 2012-2020, All Rights Reserved +# Copyright The IETF Trust 2012-2025, All Rights Reserved # -*- coding: utf-8 -*- from io import StringIO, BytesIO from PIL import Image +from django.conf import settings from django.contrib import messages from django.db.models import Q from django.http import HttpResponse, Http404 -from django.shortcuts import render, get_object_or_404, redirect +from django.shortcuts import render, redirect +from django.template.loader import render_to_string from django.utils import timezone import debug # pyflakes:ignore from ietf.ietfauth.utils import role_required -from ietf.person.models import Email, Person, Alias +from ietf.person.models import Email, Person from ietf.person.fields import select2_id_name_json -from ietf.person.forms import MergeForm -from ietf.person.utils import handle_users, merge_persons +from ietf.person.forms import MergeForm, MergeRequestForm +from ietf.person.utils import handle_users, merge_persons, lookup_persons +from ietf.utils.mail import send_mail_text def ajax_select2_search(request, model_name): @@ -62,37 +65,22 @@ def ajax_select2_search(request, model_name): page = int(request.GET.get("p", 1)) - 1 except ValueError: page = 0 - - objs = objs.distinct()[page:page + 10] + PAGE_SIZE = 10 + first_item = page * PAGE_SIZE + objs = objs.distinct()[first_item:first_item + PAGE_SIZE] return HttpResponse(select2_id_name_json(objs), content_type='application/json') def profile(request, email_or_name): - aliases = Alias.objects.filter(name=email_or_name) - persons = set(a.person for a in aliases) - - if '@' in email_or_name: - emails = Email.objects.filter(address=email_or_name) - persons.update(e.person for e in emails) - - persons = [p for p in persons if p and p.id] - if not persons: - raise Http404 - persons.sort(key=lambda p: p.id) + persons = lookup_persons(email_or_name) return render(request, 'person/profile.html', {'persons': persons, 'today': timezone.now()}) def photo(request, email_or_name): - if '@' in email_or_name: - persons = [ get_object_or_404(Email, address=email_or_name).person, ] - else: - aliases = Alias.objects.filter(name=email_or_name) - persons = list(set([ a.person for a in aliases ])) - if not persons: - raise Http404("No such person") + persons = lookup_persons(email_or_name) if len(persons) > 1: - return HttpResponse(r"\r\n".join([p.email() for p in persons]), status=300) + raise Http404("No photo found") person = persons[0] if not person.photo: raise Http404("No photo found") @@ -100,29 +88,32 @@ def photo(request, email_or_name): if not size.isdigit(): return HttpResponse("Size must be integer", status=400) size = int(size) - img = Image.open(person.photo) - img = img.resize((size, img.height*size//img.width)) - bytes = BytesIO() - try: - img.save(bytes, format='JPEG') - return HttpResponse(bytes.getvalue(), content_type='image/jpg') - except OSError: - raise Http404 + with Image.open(person.photo) as img: + img = img.resize((size, img.height*size//img.width)) + bytes = BytesIO() + try: + img.save(bytes, format='JPEG') + return HttpResponse(bytes.getvalue(), content_type='image/jpg') + except OSError: + raise Http404 @role_required("Secretariat") def merge(request): form = MergeForm() - method = 'get' + return render(request, 'person/merge.html', {'form': form}) + + +@role_required("Secretariat") +def merge_submit(request): change_details = '' warn_messages = [] source = None target = None if request.method == "GET": - form = MergeForm() if request.GET: - form = MergeForm(request.GET) + form = MergeForm(request.GET, readonly=True) if form.is_valid(): source = form.cleaned_data.get('source') target = form.cleaned_data.get('target') @@ -131,12 +122,9 @@ def merge(request): if source.user.last_login and target.user.last_login and source.user.last_login > target.user.last_login: warn_messages.append('WARNING: The most recently used login is being deleted!') change_details = handle_users(source, target, check_only=True) - method = 'post' - else: - method = 'get' if request.method == "POST": - form = MergeForm(request.POST) + form = MergeForm(request.POST, readonly=True) if form.is_valid(): source = form.cleaned_data.get('source') source_id = source.id @@ -151,11 +139,72 @@ def merge(request): messages.error(request, output) return redirect('ietf.secr.rolodex.views.view', id=target.pk) - return render(request, 'person/merge.html', { + return render(request, 'person/merge_submit.html', { 'form': form, - 'method': method, 'change_details': change_details, 'source': source, 'target': target, 'warn_messages': warn_messages, }) + + +@role_required("Secretariat") +def send_merge_request(request): + if request.method == 'GET': + merge_form = MergeForm(request.GET) + if merge_form.is_valid(): + source = merge_form.cleaned_data['source'] + target = merge_form.cleaned_data['target'] + to = [] + if source.email(): + to.append(source.email().address) + if target.email(): + to.append(target.email().address) + if source.user: + source_account = source.user.username + else: + source_account = source.email() + if target.user: + target_account = target.user.username + else: + target_account = target.email() + sender_name = request.user.person.name + subject = 'Action requested: Merging possible duplicate IETF Datatracker accounts' + context = { + 'source_account': source_account, + 'target_account': target_account, + 'sender_name': sender_name, + } + body = render_to_string('person/merge_request_email.txt', context) + initial = { + 'to': ', '.join(to), + 'frm': settings.DEFAULT_FROM_EMAIL, + 'reply_to': 'support@ietf.org', + 'subject': subject, + 'body': body, + 'by': request.user.person.pk, + } + form = MergeRequestForm(initial=initial) + else: + messages.error(request, "Error requesting merge email: " + merge_form.errors.as_text()) + return redirect("ietf.person.views.merge") + + if request.method == 'POST': + form = MergeRequestForm(request.POST) + if form.is_valid(): + extra = {"Reply-To": form.cleaned_data.get("reply_to")} + send_mail_text( + request, + form.cleaned_data.get("to"), + form.cleaned_data.get("frm"), + form.cleaned_data.get("subject"), + form.cleaned_data.get("body"), + extra=extra, + ) + + messages.success(request, "The merge confirmation email was sent.") + return redirect("ietf.person.views.merge") + + return render(request, "person/send_merge_request.html", { + "form": form, + }) diff --git a/ietf/redirects/.gitignore b/ietf/redirects/.gitignore deleted file mode 100644 index c7013ced9d..0000000000 --- a/ietf/redirects/.gitignore +++ /dev/null @@ -1,2 +0,0 @@ -/*.pyc -/settings_local.py diff --git a/ietf/redirects/fixtures/.gitignore b/ietf/redirects/fixtures/.gitignore deleted file mode 100644 index c7013ced9d..0000000000 --- a/ietf/redirects/fixtures/.gitignore +++ /dev/null @@ -1,2 +0,0 @@ -/*.pyc -/settings_local.py diff --git a/ietf/redirects/migrations/0001_initial.py b/ietf/redirects/migrations/0001_initial.py new file mode 100644 index 0000000000..7a28072a32 --- /dev/null +++ b/ietf/redirects/migrations/0001_initial.py @@ -0,0 +1,51 @@ +# Generated by Django 2.2.28 on 2023-03-20 19:22 + +from typing import List, Tuple +from django.db import migrations, models +import django.db.models.deletion +import ietf.utils.models + + +class Migration(migrations.Migration): + + initial = True + + dependencies: List[Tuple[str, str]] = [ + ] + + operations = [ + migrations.CreateModel( + name='Redirect', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('cgi', models.CharField(blank=True, max_length=50, unique=True)), + ('url', models.CharField(max_length=255)), + ('rest', models.CharField(blank=True, max_length=100)), + ('remove', models.CharField(blank=True, max_length=50)), + ], + ), + migrations.CreateModel( + name='Suffix', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('rest', models.CharField(blank=True, max_length=100)), + ('remove', models.CharField(blank=True, max_length=50)), + ], + options={ + 'verbose_name_plural': 'Suffixes', + }, + ), + migrations.CreateModel( + name='Command', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('command', models.CharField(max_length=50)), + ('url', models.CharField(blank=True, max_length=50)), + ('script', ietf.utils.models.ForeignKey(editable=False, on_delete=django.db.models.deletion.CASCADE, related_name='commands', to='redirects.Redirect')), + ('suffix', ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='redirects.Suffix')), + ], + options={ + 'unique_together': {('script', 'command')}, + }, + ), + ] diff --git a/ietf/secr/groups/__init__.py b/ietf/redirects/migrations/__init__.py similarity index 100% rename from ietf/secr/groups/__init__.py rename to ietf/redirects/migrations/__init__.py diff --git a/ietf/redirects/views.py b/ietf/redirects/views.py index af8cbf5558..af4e898261 100644 --- a/ietf/redirects/views.py +++ b/ietf/redirects/views.py @@ -48,7 +48,7 @@ def redirect(request, path="", script=""): pass # it's ok, request didn't have 'command'. except: pass # strange exception like the one described in - # https://trac.ietf.org/trac/ietfdb/ticket/179? + # https://github.com/ietf-tools/datatracker/issues/179 # just ignore the command string. if cmd is not None: remove_args.append('command') diff --git a/ietf/review/factories.py b/ietf/review/factories.py index d6780fad80..158251317e 100644 --- a/ietf/review/factories.py +++ b/ietf/review/factories.py @@ -11,6 +11,7 @@ class ReviewTeamSettingsFactory(factory.django.DjangoModelFactory): class Meta: model = ReviewTeamSettings + skip_postgeneration_save = True group = factory.SubFactory('ietf.group.factories.GroupFactory',type_id='review') reviewer_queue_policy_id = 'RotateAlphabetically' diff --git a/ietf/review/mailarch.py b/ietf/review/mailarch.py index 6ef5909a1c..61abc83aa5 100644 --- a/ietf/review/mailarch.py +++ b/ietf/review/mailarch.py @@ -6,25 +6,18 @@ # mailarchive.ietf.org import base64 -import contextlib import datetime import email.utils import hashlib -import mailbox -import tarfile -import tempfile - -from urllib.parse import urlencode -from urllib.request import urlopen +import requests import debug # pyflakes:ignore -from pyquery import PyQuery from django.conf import settings from django.utils.encoding import force_bytes, force_str -from ietf.utils.mail import get_payload_text +from ietf.utils.log import log from ietf.utils.timezone import date_today @@ -43,7 +36,7 @@ def hash_list_message_id(list_name, msgid): sha.update(force_bytes(list_name)) return force_str(base64.urlsafe_b64encode(sha.digest()).rstrip(b"=")) -def construct_query_urls(doc, team, query=None): +def construct_query_data(doc, team, query=None): list_name = list_name_from_email(team.list_email) if not list_name: return None @@ -51,82 +44,48 @@ def construct_query_urls(doc, team, query=None): if not query: query = doc.name - encoded_query = "?" + urlencode({ - "qdr": "c", # custom time frame - "start_date": (date_today() - datetime.timedelta(days=180)).isoformat(), - "email_list": list_name, - "q": "subject:({})".format(query), - "as": "1", # this is an advanced search - }) - - return { - "query": query, - "query_url": settings.MAILING_LIST_ARCHIVE_URL + "/arch/search/" + encoded_query, - "query_data_url": settings.MAILING_LIST_ARCHIVE_URL + "/arch/export/mbox/" + encoded_query, + query_data = { + 'start_date': (date_today() - datetime.timedelta(days=180)).isoformat(), + 'email_list': list_name, + 'query_value': query, + 'query': f'subject:({query})', + 'limit': '30', } + return query_data def construct_message_url(list_name, msgid): return "{}/arch/msg/{}/{}".format(settings.MAILING_LIST_ARCHIVE_URL, list_name, hash_list_message_id(list_name, msgid)) -def retrieve_messages_from_mbox(mbox_fileobj): - """Return selected content in message from mbox from mailarch.""" - res = [] - with tempfile.NamedTemporaryFile(suffix=".mbox") as mbox_file: - # mailbox.mbox needs a path, so we need to put the contents - # into a file - mbox_data = mbox_fileobj.read() - mbox_file.write(mbox_data) - mbox_file.flush() - - mbox = mailbox.mbox(mbox_file.name, create=False) - for msg in mbox: - content = "" - - for part in msg.walk(): - if part.get_content_type() == "text/plain": - charset = part.get_content_charset() or "utf-8" - content += get_payload_text(part, default_charset=charset) - - # parse a couple of things for the front end - utcdate = None - d = email.utils.parsedate_tz(msg["Date"]) - if d: - utcdate = datetime.datetime.fromtimestamp(email.utils.mktime_tz(d), datetime.timezone.utc) - - res.append({ - "from": msg["From"], - "splitfrom": email.utils.parseaddr(msg["From"]), - "subject": msg["Subject"], - "content": content.replace("\r\n", "\n").replace("\r", "\n").strip("\n"), - "message_id": email.utils.unquote(msg["Message-ID"].strip()), - "url": email.utils.unquote(msg["Archived-At"].strip()), - "date": msg["Date"], - "utcdate": (utcdate.date().isoformat(), utcdate.time().isoformat()) if utcdate else ("", ""), - }) - - return res - -def retrieve_messages(query_data_url): +def retrieve_messages(query_data): """Retrieve and return selected content from mailarch.""" - res = [] - - # This has not been rewritten to use requests.get() because get() does - # not handle file URLs out of the box, which we need for tesing - with contextlib.closing(urlopen(query_data_url, timeout=15)) as fileobj: - content_type = fileobj.info()["Content-type"] - if not content_type.startswith("application/x-tar"): - if content_type.startswith("text/html"): - r = fileobj.read(20000) - q = PyQuery(r) - div = q('div[class~="no-results"]') - if div: - raise KeyError("No results: %s -> %s" % (query_data_url, div.text(), )) - raise Exception("Export failed - this usually means no matches were found") - - with tarfile.open(fileobj=fileobj, mode='r|*') as tar: - for entry in tar: - if entry.isfile(): - mbox_fileobj = tar.extractfile(entry) - res.extend(retrieve_messages_from_mbox(mbox_fileobj)) - - return res + + headers = {'X-Api-Key': settings.MAILING_LIST_ARCHIVE_API_KEY} + try: + response = requests.post( + settings.MAILING_LIST_ARCHIVE_SEARCH_URL, + headers=headers, + json=query_data, + timeout=settings.DEFAULT_REQUESTS_TIMEOUT) + except requests.Timeout as exc: + log(f'POST request failed for [{query_data["url"]}]: {exc}') + raise RuntimeError(f'Timeout retrieving [{query_data["url"]}]') from exc + + results = [] + jresponse = response.json() + if 'results' not in jresponse or len(jresponse['results']) == 0: + raise KeyError(f'No results: {query_data["query"]}') + for msg in jresponse['results']: + # datetime is already UTC + dt = datetime.datetime.fromisoformat(msg['date']) + dt_utc = dt.replace(tzinfo=datetime.timezone.utc) + results.append({ + "from": msg["from"], + "splitfrom": email.utils.parseaddr(msg["from"]), + "subject": msg["subject"], + "content": msg["content"].replace("\r\n", "\n").replace("\r", "\n").strip("\n"), + "message_id": msg["message_id"], + "url": msg["url"], + "utcdate": (dt_utc.date().isoformat(), dt_utc.time().isoformat()), + }) + + return results diff --git a/ietf/review/migrations/0001_initial.py b/ietf/review/migrations/0001_initial.py index 5e7fde9ec7..4fc32caeef 100644 --- a/ietf/review/migrations/0001_initial.py +++ b/ietf/review/migrations/0001_initial.py @@ -1,14 +1,14 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-20 10:52 +# Generated by Django 2.2.28 on 2023-03-20 19:22 - -import datetime +from django.conf import settings from django.db import migrations, models import django.db.models.deletion +import django.utils.timezone import ietf.review.models import ietf.utils.models +import ietf.utils.timezone import ietf.utils.validators +import simple_history.models class Migration(migrations.Migration): @@ -16,109 +16,238 @@ class Migration(migrations.Migration): initial = True dependencies = [ - ('group', '0001_initial'), ('name', '0001_initial'), ('person', '0001_initial'), + migrations.swappable_dependency(settings.AUTH_USER_MODEL), + ('group', '0001_initial'), ('doc', '0001_initial'), ] operations = [ migrations.CreateModel( - name='NextReviewerInTeam', + name='UnavailablePeriod', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('next_reviewer', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), - ('team', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='group.Group')), + ('start_date', models.DateField(default=ietf.utils.timezone.date_today, help_text="Choose the start date so that you can still do a review if it's assigned just before the start date - this usually means you should mark yourself unavailable for assignment some time before you are actually away. The default is today.", null=True)), + ('end_date', models.DateField(blank=True, help_text='Leaving the end date blank means that the period continues indefinitely. You can end it later.', null=True)), + ('availability', models.CharField(choices=[('canfinish', 'Can do follow-ups'), ('unavailable', 'Completely unavailable')], max_length=30)), + ('reason', models.TextField(blank=True, default='', help_text="Provide (for the secretary's benefit) the reason why the review is unavailable", max_length=2048, verbose_name='Reason why reviewer is unavailable (Optional)')), + ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), + ('team', ietf.utils.models.ForeignKey(limit_choices_to=models.Q(_negated=True, reviewteamsettings=None), on_delete=django.db.models.deletion.CASCADE, to='group.Group')), + ], + ), + migrations.CreateModel( + name='ReviewWish', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('time', models.DateTimeField(default=django.utils.timezone.now)), + ('doc', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document')), + ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), + ('team', ietf.utils.models.ForeignKey(limit_choices_to=models.Q(_negated=True, reviewteamsettings=None), on_delete=django.db.models.deletion.CASCADE, to='group.Group')), ], options={ - 'verbose_name': 'next reviewer in team setting', - 'verbose_name_plural': 'next reviewer in team settings', + 'verbose_name_plural': 'review wishes', }, ), migrations.CreateModel( - name='ReviewerSettings', + name='ReviewTeamSettings', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('min_interval', models.IntegerField(blank=True, choices=[(7, 'Once per week'), (14, 'Once per fortnight'), (30, 'Once per month'), (61, 'Once per two months'), (91, 'Once per quarter')], null=True, verbose_name='Can review at most')), - ('filter_re', models.CharField(blank=True, help_text='Draft names matching this regular expression should not be assigned', max_length=255, validators=[ietf.utils.validators.RegexStringValidator()], verbose_name='Filter regexp')), - ('skip_next', models.IntegerField(default=0, verbose_name='Skip next assignments')), - ('remind_days_before_deadline', models.IntegerField(blank=True, help_text="To get an email reminder in case you forget to do an assigned review, enter the number of days before review deadline you want to receive it. Clear the field if you don't want a reminder.", null=True)), - ('expertise', models.TextField(blank=True, default='', help_text="Describe the reviewer's expertise in this team's area", max_length=2048, verbose_name="Reviewer's expertise in this team's area")), + ('autosuggest', models.BooleanField(default=True, verbose_name='Automatically suggest possible review requests')), + ('secr_mail_alias', models.CharField(blank=True, help_text='Email alias for all of the review team secretaries', max_length=255, verbose_name='Email alias for all of the review team secretaries')), + ('remind_days_unconfirmed_assignments', models.PositiveIntegerField(blank=True, help_text="To send a periodic email reminder to reviewers of review assignments they have neither accepted nor rejected, enter the number of days between these reminders. Clear the field if you don't want these reminders to be sent.", null=True, verbose_name='Periodic reminder of not yet accepted or rejected review assignments to reviewer every X days')), + ('group', ietf.utils.models.OneToOneField(on_delete=django.db.models.deletion.CASCADE, to='group.Group')), + ('notify_ad_when', models.ManyToManyField(blank=True, related_name='reviewteamsettings_notify_ad_set', to='name.ReviewResultName')), + ('review_results', models.ManyToManyField(default=ietf.review.models.get_default_review_results, related_name='reviewteamsettings_review_results_set', to='name.ReviewResultName')), + ('review_types', models.ManyToManyField(default=ietf.review.models.get_default_review_types, to='name.ReviewTypeName')), + ('reviewer_queue_policy', models.ForeignKey(default='RotateAlphabetically', on_delete=django.db.models.deletion.PROTECT, to='name.ReviewerQueuePolicyName')), + ], + options={ + 'verbose_name': 'Review team settings', + 'verbose_name_plural': 'Review team settings', + }, + ), + migrations.CreateModel( + name='ReviewSecretarySettings', + fields=[ + ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), + ('remind_days_before_deadline', models.IntegerField(blank=True, help_text="To get an email reminder in case a reviewer forgets to do an assigned review, enter the number of days before review deadline you want to receive it. Clear the field if you don't want a reminder.", null=True)), + ('max_items_to_show_in_reviewer_list', models.IntegerField(blank=True, help_text='Maximum number of completed items to show for one reviewer in the reviewer list view, the list is also filtered by the days to show in reviews list setting.', null=True)), + ('days_to_show_in_reviewer_list', models.IntegerField(blank=True, help_text='Maximum number of days to show in reviewer list for completed items.', null=True)), ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), - ('team', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='group.Group')), + ('team', ietf.utils.models.ForeignKey(limit_choices_to=models.Q(_negated=True, reviewteamsettings=None), on_delete=django.db.models.deletion.CASCADE, to='group.Group')), ], options={ - 'verbose_name_plural': 'reviewer settings', + 'verbose_name_plural': 'review secretary settings', }, ), migrations.CreateModel( name='ReviewRequest', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('old_id', models.IntegerField(blank=True, help_text='ID in previous review system', null=True)), - ('time', models.DateTimeField(default=datetime.datetime.now)), + ('time', models.DateTimeField(default=django.utils.timezone.now)), ('deadline', models.DateField()), ('requested_rev', models.CharField(blank=True, help_text='Fill in if a specific revision is to be reviewed, e.g. 02', max_length=16, verbose_name='requested revision')), ('comment', models.TextField(blank=True, default='', help_text='Provide any additional information to show to the review team secretary and reviewer', max_length=2048, verbose_name="Requester's comments and instructions")), - ('reviewed_rev', models.CharField(blank=True, max_length=16, verbose_name='reviewed revision')), ('doc', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='reviewrequest_set', to='doc.Document')), ('requested_by', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), - ('result', ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.ReviewResultName')), - ('review', ietf.utils.models.OneToOneField(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.Document')), - ('reviewer', ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='person.Email')), ('state', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ReviewRequestStateName')), - ('team', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='group.Group')), + ('team', ietf.utils.models.ForeignKey(limit_choices_to=models.Q(_negated=True, reviewteamsettings=None), on_delete=django.db.models.deletion.CASCADE, to='group.Group')), ('type', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ReviewTypeName')), ], ), migrations.CreateModel( - name='ReviewSecretarySettings', + name='ReviewerSettings', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('remind_days_before_deadline', models.IntegerField(blank=True, help_text="To get an email reminder in case a reviewer forgets to do an assigned review, enter the number of days before review deadline you want to receive it. Clear the field if you don't want a reminder.", null=True)), + ('min_interval', models.IntegerField(blank=True, choices=[(7, 'Once per week'), (14, 'Once per fortnight'), (30, 'Once per month'), (61, 'Once per two months'), (91, 'Once per quarter')], null=True, verbose_name='Can review at most')), + ('filter_re', models.CharField(blank=True, help_text='Internet-Draft names matching this regular expression should not be assigned', max_length=255, validators=[ietf.utils.validators.RegexStringValidator()], verbose_name='Filter regexp')), + ('skip_next', models.IntegerField(default=0, verbose_name='Skip next assignments')), + ('remind_days_before_deadline', models.IntegerField(blank=True, help_text="To get an email reminder in case you forget to do an assigned review, enter the number of days before review deadline you want to receive it. Clear the field if you don't want this reminder.", null=True)), + ('remind_days_open_reviews', models.PositiveIntegerField(blank=True, help_text="To get a periodic email reminder of all your open reviews, enter the number of days between these reminders. Clear the field if you don't want these reminders.", null=True, verbose_name='Periodic reminder of open reviews every X days')), + ('request_assignment_next', models.BooleanField(default=False, help_text='If you would like to be assigned to a review as soon as possible, select this option. It is automatically reset once you receive any assignment.', verbose_name='Select me next for an assignment')), + ('expertise', models.TextField(blank=True, default='', help_text="Describe the reviewer's expertise in this team's area", max_length=2048, verbose_name="Reviewer's expertise in this team's area")), ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), - ('team', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='group.Group')), + ('team', ietf.utils.models.ForeignKey(limit_choices_to=models.Q(_negated=True, reviewteamsettings=None), on_delete=django.db.models.deletion.CASCADE, to='group.Group')), ], options={ - 'verbose_name_plural': 'review secretary settings', + 'verbose_name_plural': 'reviewer settings', }, ), migrations.CreateModel( - name='ReviewTeamSettings', + name='ReviewAssignment', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('autosuggest', models.BooleanField(default=True, verbose_name='Automatically suggest possible review requests')), - ('group', ietf.utils.models.OneToOneField(on_delete=django.db.models.deletion.CASCADE, to='group.Group')), - ('review_results', models.ManyToManyField(default=ietf.review.models.get_default_review_results, to='name.ReviewResultName')), - ('review_types', models.ManyToManyField(default=ietf.review.models.get_default_review_types, to='name.ReviewTypeName')), + ('assigned_on', models.DateTimeField(blank=True, null=True)), + ('completed_on', models.DateTimeField(blank=True, null=True)), + ('reviewed_rev', models.CharField(blank=True, max_length=16, verbose_name='reviewed revision')), + ('mailarch_url', models.URLField(blank=True, null=True)), + ('result', ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.ReviewResultName')), + ('review', ietf.utils.models.OneToOneField(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.Document')), + ('review_request', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='review.ReviewRequest')), + ('reviewer', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Email')), + ('state', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ReviewAssignmentStateName')), ], - options={ - 'verbose_name': 'Review team settings', - 'verbose_name_plural': 'Review team settings', - }, ), migrations.CreateModel( - name='ReviewWish', + name='NextReviewerInTeam', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('time', models.DateTimeField(default=datetime.datetime.now)), - ('doc', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document')), - ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), - ('team', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='group.Group')), + ('next_reviewer', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), + ('team', ietf.utils.models.ForeignKey(limit_choices_to=models.Q(_negated=True, reviewteamsettings=None), on_delete=django.db.models.deletion.CASCADE, to='group.Group')), ], options={ - 'verbose_name_plural': 'review wishes', + 'verbose_name': 'next reviewer in team setting', + 'verbose_name_plural': 'next reviewer in team settings', }, ), migrations.CreateModel( - name='UnavailablePeriod', + name='HistoricalUnavailablePeriod', fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('start_date', models.DateField(default=datetime.date.today, help_text="Choose the start date so that you can still do a review if it's assigned just before the start date - this usually means you should mark yourself unavailable for assignment some time before you are actually away.", null=True)), + ('id', models.IntegerField(auto_created=True, blank=True, db_index=True, verbose_name='ID')), + ('history_change_reason', models.TextField(null=True)), + ('start_date', models.DateField(default=ietf.utils.timezone.date_today, help_text="Choose the start date so that you can still do a review if it's assigned just before the start date - this usually means you should mark yourself unavailable for assignment some time before you are actually away. The default is today.", null=True)), ('end_date', models.DateField(blank=True, help_text='Leaving the end date blank means that the period continues indefinitely. You can end it later.', null=True)), ('availability', models.CharField(choices=[('canfinish', 'Can do follow-ups'), ('unavailable', 'Completely unavailable')], max_length=30)), - ('person', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Person')), - ('team', ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='group.Group')), + ('reason', models.TextField(blank=True, default='', help_text="Provide (for the secretary's benefit) the reason why the review is unavailable", max_length=2048, verbose_name='Reason why reviewer is unavailable (Optional)')), + ('history_id', models.AutoField(primary_key=True, serialize=False)), + ('history_date', models.DateTimeField(db_index=True)), + ('history_type', models.CharField(choices=[('+', 'Created'), ('~', 'Changed'), ('-', 'Deleted')], max_length=1)), + ('history_user', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='+', to=settings.AUTH_USER_MODEL)), + ('person', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='person.Person')), + ('team', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, limit_choices_to=models.Q(_negated=True, reviewteamsettings=None), null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='group.Group')), ], + options={ + 'verbose_name': 'historical unavailable period', + 'verbose_name_plural': 'historical unavailable periods', + 'ordering': ('-history_date', '-history_id'), + 'get_latest_by': ('history_date', 'history_id'), + }, + bases=(simple_history.models.HistoricalChanges, models.Model), + ), + migrations.CreateModel( + name='HistoricalReviewRequest', + fields=[ + ('id', models.IntegerField(auto_created=True, blank=True, db_index=True, verbose_name='ID')), + ('history_change_reason', models.TextField(null=True)), + ('time', models.DateTimeField(default=django.utils.timezone.now)), + ('deadline', models.DateField()), + ('requested_rev', models.CharField(blank=True, help_text='Fill in if a specific revision is to be reviewed, e.g. 02', max_length=16, verbose_name='requested revision')), + ('comment', models.TextField(blank=True, default='', help_text='Provide any additional information to show to the review team secretary and reviewer', max_length=2048, verbose_name="Requester's comments and instructions")), + ('history_id', models.AutoField(primary_key=True, serialize=False)), + ('history_date', models.DateTimeField(db_index=True)), + ('history_type', models.CharField(choices=[('+', 'Created'), ('~', 'Changed'), ('-', 'Deleted')], max_length=1)), + ('doc', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='doc.Document')), + ('history_user', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='+', to=settings.AUTH_USER_MODEL)), + ('requested_by', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='person.Person')), + ('state', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='name.ReviewRequestStateName')), + ('team', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, limit_choices_to=models.Q(_negated=True, reviewteamsettings=None), null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='group.Group')), + ('type', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='name.ReviewTypeName')), + ], + options={ + 'verbose_name': 'historical review request', + 'verbose_name_plural': 'historical review requests', + 'ordering': ('-history_date', '-history_id'), + 'get_latest_by': ('history_date', 'history_id'), + }, + bases=(simple_history.models.HistoricalChanges, models.Model), + ), + migrations.CreateModel( + name='HistoricalReviewerSettings', + fields=[ + ('id', models.IntegerField(auto_created=True, blank=True, db_index=True, verbose_name='ID')), + ('history_change_reason', models.TextField(null=True)), + ('min_interval', models.IntegerField(blank=True, choices=[(7, 'Once per week'), (14, 'Once per fortnight'), (30, 'Once per month'), (61, 'Once per two months'), (91, 'Once per quarter')], null=True, verbose_name='Can review at most')), + ('filter_re', models.CharField(blank=True, help_text='Internet-Draft names matching this regular expression should not be assigned', max_length=255, validators=[ietf.utils.validators.RegexStringValidator()], verbose_name='Filter regexp')), + ('skip_next', models.IntegerField(default=0, verbose_name='Skip next assignments')), + ('remind_days_before_deadline', models.IntegerField(blank=True, help_text="To get an email reminder in case you forget to do an assigned review, enter the number of days before review deadline you want to receive it. Clear the field if you don't want this reminder.", null=True)), + ('remind_days_open_reviews', models.PositiveIntegerField(blank=True, help_text="To get a periodic email reminder of all your open reviews, enter the number of days between these reminders. Clear the field if you don't want these reminders.", null=True, verbose_name='Periodic reminder of open reviews every X days')), + ('request_assignment_next', models.BooleanField(default=False, help_text='If you would like to be assigned to a review as soon as possible, select this option. It is automatically reset once you receive any assignment.', verbose_name='Select me next for an assignment')), + ('expertise', models.TextField(blank=True, default='', help_text="Describe the reviewer's expertise in this team's area", max_length=2048, verbose_name="Reviewer's expertise in this team's area")), + ('history_id', models.AutoField(primary_key=True, serialize=False)), + ('history_date', models.DateTimeField(db_index=True)), + ('history_type', models.CharField(choices=[('+', 'Created'), ('~', 'Changed'), ('-', 'Deleted')], max_length=1)), + ('history_user', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='+', to=settings.AUTH_USER_MODEL)), + ('person', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='person.Person')), + ('team', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, limit_choices_to=models.Q(_negated=True, reviewteamsettings=None), null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='group.Group')), + ], + options={ + 'verbose_name': 'historical reviewer settings', + 'verbose_name_plural': 'historical reviewer settings', + 'ordering': ('-history_date', '-history_id'), + 'get_latest_by': ('history_date', 'history_id'), + }, + bases=(simple_history.models.HistoricalChanges, models.Model), + ), + migrations.CreateModel( + name='HistoricalReviewAssignment', + fields=[ + ('id', models.IntegerField(auto_created=True, blank=True, db_index=True, verbose_name='ID')), + ('history_change_reason', models.TextField(null=True)), + ('assigned_on', models.DateTimeField(blank=True, null=True)), + ('completed_on', models.DateTimeField(blank=True, null=True)), + ('reviewed_rev', models.CharField(blank=True, max_length=16, verbose_name='reviewed revision')), + ('mailarch_url', models.URLField(blank=True, null=True)), + ('history_id', models.AutoField(primary_key=True, serialize=False)), + ('history_date', models.DateTimeField(db_index=True)), + ('history_type', models.CharField(choices=[('+', 'Created'), ('~', 'Changed'), ('-', 'Deleted')], max_length=1)), + ('history_user', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='+', to=settings.AUTH_USER_MODEL)), + ('result', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='name.ReviewResultName')), + ('review', models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='doc.Document')), + ('review_request', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='review.ReviewRequest')), + ('reviewer', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='person.Email')), + ('state', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='name.ReviewAssignmentStateName')), + ], + options={ + 'verbose_name': 'historical review assignment', + 'verbose_name_plural': 'historical review assignments', + 'ordering': ('-history_date', '-history_id'), + 'get_latest_by': ('history_date', 'history_id'), + }, + bases=(simple_history.models.HistoricalChanges, models.Model), + ), + migrations.AddConstraint( + model_name='reviewersettings', + constraint=models.UniqueConstraint(fields=('team', 'person'), name='unique_reviewer_settings_per_team_person'), ), ] diff --git a/ietf/review/migrations/0002_reviewteamsettings_allow_reviewer_to_reject_after_deadline.py b/ietf/review/migrations/0002_reviewteamsettings_allow_reviewer_to_reject_after_deadline.py new file mode 100644 index 0000000000..08990fbcc0 --- /dev/null +++ b/ietf/review/migrations/0002_reviewteamsettings_allow_reviewer_to_reject_after_deadline.py @@ -0,0 +1,18 @@ +# Generated by Django 2.2.28 on 2023-03-25 06:34 + +from django.db import migrations, models + + +class Migration(migrations.Migration): + + dependencies = [ + ('review', '0001_initial'), + ] + + operations = [ + migrations.AddField( + model_name='reviewteamsettings', + name='allow_reviewer_to_reject_after_deadline', + field=models.BooleanField(default=False, verbose_name='Allow reviewer to reject request after deadline.'), + ), + ] diff --git a/ietf/review/migrations/0002_unavailableperiod_reason.py b/ietf/review/migrations/0002_unavailableperiod_reason.py deleted file mode 100644 index e4c81ad7cd..0000000000 --- a/ietf/review/migrations/0002_unavailableperiod_reason.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.14 on 2018-07-23 15:11 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0001_initial'), - ] - - operations = [ - migrations.AddField( - model_name='unavailableperiod', - name='reason', - field=models.TextField(blank=True, default='', help_text="Provide (for the secretary's benefit) the reason why the review is unavailable", max_length=2048, verbose_name='Reason why reviewer is unavailable (Optional)'), - ), - ] diff --git a/ietf/review/migrations/0003_add_notify_ad_when.py b/ietf/review/migrations/0003_add_notify_ad_when.py deleted file mode 100644 index 78e0d79595..0000000000 --- a/ietf/review/migrations/0003_add_notify_ad_when.py +++ /dev/null @@ -1,28 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.16 on 2018-11-02 10:10 - - -from django.db import migrations, models -import ietf.review.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0004_add_prefix_to_doctypenames'), - ('review', '0002_unavailableperiod_reason'), - ] - - operations = [ - migrations.AddField( - model_name='reviewteamsettings', - name='notify_ad_when', - field=models.ManyToManyField(related_name='reviewteamsettings_notify_ad_set', to='name.ReviewResultName'), - ), - migrations.AlterField( - model_name='reviewteamsettings', - name='review_results', - field=models.ManyToManyField(default=ietf.review.models.get_default_review_results, related_name='reviewteamsettings_review_results_set', to='name.ReviewResultName'), - ), - ] diff --git a/ietf/review/migrations/0004_reviewteamsettings_secr_mail_alias.py b/ietf/review/migrations/0004_reviewteamsettings_secr_mail_alias.py deleted file mode 100644 index 96fb0ba1cf..0000000000 --- a/ietf/review/migrations/0004_reviewteamsettings_secr_mail_alias.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.16 on 2018-11-03 03:10 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0003_add_notify_ad_when'), - ] - - operations = [ - migrations.AddField( - model_name='reviewteamsettings', - name='secr_mail_alias', - field=models.CharField(blank=True, help_text='Email alias for all of the review team secretaries', max_length=255, verbose_name='Email alias for all of the review team secretaries'), - ), - ] diff --git a/ietf/review/migrations/0005_set_secdir_notify_ad_when.py b/ietf/review/migrations/0005_set_secdir_notify_ad_when.py deleted file mode 100644 index 66b768a50d..0000000000 --- a/ietf/review/migrations/0005_set_secdir_notify_ad_when.py +++ /dev/null @@ -1,24 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.16 on 2018-11-02 10:20 - - -from django.db import migrations - -def forward(apps, schema_editor): - ReviewTeamSettings = apps.get_model('review','ReviewTeamSettings') - ReviewTeamSettings.objects.get(group__acronym='secdir').notify_ad_when.set(['serious-issues', 'issues', 'not-ready']) - -def reverse(apps, schema_editor): - ReviewTeamSettings = apps.get_model('review','ReviewTeamSettings') - ReviewTeamSettings.objects.get(group__acronym='secdir').notify_ad_when.set([]) - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0004_reviewteamsettings_secr_mail_alias'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/review/migrations/0006_historicalreviewersettings.py b/ietf/review/migrations/0006_historicalreviewersettings.py deleted file mode 100644 index 4129ab409f..0000000000 --- a/ietf/review/migrations/0006_historicalreviewersettings.py +++ /dev/null @@ -1,48 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.16 on 2018-11-09 08:31 - - -from django.conf import settings -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models -import ietf.utils.validators -import simple_history.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('group', '0003_groupfeatures_data'), - migrations.swappable_dependency(settings.AUTH_USER_MODEL), - ('person', '0008_auto_20181014_1448'), - ('review', '0005_set_secdir_notify_ad_when'), - ] - - operations = [ - migrations.CreateModel( - name='HistoricalReviewerSettings', - fields=[ - ('id', models.IntegerField(auto_created=True, blank=True, db_index=True, verbose_name='ID')), - ('min_interval', models.IntegerField(blank=True, choices=[(7, 'Once per week'), (14, 'Once per fortnight'), (30, 'Once per month'), (61, 'Once per two months'), (91, 'Once per quarter')], null=True, verbose_name='Can review at most')), - ('filter_re', models.CharField(blank=True, help_text='Draft names matching this regular expression should not be assigned', max_length=255, validators=[ietf.utils.validators.RegexStringValidator()], verbose_name='Filter regexp')), - ('skip_next', models.IntegerField(default=0, verbose_name='Skip next assignments')), - ('remind_days_before_deadline', models.IntegerField(blank=True, help_text="To get an email reminder in case you forget to do an assigned review, enter the number of days before review deadline you want to receive it. Clear the field if you don't want a reminder.", null=True)), - ('expertise', models.TextField(blank=True, default='', help_text="Describe the reviewer's expertise in this team's area", max_length=2048, verbose_name="Reviewer's expertise in this team's area")), - ('history_id', models.AutoField(primary_key=True, serialize=False)), - ('history_change_reason', models.CharField(max_length=100, null=True)), - ('history_date', models.DateTimeField()), - ('history_type', models.CharField(choices=[('+', 'Created'), ('~', 'Changed'), ('-', 'Deleted')], max_length=1)), - ('history_user', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='+', to=settings.AUTH_USER_MODEL)), - ('person', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='person.Person')), - ('team', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='group.Group')), - ], - options={ - 'ordering': ('-history_date', '-history_id'), - 'get_latest_by': 'history_date', - 'verbose_name': 'historical reviewer settings', - }, - bases=(simple_history.models.HistoricalChanges, models.Model), - ), - ] diff --git a/ietf/review/migrations/0007_allow_notify_ad_when_to_be_blank.py b/ietf/review/migrations/0007_allow_notify_ad_when_to_be_blank.py deleted file mode 100644 index 11da10b36d..0000000000 --- a/ietf/review/migrations/0007_allow_notify_ad_when_to_be_blank.py +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.17 on 2018-12-06 13:16 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0006_historicalreviewersettings'), - ] - - operations = [ - migrations.AlterField( - model_name='reviewteamsettings', - name='notify_ad_when', - field=models.ManyToManyField(blank=True, related_name='reviewteamsettings_notify_ad_set', to='name.ReviewResultName'), - ), - ] diff --git a/ietf/review/migrations/0008_remove_reviewrequest_old_id.py b/ietf/review/migrations/0008_remove_reviewrequest_old_id.py deleted file mode 100644 index 7f3e0b24ca..0000000000 --- a/ietf/review/migrations/0008_remove_reviewrequest_old_id.py +++ /dev/null @@ -1,20 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.17 on 2019-01-03 12:34 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0007_allow_notify_ad_when_to_be_blank'), - ] - - operations = [ - migrations.RemoveField( - model_name='reviewrequest', - name='old_id', - ), - ] diff --git a/ietf/review/migrations/0009_refactor_review_request.py b/ietf/review/migrations/0009_refactor_review_request.py deleted file mode 100644 index d3b22e138b..0000000000 --- a/ietf/review/migrations/0009_refactor_review_request.py +++ /dev/null @@ -1,68 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.18 on 2019-01-04 14:27 - - -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0009_move_non_url_externalurls_to_uploaded_filename'), - ('name', '0005_reviewassignmentstatename'), - ('person', '0008_auto_20181014_1448'), - ('review', '0008_remove_reviewrequest_old_id'), - ] - - operations = [ - migrations.CreateModel( - name='ReviewAssignment', - fields=[ - ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), - ('assigned_on', models.DateTimeField(blank=True, null=True)), - ('completed_on', models.DateTimeField(blank=True, null=True)), - ('reviewed_rev', models.CharField(blank=True, max_length=16, verbose_name='reviewed revision')), - ('mailarch_url', models.URLField(blank=True, null=True)), - ('result', ietf.utils.models.ForeignKey(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='name.ReviewResultName')), - ('review', ietf.utils.models.OneToOneField(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.Document')), - ], - ), - migrations.RenameField( - model_name='reviewrequest', - old_name='result', - new_name='unused_result', - ), - migrations.RenameField( - model_name='reviewrequest', - old_name='review', - new_name='unused_review', - ), - migrations.RenameField( - model_name='reviewrequest', - old_name='reviewed_rev', - new_name='unused_reviewed_rev', - ), - migrations.RenameField( - model_name='reviewrequest', - old_name='reviewer', - new_name='unused_reviewer', - ), - migrations.AddField( - model_name='reviewassignment', - name='review_request', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='review.ReviewRequest'), - ), - migrations.AddField( - model_name='reviewassignment', - name='reviewer', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='person.Email'), - ), - migrations.AddField( - model_name='reviewassignment', - name='state', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='name.ReviewAssignmentStateName'), - ), - ] diff --git a/ietf/review/migrations/0010_populate_review_assignments.py b/ietf/review/migrations/0010_populate_review_assignments.py deleted file mode 100644 index 83f87325bd..0000000000 --- a/ietf/review/migrations/0010_populate_review_assignments.py +++ /dev/null @@ -1,125 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.18 on 2019-01-04 14:34 - - -import sys - -from tqdm import tqdm - -from django.db import migrations - -def assigned_time(model, request): - e = model.objects.filter(doc=request.doc, type="assigned_review_request", review_request=request).order_by('-time', '-id').first() - return e.time if e and e.time else None - -def done_time(model, request): - if request.unused_review and request.unused_review.time: - return request.unused_review.time - e = model.objects.filter(doc=request.doc, type="closed_review_request", review_request=request).order_by('-time', '-id').first() - time = e.time if e and e.time else None - return time if time else request.time - -def map_request_state_to_assignment_state(req_state_id): - if req_state_id == 'requested': - return 'assigned' - elif req_state_id in ('no-review-document', 'no-review-version'): - return 'withdrawn' - else: - return req_state_id - -def forward(apps, schema_editor): - ReviewRequest = apps.get_model('review', 'ReviewRequest') - ReviewAssignment = apps.get_model('review', 'ReviewAssignment') - Document = apps.get_model('doc', 'Document') - ReviewRequestDocEvent = apps.get_model('doc','ReviewRequestDocEvent') - ReviewAssignmentDocEvent = apps.get_model('doc','ReviewAssignmentDocEvent') - - sys.stderr.write('\n') # introduce a newline before tqdm starts writing - for request in tqdm(ReviewRequest.objects.exclude(unused_reviewer__isnull=True)): - assignment_state = map_request_state_to_assignment_state(request.state_id) - if not (assignment_state in ('assigned', 'accepted', 'completed', 'no-response', 'overtaken', 'part-completed', 'rejected', 'withdrawn', 'unknown')): - print(("Trouble with review_request",request.pk,"with state",request.state_id)) - exit(-1) - ReviewAssignment.objects.create( - review_request = request, - state_id = assignment_state, - reviewer = request.unused_reviewer, - assigned_on = assigned_time(ReviewRequestDocEvent, request), - completed_on = done_time(ReviewRequestDocEvent, request), - review = request.unused_review, - reviewed_rev = request.unused_reviewed_rev, - result = request.unused_result, - mailarch_url = request.unused_review and request.unused_review.external_url, - ) - Document.objects.filter(type_id='review').update(external_url='') - ReviewRequest.objects.filter(state_id__in=('accepted', 'rejected', 'no-response', 'part-completed', 'completed', 'unknown')).update(state_id='assigned') - ReviewRequest.objects.filter(state_id='requested',unused_reviewer__isnull=False).update(state_id='assigned') - - for req_event in tqdm(ReviewRequestDocEvent.objects.filter(type="assigned_review_request",review_request__unused_reviewer__isnull=False)): - ReviewAssignmentDocEvent.objects.create( - time = req_event.time, - type = req_event.type, - by = req_event.by, - doc = req_event.doc, - rev = req_event.rev, - desc = req_event.desc, - review_assignment = req_event.review_request.reviewassignment_set.first(), - state_id = 'assigned' - ) - - for req_event in tqdm(ReviewRequestDocEvent.objects.filter(type="closed_review_request", - state_id__in=('completed', 'no-response', 'part-completed', 'rejected', 'unknown', 'withdrawn'), - review_request__unused_reviewer__isnull=False)): - ReviewAssignmentDocEvent.objects.create( - time = req_event.time, - type = req_event.type, - by = req_event.by, - doc = req_event.doc, - rev = req_event.rev, - desc = req_event.desc, - review_assignment = req_event.review_request.reviewassignment_set.first(), - state_id = req_event.state_id - ) - - ReviewRequestDocEvent.objects.filter(type="closed_review_request", - state_id__in=('completed', 'no-response', 'part-completed', 'rejected', 'unknown', 'withdrawn'), - review_request__unused_reviewer__isnull=False).delete() - - -def reverse(apps, schema_editor): - ReviewAssignment = apps.get_model('review', 'ReviewAssignment') - ReviewRequestDocEvent = apps.get_model('doc','ReviewRequestDocEvent') - ReviewAssignmentDocEvent = apps.get_model('doc','ReviewAssignmentDocEvent') - - sys.stderr.write('\n') - for assignment in tqdm(ReviewAssignment.objects.all()): - if assignment.review_request.unused_review: - assignment.review_request.unused_review.external_url = assignment.mailarch_url - assignment.review_request.unused_review.save() - assignment.review_request.state_id = assignment.state_id - assignment.review_request.save() - - for asgn_event in tqdm(ReviewAssignmentDocEvent.objects.filter(state_id__in=('completed', 'no-response', 'part-completed', 'rejected', 'unknown', 'withdrawn'))): - ReviewRequestDocEvent.objects.create( - time = asgn_event.time, - type = asgn_event.type, - by = asgn_event.by, - doc = asgn_event.doc, - rev = asgn_event.rev, - desc = asgn_event.desc, - review_request = asgn_event.review_assignment.review_request, - state_id = asgn_event.state_id - ) - ReviewAssignmentDocEvent.objects.all().delete() - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0009_refactor_review_request'), - ('doc','0011_reviewassignmentdocevent') - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/review/migrations/0011_review_document2_fk.py b/ietf/review/migrations/0011_review_document2_fk.py deleted file mode 100644 index 1fca4fd4e8..0000000000 --- a/ietf/review/migrations/0011_review_document2_fk.py +++ /dev/null @@ -1,59 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-08 11:58 - - -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0015_1_add_fk_to_document_id'), - ('review', '0010_populate_review_assignments'), - ] - - operations = [ - migrations.AddField( - model_name='reviewrequest', - name='doc2', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, related_name='reviewrequest_set', to='doc.Document', to_field=b'id'), - ), - migrations.AddField( - model_name='reviewwish', - name='doc2', - field=ietf.utils.models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field=b'id'), - ), - migrations.AddField( - model_name='reviewrequest', - name='unused_review2', - field=ietf.utils.models.OneToOneField(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field='id'), - ), - migrations.AddField( - model_name='reviewassignment', - name='review2', - field=ietf.utils.models.OneToOneField(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, to='doc.Document', to_field='id'), - ), - migrations.AlterField( - model_name='reviewrequest', - name='doc', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='old_revreq', to='doc.Document', to_field=b'name'), - ), - migrations.AlterField( - model_name='reviewwish', - name='doc', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='old_revwish', to='doc.Document', to_field=b'name'), - ), - migrations.AlterField( - model_name='reviewrequest', - name='unused_review', - field=ietf.utils.models.OneToOneField(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='old_unused_review', to='doc.Document', to_field=b'name'), - ), - migrations.AlterField( - model_name='reviewassignment', - name='review', - field=ietf.utils.models.OneToOneField(blank=True, null=True, on_delete=django.db.models.deletion.CASCADE, related_name='old_reviewassignment', to='doc.Document', to_field=b'name'), - ), - ] diff --git a/ietf/review/migrations/0012_remove_old_document_field.py b/ietf/review/migrations/0012_remove_old_document_field.py deleted file mode 100644 index 2c5b5c2776..0000000000 --- a/ietf/review/migrations/0012_remove_old_document_field.py +++ /dev/null @@ -1,48 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-20 09:53 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0011_review_document2_fk'), - ] - - operations = [ - migrations.RemoveField( - model_name='reviewrequest', - name='doc', - ), - migrations.RemoveField( - model_name='reviewrequest', - name='unused_reviewer', - ), - migrations.RemoveField( - model_name='reviewrequest', - name='unused_review', - ), - migrations.RemoveField( - model_name='reviewrequest', - name='unused_review2', - ), - migrations.RemoveField( - model_name='reviewrequest', - name='unused_reviewed_rev', - ), - migrations.RemoveField( - model_name='reviewrequest', - name='unused_result', - ), - migrations.RemoveField( - model_name='reviewwish', - name='doc', - ), - migrations.RemoveField( - model_name='reviewassignment', - name='review', - ), - ] diff --git a/ietf/review/migrations/0013_rename_field_document2.py b/ietf/review/migrations/0013_rename_field_document2.py deleted file mode 100644 index 0577b0d31c..0000000000 --- a/ietf/review/migrations/0013_rename_field_document2.py +++ /dev/null @@ -1,32 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-05-21 05:31 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - dependencies = [ - ('doc', '0019_rename_field_document2'), - ('review', '0012_remove_old_document_field'), - ] - - operations = [ - migrations.RenameField( - model_name='reviewrequest', - old_name='doc2', - new_name='doc', - ), - migrations.RenameField( - model_name='reviewwish', - old_name='doc2', - new_name='doc', - ), - migrations.RenameField( - model_name='reviewassignment', - old_name='review2', - new_name='review', - ), - ] diff --git a/ietf/review/migrations/0014_document_primary_key_cleanup.py b/ietf/review/migrations/0014_document_primary_key_cleanup.py deleted file mode 100644 index 20c28ace6d..0000000000 --- a/ietf/review/migrations/0014_document_primary_key_cleanup.py +++ /dev/null @@ -1,28 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.20 on 2019-06-10 03:47 - - -from django.db import migrations -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0013_rename_field_document2'), - ] - - operations = [ - migrations.AlterField( - model_name='reviewrequest', - name='doc', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='reviewrequest_set', to='doc.Document'), - ), - migrations.AlterField( - model_name='reviewwish', - name='doc', - field=ietf.utils.models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='doc.Document'), - ), - ] diff --git a/ietf/review/migrations/0015_populate_completed_on_for_rejected.py b/ietf/review/migrations/0015_populate_completed_on_for_rejected.py deleted file mode 100644 index feed8ea1c7..0000000000 --- a/ietf/review/migrations/0015_populate_completed_on_for_rejected.py +++ /dev/null @@ -1,29 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.24 on 2019-09-30 08:17 - - -from django.db import migrations - -def forward(apps, schema_editor): - ReviewAssignmentDocEvent = apps.get_model('doc','ReviewAssignmentDocEvent') - for event in ReviewAssignmentDocEvent.objects.filter(type="closed_review_assignment",state_id='rejected',review_assignment__completed_on__isnull=True): - event.review_assignment.completed_on = event.time - event.review_assignment.save() - - -def reverse(apps, schema_editor): - # There's no harm in leaving the newly set completed_on values even if this is rolled back - pass - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0014_document_primary_key_cleanup'), - ('doc', '0026_add_draft_rfceditor_state'), - ] - - operations = [ - migrations.RunPython(forward, reverse) - ] diff --git a/ietf/review/migrations/0016_add_remind_days_open_reviews.py b/ietf/review/migrations/0016_add_remind_days_open_reviews.py deleted file mode 100644 index 57e9911475..0000000000 --- a/ietf/review/migrations/0016_add_remind_days_open_reviews.py +++ /dev/null @@ -1,26 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.23 on 2019-09-05 05:03 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0015_populate_completed_on_for_rejected'), - ] - - operations = [ - migrations.AddField( - model_name='historicalreviewersettings', - name='remind_days_open_reviews', - field=models.PositiveIntegerField(blank=True, name="Periodic reminder of open reviews every X days", help_text="To get a periodic email reminder of all your open reviews, enter the number of days between these reminders. Clear the field if you don't want these reminders.", null=True), - ), - migrations.AddField( - model_name='reviewersettings', - name='remind_days_open_reviews', - field=models.PositiveIntegerField(blank=True, verbose_name="Periodic reminder of open reviews every X days", help_text="To get a periodic email reminder of all your open reviews, enter the number of days between these reminders. Clear the field if you don't want these reminders.", null=True), - ), - ] diff --git a/ietf/review/migrations/0017_add_review_team_remind_days_unconfirmed_assignments.py b/ietf/review/migrations/0017_add_review_team_remind_days_unconfirmed_assignments.py deleted file mode 100644 index af48a42435..0000000000 --- a/ietf/review/migrations/0017_add_review_team_remind_days_unconfirmed_assignments.py +++ /dev/null @@ -1,22 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.23 on 2019-10-01 04:40 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0016_add_remind_days_open_reviews'), - ] - - operations = [ - migrations.AddField( - model_name='reviewteamsettings', - name='remind_days_unconfirmed_assignments', - field=models.PositiveIntegerField(blank=True, help_text="To send a periodic email reminder to reviewers of review assignments that are not accepted yet, enter the number of days between these reminders. Clear the field if you don't want these reminders to be sent.", null=True, verbose_name='Periodic reminder of not yet accepted or rejected review assignments to reviewer every X days'), - ), - - ] diff --git a/ietf/review/migrations/0018_auto_20191015_1014.py b/ietf/review/migrations/0018_auto_20191015_1014.py deleted file mode 100644 index f61c755024..0000000000 --- a/ietf/review/migrations/0018_auto_20191015_1014.py +++ /dev/null @@ -1,36 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.25 on 2019-10-15 10:14 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0017_add_review_team_remind_days_unconfirmed_assignments'), - ] - - operations = [ - migrations.AlterField( - model_name='historicalreviewersettings', - name='remind_days_before_deadline', - field=models.IntegerField(blank=True, help_text="To get an email reminder in case you forget to do an assigned review, enter the number of days before review deadline you want to receive it. Clear the field if you don't want this reminder.", null=True), - ), - migrations.AlterField( - model_name='historicalreviewersettings', - name='remind_days_open_reviews', - field=models.PositiveIntegerField(blank=True, help_text="To get a periodic email reminder of all your open reviews, enter the number of days between these reminders. Clear the field if you don't want these reminders.", null=True, verbose_name='Periodic reminder of open reviews every X days'), - ), - migrations.AlterField( - model_name='reviewersettings', - name='remind_days_before_deadline', - field=models.IntegerField(blank=True, help_text="To get an email reminder in case you forget to do an assigned review, enter the number of days before review deadline you want to receive it. Clear the field if you don't want this reminder.", null=True), - ), - migrations.AlterField( - model_name='reviewteamsettings', - name='remind_days_unconfirmed_assignments', - field=models.PositiveIntegerField(blank=True, help_text="To send a periodic email reminder to reviewers of review assignments they have neither accepted nor rejected, enter the number of days between these reminders. Clear the field if you don't want these reminders to be sent.", null=True, verbose_name='Periodic reminder of not yet accepted or rejected review assignments to reviewer every X days'), - ), - ] diff --git a/ietf/review/migrations/0019_auto_20191023_0829.py b/ietf/review/migrations/0019_auto_20191023_0829.py deleted file mode 100644 index c5e84573c1..0000000000 --- a/ietf/review/migrations/0019_auto_20191023_0829.py +++ /dev/null @@ -1,22 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.25 on 2019-10-23 08:29 - - -import datetime -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0018_auto_20191015_1014'), - ] - - operations = [ - migrations.AlterField( - model_name='unavailableperiod', - name='start_date', - field=models.DateField(default=datetime.date.today, help_text="Choose the start date so that you can still do a review if it's assigned just before the start date - this usually means you should mark yourself unavailable for assignment some time before you are actually away. The default is today.", null=True), - ), - ] diff --git a/ietf/review/migrations/0020_auto_20191115_2059.py b/ietf/review/migrations/0020_auto_20191115_2059.py deleted file mode 100644 index 025e3d2d32..0000000000 --- a/ietf/review/migrations/0020_auto_20191115_2059.py +++ /dev/null @@ -1,26 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.26 on 2019-11-15 20:59 - - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0019_auto_20191023_0829'), - ] - - operations = [ - migrations.AddField( - model_name='reviewsecretarysettings', - name='days_to_show_in_reviewer_list', - field=models.IntegerField(blank=True, help_text='Maximum number of days to show in reviewer list for completed items.', null=True), - ), - migrations.AddField( - model_name='reviewsecretarysettings', - name='max_items_to_show_in_reviewer_list', - field=models.IntegerField(blank=True, help_text='Maximum number of completed items to show for one reviewer in the reviewer list view, the list is also filtered by the days to show in reviews list setting.', null=True), - ), - ] diff --git a/ietf/review/migrations/0021_add_additional_history.py b/ietf/review/migrations/0021_add_additional_history.py deleted file mode 100644 index 58dc709ec8..0000000000 --- a/ietf/review/migrations/0021_add_additional_history.py +++ /dev/null @@ -1,224 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.23 on 2019-11-19 04:36 - - -import datetime -from django.conf import settings -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models -import ietf.utils.validators -import simple_history.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('person', '0009_auto_20190118_0725'), - migrations.swappable_dependency(settings.AUTH_USER_MODEL), - ('doc', '0026_add_draft_rfceditor_state'), - ('name', '0007_fix_m2m_slug_id_length'), - ('group', '0019_rename_field_document2'), - ('review', '0020_auto_20191115_2059'), - ] - - operations = [ - migrations.CreateModel( - name='HistoricalReviewAssignment', - fields=[ - ('id', models.IntegerField(auto_created=True, blank=True, db_index=True, verbose_name='ID')), - ('history_change_reason', models.TextField(null=True)), - ('assigned_on', models.DateTimeField(blank=True, null=True)), - ('completed_on', models.DateTimeField(blank=True, null=True)), - ('reviewed_rev', models.CharField(blank=True, max_length=16, verbose_name='reviewed revision')), - ('mailarch_url', models.URLField(blank=True, null=True)), - ('history_id', models.AutoField(primary_key=True, serialize=False)), - ('history_date', models.DateTimeField()), - ('history_type', models.CharField(choices=[('+', 'Created'), ('~', 'Changed'), ('-', 'Deleted')], max_length=1)), - ('history_user', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='+', to=settings.AUTH_USER_MODEL)), - ('result', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='name.ReviewResultName')), - ('review', models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='doc.Document')), - ], - options={ - 'verbose_name': 'historical review assignment', - 'ordering': ('-history_date', '-history_id'), - 'get_latest_by': 'history_date', - }, - bases=(simple_history.models.HistoricalChanges, models.Model), - ), - migrations.CreateModel( - name='HistoricalReviewRequest', - fields=[ - ('id', models.IntegerField(auto_created=True, blank=True, db_index=True, verbose_name='ID')), - ('history_change_reason', models.TextField(null=True)), - ('time', models.DateTimeField(default=datetime.datetime.now)), - ('deadline', models.DateField()), - ('requested_rev', models.CharField(blank=True, help_text='Fill in if a specific revision is to be reviewed, e.g. 02', max_length=16, verbose_name='requested revision')), - ('comment', models.TextField(blank=True, default='', help_text='Provide any additional information to show to the review team secretary and reviewer', max_length=2048, verbose_name="Requester's comments and instructions")), - ('history_id', models.AutoField(primary_key=True, serialize=False)), - ('history_date', models.DateTimeField()), - ('history_type', models.CharField(choices=[('+', 'Created'), ('~', 'Changed'), ('-', 'Deleted')], max_length=1)), - ('doc', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='doc.Document')), - ('history_user', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='+', to=settings.AUTH_USER_MODEL)), - ('requested_by', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='person.Person')), - ('state', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='name.ReviewRequestStateName')), - ('team', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='group.Group')), - ('type', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='name.ReviewTypeName')), - ], - options={ - 'verbose_name': 'historical review request', - 'ordering': ('-history_date', '-history_id'), - 'get_latest_by': 'history_date', - }, - bases=(simple_history.models.HistoricalChanges, models.Model), - ), - migrations.CreateModel( - name='HistoricalUnavailablePeriod', - fields=[ - ('id', models.IntegerField(auto_created=True, blank=True, db_index=True, verbose_name='ID')), - ('history_change_reason', models.TextField(null=True)), - ('start_date', models.DateField(default=datetime.date.today, help_text="Choose the start date so that you can still do a review if it's assigned just before the start date - this usually means you should mark yourself unavailable for assignment some time before you are actually away. The default is today.", null=True)), - ('end_date', models.DateField(blank=True, help_text='Leaving the end date blank means that the period continues indefinitely. You can end it later.', null=True)), - ('availability', models.CharField(choices=[('canfinish', 'Can do follow-ups'), ('unavailable', 'Completely unavailable')], max_length=30)), - ('reason', models.TextField(blank=True, default='', help_text="Provide (for the secretary's benefit) the reason why the review is unavailable", max_length=2048, verbose_name='Reason why reviewer is unavailable (Optional)')), - ('history_id', models.AutoField(primary_key=True, serialize=False)), - ('history_date', models.DateTimeField()), - ('history_type', models.CharField(choices=[('+', 'Created'), ('~', 'Changed'), ('-', 'Deleted')], max_length=1)), - ('history_user', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='+', to=settings.AUTH_USER_MODEL)), - ('person', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='person.Person')), - ('team', ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='group.Group')), - ], - options={ - 'verbose_name': 'historical unavailable period', - 'ordering': ('-history_date', '-history_id'), - 'get_latest_by': 'history_date', - }, - bases=(simple_history.models.HistoricalChanges, models.Model), - ), - migrations.AlterField( - model_name='historicalreviewersettings', - name='expertise', - field=models.TextField(blank=True, default='', help_text="Describe the reviewer's expertise in this team's area", max_length=2048, verbose_name="Reviewer's expertise in this team's area"), - ), - migrations.AlterField( - model_name='historicalreviewersettings', - name='filter_re', - field=models.CharField(blank=True, help_text='Draft names matching this regular expression should not be assigned', max_length=255, validators=[ietf.utils.validators.RegexStringValidator()], verbose_name='Filter regexp'), - ), - migrations.AlterField( - model_name='historicalreviewersettings', - name='min_interval', - field=models.IntegerField(blank=True, choices=[(7, 'Once per week'), (14, 'Once per fortnight'), (30, 'Once per month'), (61, 'Once per two months'), (91, 'Once per quarter')], null=True, verbose_name='Can review at most'), - ), - migrations.AlterField( - model_name='historicalreviewersettings', - name='remind_days_before_deadline', - field=models.IntegerField(blank=True, help_text="To get an email reminder in case you forget to do an assigned review, enter the number of days before review deadline you want to receive it. Clear the field if you don't want this reminder.", null=True), - ), - migrations.AlterField( - model_name='historicalreviewersettings', - name='remind_days_open_reviews', - field=models.PositiveIntegerField(blank=True, help_text="To get a periodic email reminder of all your open reviews, enter the number of days between these reminders. Clear the field if you don't want these reminders.", null=True, verbose_name='Periodic reminder of open reviews every X days'), - ), - migrations.AlterField( - model_name='historicalreviewersettings', - name='skip_next', - field=models.IntegerField(default=0, verbose_name='Skip next assignments'), - ), - migrations.AlterField( - model_name='reviewassignment', - name='reviewed_rev', - field=models.CharField(blank=True, max_length=16, verbose_name='reviewed revision'), - ), - migrations.AlterField( - model_name='reviewersettings', - name='expertise', - field=models.TextField(blank=True, default='', help_text="Describe the reviewer's expertise in this team's area", max_length=2048, verbose_name="Reviewer's expertise in this team's area"), - ), - migrations.AlterField( - model_name='reviewersettings', - name='filter_re', - field=models.CharField(blank=True, help_text='Draft names matching this regular expression should not be assigned', max_length=255, validators=[ietf.utils.validators.RegexStringValidator()], verbose_name='Filter regexp'), - ), - migrations.AlterField( - model_name='reviewersettings', - name='min_interval', - field=models.IntegerField(blank=True, choices=[(7, 'Once per week'), (14, 'Once per fortnight'), (30, 'Once per month'), (61, 'Once per two months'), (91, 'Once per quarter')], null=True, verbose_name='Can review at most'), - ), - migrations.AlterField( - model_name='reviewersettings', - name='remind_days_before_deadline', - field=models.IntegerField(blank=True, help_text="To get an email reminder in case you forget to do an assigned review, enter the number of days before review deadline you want to receive it. Clear the field if you don't want this reminder.", null=True), - ), - migrations.AlterField( - model_name='reviewersettings', - name='skip_next', - field=models.IntegerField(default=0, verbose_name='Skip next assignments'), - ), - migrations.AlterField( - model_name='reviewrequest', - name='comment', - field=models.TextField(blank=True, default='', help_text='Provide any additional information to show to the review team secretary and reviewer', max_length=2048, verbose_name="Requester's comments and instructions"), - ), - migrations.AlterField( - model_name='reviewrequest', - name='requested_rev', - field=models.CharField(blank=True, help_text='Fill in if a specific revision is to be reviewed, e.g. 02', max_length=16, verbose_name='requested revision'), - ), - migrations.AlterField( - model_name='reviewsecretarysettings', - name='remind_days_before_deadline', - field=models.IntegerField(blank=True, help_text="To get an email reminder in case a reviewer forgets to do an assigned review, enter the number of days before review deadline you want to receive it. Clear the field if you don't want a reminder.", null=True), - ), - migrations.AlterField( - model_name='reviewteamsettings', - name='autosuggest', - field=models.BooleanField(default=True, verbose_name='Automatically suggest possible review requests'), - ), - migrations.AlterField( - model_name='reviewteamsettings', - name='remind_days_unconfirmed_assignments', - field=models.PositiveIntegerField(blank=True, help_text="To send a periodic email reminder to reviewers of review assignments they have neither accepted nor rejected, enter the number of days between these reminders. Clear the field if you don't want these reminders to be sent.", null=True, verbose_name='Periodic reminder of not yet accepted or rejected review assignments to reviewer every X days'), - ), - migrations.AlterField( - model_name='reviewteamsettings', - name='secr_mail_alias', - field=models.CharField(blank=True, help_text='Email alias for all of the review team secretaries', max_length=255, verbose_name='Email alias for all of the review team secretaries'), - ), - migrations.AlterField( - model_name='unavailableperiod', - name='availability', - field=models.CharField(choices=[('canfinish', 'Can do follow-ups'), ('unavailable', 'Completely unavailable')], max_length=30), - ), - migrations.AlterField( - model_name='unavailableperiod', - name='end_date', - field=models.DateField(blank=True, help_text='Leaving the end date blank means that the period continues indefinitely. You can end it later.', null=True), - ), - migrations.AlterField( - model_name='unavailableperiod', - name='reason', - field=models.TextField(blank=True, default='', help_text="Provide (for the secretary's benefit) the reason why the review is unavailable", max_length=2048, verbose_name='Reason why reviewer is unavailable (Optional)'), - ), - migrations.AlterField( - model_name='unavailableperiod', - name='start_date', - field=models.DateField(default=datetime.date.today, help_text="Choose the start date so that you can still do a review if it's assigned just before the start date - this usually means you should mark yourself unavailable for assignment some time before you are actually away. The default is today.", null=True), - ), - migrations.AddField( - model_name='historicalreviewassignment', - name='review_request', - field=ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='review.ReviewRequest'), - ), - migrations.AddField( - model_name='historicalreviewassignment', - name='reviewer', - field=ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='person.Email'), - ), - migrations.AddField( - model_name='historicalreviewassignment', - name='state', - field=ietf.utils.models.ForeignKey(blank=True, db_constraint=False, null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='name.ReviewAssignmentStateName'), - ), - ] diff --git a/ietf/review/migrations/0022_reviewer_queue_policy_and_request_assignment_next.py b/ietf/review/migrations/0022_reviewer_queue_policy_and_request_assignment_next.py deleted file mode 100644 index 9d705ea143..0000000000 --- a/ietf/review/migrations/0022_reviewer_queue_policy_and_request_assignment_next.py +++ /dev/null @@ -1,37 +0,0 @@ -# Copyright The IETF Trust 2019-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.23 on 2019-11-18 08:50 - - -from django.db import migrations, models -import django.db.models.deletion - - -class Migration(migrations.Migration): - - dependencies = [ - ('name', '0008_reviewerqueuepolicyname'), - ('review', '0021_add_additional_history'), - ] - - operations = [ - migrations.AddField( - model_name='reviewteamsettings', - name='reviewer_queue_policy', - field=models.ForeignKey(default='RotateAlphabetically', on_delete=django.db.models.deletion.PROTECT, to='name.ReviewerQueuePolicyName'), - ), - migrations.AddField( - model_name='historicalreviewersettings', - name='request_assignment_next', - field=models.BooleanField(default=False, - help_text='If you would like to be assigned to a review as soon as possible, select this option. It is automatically reset once you receive any assignment.', - verbose_name='Select me next for an assignment'), - ), - migrations.AddField( - model_name='reviewersettings', - name='request_assignment_next', - field=models.BooleanField(default=False, - help_text='If you would like to be assigned to a review as soon as possible, select this option. It is automatically reset once you receive any assignment.', - verbose_name='Select me next for an assignment'), - ), - ] diff --git a/ietf/review/migrations/0023_historicalreviewersettings_change_reason_text_field.py b/ietf/review/migrations/0023_historicalreviewersettings_change_reason_text_field.py deleted file mode 100644 index 543fa34517..0000000000 --- a/ietf/review/migrations/0023_historicalreviewersettings_change_reason_text_field.py +++ /dev/null @@ -1,21 +0,0 @@ -# -*- coding: utf-8 -*- -# Copyright The IETF Trust 2020, All Rights Reserved -# Generated by Django 1.11.26 on 2019-12-21 11:52 -from __future__ import unicode_literals - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0022_reviewer_queue_policy_and_request_assignment_next'), - ] - - operations = [ - migrations.AlterField( - model_name='historicalreviewersettings', - name='history_change_reason', - field=models.TextField(null=True), - ), - ] diff --git a/ietf/review/migrations/0024_auto_20200520_0017.py b/ietf/review/migrations/0024_auto_20200520_0017.py deleted file mode 100644 index 1b1cba31fe..0000000000 --- a/ietf/review/migrations/0024_auto_20200520_0017.py +++ /dev/null @@ -1,60 +0,0 @@ -# Generated by Django 2.0.13 on 2020-05-20 00:17 - -from django.db import migrations, models -import django.db.models.deletion -import ietf.utils.models - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0023_historicalreviewersettings_change_reason_text_field'), - ] - - operations = [ - migrations.AlterField( - model_name='historicalreviewersettings', - name='team', - field=ietf.utils.models.ForeignKey(blank=True, db_constraint=False, limit_choices_to=models.Q(_negated=True, reviewteamsettings=None), null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='group.Group'), - ), - migrations.AlterField( - model_name='historicalreviewrequest', - name='team', - field=ietf.utils.models.ForeignKey(blank=True, db_constraint=False, limit_choices_to=models.Q(_negated=True, reviewteamsettings=None), null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='group.Group'), - ), - migrations.AlterField( - model_name='historicalunavailableperiod', - name='team', - field=ietf.utils.models.ForeignKey(blank=True, db_constraint=False, limit_choices_to=models.Q(_negated=True, reviewteamsettings=None), null=True, on_delete=django.db.models.deletion.DO_NOTHING, related_name='+', to='group.Group'), - ), - migrations.AlterField( - model_name='nextreviewerinteam', - name='team', - field=ietf.utils.models.ForeignKey(limit_choices_to=models.Q(_negated=True, reviewteamsettings=None), on_delete=django.db.models.deletion.CASCADE, to='group.Group'), - ), - migrations.AlterField( - model_name='reviewersettings', - name='team', - field=ietf.utils.models.ForeignKey(limit_choices_to=models.Q(_negated=True, reviewteamsettings=None), on_delete=django.db.models.deletion.CASCADE, to='group.Group'), - ), - migrations.AlterField( - model_name='reviewrequest', - name='team', - field=ietf.utils.models.ForeignKey(limit_choices_to=models.Q(_negated=True, reviewteamsettings=None), on_delete=django.db.models.deletion.CASCADE, to='group.Group'), - ), - migrations.AlterField( - model_name='reviewsecretarysettings', - name='team', - field=ietf.utils.models.ForeignKey(limit_choices_to=models.Q(_negated=True, reviewteamsettings=None), on_delete=django.db.models.deletion.CASCADE, to='group.Group'), - ), - migrations.AlterField( - model_name='reviewwish', - name='team', - field=ietf.utils.models.ForeignKey(limit_choices_to=models.Q(_negated=True, reviewteamsettings=None), on_delete=django.db.models.deletion.CASCADE, to='group.Group'), - ), - migrations.AlterField( - model_name='unavailableperiod', - name='team', - field=ietf.utils.models.ForeignKey(limit_choices_to=models.Q(_negated=True, reviewteamsettings=None), on_delete=django.db.models.deletion.CASCADE, to='group.Group'), - ), - ] diff --git a/ietf/review/migrations/0025_repair_assignments.py b/ietf/review/migrations/0025_repair_assignments.py deleted file mode 100644 index 49ce65d02d..0000000000 --- a/ietf/review/migrations/0025_repair_assignments.py +++ /dev/null @@ -1,378 +0,0 @@ -# Copyright The IETF Trust 2020 All Rights Reserved - -import os - -from django.conf import settings -from django.db import migrations - - -class Helper(object): - - def __init__(self, review_path, comments_by, document_class): - self.review_path = review_path - self.comments_by = comments_by - self.document_class = document_class - - def remove_file(self,name): - filename = os.path.join(self.review_path, '{}.txt'.format(name)) - os.remove(filename) - - def rename_file(self, old_name, new_name): - old_filename = os.path.join(self.review_path, '{}.txt'.format(old_name)) - new_filename = os.path.join(self.review_path, '{}.txt'.format(new_name)) - os.rename(old_filename, new_filename) - - def add_comment(self, name, comment): - doc = self.document_class.objects.get(name=name) - doc.docevent_set.create( - type = 'added_comment', - by = self.comments_by, - rev = doc.rev, - desc = comment, - ) - -def forward(apps, schema_editor): - ReviewAssignment = apps.get_model('review','ReviewAssignment') - Document = apps.get_model('doc','Document') - Person = apps.get_model('person','Person') - - # The calculation of review_path makes the assumption that DOCUMENT_PATH_PATTERN only uses - # things that are invariant for review documents. For production, as of this commit, that's - # DOCUMENT_PATH_PATTERN = '/a/www/ietf-ftp/{doc.type_id}/' - - helper = Helper( - review_path = settings.DOCUMENT_PATH_PATTERN.format(doc=Document.objects.filter(type_id='review').last()), - comments_by = Person.objects.get(name='(System)'), - document_class = Document, - ) - - # review-allan-5g-fmc-encapsulation-04-tsvart-lc-black is a double-submit - # In [120]: for d in Document.objects.filter(name__contains='review-allan-5g-fmc-encapsulation-04-tsvart - # ...: ').order_by('name'): - # ...: print(d.name, d.time) - # ...: - # review-allan-5g-fmc-encapsulation-04-tsvart-lc-black-2020-06-30 2020-06-30 11:06:30 - # review-allan-5g-fmc-encapsulation-04-tsvart-lc-black-2020-06-30-2 2020-06-30 11:06:30 - # (I've put some more detail below on this as my understanding of double-submit improved) - # The recommendation is to point the reviewassignment at the first submission, and delete the -2. - # - a = ReviewAssignment.objects.get(review_request__doc__name='draft-allan-5g-fmc-encapsulation', review_request__team__acronym='tsvart') - a.review = Document.objects.get(name='review-allan-5g-fmc-encapsulation-04-tsvart-lc-black-2020-06-30') - a.save() - Document.objects.filter(name='review-allan-5g-fmc-encapsulation-04-tsvart-lc-black-2020-06-30-2').delete() - helper.remove_file('review-allan-5g-fmc-encapsulation-04-tsvart-lc-black-2020-06-30-2') - helper.add_comment('draft-allan-5g-fmc-encapsulation', 'Removed an unintended duplicate version of the tsvart lc review') - - - # This one is just simply disconnected. No duplicates or anything to remove. - a = ReviewAssignment.objects.get(review_request__doc__name='draft-ietf-6lo-minimal-fragment',review_request__team__acronym='opsdir') - r = Document.objects.get(name='review-ietf-6lo-minimal-fragment-09-opsdir-lc-banks-2020-01-31') - a.review = r - a.state_id = 'completed' - a.result_id = 'nits' - a.reviewed_rev = '09' - a.completed_on = r.time - a.save() - helper.add_comment('draft-ietf-6lo-minimal-fragment', 'Reconnected opsdir lc review') - - # This review took place when we were spinning up the review tool. I suspect there were bugs at the time that we no longer have insight into. - # These two do not exist on disk - # review-ietf-6lo-privacy-considerations-04-secdir-lc-kaduk-2016-11-30 - # review-ietf-6lo-privacy-considerations-04-secdir-lc-kaduk-2016-11-30-2 - # These are identical except for 0d0a vs 0a and newline at end of file - # review-ietf-6lo-privacy-considerations-04-secdir-lc-kaduk-2016-11-30-3 - # review-ietf-6lo-privacy-considerations-04-secdir-lc-kaduk-2016-12-01 - # -12-01 is already reachable from the assignment. I suggest: - Document.objects.filter(name__startswith='review-ietf-6lo-privacy-considerations-04-secdir-lc-kaduk-2016-11-30').delete() - helper.remove_file('review-ietf-6lo-privacy-considerations-04-secdir-lc-kaduk-2016-11-30-3') - helper.add_comment('draft-ietf-6lo-privacy-considerations','Removed unintended duplicates of secdir lc review') - - a = ReviewAssignment.objects.get(review_request__doc__name='draft-ietf-bess-nsh-bgp-control-plane',review_request__team__acronym='rtgdir',reviewer__person__name__icontains='singh') - r = Document.objects.get(name='review-ietf-bess-nsh-bgp-control-plane-13-rtgdir-lc-singh-2020-01-29') - a.review = r - a.state_id = 'completed' - a.reviewed_rev='13' - a.result_id='issues' - a.completed_on=r.time - a.save() - helper.add_comment('draft-ietf-bess-nsh-bgp-control-plane','Reconnected rtgdir lc review') - - # In [121]: for d in Document.objects.filter(name__contains='review-ietf-capport-architecture-08-genart-lc-halpern').order_by('name'): - # ...: print(d.name, d.time) - # review-ietf-capport-architecture-08-genart-lc-halpern-2020-05-16 2020-05-16 15:34:35 - # review-ietf-capport-architecture-08-genart-lc-halpern-2020-05-16-2 2020-05-16 15:35:55 - # Not the same as the other double-submits, but likely a failure on the first submit midway through processing that led to the second. - # rjsparks@ietfa:/a/www/ietf-ftp/review> ls review-ietf-capport-architecture-08-genart-lc-halpern* - # review-ietf-capport-architecture-08-genart-lc-halpern-2020-05-16-2.txt - # Only -2 exists on disk - # -2 is what is currently pointed to by the review assignment. - helper.rename_file('review-ietf-capport-architecture-08-genart-lc-halpern-2020-05-16-2','review-ietf-capport-architecture-08-genart-lc-halpern-2020-05-16') - a = ReviewAssignment.objects.get(review_request__doc__name='draft-ietf-capport-architecture',review_request__type_id='lc',reviewer__person__name='Joel M. Halpern') - a.review = Document.objects.get(name='review-ietf-capport-architecture-08-genart-lc-halpern-2020-05-16') - a.save() - Document.objects.filter(name='review-ietf-capport-architecture-08-genart-lc-halpern-2020-05-16-2').delete() - helper.add_comment('draft-ietf-capport-architecture','Removed an unintended duplicate version of the genart lc review') - # Any external references to -05-16-2 will now break, but I'm reticent to start down the path of putting something sensical there. - # Right now, any review document not pointed to by a reviewassignment is a database error, and the view code assumes it won't happen. - # We can make it more robust against crashing, but I don't think we should try to adapt the models to model the corruption. - - - # Opsdir last call review of draft-ietf-cbor-array-tags-07 - # Got sent to the opsdir list (which is has a private archive) twice - no resulting thread from either - # rjsparks@ietfa:/a/www/ietf-ftp/review> ls review-ietf-cbor-array-tags-07-opsdir-lc-dunbar* - # review-ietf-cbor-array-tags-07-opsdir-lc-dunbar-2019-08-26-2.txt - # review-ietf-cbor-array-tags-07-opsdir-lc-dunbar-2019-08-26.txt - # In [122]: for d in Document.objects.filter(name__contains='review-ietf-cbor-array-tags-07-opsdir-lc-dunbar').order_by('name'): - # ...: print(d.name, d.time) - # review-ietf-cbor-array-tags-07-opsdir-lc-dunbar-2019-08-26 2019-08-26 14:13:29 - # review-ietf-cbor-array-tags-07-opsdir-lc-dunbar-2019-08-26-2 2019-08-26 14:13:29 - # This is a double-submit. - # The ReviewAssignment already points to 08-26 - Document.objects.filter(name='review-ietf-cbor-array-tags-07-opsdir-lc-dunbar-2019-08-26-2').delete() - helper.remove_file('review-ietf-cbor-array-tags-07-opsdir-lc-dunbar-2019-08-26-2') - helper.add_comment('draft-ietf-cbor-array-tags','Removed unintended duplicate of opsdir lc review') - - # In [73]: for d in Document.objects.filter(name__startswith='review-ietf-detnet-mpls-over-udp-ip'): - # ...: print(d.name, d.time) - # - # review-ietf-detnet-mpls-over-udp-ip-06-genart-lc-holmberg-2020-09-01 2020-09-01 13:47:55 - # review-ietf-detnet-mpls-over-udp-ip-06-genart-lc-holmberg-2020-09-01-2 2020-09-01 13:47:55 - # review-ietf-detnet-mpls-over-udp-ip-06-opsdir-lc-romascanu-2020-09-03 2020-09-03 02:49:33 - # review-ietf-detnet-mpls-over-udp-ip-06-opsdir-lc-romascanu-2020-09-03-2 2020-09-03 02:49:33 - # - # Both of those are places where the submit button got hit twice in rapid succession. - # Messages went to the list twice. No threads were started - # The review assignments currently point to the -2 versions. I think we change them to point to the not -2 versions and delete the -2 version documents. - # - a = ReviewAssignment.objects.get(review__name='review-ietf-detnet-mpls-over-udp-ip-06-genart-lc-holmberg-2020-09-01-2') - a.review = Document.objects.get(name='review-ietf-detnet-mpls-over-udp-ip-06-genart-lc-holmberg-2020-09-01') - a.save() - Document.objects.filter(name='review-ietf-detnet-mpls-over-udp-ip-06-genart-lc-holmberg-2020-09-01-2').delete() - helper.remove_file('review-ietf-detnet-mpls-over-udp-ip-06-genart-lc-holmberg-2020-09-01-2') - # - a = ReviewAssignment.objects.get(review__name='review-ietf-detnet-mpls-over-udp-ip-06-opsdir-lc-romascanu-2020-09-03-2') - a.review = Document.objects.get(name='review-ietf-detnet-mpls-over-udp-ip-06-opsdir-lc-romascanu-2020-09-03') - a.save() - Document.objects.filter(name='review-ietf-detnet-mpls-over-udp-ip-06-opsdir-lc-romascanu-2020-09-03-2').delete() - helper.remove_file('review-ietf-detnet-mpls-over-udp-ip-06-opsdir-lc-romascanu-2020-09-03-2') - helper.add_comment('draft-ietf-detnet-mpls-over-udp-ip','Removed unintended duplicate of opsdir and genart lc reviews') - - # This draft had a contentious last call (not because of this review) - # rjsparks@ietfa:/a/www/ietf-ftp/review> ls -l review-ietf-dprive-rfc7626-bis-03-genart* - # -rw-r--r-- 1 wwwrun www 1087 Dec 4 2019 review-ietf-dprive-rfc7626-bis-03-genart-lc-shirazipour-2019-12-04-2.txt - # -rw-r--r-- 1 wwwrun www 1087 Dec 4 2019 review-ietf-dprive-rfc7626-bis-03-genart-lc-shirazipour-2019-12-04-3.txt - # -rw-r--r-- 1 wwwrun www 1087 Dec 4 2019 review-ietf-dprive-rfc7626-bis-03-genart-lc-shirazipour-2019-12-04.txt - # These files are identical. - # In [75]: Document.objects.filter(name__startswith='review-ietf-dprive-rfc7626-bis-03-genart-lc-shirazipour-2019-12-04').values_list('time',flat=True) - # Out[75]: - # So again, the submit button got hit several times in rapid succession - # Interestingly, this was a case where the review was sent to the list first and then Meral told the datatracker about it, so it's only on the list(s) once. - # I think we change the assignment to point to 12-04 and delete -2 and -3. - a = ReviewAssignment.objects.get(review__name='review-ietf-dprive-rfc7626-bis-03-genart-lc-shirazipour-2019-12-04-3') - a.review = Document.objects.get(name='review-ietf-dprive-rfc7626-bis-03-genart-lc-shirazipour-2019-12-04') - a.save() - Document.objects.filter(name__startswith='review-ietf-dprive-rfc7626-bis-03-genart-lc-shirazipour-2019-12-04-').delete() - helper.remove_file('review-ietf-dprive-rfc7626-bis-03-genart-lc-shirazipour-2019-12-04-2') - helper.remove_file('review-ietf-dprive-rfc7626-bis-03-genart-lc-shirazipour-2019-12-04-3') - helper.add_comment('draft-ietf-dprive-rfc7626-bis','Removed unintended duplicates of genart lc review') - - # In [76]: Document.objects.filter(name__startswith='review-ietf-emu-rfc5448bis-06-secdir-lc-rose') - # Out[76]: , ]> - # rjsparks@ietfa:/a/www/ietf-ftp/review> ls review-ietf-emu-rfc5448bis-06-secdir-lc-rose* - # review-ietf-emu-rfc5448bis-06-secdir-lc-rose-2020-02-06.txt - # review assignment points to 02-06. - # Suggest we delete -01-27 Document object. There's nothing matching on disk to remove. - Document.objects.filter(name='review-ietf-emu-rfc5448bis-06-secdir-lc-rose-2020-01-27').delete() - helper.add_comment('draft-ietf-emu-rfc5448bis','Removed duplicate secdir lc review') - - # rjsparks@ietfa:/a/www/ietf-ftp/review> ls review-ietf-mmusic-t140-usage-data-channel-11-tsvart* - # review-ietf-mmusic-t140-usage-data-channel-11-tsvart-lc-scharf-2020-01-28.txt - # review-ietf-mmusic-t140-usage-data-channel-11-tsvart-lc-scharf-2020-03-27.txt - # The second was derived from the list post, so it has "Reviewer: Michael Scharf Review result: Ready with Nits" added and is reflowed, but is otherwise identical. - # In [80]: Document.objects.filter(name__startswith='review-ietf-mmusic-t140-usage - # ...: -data-channel-11-tsvart-lc-scharf') - # Out[80]: , ]> - # the 03-27 version was a resubmission by a secretary (Wes). The assignment currently points to 03-07, and it points to the 01-29 (! date there is in UTC) list entry. - # 01-29 is in the window of confusion. - # Suggest we just delete the -01-28 document object. - Document.objects.filter(name='review-ietf-mmusic-t140-usage-data-channel-11-tsvart-lc-scharf-2020-01-28').delete() - helper.remove_file('review-ietf-mmusic-t140-usage-data-channel-11-tsvart-lc-scharf-2020-01-28') - helper.add_comment('draft-ietf-mmusic-t140-usage-data-channel','Removed duplicate tsvart lc review') - - # In [81]: Document.objects.filter(name__startswith='review-ietf-mpls-spl-terminology-03-opsdir-lc-jaeggli').values_list('time',flat=True) - # Out[81]: - # Another double-submit. - # The review assignment already points to -08-15. Suggest we delete -08-15-2 - Document.objects.filter(name='review-ietf-mpls-spl-terminology-03-opsdir-lc-jaeggli-2020-08-15-2').delete() - helper.remove_file('review-ietf-mpls-spl-terminology-03-opsdir-lc-jaeggli-2020-08-15-2') - helper.add_comment('draft-ietf-mpls-spl-terminology','Removed unintended duplicate opsdir lc review') - - - # In [82]: Document.objects.filter(name__startswith='review-ietf-netconf-subscribed-notifications-23-rtgdir').values_list('time',flat=True) - # Out[82]: - # Another double-submit. This time the second won and the review assignments points to -2. I suggest we change it to point to the base and delete -2. - a = ReviewAssignment.objects.get(review__name='review-ietf-netconf-subscribed-notifications-23-rtgdir-lc-singh-2019-04-16-2') - a.review = Document.objects.get(name='review-ietf-netconf-subscribed-notifications-23-rtgdir-lc-singh-2019-04-16') - a.save() - Document.objects.filter(name='review-ietf-netconf-subscribed-notifications-23-rtgdir-lc-singh-2019-04-16-2').delete() - helper.remove_file('review-ietf-netconf-subscribed-notifications-23-rtgdir-lc-singh-2019-04-16-2') - helper.add_comment('draft-ietf-netconf-subscribed-notifications','Removed unintended duplicate rtgdir lc review') - - # In [84]: Document.objects.filter(name__contains='ietf-netconf-yang-push-22-genart-lc-bryant').values_list('time',flat=True) - # Out[84]: - # In [85]: ReviewAssignment.objects.get(review__name__contains='ietf-netconf-yang- - # ...: push-22-genart-lc-bryant').review - # Out[85]: - # Same as above. - a = ReviewAssignment.objects.get(review__name='review-ietf-netconf-yang-push-22-genart-lc-bryant-2019-04-10-2') - a.review = Document.objects.get(name='review-ietf-netconf-yang-push-22-genart-lc-bryant-2019-04-10') - a.save() - Document.objects.filter(name='review-ietf-netconf-yang-push-22-genart-lc-bryant-2019-04-10-2').delete() - helper.remove_file('review-ietf-netconf-yang-push-22-genart-lc-bryant-2019-04-10-2') - helper.add_comment('draft-ietf-netconf-yang-push','Removed unintended duplicate genart lc review') - - # In [92]: for d in Document.objects.filter(name__contains='ietf-nfsv4-rpcrdma-cm-pvt-data-06').order_by('name'): - # ...: print(d.name, d.external_url) - # ...: - # review-ietf-nfsv4-rpcrdma-cm-pvt-data-06-genart-lc-nandakumar-2020-01-27 https://mailarchive.ietf.org/arch/msg/gen-art/b'rGU9fbpAGtmz55Rcdfnl9ZsqMIo' - # review-ietf-nfsv4-rpcrdma-cm-pvt-data-06-genart-lc-nandakumar-2020-01-30 https://mailarchive.ietf.org/arch/msg/gen-art/rGU9fbpAGtmz55Rcdfnl9ZsqMIo - # review-ietf-nfsv4-rpcrdma-cm-pvt-data-06-opsdir-lc-comstedt-2020-01-27 https://mailarchive.ietf.org/arch/msg/ops-dir/b'BjEE4Y0ZDRALgueoS_lbL5U06js' - # review-ietf-nfsv4-rpcrdma-cm-pvt-data-06-opsdir-lc-comstedt-2020-02-10 https://mailarchive.ietf.org/arch/msg/ops-dir/BjEE4Y0ZDRALgueoS_lbL5U06js - # review-ietf-nfsv4-rpcrdma-cm-pvt-data-06-secdir-lc-sheffer-2020-01-26 https://mailarchive.ietf.org/arch/msg/secdir/hY6OTDbplzp9uONAvEjkcfa-N4A - # This straddled the period of confusion. - # For genart, Jean completed the review for the reviewer twice - # for opsdir the reviewer submitted on different dates. - # In both cases, the review only went to the list once, (the links above that are b'<>' are broken anyhow). - # rjsparks@ietfa:/a/www/ietf-ftp/review> ls review-ietf-nfsv4-rpcrdma-cm-pvt-data-06-genart-lc-nandakumar* - # review-ietf-nfsv4-rpcrdma-cm-pvt-data-06-genart-lc-nandakumar-2020-01-30.txt - # rjsparks@ietfa:/a/www/ietf-ftp/review> ls review-ietf-nfsv4-rpcrdma-cm-pvt-data-06-opsdir-lc-comstedt* - # review-ietf-nfsv4-rpcrdma-cm-pvt-data-06-opsdir-lc-comstedt-2020-02-10.txt - # The review assignment objects point to the ones that are backed by disk files. I suggest we delete the others. - Document.objects.filter(name='review-ietf-nfsv4-rpcrdma-cm-pvt-data-06-genart-lc-nandakumar-2020-01-27').delete() - Document.objects.filter(name='review-ietf-nfsv4-rpcrdma-cm-pvt-data-06-opsdir-lc-comstedt-2020-01-27').delete() - ReviewAssignment.objects.filter(review_request__doc__name='draft-ietf-nfsv4-rpcrdma-cm-pvt-data').exclude(review__type_id='review').delete() - helper.add_comment('draft-ietf-nfsv4-rpcrdma-cm-pvt-data','Removed unintended duplicate genart and opsdir lc reviews') - - # In [101]: ReviewAssignment.objects.filter(review__name__contains='review-ietf-pce-pcep-flowspec-09-tsvart').values_list('review__name',flat=True) - # Out[101]: - # In [102]: for d in Document.objects.filter(name__contains='review-ietf-pce-pcep-flowspec-09-tsvart').order_by('name'): - # ...: print(d.name, d.time) - # review-ietf-pce-pcep-flowspec-09-tsvart-lc-touch-2020-07-03 2020-07-03 19:20:49 - # review-ietf-pce-pcep-flowspec-09-tsvart-lc-touch-2020-07-03-2 2020-07-03 19:20:49 - # Same as double-submits above. - a = ReviewAssignment.objects.get(review__name='review-ietf-pce-pcep-flowspec-09-tsvart-lc-touch-2020-07-03-2') - a.review = Document.objects.get(name='review-ietf-pce-pcep-flowspec-09-tsvart-lc-touch-2020-07-03') - a.save() - Document.objects.filter(name='review-ietf-pce-pcep-flowspec-09-tsvart-lc-touch-2020-07-03-2').delete() - helper.remove_file('review-ietf-pce-pcep-flowspec-09-tsvart-lc-touch-2020-07-03-2') - helper.add_comment('draft-ietf-pce-pcep-flowspec','Removed unintended duplicate tsvart lc review') - - # draft-ietf-pim-msdp-yang history is a bit more complicated. - # It is currently (23Sep2020) in AUTH48... - # There is a combination of secretaries trying to route around the confusion and a reviewer accidentally updating the wrong review. - # Everything went to lists. - # In [110]: for rq in ReviewRequest.objects.filter(doc__name__contains='ietf-pim-msdp-yang'): - # ...: print(rq) - # Early review on draft-ietf-pim-msdp-yang by YANG Doctors Assigned - # Last Call review on draft-ietf-pim-msdp-yang by YANG Doctors Assigned - # Last Call review on draft-ietf-pim-msdp-yang by Routing Area Directorate Assigned - # Last Call review on draft-ietf-pim-msdp-yang by General Area Review Team (Gen-ART) Overtaken by Events - # Last Call review on draft-ietf-pim-msdp-yang by Security Area Directorate Overtaken by Events - # Last Call review on draft-ietf-pim-msdp-yang by Ops Directorate Assigned - # Last Call review on draft-ietf-pim-msdp-yang by Transport Area Review Team Team Will not Review Document - # Telechat review on draft-ietf-pim-msdp-yang by Security Area Directorate Team Will not Review Version - # In [107]: for a in ReviewAssignment.objects.filter(review__name__contains='review-ietf-pim-msdp-yang').order_by('review__name'): - # ...: print (a) - # Assignment for Reshad Rahman (Completed) : yangdoctors Early of draft-ietf-pim-msdp-yang - # Assignment for Yingzhen Qu (Completed) : rtgdir Last Call of draft-ietf-pim-msdp-yang - # Assignment for Reshad Rahman (Completed) : yangdoctors Last Call of draft-ietf-pim-msdp-yang - # In [105]: for d in Document.objects.filter(name__contains='review-ietf-pim-msdp-yang').order_by('name'): - # ...: print(d.name, d.time) - # review-ietf-pim-msdp-yang-01-yangdoctors-early-rahman-2018-01-12 2020-02-12 15:00:06 - # review-ietf-pim-msdp-yang-08-rtgdir-lc-qu-2020-01-20 2020-01-20 15:00:00 - # review-ietf-pim-msdp-yang-12-secdir-lc-roca-2020-01-29 2020-01-29 00:18:10 - # review-ietf-pim-msdp-yang-12-yangdoctors-lc-rahman-2020-01-28 2020-01-28 19:05:44 - # review-ietf-pim-msdp-yang-16-yangdoctors-lc-rahman-2020-03-20 2020-03-20 06:16:25 - # But - # In [113]: for a in ReviewAssignment.objects.filter(review_request__doc__name='draft-ietf-pim-msdp-yang'): - # ...: print(a) - # Assignment for Reshad Rahman (Completed) : yangdoctors Early of draft-ietf-pim-msdp-yang - # Assignment for Reshad Rahman (Completed) : yangdoctors Last Call of draft-ietf-pim-msdp-yang - # Assignment for Yingzhen Qu (Completed) : rtgdir Last Call of draft-ietf-pim-msdp-yang - # Assignment for Meral Shirazipour (No Response) : genart Last Call of draft-ietf-pim-msdp-yang - # Assignment for Vincent Roca (No Response) : secdir Last Call of draft-ietf-pim-msdp-yang - # Assignment for Shwetha Bhandari (Accepted) : opsdir Last Call of draft-ietf-pim-msdp-yang - # So, the secdir assignment that exists needs to have the review added and its state changed. - a = ReviewAssignment.objects.get(review_request__doc__name='draft-ietf-pim-msdp-yang',review_request__type_id='lc',reviewer__person__name="Vincent Roca") - r = Document.objects.get(name='review-ietf-pim-msdp-yang-12-secdir-lc-roca-2020-01-29') - a.review = r - a.state_id = 'completed' - a.completed_on = r.time - a.reviewed_rev = '12' - a.save() - # A new ReviewAssignment needs to be added to point to the yangdoctor review of -12 - a16 = ReviewAssignment.objects.get(review__name='review-ietf-pim-msdp-yang-16-yangdoctors-lc-rahman-2020-03-20') - r12 = Document.objects.get(name='review-ietf-pim-msdp-yang-12-yangdoctors-lc-rahman-2020-01-28') - ReviewAssignment.objects.create( - review_request = a16.review_request, - state_id = 'completed', - reviewer = a16.reviewer, - # Intentionally not making up assigned_on - completed_on = r12.time, - review = r12, - reviewed_rev = '12', - result_id = 'ready-issues', - mailarch_url = r12.external_url, - ) - # The secdir review request state is not the best, but it's not worth changing that history. - # Intentionally not changing the state of the opsdir assignment - # No suggestions to delete anything here. - helper.add_comment('draft-ietf-pim-msdp-yang', 'Reconnected secdir lc review and changed assignment state to completed. Reconnected yangdoctors review of -12.') - - # review-ietf-spring-srv6-network-programming-17-opsdir-lc-romascanu-2020-08-20 2020-08-20 02:43:08 - # review-ietf-spring-srv6-network-programming-17-opsdir-lc-romascanu-2020-08-20-2 2020-08-20 02:43:08 - # Assignment currently points to -2. Another double-submit. Resolve as above. - a = ReviewAssignment.objects.get(review__name='review-ietf-spring-srv6-network-programming-17-opsdir-lc-romascanu-2020-08-20-2') - a.review = Document.objects.get(name='review-ietf-spring-srv6-network-programming-17-opsdir-lc-romascanu-2020-08-20') - a.save() - Document.objects.filter(name='review-ietf-spring-srv6-network-programming-17-opsdir-lc-romascanu-2020-08-20-2').delete() - helper.remove_file('review-ietf-spring-srv6-network-programming-17-opsdir-lc-romascanu-2020-08-20-2') - helper.add_comment('draft-ietf-spring-srv6-network-programming','Removed unintended duplicate opsdir lc review') - - # In [116]: ReviewAssignment.objects.filter(review__name__contains='review-ietf-stir-passport-divert-07- - # ...: opsdir').values_list('review__name',flat=True) - # Out[116]: - # In [117]: for d in Document.objects.filter(name__contains='review-ietf-stir-passport-divert-07-opsdir').order_by('name'): - # ...: print(d.name, d.time) - # review-ietf-stir-passport-divert-07-opsdir-lc-dunbar-2019-12-02 2019-12-02 14:59:57 - # review-ietf-stir-passport-divert-07-opsdir-lc-dunbar-2019-12-02-2 2019-12-02 14:59:57 - # Another double-submit. Same treatment as above. - a = ReviewAssignment.objects.get(review__name='review-ietf-stir-passport-divert-07-opsdir-lc-dunbar-2019-12-02-2') - a.review = Document.objects.get(name='review-ietf-stir-passport-divert-07-opsdir-lc-dunbar-2019-12-02') - a.save() - Document.objects.filter(name='review-ietf-stir-passport-divert-07-opsdir-lc-dunbar-2019-12-02-2').delete() - helper.remove_file('review-ietf-stir-passport-divert-07-opsdir-lc-dunbar-2019-12-02-2') - helper.add_comment('draft-ietf-stir-passport-divert','Removed unintended duplicate opsdir lc review') - - # After the above... - # In [57]: for d in Document.objects.filter(type_id='review',reviewassignment__isnull=True): - # ...: print (d) - # review-ietf-dots-architecture-15-tsvart-lc-tuexen-2020-01-27 - # There are no files on disk matching that and nothing references it. - Document.objects.filter(name='review-ietf-dots-architecture-15-tsvart-lc-tuexen-2020-01-27').delete() - -def reverse(apps, schema_editor): - # There is no point in trying to return to the broken objects - pass - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0024_auto_20200520_0017'), - ('doc','0036_orgs_vs_repos'), - ('person','0016_auto_20200807_0750'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/review/migrations/0026_repair_more_assignments.py b/ietf/review/migrations/0026_repair_more_assignments.py deleted file mode 100644 index 50c24f03be..0000000000 --- a/ietf/review/migrations/0026_repair_more_assignments.py +++ /dev/null @@ -1,98 +0,0 @@ -# Copyright The IETF Trust 2020 All Rights Reserved - -import os - -from django.conf import settings -from django.db import migrations - -class Helper(object): - - def __init__(self, review_path, comments_by, document_class): - self.review_path = review_path - self.comments_by = comments_by - self.document_class = document_class - - def remove_file(self,name): - filename = os.path.join(self.review_path, '{}.txt'.format(name)) - os.remove(filename) - - def rename_file(self, old_name, new_name): - old_filename = os.path.join(self.review_path, '{}.txt'.format(old_name)) - new_filename = os.path.join(self.review_path, '{}.txt'.format(new_name)) - os.rename(old_filename, new_filename) - - def add_comment(self, name, comment): - doc = self.document_class.objects.get(name=name) - doc.docevent_set.create( - type = 'added_comment', - by = self.comments_by, - rev = doc.rev, - desc = comment, - ) - -def forward(apps,schema_editor): - Document = apps.get_model('doc','Document') - Person = apps.get_model('person','Person') - - # The calculation of review_path makes the assumption that DOCUMENT_PATH_PATTERN only uses - # things that are invariant for review documents. For production, as of this commit, that's - # DOCUMENT_PATH_PATTERN = '/a/www/ietf-ftp/{doc.type_id}/'. There are plans to change that pattern - # soon to '/a/ietfdata/doc/{doc.type_id}/' - - helper = Helper( - review_path = settings.DOCUMENT_PATH_PATTERN.format(doc=Document.objects.filter(type_id='review').last()), - comments_by = Person.objects.get(name='(System)'), - document_class = Document, - ) - - # In [2]: for d in Document.objects.filter(name__startswith='review-ietf-capport-api-07-opsdir'): - # ...: print(d.name,d.time) - # review-ietf-capport-api-07-opsdir-lc-dunbar-2020-05-09 2020-05-09 14:59:40 - # review-ietf-capport-api-07-opsdir-lc-dunbar-2020-05-09-2 2020-05-09 15:06:44 - # This is similar to draft-ietf-capport-architecture-08-genart-lc-halpern... - # Only -2 exists on disk. - # But the Document for ...-2020-05-09 has not type or state - it was very incompletely set up - deleting it results in: - # (3, - # {'community.CommunityList_added_docs': 0, - # 'community.SearchRule_name_contains_index': 0, - # 'doc.RelatedDocument': 0, - # 'doc.DocumentAuthor': 0, - # 'doc.Document_states': 0, - # 'doc.Document_tags': 0, - # 'doc.Document_formal_languages': 0, - # 'doc.DocumentURL': 0, - # 'doc.DocExtResource': 0, - # 'doc.DocAlias_docs': 1, - # 'doc.DocReminder': 0, - # 'group.GroupMilestone_docs': 0, - # 'group.GroupMilestoneHistory_docs': 0, - # 'liaisons.LiaisonStatementAttachment': 0, - # 'meeting.SessionPresentation': 0, - # 'message.Message_related_docs': 0, - # 'review.ReviewWish': 0, - # 'doc.DocEvent': 1, - # 'doc.Document': 1}) - # Repairing that back to remove the -2 will hide more information than simply removing the incompletely set up document object. - # But the -2 document currently has no type (see #3145) - Document.objects.get(name='review-ietf-capport-api-07-opsdir-lc-dunbar-2020-05-09').delete() - review = Document.objects.get(name='review-ietf-capport-api-07-opsdir-lc-dunbar-2020-05-09-2') - review.type_id='review' - review.save() - helper.add_comment('draft-ietf-capport-api','Removed an unintended duplicate version of the opsdir lc review') - -def reverse(apps,schema_editor): - # There is no point in returning to the broken version - pass - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0025_repair_assignments'), - ('doc','0039_auto_20201109_0439'), - ('person','0018_auto_20201109_0439'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - ] diff --git a/ietf/review/migrations/0027_unique_constraint_for_reviewersettings.py b/ietf/review/migrations/0027_unique_constraint_for_reviewersettings.py deleted file mode 100644 index ecd192e1bc..0000000000 --- a/ietf/review/migrations/0027_unique_constraint_for_reviewersettings.py +++ /dev/null @@ -1,89 +0,0 @@ -# Generated by Django 2.2.17 on 2021-01-24 07:24 - -from django.db import migrations, models - - - -def forward(apps, schema_editor): - """Forward migration - - Attempts to reconcile and remove duplicate ReviewerSettings - """ - ReviewerSettings = apps.get_model('review', 'ReviewerSettings') - HistoricalReviewerSettings = apps.get_model('review', 'HistoricalReviewerSettings') - - def reconcile_duplicate_settings(duplicate_settings, histories): - """Helper to decide how to handle duplicate settings""" - team = duplicate_settings[0].team - person = duplicate_settings[0].person - assert(all((s.person == person and s.team == team for s in duplicate_settings))) - - print('\n>> Found duplicate settings for {}'.format(duplicate_settings[0])) - # In the DB as of Dec 2020, the only duplicate settings sets were pairs where the - # earlier PK had a change history and the latter PK did not. Based on this, assuming - # a change history indicates that the settings are important. If only one has history, - # use that one. If multiple have a history, throw an error. - settings_with_history = [s for s in duplicate_settings if histories[s.pk].count() > 0] - if len(settings_with_history) == 0: - duplicate_settings.sort(key=lambda s: s.pk) - keep = duplicate_settings[-1] - reason = 'chosen by pk' - elif len(settings_with_history) == 1: - keep = settings_with_history[0] - reason = 'chosen because has change history' - else: - # Don't try to guess what to do if multiple settings have change histories - raise RuntimeError( - 'Multiple ReviewerSettings with change history for {}. Please resolve manually.'.format( - settings_with_history[0] - ) - ) - - print('>> Keeping pk={} ({})'.format(keep.pk, reason)) - for settings in duplicate_settings: - if settings.pk != keep.pk: - print('>> Deleting pk={}'.format(settings.pk)) - settings.delete() - - # forward migration starts here - if ReviewerSettings.objects.count() == 0: - return # nothing to do - - records = dict() # list of records, keyed by (person_id, team_id) - for rs in ReviewerSettings.objects.all().order_by('pk'): - key = (rs.person_id, rs.team_id) - if key in records: - records[key].append(rs) - else: - records[key] = [rs] - - for duplicate_settings in records.values(): - if len(duplicate_settings) > 1: - histories = dict() - for ds in duplicate_settings: - histories[ds.pk] = HistoricalReviewerSettings.objects.filter( - id=ds.pk - ) - reconcile_duplicate_settings(duplicate_settings, histories) - -def reverse(apps, schema_editor): - """Reverse migration - - Does nothing, but no harm in reverse migration. - """ - pass - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0026_repair_more_assignments'), - ] - - operations = [ - migrations.RunPython(forward, reverse), - migrations.AddConstraint( - model_name='reviewersettings', - constraint=models.UniqueConstraint(fields=('team', 'person'), name='unique_reviewer_settings_per_team_person'), - ), - ] diff --git a/ietf/review/migrations/0028_auto_20220513_1456.py b/ietf/review/migrations/0028_auto_20220513_1456.py deleted file mode 100644 index 44cb9ae3b9..0000000000 --- a/ietf/review/migrations/0028_auto_20220513_1456.py +++ /dev/null @@ -1,49 +0,0 @@ -# Generated by Django 2.2.28 on 2022-05-13 14:56 - -from django.db import migrations, models - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0027_unique_constraint_for_reviewersettings'), - ] - - operations = [ - migrations.AlterModelOptions( - name='historicalreviewassignment', - options={'get_latest_by': ('history_date', 'history_id'), 'ordering': ('-history_date', '-history_id'), 'verbose_name': 'historical review assignment', 'verbose_name_plural': 'historical review assignments'}, - ), - migrations.AlterModelOptions( - name='historicalreviewersettings', - options={'get_latest_by': ('history_date', 'history_id'), 'ordering': ('-history_date', '-history_id'), 'verbose_name': 'historical reviewer settings', 'verbose_name_plural': 'historical reviewer settings'}, - ), - migrations.AlterModelOptions( - name='historicalreviewrequest', - options={'get_latest_by': ('history_date', 'history_id'), 'ordering': ('-history_date', '-history_id'), 'verbose_name': 'historical review request', 'verbose_name_plural': 'historical review requests'}, - ), - migrations.AlterModelOptions( - name='historicalunavailableperiod', - options={'get_latest_by': ('history_date', 'history_id'), 'ordering': ('-history_date', '-history_id'), 'verbose_name': 'historical unavailable period', 'verbose_name_plural': 'historical unavailable periods'}, - ), - migrations.AlterField( - model_name='historicalreviewassignment', - name='history_date', - field=models.DateTimeField(db_index=True), - ), - migrations.AlterField( - model_name='historicalreviewersettings', - name='history_date', - field=models.DateTimeField(db_index=True), - ), - migrations.AlterField( - model_name='historicalreviewrequest', - name='history_date', - field=models.DateTimeField(db_index=True), - ), - migrations.AlterField( - model_name='historicalunavailableperiod', - name='history_date', - field=models.DateTimeField(db_index=True), - ), - ] diff --git a/ietf/review/migrations/0029_use_timezone_now_for_review_models.py b/ietf/review/migrations/0029_use_timezone_now_for_review_models.py deleted file mode 100644 index a8a0558d4c..0000000000 --- a/ietf/review/migrations/0029_use_timezone_now_for_review_models.py +++ /dev/null @@ -1,29 +0,0 @@ -# Generated by Django 2.2.28 on 2022-07-12 11:24 - -from django.db import migrations, models -import django.utils.timezone - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0028_auto_20220513_1456'), - ] - - operations = [ - migrations.AlterField( - model_name='historicalreviewrequest', - name='time', - field=models.DateTimeField(default=django.utils.timezone.now), - ), - migrations.AlterField( - model_name='reviewrequest', - name='time', - field=models.DateTimeField(default=django.utils.timezone.now), - ), - migrations.AlterField( - model_name='reviewwish', - name='time', - field=models.DateTimeField(default=django.utils.timezone.now), - ), - ] diff --git a/ietf/review/migrations/0030_use_date_today_helper.py b/ietf/review/migrations/0030_use_date_today_helper.py deleted file mode 100644 index b006d276f4..0000000000 --- a/ietf/review/migrations/0030_use_date_today_helper.py +++ /dev/null @@ -1,24 +0,0 @@ -# Generated by Django 2.2.28 on 2022-10-18 15:43 - -from django.db import migrations, models -import ietf.utils.timezone - - -class Migration(migrations.Migration): - - dependencies = [ - ('review', '0029_use_timezone_now_for_review_models'), - ] - - operations = [ - migrations.AlterField( - model_name='historicalunavailableperiod', - name='start_date', - field=models.DateField(default=ietf.utils.timezone.date_today, help_text="Choose the start date so that you can still do a review if it's assigned just before the start date - this usually means you should mark yourself unavailable for assignment some time before you are actually away. The default is today.", null=True), - ), - migrations.AlterField( - model_name='unavailableperiod', - name='start_date', - field=models.DateField(default=ietf.utils.timezone.date_today, help_text="Choose the start date so that you can still do a review if it's assigned just before the start date - this usually means you should mark yourself unavailable for assignment some time before you are actually away. The default is today.", null=True), - ), - ] diff --git a/ietf/review/models.py b/ietf/review/models.py index f0ec780948..d1b50c7117 100644 --- a/ietf/review/models.py +++ b/ietf/review/models.py @@ -34,7 +34,7 @@ class ReviewerSettings(models.Model): min_interval = models.IntegerField(verbose_name="Can review at most", choices=INTERVALS, blank=True, null=True) filter_re = models.CharField(max_length=255, verbose_name="Filter regexp", blank=True, validators=[validate_regular_expression_string, ], - help_text="Draft names matching this regular expression should not be assigned") + help_text="Internet-Draft names matching this regular expression should not be assigned") skip_next = models.IntegerField(default=0, verbose_name="Skip next assignments") remind_days_before_deadline = models.IntegerField(null=True, blank=True, help_text="To get an email reminder in case you forget to do an assigned review, enter the number of days before review deadline you want to receive it. Clear the field if you don't want this reminder.") remind_days_open_reviews = models.PositiveIntegerField(null=True, blank=True, verbose_name="Periodic reminder of open reviews every X days", help_text="To get a periodic email reminder of all your open reviews, enter the number of days between these reminders. Clear the field if you don't want these reminders.") @@ -143,6 +143,10 @@ def all_completed_assignments_for_doc(self): def request_closed_time(self): return self.doc.request_closed_time(self) or self.time + def add_history(self, description): + self._change_reason = description + self.save() + class ReviewAssignment(models.Model): """ One of possibly many reviews assigned in response to a ReviewRequest """ history = HistoricalRecords(history_change_reason_field=models.TextField(null=True)) @@ -191,6 +195,7 @@ class ReviewTeamSettings(models.Model): """Holds configuration specific to groups that are review teams""" group = OneToOneField(Group) autosuggest = models.BooleanField(default=True, verbose_name="Automatically suggest possible review requests") + allow_reviewer_to_reject_after_deadline = models.BooleanField(default=False, verbose_name="Allow reviewer to reject request after deadline.") reviewer_queue_policy = models.ForeignKey(ReviewerQueuePolicyName, default='RotateAlphabetically', on_delete=models.PROTECT) review_types = models.ManyToManyField(ReviewTypeName, default=get_default_review_types) review_results = models.ManyToManyField(ReviewResultName, default=get_default_review_results, related_name='reviewteamsettings_review_results_set') diff --git a/ietf/review/policies.py b/ietf/review/policies.py index 8799cd3d50..91398a1b24 100644 --- a/ietf/review/policies.py +++ b/ietf/review/policies.py @@ -1,4 +1,4 @@ -# Copyright The IETF Trust 2019-2021, All Rights Reserved +# Copyright The IETF Trust 2019-2023, All Rights Reserved import re @@ -7,9 +7,10 @@ from django.utils import timezone from simple_history.utils import bulk_update_with_history -from ietf.doc.models import DocumentAuthor, DocAlias +from ietf.doc.models import DocumentAuthor from ietf.doc.utils import extract_complete_replaces_ancestor_mapping_for_docs from ietf.group.models import Role +from ietf.name.models import ReviewAssignmentStateName from ietf.person.models import Person import debug # pyflakes:ignore from ietf.review.models import NextReviewerInTeam, ReviewerSettings, ReviewWish, ReviewRequest, \ @@ -23,7 +24,7 @@ """ This file contains policies regarding reviewer queues. The policies are documented in more detail on: -https://trac.ietf.org/trac/ietfdb/wiki/ReviewerQueuePolicy +https://github.com/ietf-tools/datatracker/wiki/ReviewerQueuePolicy Terminology used here should match terminology used in that document. """ @@ -55,8 +56,6 @@ def persons_with_previous_review(team, review_req, possible_person_ids, state_id reviewassignment__state=state_id, team=team, ).distinct() - if review_req.pk is not None: - has_reviewed_previous = has_reviewed_previous.exclude(pk=review_req.pk) has_reviewed_previous = set( has_reviewed_previous.values_list("reviewassignment__reviewer__person", flat=True)) return has_reviewed_previous @@ -70,7 +69,14 @@ def assign_reviewer(self, review_req, reviewer, add_skip): """Assign a reviewer to a request and update policy state accordingly""" # Update policy state first - needed by LRU policy to correctly compute whether assignment was in-order self.update_policy_state_for_assignment(review_req, reviewer.person, add_skip) - return review_req.reviewassignment_set.create(state_id='assigned', reviewer=reviewer, assigned_on=timezone.now()) + assignment = review_req.reviewassignment_set.filter(reviewer=reviewer).first() + if assignment: + assignment.state = ReviewAssignmentStateName.objects.get(slug='assigned', used=True) + assignment.assigned_on = timezone.now() + assignment.save() + return assignment + else: + return review_req.reviewassignment_set.create(state_id='assigned', reviewer=reviewer, assigned_on=timezone.now()) def default_reviewer_rotation_list(self, include_unavailable=False): """ Return a list of reviewers (Person objects) in the default reviewer rotation for a policy. @@ -84,7 +90,7 @@ def default_reviewer_rotation_list(self, include_unavailable=False): rotation_list = self._filter_unavailable_reviewers(rotation_list) return rotation_list - def return_reviewer_to_rotation_top(self, reviewer_person): + def set_wants_to_be_next(self, reviewer_person): """ Return a reviewer to the top of the rotation, e.g. because they rejected a review, and should retroactively not have been rotated over. @@ -125,12 +131,15 @@ def _update_skip_next(self, rotation_pks, assignee_person): assignee_index = rotation_pks.index(assignee_person.pk) skipped = rotation_pks[0:assignee_index] skipped_settings = self.team.reviewersettings_set.filter(person__in=skipped) # list of PKs is valid here + changed = [] for ss in skipped_settings: - ss.skip_next = max(0, ss.skip_next - 1) # ensure we don't go negative - bulk_update_with_history(skipped_settings, + if ss.skip_next > 0: + ss.skip_next = max(0, ss.skip_next - 1) # ensure we don't go negative + ss._change_reason = "Skip count decremented" + changed.append(ss) + bulk_update_with_history(changed, ReviewerSettings, - ['skip_next'], - default_change_reason='skipped') + ['skip_next']) def _assignment_in_order(self, rotation_pks, assignee_person): """Is this an in-order assignment?""" @@ -168,24 +177,27 @@ def setup_reviewer_field(self, field, review_req): PersonEmailChoiceField(label="Assign Reviewer", empty_label="(None)") """ - # Collect a set of person IDs for people who have either not responded - # to or outright rejected reviewing this document in the past + # Collect a set of person IDs for people who have not responded + # to this document in the past rejecting_reviewer_ids = review_req.doc.reviewrequest_set.filter( - reviewassignment__state__slug__in=('rejected', 'no-response') + reviewassignment__state__slug='no-response' ).values_list( 'reviewassignment__reviewer__person_id', flat=True ) - # Query the Email objects for reviewers who haven't rejected or + # Query the Email objects for reviewers who haven't # not responded to this document in the past field.queryset = field.queryset.filter( role__name="reviewer", role__group=review_req.team ).exclude( person_id__in=rejecting_reviewer_ids ) - one_assignment = (review_req.reviewassignment_set - .exclude(state__slug__in=('rejected', 'no-response')) - .first()) + one_assignment = None + if review_req.pk is not None: + # cannot use reviewassignment_set relation until review_req has been created + one_assignment = (review_req.reviewassignment_set + .exclude(state__slug__in=('rejected', 'no-response')) + .first()) if one_assignment: field.initial = one_assignment.reviewer_id @@ -253,12 +265,15 @@ def _filter_unavailable_reviewers(self, reviewers, review_req=None): def _clear_request_next_assignment(self, person): s = self._reviewer_settings_for(person) - s.request_assignment_next = False - s.save() + if s.request_assignment_next: + s.request_assignment_next = False + s._change_reason = "Clearing request next assignment" + s.save() def _add_skip(self, person): s = self._reviewer_settings_for(person) s.skip_next += 1 + s._change_reason = "Incrementing skip count" s.save() def _reviewer_settings_for(self, person): @@ -284,8 +299,6 @@ def __init__(self, email_queryset, review_req, rotation_list): def _collect_context(self): """Collect all relevant data about this team, document and review request.""" - self.doc_aliases = DocAlias.objects.filter(docs=self.doc).values_list("name", flat=True) - # This data is collected as a dict, keys being person IDs, values being numbers/objects. self.rotation_index = {p.pk: i for i, p in enumerate(self.rotation_list)} self.reviewer_settings = self._reviewer_settings_for_person_ids(self.possible_person_ids) @@ -351,8 +364,7 @@ def format_period(p): add_boolean_score(+1, email.person_id in self.wish_to_review, "wishes to review document") add_boolean_score(-1, email.person_id in self.connections, self.connections.get(email.person_id)) # reviewer is somehow connected: bad - add_boolean_score(-1, settings.filter_re and any( - re.search(settings.filter_re, n) for n in self.doc_aliases), "filter regexp matches") + add_boolean_score(-1, settings.filter_re and re.search(settings.filter_re, self.doc.name), "filter regexp matches") # minimum interval between reviews days_needed = self.days_needed_for_reviewers.get(email.person_id, 0) @@ -472,12 +484,13 @@ def default_reviewer_rotation_list(self, include_unavailable=False): return reviewers[next_reviewer_index:] + reviewers[:next_reviewer_index] - def return_reviewer_to_rotation_top(self, reviewer_person): + def set_wants_to_be_next(self, reviewer_person): # As RotateAlphabetically does not keep a full rotation list, # returning someone to a particular order is complex. # Instead, the "assign me next" flag is set. settings = self._reviewer_settings_for(reviewer_person) settings.request_assignment_next = True + settings._change_reason = "Setting request next assignment" settings.save() def _update_skip_next(self, rotation_pks, assignee_person): @@ -517,20 +530,22 @@ def _update_skip_next(self, rotation_pks, assignee_person): min_skip_next = min([rs.skip_next for rs in rotation_settings.values()]) next_reviewer_index = None + changed = [] for index, pk in enumerate(unfolded_rotation_pks): rs = rotation_settings.get(pk) if (rs is None) or (rs.skip_next == min_skip_next): next_reviewer_index = index break else: - rs.skip_next = max(0, rs.skip_next - 1) # ensure never negative + if rs.skip_next > 0: + rs.skip_next = max(0, rs.skip_next - 1) # ensure never negative + rs._change_reason = "Skip count decremented" + changed.append(rs) log.assertion('next_reviewer_index is not None') # some entry in the list must have the minimum value - - bulk_update_with_history(rotation_settings.values(), + bulk_update_with_history(changed, ReviewerSettings, - ['skip_next'], - default_change_reason='skipped') + ['skip_next']) next_reviewer_pk = unfolded_rotation_pks[next_reviewer_index] NextReviewerInTeam.objects.update_or_create( @@ -566,11 +581,14 @@ def default_reviewer_rotation_list(self, include_unavailable=False): rotation_list += reviewers_with_assignment return rotation_list - def return_reviewer_to_rotation_top(self, reviewer_person): + def set_wants_to_be_next(self, reviewer_person): # Reviewer rotation for this policy ignores rejected/withdrawn # reviews, so it automatically adjusts the position of someone # who rejected a review and no further action is needed. - pass + settings = self._reviewer_settings_for(reviewer_person) + settings.request_assignment_next = True + settings._change_reason = "Setting request next assignment" + settings.save() QUEUE_POLICY_NAME_MAPPING = { diff --git a/ietf/review/tasks.py b/ietf/review/tasks.py new file mode 100644 index 0000000000..5d8afa6943 --- /dev/null +++ b/ietf/review/tasks.py @@ -0,0 +1,43 @@ +# Copyright The IETF Trust 2024, All Rights Reserved +# +# Celery task definitions +# +from celery import shared_task + +from ietf.review.utils import ( + review_assignments_needing_reviewer_reminder, email_reviewer_reminder, + review_assignments_needing_secretary_reminder, email_secretary_reminder, + send_unavailability_period_ending_reminder, send_reminder_all_open_reviews, + send_review_reminder_overdue_assignment, send_reminder_unconfirmed_assignments) +from ietf.utils.log import log +from ietf.utils.timezone import date_today, DEADLINE_TZINFO + + +@shared_task +def send_review_reminders_task(): + today = date_today(DEADLINE_TZINFO) + + for assignment in review_assignments_needing_reviewer_reminder(today): + email_reviewer_reminder(assignment) + log("Emailed reminder to {} for review of {} in {} (req. id {})".format(assignment.reviewer.address, assignment.review_request.doc_id, assignment.review_request.team.acronym, assignment.review_request.pk)) + + for assignment, secretary_role in review_assignments_needing_secretary_reminder(today): + email_secretary_reminder(assignment, secretary_role) + review_req = assignment.review_request + log("Emailed reminder to {} for review of {} in {} (req. id {})".format(secretary_role.email.address, review_req.doc_id, review_req.team.acronym, review_req.pk)) + + period_end_reminders_sent = send_unavailability_period_ending_reminder(today) + for msg in period_end_reminders_sent: + log(msg) + + overdue_reviews_reminders_sent = send_review_reminder_overdue_assignment(today) + for msg in overdue_reviews_reminders_sent: + log(msg) + + open_reviews_reminders_sent = send_reminder_all_open_reviews(today) + for msg in open_reviews_reminders_sent: + log(msg) + + unconfirmed_assignment_reminders_sent = send_reminder_unconfirmed_assignments(today) + for msg in unconfirmed_assignment_reminders_sent: + log(msg) diff --git a/ietf/review/tests.py b/ietf/review/tests.py index f76d03f56d..5dc8f11e8e 100644 --- a/ietf/review/tests.py +++ b/ietf/review/tests.py @@ -1,19 +1,27 @@ # Copyright The IETF Trust 2019-2020, All Rights Reserved # -*- coding: utf-8 -*- import datetime +from unittest import mock +import debug # pyflakes:ignore + +from pyquery import PyQuery from ietf.group.factories import RoleFactory +from ietf.doc.factories import WgDraftFactory from ietf.utils.mail import empty_outbox, get_payload_text, outbox from ietf.utils.test_utils import TestCase, reload_db_objects +from ietf.utils.test_utils import login_testing_unauthorized, unicontent from ietf.utils.timezone import date_today, datetime_from_date from .factories import ReviewAssignmentFactory, ReviewRequestFactory, ReviewerSettingsFactory from .mailarch import hash_list_message_id from .models import ReviewerSettings, ReviewSecretarySettings, ReviewTeamSettings, UnavailablePeriod +from .tasks import send_review_reminders_task from .utils import (email_secretary_reminder, review_assignments_needing_secretary_reminder, email_reviewer_reminder, review_assignments_needing_reviewer_reminder, send_reminder_unconfirmed_assignments, send_review_reminder_overdue_assignment, send_reminder_all_open_reviews, send_unavailability_period_ending_reminder, ORIGIN_DATE_PERIODIC_REMINDERS) +from django.urls import reverse as urlreverse class HashTest(TestCase): @@ -508,3 +516,103 @@ def test_send_reminder_all_open_reviews(self): self.assertTrue(self.reviewer.email_address() in log[0]) self.assertTrue('1 open review' in log[0]) +class AddReviewCommentTestCase(TestCase): + def test_review_add_comment(self): + draft = WgDraftFactory(name='draft-ietf-mars-test',group__acronym='mars') + review_req = ReviewRequestFactory(doc=draft, state_id='assigned') + ReviewAssignmentFactory(review_request=review_req, state_id='assigned') + + url_post = urlreverse('ietf.doc.views_review.add_request_comment', kwargs=dict(name=draft.name, request_id=review_req.pk)) + url_page = urlreverse('ietf.doc.views_review.review_request', kwargs=dict(name=draft.name, request_id=review_req.pk)) + + login_testing_unauthorized(self, "secretary", url_post) + + # Check that we do not have entry on the page + r = self.client.get(url_page) + self.assertEqual(r.status_code, 200) + # Needs to have history + self.assertContains(r, 'History') + # But can't have the comment we are goint to add. + self.assertNotContains(r, 'This is a test.') + + # Get the form + r = self.client.get(url_post) + self.assertEqual(r.status_code, 200) + q = PyQuery(unicontent(r)) + self.assertEqual(len(q('form textarea[name=comment]')), 1) + + # Post the comment + r = self.client.post(url_post, dict(comment="This is a test.")) + self.assertEqual(r.status_code, 302) + + # Get the main page again + r = self.client.get(url_page) + self.assertEqual(r.status_code, 200) + # Needs to have history + self.assertContains(r, 'History') + # But can't have the comment we are goint to add. + self.assertContains(r, 'This is a test.') + + +class TaskTests(TestCase): + # hyaaa it's mockzilla + @mock.patch("ietf.review.tasks.date_today") + @mock.patch("ietf.review.tasks.review_assignments_needing_reviewer_reminder") + @mock.patch("ietf.review.tasks.email_reviewer_reminder") + @mock.patch("ietf.review.tasks.review_assignments_needing_secretary_reminder") + @mock.patch("ietf.review.tasks.email_secretary_reminder") + @mock.patch("ietf.review.tasks.send_unavailability_period_ending_reminder") + @mock.patch("ietf.review.tasks.send_reminder_all_open_reviews") + @mock.patch("ietf.review.tasks.send_review_reminder_overdue_assignment") + @mock.patch("ietf.review.tasks.send_reminder_unconfirmed_assignments") + def test_send_review_reminders_task( + self, + mock_send_reminder_unconfirmed_assignments, + mock_send_review_reminder_overdue_assignment, + mock_send_reminder_all_open_reviews, + mock_send_unavailability_period_ending_reminder, + mock_email_secretary_reminder, + mock_review_assignments_needing_secretary_reminder, + mock_email_reviewer_reminder, + mock_review_assignments_needing_reviewer_reminder, + mock_date_today, + ): + """Test that send_review_reminders calls functions correctly + + Does not test individual methods, just that they are called as expected. + """ + mock_today = object() + assignment = ReviewAssignmentFactory() + secretary_role = RoleFactory(name_id="secr") + + mock_date_today.return_value = mock_today + mock_review_assignments_needing_reviewer_reminder.return_value = [assignment] + mock_review_assignments_needing_secretary_reminder.return_value = [[assignment, secretary_role]] + mock_send_unavailability_period_ending_reminder.return_value = ["pretending I sent a period end reminder"] + mock_send_review_reminder_overdue_assignment.return_value = ["pretending I sent an overdue reminder"] + mock_send_reminder_all_open_reviews.return_value = ["pretending I sent an open review reminder"] + mock_send_reminder_unconfirmed_assignments.return_value = ["pretending I sent an unconfirmed reminder"] + + send_review_reminders_task() + + self.assertEqual(mock_review_assignments_needing_reviewer_reminder.call_count, 1) + self.assertEqual(mock_review_assignments_needing_reviewer_reminder.call_args[0], (mock_today,)) + self.assertEqual(mock_email_reviewer_reminder.call_count, 1) + self.assertEqual(mock_email_reviewer_reminder.call_args[0], (assignment,)) + + self.assertEqual(mock_review_assignments_needing_secretary_reminder.call_count, 1) + self.assertEqual(mock_review_assignments_needing_secretary_reminder.call_args[0], (mock_today,)) + self.assertEqual(mock_email_secretary_reminder.call_count, 1) + self.assertEqual(mock_email_secretary_reminder.call_args[0], (assignment, secretary_role)) + + self.assertEqual(mock_send_unavailability_period_ending_reminder.call_count, 1) + self.assertEqual(mock_send_unavailability_period_ending_reminder.call_args[0], (mock_today,)) + + self.assertEqual(mock_send_review_reminder_overdue_assignment.call_count, 1) + self.assertEqual(mock_send_review_reminder_overdue_assignment.call_args[0], (mock_today,)) + + self.assertEqual(mock_send_reminder_all_open_reviews.call_count, 1) + self.assertEqual(mock_send_reminder_all_open_reviews.call_args[0], (mock_today,)) + + self.assertEqual(mock_send_reminder_unconfirmed_assignments.call_count, 1) + self.assertEqual(mock_send_reminder_unconfirmed_assignments.call_args[0], (mock_today,)) diff --git a/ietf/review/tests_policies.py b/ietf/review/tests_policies.py index b675b2384b..1eb57a8bec 100644 --- a/ietf/review/tests_policies.py +++ b/ietf/review/tests_policies.py @@ -1,4 +1,4 @@ -# Copyright The IETF Trust 2016-2021, All Rights Reserved +# Copyright The IETF Trust 2016-2023, All Rights Reserved import debug # pyflakes:ignore import datetime @@ -115,7 +115,7 @@ def reviewer_settings_for(self, person): return (ReviewerSettings.objects.filter(team=self.team, person=person).first() or ReviewerSettings(team=self.team, person=person)) - def test_return_reviewer_to_rotation_top(self): + def test_set_wants_to_be_next(self): # Subclass must implement this raise NotImplementedError @@ -471,9 +471,6 @@ def test_setup_reviewer_field(self): addresses = list( map( lambda choice: choice[0], field.choices ) ) - self.assertNotIn( - str(rejected_reviewer.email()), addresses, - "Reviews should not suggest people who have rejected this request in the past") self.assertNotIn( str(no_response_reviewer.email()), addresses, "Reviews should not suggest people who have not responded to this request in the past.") @@ -507,10 +504,10 @@ def test_default_reviewer_rotation_list_with_nextreviewerinteam(self): rotation = self.policy.default_reviewer_rotation_list() self.assertEqual(rotation, available_reviewers[2:] + available_reviewers[:1]) - def test_return_reviewer_to_rotation_top(self): + def test_set_wants_to_be_next(self): reviewer = self.append_reviewer() - self.policy.return_reviewer_to_rotation_top(reviewer) - self.assertTrue(ReviewerSettings.objects.get(person=reviewer).request_assignment_next) + self.policy.set_wants_to_be_next(reviewer) + self.assertTrue(self.reviewer_settings_for(reviewer).request_assignment_next) def test_update_policy_state_for_assignment(self): # make a bunch of reviewers @@ -723,9 +720,10 @@ def test_default_review_rotation_list_uses_assigned_on_date(self): self.assertEqual(self.policy.default_reviewer_rotation_list(), available_reviewers[2:] + [first_reviewer, second_reviewer]) - def test_return_reviewer_to_rotation_top(self): - # Should do nothing, this is implicit in this policy, no state change is needed. - self.policy.return_reviewer_to_rotation_top(self.append_reviewer()) + def test_set_wants_to_be_next(self): + reviewer = self.append_reviewer() + self.policy.set_wants_to_be_next(reviewer) + self.assertTrue(self.reviewer_settings_for(reviewer).request_assignment_next) def test_assign_reviewer_updates_skip_next_without_add_skip(self): """Skipping reviewers with add_skip=False should update skip_counts properly""" @@ -835,7 +833,7 @@ def test_determine_ranking(self): self.assertEqual(len(ranking), 2) self.assertEqual(ranking[0]['email'], reviewer_high.email()) self.assertEqual(ranking[1]['email'], reviewer_low.email()) - # These scores follow the ordering of https://trac.ietf.org/trac/ietfdb/wiki/ReviewerQueuePolicy, + # These scores follow the ordering of https://github.com/ietf-tools/datatracker/wiki/ReviewerQueuePolicy, self.assertEqual(ranking[0]['scores'], [ 1, 1, 1, 1, 1, 1, 0, 0, -1]) self.assertEqual(ranking[1]['scores'], [-1, -1, -1, -1, -1, -1, -91, -2, 0]) self.assertEqual(ranking[0]['label'], 'Test Reviewer-high: unavailable indefinitely (Can do follow-ups); requested to be selected next for assignment; reviewed document before; wishes to review document; #2; 1 no response, 1 partially complete, 1 fully completed') diff --git a/ietf/review/utils.py b/ietf/review/utils.py index 032b1fd418..61494738d3 100644 --- a/ietf/review/utils.py +++ b/ietf/review/utils.py @@ -1,4 +1,4 @@ -# Copyright The IETF Trust 2016-2020, All Rights Reserved +# Copyright The IETF Trust 2016-2023, All Rights Reserved # -*- coding: utf-8 -*- @@ -50,6 +50,8 @@ def can_request_review_of_doc(user, doc): if not user.is_authenticated: return False + # This is in a strange place as it has nothing to do with the user + # but this utility is used in too many places to move this quickly. if doc.type_id == 'draft' and doc.get_state_slug() != 'active': return False @@ -79,6 +81,11 @@ def review_assignments_to_list_for_docs(docs): return extract_revision_ordered_review_assignments_for_documents_and_replaced(assignment_qs, doc_names) +def review_requests_to_list_for_docs(docs): + review_requests_qs = ReviewRequest.objects.filter(Q(state_id='requested')) + doc_names = [d.name for d in docs] + return extract_revision_ordered_review_requests_for_documents_and_replaced(review_requests_qs, doc_names) + def augment_review_requests_with_events(review_reqs): req_dict = { r.pk: r for r in review_reqs } for e in ReviewRequestDocEvent.objects.filter(review_request__in=review_reqs, type__in=["assigned_review_request", "closed_review_request"]).order_by("time"): @@ -308,7 +315,7 @@ def email_review_assignment_change(request, review_assignment, subject, msg, by, doc=review_assignment.review_request.doc, group=review_assignment.review_request.team, review_assignment=review_assignment, - skip_review_secretary=not notify_secretary, + skip_secretary=not notify_secretary, skip_review_reviewer=not notify_reviewer, skip_review_requested_by=not notify_requested_by, ) @@ -328,11 +335,11 @@ def email_review_request_change(request, review_req, subject, msg, by, notify_se was done by that party.""" (to, cc) = gather_address_lists( 'review_req_changed', - skipped_recipients=[Person.objects.get(name="(System)").formatted_email(), by.email_address()], + skipped_recipients=[Person.objects.get(name="(System)").formatted_email()], doc=review_req.doc, group=review_req.team, review_request=review_req, - skip_review_secretary=not notify_secretary, + skip_secretary=not notify_secretary, skip_review_reviewer=not notify_reviewer, skip_review_requested_by=not notify_requested_by, ) @@ -382,8 +389,13 @@ def assign_review_request_to_reviewer(request, review_req, reviewer, add_skip=Fa # with a different view on a ReviewAssignment. log.assertion('reviewer is not None') - if review_req.reviewassignment_set.filter(reviewer=reviewer).exists(): - return + # cannot reference reviewassignment_set relation until pk exists + if review_req.pk is not None: + reviewassignment_set = review_req.reviewassignment_set.filter(reviewer=reviewer) + if (reviewassignment_set.exists() and not + (reviewassignment_set.filter(state_id='rejected').exists() or + reviewassignment_set.filter(state_id='withdrawn').exists())): + return # Note that assigning a review no longer unassigns other reviews @@ -430,9 +442,9 @@ def assign_review_request_to_reviewer(request, review_req, reviewer, add_skip=Fa email_review_request_change( request, review_req, - "%s %s assignment: %s" % (review_req.team.acronym.capitalize(), review_req.type.name,review_req.doc.name), + "For %s, %s %s review by %s assigned: %s" % (reviewer.person.name, review_req.team.acronym.capitalize(), review_req.type.name, review_req.deadline, review_req.doc.name), msg , - by=request.user.person, notify_secretary=False, notify_reviewer=True, notify_requested_by=False) + by=request.user.person, notify_secretary=True, notify_reviewer=True, notify_requested_by=True) def close_review_request(request, review_req, close_state, close_comment=''): @@ -588,17 +600,21 @@ def blocks(existing, request): and existing.reviewassignment_set.filter(state_id__in=("assigned", "accepted")).exists() and (not existing.requested_rev or existing.requested_rev == request.doc.rev)) request_closed = existing.state_id not in ('requested','assigned') + # Is there a review request for this document already in system + requested = existing.state_id in ('requested') and (not existing.requested_rev or existing.requested_rev == request.doc.rev) # at least one assignment was completed for the requested version or the current doc version if no specific version was requested: some_assignment_completed = existing.reviewassignment_set.filter(reviewed_rev=existing.requested_rev or existing.doc.rev, state_id='completed').exists() - return any([no_review_document, no_review_rev, pending, request_closed, some_assignment_completed]) + return any([no_review_document, no_review_rev, pending, request_closed, requested, some_assignment_completed]) res = [r for r in requests.values() if not any(blocks(e, r) for e in existing_requests[r.doc_id])] res.sort(key=lambda r: (r.deadline, r.doc_id), reverse=True) return res -def extract_revision_ordered_review_assignments_for_documents_and_replaced(review_assignment_queryset, names): +def extract_revision_ordered_review_assignments_for_documents_and_replaced( + review_assignment_queryset, names +): """Extracts all review assignments for document names (including replaced ancestors), return them neatly sorted.""" names = set(names) @@ -607,8 +623,13 @@ def extract_revision_ordered_review_assignments_for_documents_and_replaced(revie assignments_for_each_doc = defaultdict(list) replacement_name_set = set(e for l in replaces.values() for e in l) | names - for r in ( review_assignment_queryset.filter(review_request__doc__name__in=replacement_name_set) - .order_by("-reviewed_rev","-assigned_on", "-id").iterator()): + for r in ( + review_assignment_queryset.filter( + review_request__doc__name__in=replacement_name_set + ) + .order_by("-reviewed_rev", "-assigned_on", "-id") + .iterator(chunk_size=2000) # chunk_size not tested, using pre-Django 5 default value + ): assignments_for_each_doc[r.review_request.doc.name].append(r) # now collect in breadth-first order to keep the revision order intact @@ -646,7 +667,10 @@ def extract_revision_ordered_review_assignments_for_documents_and_replaced(revie return res -def extract_revision_ordered_review_requests_for_documents_and_replaced(review_request_queryset, names): + +def extract_revision_ordered_review_requests_for_documents_and_replaced( + review_request_queryset, names +): """Extracts all review requests for document names (including replaced ancestors), return them neatly sorted.""" names = set(names) @@ -654,7 +678,13 @@ def extract_revision_ordered_review_requests_for_documents_and_replaced(review_r replaces = extract_complete_replaces_ancestor_mapping_for_docs(names) requests_for_each_doc = defaultdict(list) - for r in review_request_queryset.filter(doc__name__in=set(e for l in replaces.values() for e in l) | names).order_by("-time", "-id").iterator(): + for r in ( + review_request_queryset.filter( + doc__name__in=set(e for l in replaces.values() for e in l) | names + ) + .order_by("-time", "-id") + .iterator(chunk_size=2000) # chunk_size not tested, using pre-Django 5 default value + ): requests_for_each_doc[r.doc.name].append(r) # now collect in breadth-first order to keep the revision order intact diff --git a/ietf/secr/announcement/forms.py b/ietf/secr/announcement/forms.py index 3aacbfe622..91004ea270 100644 --- a/ietf/secr/announcement/forms.py +++ b/ietf/secr/announcement/forms.py @@ -14,93 +14,133 @@ # Globals # --------------------------------------------- -TO_LIST = ('IETF Announcement List ', - 'I-D Announcement List ', - 'RFP Announcement List ', - 'The IESG ', - 'Working Group Chairs ', - 'BOF Chairs ', - 'Other...') +TO_LIST = ( + "IETF Announcement List ", + "I-D Announcement List ", + "RFP Announcement List ", + "The IESG ", + "Working Group Chairs ", + "BOF Chairs ", + "Other...", +) # --------------------------------------------- # Helper Functions # --------------------------------------------- + def get_from_choices(user): - ''' + """ This function returns a choices tuple containing all the Announced From choices. Including leadership chairs and other entities. - ''' + """ addresses = [] - if has_role(user,'Secretariat'): - addresses = AnnouncementFrom.objects.values_list('address', flat=True).order_by('address').distinct() + if has_role(user, "Secretariat"): + addresses = ( + AnnouncementFrom.objects.values_list("address", flat=True) + .order_by("address") + .distinct() + ) else: for role in user.person.role_set.all(): - addresses.extend(AnnouncementFrom.objects.filter(name=role.name, group=role.group).values_list('address', flat=True).order_by('address')) + addresses.extend( + AnnouncementFrom.objects.filter(name=role.name, group=role.group) + .values_list("address", flat=True) + .order_by("address") + ) nomcom_choices = get_nomcom_choices(user) if nomcom_choices: addresses = list(addresses) + nomcom_choices - - return list(zip(addresses, addresses)) + + choices = list(zip(addresses, addresses)) + if len(choices) > 1: + choices.insert(0, ("", "(Choose an option)")) + return choices def get_nomcom_choices(user): - ''' + """ Returns the list of nomcom email addresses for given user - ''' - nomcoms = Role.objects.filter(name="chair", - group__acronym__startswith="nomcom", - group__state="active", - group__type="nomcom", - person=user.person) + """ + nomcoms = Role.objects.filter( + name="chair", + group__acronym__startswith="nomcom", + group__state="active", + group__type="nomcom", + person=user.person, + ) addresses = [] for nomcom in nomcoms: year = nomcom.group.acronym[-4:] - addresses.append('NomCom Chair %s ' % (year,year)) + addresses.append("NomCom Chair %s " % (year, year)) return addresses - + def get_to_choices(): - return list(zip(TO_LIST,TO_LIST)) + return list(zip(TO_LIST, TO_LIST)) # --------------------------------------------- # Forms # --------------------------------------------- + class AnnounceForm(forms.ModelForm): - nomcom = forms.ModelChoiceField(queryset=Group.objects.filter(acronym__startswith='nomcom',type='nomcom',state='active'),required=False) + nomcom = forms.ModelChoiceField( + queryset=Group.objects.filter( + acronym__startswith="nomcom", type="nomcom", state="active" + ), + required=False, + ) to_custom = MultiEmailField(required=False) class Meta: model = Message - fields = ('nomcom', 'to','to_custom','frm','cc','bcc','reply_to','subject','body') + fields = ( + "nomcom", + "to", + "to_custom", + "frm", + "cc", + "bcc", + "reply_to", + "subject", + "body", + ) + labels = {"frm": "From"} + help_texts = { + "to": "Select name OR select Other... and enter email below", + "cc": "Use comma separated lists for emails (Cc, Bcc, Reply To)", + } def __init__(self, *args, **kwargs): - if 'hidden' in kwargs: - self.hidden = kwargs.pop('hidden') + if "hidden" in kwargs: + self.hidden = kwargs.pop("hidden") else: self.hidden = False - user = kwargs.pop('user') + user = kwargs.pop("user") person = user.person super(AnnounceForm, self).__init__(*args, **kwargs) - self.fields['to'].widget = forms.Select(choices=get_to_choices()) - self.fields['to'].help_text = 'Select name OR select Other... and enter email below' - self.fields['cc'].help_text = 'Use comma separated lists for emails (Cc, Bcc, Reply To)' - self.fields['frm'].widget = forms.Select(choices=get_from_choices(user)) - self.fields['frm'].label = 'From' - self.fields['reply_to'].required = True - self.fields['nomcom'].label = 'NomCom message:' - nomcom_roles = person.role_set.filter(group__in=self.fields['nomcom'].queryset,name='chair') - secr_roles = person.role_set.filter(group__acronym='secretariat',name='secr') + self.fields["to"].widget = forms.Select(choices=get_to_choices()) + self.fields["frm"].widget = forms.Select(choices=get_from_choices(user)) + self.fields["reply_to"].required = True + # nomcom field is defined declaratively so label and help_text must be set here + self.fields["nomcom"].label = "NomCom message:" + self.fields["nomcom"].help_text = ( + "If this is a NomCom announcement specify which NomCom group here" + ) + nomcom_roles = person.role_set.filter( + group__in=self.fields["nomcom"].queryset, name="chair" + ) + secr_roles = person.role_set.filter(group__acronym="secretariat", name="secr") if nomcom_roles: - self.initial['nomcom'] = nomcom_roles[0].group.pk + self.initial["nomcom"] = nomcom_roles[0].group.pk if not nomcom_roles and not secr_roles: - self.fields['nomcom'].widget = forms.HiddenInput() - + self.fields["nomcom"].widget = forms.HiddenInput() + if self.hidden: for key in list(self.fields.keys()): self.fields[key].widget = forms.HiddenInput() @@ -110,25 +150,29 @@ def clean(self): data = self.cleaned_data if self.errors: return self.cleaned_data - if data['to'] == 'Other...' and not data['to_custom']: + if data["to"] == "Other..." and not data["to_custom"]: raise forms.ValidationError('You must enter a "To" email address') - for k in ['to', 'frm', 'cc',]: + for k in [ + "to", + "frm", + "cc", + ]: data[k] = unescape(data[k]) return data def save(self, *args, **kwargs): - user = kwargs.pop('user') + user = kwargs.pop("user") message = super(AnnounceForm, self).save(commit=False) message.by = user.person - if self.cleaned_data['to'] == 'Other...': - message.to = self.cleaned_data['to_custom'] - if kwargs['commit']: + if self.cleaned_data["to"] == "Other...": + message.to = self.cleaned_data["to_custom"] + if kwargs["commit"]: message.save() # handle nomcom message - nomcom = self.cleaned_data.get('nomcom',False) + nomcom = self.cleaned_data.get("nomcom", False) if nomcom: message.related_groups.add(nomcom) - return message \ No newline at end of file + return message diff --git a/ietf/secr/announcement/tests.py b/ietf/secr/announcement/tests.py index c50e997f97..f08e824397 100644 --- a/ietf/secr/announcement/tests.py +++ b/ietf/secr/announcement/tests.py @@ -6,7 +6,7 @@ from django.urls import reverse -import debug # pyflakes:ignore +import debug # pyflakes:ignore from ietf.utils.test_utils import TestCase from ietf.group.factories import RoleFactory @@ -17,97 +17,102 @@ from ietf.message.models import AnnouncementFrom from ietf.utils.mail import outbox, empty_outbox -SECR_USER='secretary' -WG_USER='' -AD_USER='' +SECR_USER = "secretary" +WG_USER = "" +AD_USER = "" + class SecrAnnouncementTestCase(TestCase): def setUp(self): super().setUp() - chair = RoleName.objects.get(slug='chair') - secr = RoleName.objects.get(slug='secr') - ietf = Group.objects.get(acronym='ietf') - iab = Group.objects.get(acronym='iab') - secretariat = Group.objects.get(acronym='secretariat') - AnnouncementFrom.objects.create(name=secr,group=secretariat,address='IETF Secretariat ') - AnnouncementFrom.objects.create(name=chair,group=ietf,address='IETF Chair ') - AnnouncementFrom.objects.create(name=chair,group=iab,address='IAB Chair ') + chair = RoleName.objects.get(slug="chair") + secr = RoleName.objects.get(slug="secr") + ietf = Group.objects.get(acronym="ietf") + iab = Group.objects.get(acronym="iab") + secretariat = Group.objects.get(acronym="secretariat") + AnnouncementFrom.objects.create( + name=secr, + group=secretariat, + address="IETF Secretariat ", + ) + AnnouncementFrom.objects.create( + name=chair, group=ietf, address="IETF Chair " + ) + AnnouncementFrom.objects.create( + name=chair, group=iab, address="IAB Chair " + ) def test_main(self): "Main Test" - url = reverse('ietf.secr.announcement.views.main') + url = reverse("ietf.secr.announcement.views.main") self.client.login(username="secretary", password="secretary+password") r = self.client.get(url) self.assertEqual(r.status_code, 200) - + def test_main_announce_from(self): - url = reverse('ietf.secr.announcement.views.main') + url = reverse("ietf.secr.announcement.views.main") # Secretariat self.client.login(username="secretary", password="secretary+password") r = self.client.get(url) self.assertEqual(r.status_code, 200) q = PyQuery(r.content) - self.assertEqual(len(q('#id_frm option')),3) + self.assertEqual(len(q("#id_frm option")), 4) # IAB Chair self.client.login(username="iab-chair", password="iab-chair+password") r = self.client.get(url) self.assertEqual(r.status_code, 200) q = PyQuery(r.content) - self.assertEqual(len(q('#id_frm option')),1) - self.assertTrue('' in q('#id_frm option').val()) + self.assertEqual(len(q("#id_frm option")), 1) + self.assertTrue("" in q("#id_frm option").val()) # IETF Chair self.client.login(username="ietf-chair", password="ietf-chair+password") r = self.client.get(url) self.assertEqual(r.status_code, 200) q = PyQuery(r.content) - self.assertEqual(len(q('#id_frm option')),1) - self.assertTrue('' in q('#id_frm option').val()) + self.assertEqual(len(q("#id_frm option")), 1) + self.assertTrue("" in q("#id_frm option").val()) + class UnauthorizedAnnouncementCase(TestCase): def test_unauthorized(self): "Unauthorized Test" - url = reverse('ietf.secr.announcement.views.main') - person = RoleFactory(name_id='chair',group__acronym='mars').person - self.client.login(username=person.user.username, password=person.user.username+"+password") + url = reverse("ietf.secr.announcement.views.main") + person = RoleFactory(name_id="chair", group__acronym="mars").person + self.client.login( + username=person.user.username, password=person.user.username + "+password" + ) r = self.client.get(url) self.assertEqual(r.status_code, 403) - + + class SubmitAnnouncementCase(TestCase): - def test_invalid_submit(self): - "Invalid Submit" - url = reverse('ietf.secr.announcement.views.main') - post_data = {'id_subject':''} - self.client.login(username="secretary", password="secretary+password") - r = self.client.post(url,post_data) - self.assertEqual(r.status_code, 200) - q = PyQuery(r.content) - self.assertTrue(len(q('form ul.errorlist')) > 0) - def test_valid_submit(self): "Valid Submit" nomcom_test_data() empty_outbox() - url = reverse('ietf.secr.announcement.views.main') - confirm_url = reverse('ietf.secr.announcement.views.confirm') - nomcom = Group.objects.get(type='nomcom') - post_data = {'nomcom': nomcom.pk, - 'to':'Other...', - 'to_custom':'rcross@amsl.com', - 'frm':'IETF Secretariat <ietf-secretariat@ietf.org>', - 'reply_to':'secretariat@ietf.org', - 'subject':'Test Subject', - 'body':'This is a test.'} + url = reverse("ietf.secr.announcement.views.main") + confirm_url = reverse("ietf.secr.announcement.views.confirm") + nomcom = Group.objects.get(type="nomcom") + post_data = { + "nomcom": nomcom.pk, + "to": "Other...", + "to_custom": "phil@example.com", + "frm": "IETF Secretariat <ietf-secretariat@ietf.org>", + "reply_to": "secretariat@ietf.org", + "subject": "Test Subject", + "body": "This is a test.", + } self.client.login(username="secretary", password="secretary+password") - response = self.client.post(url,post_data) - self.assertContains(response, 'Confirm Announcement') - response = self.client.post(confirm_url,post_data,follow=True) + response = self.client.post(url, post_data) + self.assertContains(response, "Confirm Announcement") + response = self.client.post(confirm_url, post_data, follow=True) self.assertRedirects(response, url) - self.assertEqual(len(outbox),1) - self.assertEqual(outbox[0]['subject'],'Test Subject') - self.assertEqual(outbox[0]['to'],'') - message = Message.objects.filter(by__user__username='secretary').last() - self.assertEqual(message.subject,'Test Subject') + self.assertEqual(len(outbox), 1) + self.assertEqual(outbox[0]["subject"], "Test Subject") + self.assertEqual(outbox[0]["to"], "") + message = Message.objects.filter(by__user__username="secretary").last() + self.assertEqual(message.subject, "Test Subject") self.assertTrue(nomcom in message.related_groups.all()) diff --git a/ietf/secr/announcement/urls.py b/ietf/secr/announcement/urls.py index 3c3c05a09c..dc534f64ae 100644 --- a/ietf/secr/announcement/urls.py +++ b/ietf/secr/announcement/urls.py @@ -1,8 +1,7 @@ - from ietf.secr.announcement import views from ietf.utils.urls import url urlpatterns = [ - url(r'^$', views.main), - url(r'^confirm/$', views.confirm), + url(r"^$", views.main), + url(r"^confirm/$", views.confirm), ] diff --git a/ietf/secr/announcement/views.py b/ietf/secr/announcement/views.py index 42de089c59..5617ae9e6f 100644 --- a/ietf/secr/announcement/views.py +++ b/ietf/secr/announcement/views.py @@ -18,86 +18,93 @@ # Helper Functions # ------------------------------------------------- def check_access(user): - ''' + """ This function takes a Django User object and returns true if the user has access to the Announcement app. - ''' + """ if hasattr(user, "person"): person = user.person if has_role(user, "Secretariat"): return True - + for role in person.role_set.all(): - if AnnouncementFrom.objects.filter(name=role.name,group=role.group): + if AnnouncementFrom.objects.filter(name=role.name, group=role.group): return True - if Role.objects.filter(name="chair", - group__acronym__startswith="nomcom", - group__state="active", - group__type="nomcom", - person=person): + if Role.objects.filter( + name="chair", + group__acronym__startswith="nomcom", + group__state="active", + group__type="nomcom", + person=person, + ): return True return False + # -------------------------------------------------- # STANDARD VIEW FUNCTIONS # -------------------------------------------------- # this seems to cause some kind of circular problem # @check_for_cancel(reverse('home')) @login_required -@check_for_cancel('../') +@check_for_cancel("../") def main(request): - ''' + """ Main view for Announcement tool. Authrozied users can fill out email details: header, body, etc and send. - ''' + """ if not check_access(request.user): - permission_denied(request, 'Restricted to: Secretariat, IAD, or chair of IETF, IAB, RSOC, RSE, IAOC, ISOC, NomCom.') + permission_denied( + request, + "Restricted to: Secretariat, IAD, or chair of IETF, IAB, RSOC, RSE, IAOC, ISOC, NomCom.", + ) - form = AnnounceForm(request.POST or None,user=request.user) + form = AnnounceForm(request.POST or None, user=request.user) if form.is_valid(): # recast as hidden form for next page of process form = AnnounceForm(request.POST, user=request.user, hidden=True) - if form.data['to'] == 'Other...': - to = form.data['to_custom'] + if form.data["to"] == "Other...": + to = form.data["to_custom"] else: - to = form.data['to'] + to = form.data["to"] - return render(request, 'announcement/confirm.html', { - 'message': form.data, - 'to': to, - 'form': form}, + return render( + request, + "announcement/confirm.html", + {"message": form.data, "to": to, "form": form}, ) - return render(request, 'announcement/main.html', { 'form': form} ) + return render(request, "announcement/index.html", {"form": form}) + @login_required -@check_for_cancel('../') +@check_for_cancel("../") def confirm(request): if not check_access(request.user): - permission_denied(request, 'Restricted to: Secretariat, IAD, or chair of IETF, IAB, RSOC, RSE, IAOC, ISOC, NomCom.') + permission_denied( + request, + "Restricted to: Secretariat, IAD, or chair of IETF, IAB, RSOC, RSE, IAOC, ISOC, NomCom.", + ) - if request.method == 'POST': + if request.method == "POST": form = AnnounceForm(request.POST, user=request.user) - if request.method == 'POST': - message = form.save(user=request.user,commit=True) - extra = {'Reply-To': message.get('reply_to') } - send_mail_text(None, - message.to, - message.frm, - message.subject, - message.body, - cc=message.cc, - bcc=message.bcc, - extra=extra, - ) - - messages.success(request, 'The announcement was sent.') - return redirect('ietf.secr.announcement.views.main') - - - - + if request.method == "POST": + message = form.save(user=request.user, commit=True) + extra = {"Reply-To": message.get("reply_to")} + send_mail_text( + None, + message.to, + message.frm, + message.subject, + message.body, + cc=message.cc, + bcc=message.bcc, + extra=extra, + ) + + messages.success(request, "The announcement was sent.") + return redirect("ietf.secr.announcement.views.main") diff --git a/ietf/secr/areas/forms.py b/ietf/secr/areas/forms.py deleted file mode 100644 index 2a34408ad4..0000000000 --- a/ietf/secr/areas/forms.py +++ /dev/null @@ -1,49 +0,0 @@ -from django import forms - -from ietf.person.models import Person, Email - -import re - -STATE_CHOICES = ( - (1, 'Active'), - (2, 'Concluded') -) - - - -class AreaDirectorForm(forms.Form): - ad_name = forms.CharField(max_length=100,label='Name',help_text="To see a list of people type the first name, or last name, or both.") - #login = forms.EmailField(max_length=75,help_text="This should be the person's primary email address.") - #email = forms.ChoiceField(help_text="This should be the person's primary email address.") - email = forms.CharField(help_text="Select the email address to associate with this AD Role") - - # set css class=name-autocomplete for name field (to provide select list) - def __init__(self, *args, **kwargs): - super(AreaDirectorForm, self).__init__(*args, **kwargs) - self.fields['ad_name'].widget.attrs['class'] = 'name-autocomplete' - self.fields['email'].widget = forms.Select(choices=[]) - - def clean_ad_name(self): - name = self.cleaned_data.get('ad_name', '') - # check for tag within parenthesis to ensure name was selected from the list - m = re.search(r'\((\d+)\)', name) - if not name or not m: - raise forms.ValidationError("You must select an entry from the list!") - try: - id = m.group(1) - person = Person.objects.get(id=id) - except Person.DoesNotExist: - raise forms.ValidationError("ERROR finding Person with ID: %s" % id) - return person - - def clean_email(self): - # this ChoiceField gets populated by javascript so skip regular validation - # which raises an error - email = self.cleaned_data['email'] - if not email: - raise forms.ValidationError("You must select an email. If none are listed you'll need to add one first.") - try: - obj = Email.objects.get(address=email) - except Email.DoesNotExist: - raise forms.ValidationError("Can't find this email.") - return obj diff --git a/ietf/secr/areas/tests.py b/ietf/secr/areas/tests.py deleted file mode 100644 index f7341dfa9e..0000000000 --- a/ietf/secr/areas/tests.py +++ /dev/null @@ -1,34 +0,0 @@ -from django.urls import reverse - -from ietf.group.factories import GroupFactory, GroupEventFactory -from ietf.group.models import Group, GroupEvent -from ietf.person.models import Person -from ietf.utils.test_utils import TestCase - - -SECR_USER='secretary' - -def augment_data(): - system = Person.objects.get(name="(System)") - area = Group.objects.get(acronym='farfut') - GroupEvent.objects.create(group=area, - type='started', - by=system) - -class SecrAreasTestCase(TestCase): - def test_main(self): - "Main Test" - GroupFactory(type_id='area') - url = reverse('ietf.secr.areas.views.list_areas') - self.client.login(username="secretary", password="secretary+password") - response = self.client.get(url) - self.assertEqual(response.status_code, 200) - - def test_view(self): - "View Test" - area = GroupEventFactory(type='started',group__type_id='area').group - url = reverse('ietf.secr.areas.views.view', kwargs={'name':area.acronym}) - self.client.login(username="secretary", password="secretary+password") - response = self.client.get(url) - self.assertEqual(response.status_code, 200) - diff --git a/ietf/secr/areas/urls.py b/ietf/secr/areas/urls.py deleted file mode 100644 index cbb9a4a2f3..0000000000 --- a/ietf/secr/areas/urls.py +++ /dev/null @@ -1,12 +0,0 @@ - -from ietf.secr.areas import views -from ietf.utils.urls import url - -urlpatterns = [ - url(r'^$', views.list_areas), - url(r'^getemails', views.getemails), - url(r'^getpeople', views.getpeople), - url(r'^(?P[A-Za-z0-9.-]+)/$', views.view), - url(r'^(?P[A-Za-z0-9.-]+)/people/$', views.people), - url(r'^(?P[A-Za-z0-9.-]+)/people/modify/$', views.modify), -] diff --git a/ietf/secr/areas/views.py b/ietf/secr/areas/views.py deleted file mode 100644 index 93e438cf8d..0000000000 --- a/ietf/secr/areas/views.py +++ /dev/null @@ -1,203 +0,0 @@ -import json - -from django.contrib import messages -from django.http import HttpResponse -from django.shortcuts import render, get_object_or_404, redirect - -from ietf.group.models import Group, GroupEvent, Role -from ietf.group.utils import save_group_in_history -from ietf.ietfauth.utils import role_required -from ietf.person.models import Person -from ietf.secr.areas.forms import AreaDirectorForm - -# -------------------------------------------------- -# AJAX FUNCTIONS -# -------------------------------------------------- -def getpeople(request): - """ - Ajax function to find people. Takes one or two terms (ignores rest) and - returns JSON format response: first name, last name, primary email, tag - """ - result = [] - term = request.GET.get('term','') - - qs = Person.objects.filter(name__icontains=term) - for item in qs: - full = '%s - (%s)' % (item.name,item.id) - result.append(full) - - return HttpResponse(json.dumps(result), content_type='application/javascript') - -def getemails(request): - """ - Ajax function to get emails for given Person Id. Used for adding Area ADs. - returns JSON format response: [{id:email, value:email},...] - """ - - results=[] - id = request.GET.get('id','') - person = Person.objects.get(id=id) - for item in person.email_set.filter(active=True): - d = {'id': item.address, 'value': item.address} - results.append(d) - - return HttpResponse(json.dumps(results), content_type='application/javascript') - -# -------------------------------------------------- -# STANDARD VIEW FUNCTIONS -# -------------------------------------------------- - - -@role_required('Secretariat') -def list_areas(request): - """ - List IETF Areas - - **Templates:** - - * ``areas/list.html`` - - **Template Variables:** - - * results - - """ - - results = Group.objects.filter(type="area").order_by('name') - - return render(request, 'areas/list.html', { 'results': results} ) - -@role_required('Secretariat') -def people(request, name): - """ - Edit People associated with Areas, Area Directors. - - # Legacy ------------------ - When a new Director is first added they get a user_level of 4, read-only. - Then, when Director is made active (Enable Voting) user_level = 1. - - # New --------------------- - First Director's are assigned the Role 'pre-ad' Incoming Area director - Then they get 'ad' role - - **Templates:** - - * ``areas/people.html`` - - **Template Variables:** - - * directors, area - - """ - area = get_object_or_404(Group, type='area', acronym=name) - - if request.method == 'POST': - if request.POST.get('submit', '') == "Add": - form = AreaDirectorForm(request.POST) - if form.is_valid(): - email = form.cleaned_data['email'] - person = form.cleaned_data['ad_name'] - - # save group - save_group_in_history(area) - - # create role - Role.objects.create(name_id='pre-ad',group=area,email=email,person=person) - - if not email.origin or email.origin == person.user.username: - email.origin = "role: %s %s" % (area.acronym, 'pre-ad') - email.save() - - messages.success(request, 'New Area Director added successfully!') - return redirect('ietf.secr.areas.views.view', name=name) - else: - form = AreaDirectorForm() - - directors = area.role_set.filter(name__slug__in=('ad','pre-ad')) - return render(request, 'areas/people.html', { - 'area': area, - 'form': form, - 'directors': directors}, - ) - -@role_required('Secretariat') -def modify(request, name): - """ - Handle state changes of Area Directors (enable voting, retire) - # Legacy -------------------------- - Enable Voting actions - - user_level = 1 - - create TelechatUser object - Per requirements, the Retire button shall perform the following DB updates - - update iesg_login row, user_level = 2 (per Matt Feb 7, 2011) - - remove telechat_user row (to revoke voting rights) - - update IETFWG(groups) set area_director = TBD - - remove area_director row - # New ------------------------------ - Enable Voting: change Role from 'pre-ad' to 'ad' - Retire: save in history, delete role record, set group assn to TBD - - **Templates:** - - * none - - Redirects to view page on success. - """ - - area = get_object_or_404(Group, type='area', acronym=name) - - # should only get here with POST method - if request.method == 'POST': - # setup common request variables - tag = request.POST.get('tag', '') - person = Person.objects.get(id=tag) - - # save group - save_group_in_history(area) - - # handle retire request - if request.POST.get('submit', '') == "Retire": - role = Role.objects.get(group=area,name__in=('ad','pre-ad'),person=person) - role.delete() - - # update groups that have this AD as primary AD - Role.objects.filter(name__in=('ad','pre-ad'),person=person,group__type='wg',group__state__in=('active','bof')).delete() - - messages.success(request, 'The Area Director has been retired successfully!') - - # handle voting request - if request.POST.get('submit', '') == "Enable Voting": - role = Role.objects.get(group=area,name__slug='pre-ad',person=person) - role.name_id = 'ad' - role.save() - - messages.success(request, 'Voting rights have been granted successfully!') - - return redirect('ietf.secr.areas.views.view', name=name) - -@role_required('Secretariat') -def view(request, name): - """ - View Area information. - - **Templates:** - - * ``areas/view.html`` - - **Template Variables:** - - * area, directors - - """ - area = get_object_or_404(Group, type='area', acronym=name) - try: - area.start_date = area.groupevent_set.order_by('time')[0].time - area.concluded_date = area.groupevent_set.get(type='concluded').time - except GroupEvent.DoesNotExist: - pass - directors = area.role_set.filter(name__slug__in=('ad','pre-ad')) - - return render(request, 'areas/view.html', { - 'area': area, - 'directors': directors}, - ) diff --git a/ietf/secr/console/tests.py b/ietf/secr/console/tests.py deleted file mode 100644 index 9de5707955..0000000000 --- a/ietf/secr/console/tests.py +++ /dev/null @@ -1,27 +0,0 @@ -# Copyright The IETF Trust 2013-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -""" -This file demonstrates two different styles of tests (one doctest and one -unittest). These will both pass when you run "manage.py test". - -Replace these with more appropriate tests for your application. -""" - -from ietf.utils.test_utils import TestCase - -class SimpleTest(TestCase): - def test_basic_addition(self): - """ - Tests that 1 + 1 always equals 2. - """ - self.assertEqual(1 + 1, 2) - -__test__ = {"doctest": """ -Another way to test that 1 + 1 is equal to 2. - ->>> 1 + 1 == 2 -True -"""} - diff --git a/ietf/secr/console/urls.py b/ietf/secr/console/urls.py deleted file mode 100644 index 5c76a6867f..0000000000 --- a/ietf/secr/console/urls.py +++ /dev/null @@ -1,7 +0,0 @@ - -from ietf.secr.console import views -from ietf.utils.urls import url - -urlpatterns = [ - url(r'^$', views.main), -] diff --git a/ietf/secr/console/views.py b/ietf/secr/console/views.py deleted file mode 100644 index 9e582fd089..0000000000 --- a/ietf/secr/console/views.py +++ /dev/null @@ -1,17 +0,0 @@ - -from django.shortcuts import render - -from ietf.doc.models import DocEvent -from ietf.ietfauth.utils import role_required - -@role_required('Secretariat') -def main(request): - ''' - Main view for the Console - ''' - - latest_docevent = DocEvent.objects.all().order_by('-time')[0] - - return render(request, 'console/main.html', { - 'latest_docevent': latest_docevent}, - ) diff --git a/ietf/secr/groups/forms.py b/ietf/secr/groups/forms.py deleted file mode 100644 index 7b9611515e..0000000000 --- a/ietf/secr/groups/forms.py +++ /dev/null @@ -1,117 +0,0 @@ -import re - -from django import forms -from django.db.models import Count - -from ietf.group.models import Group, Role -from ietf.name.models import GroupStateName, GroupTypeName, RoleName -from ietf.person.models import Person, Email - - -# --------------------------------------------- -# Select Choices -# --------------------------------------------- -SEARCH_MEETING_CHOICES = (('',''),('NO','NO'),('YES','YES')) - -# --------------------------------------------- -# Functions -# --------------------------------------------- -def get_person(name): - ''' - This function takes a string which is in the name autocomplete format "name - (id)" - and returns a person object - ''' - - match = re.search(r'\((\d+)\)', name) - if not match: - return None - id = match.group(1) - try: - person = Person.objects.get(id=id) - except (Person.ObjectDoesNoExist, Person.MultipleObjectsReturned): - return None - return person - -def get_parent_group_choices(): - area_choices = [(g.id, g.name) for g in Group.objects.filter(type='area',state='active')] - other_parents = Group.objects.annotate(children=Count('group')).filter(children__gt=0).order_by('name').exclude(type='area') - other_choices = [(g.id, g.name) for g in other_parents] - choices = (('Working Group Areas',area_choices),('Other',other_choices)) - return choices - -# --------------------------------------------- -# Forms -# --------------------------------------------- - -class DescriptionForm (forms.Form): - description = forms.CharField(widget=forms.Textarea(attrs={'rows':'20'}),required=True, strip=False) - - - -class RoleForm(forms.Form): - name = forms.ModelChoiceField(RoleName.objects.filter(slug__in=('chair','editor','secr','techadv')),empty_label=None) - person = forms.CharField(max_length=50,widget=forms.TextInput(attrs={'class':'name-autocomplete'}),help_text="To see a list of people type the first name, or last name, or both.") - email = forms.CharField(widget=forms.Select(),help_text="Select an email") - group_acronym = forms.CharField(widget=forms.HiddenInput(),required=False) - - def __init__(self, *args, **kwargs): - self.group = kwargs.pop('group') - super(RoleForm, self).__init__(*args,**kwargs) - # this form is re-used in roles app, use different roles in select - if self.group.features.custom_group_roles: - self.fields['name'].queryset = RoleName.objects.all() - - # check for id within parenthesis to ensure name was selected from the list - def clean_person(self): - person = self.cleaned_data.get('person', '') - m = re.search(r'(\d+)', person) - if person and not m: - raise forms.ValidationError("You must select an entry from the list!") - - # return person object - return get_person(person) - - # check that email exists and return the Email object - def clean_email(self): - email = self.cleaned_data['email'] - try: - obj = Email.objects.get(address=email) - except Email.ObjectDoesNoExist: - raise forms.ValidationError("Email address not found!") - - # return email object - return obj - - def clean(self): - # here we abort if there are any errors with individual fields - # One predictable problem is that the user types a name rather then - # selecting one from the list, as instructed to do. We need to abort - # so the error is displayed before trying to call get_person() - if any(self.errors): - # Don't bother validating the formset unless each form is valid on its own - return - super(RoleForm, self).clean() - cleaned_data = self.cleaned_data - person = cleaned_data['person'] - email = cleaned_data['email'] - name = cleaned_data['name'] - group_acronym = cleaned_data['group_acronym'] - - if email.person != person: - raise forms.ValidationError('ERROR: The person associated with the chosen email address is different from the chosen person') - - if Role.objects.filter(name=name,group=self.group,person=person,email=email): - raise forms.ValidationError('ERROR: This is a duplicate entry') - - if not group_acronym: - raise forms.ValidationError('You must select a group.') - - return cleaned_data - -class SearchForm(forms.Form): - group_acronym = forms.CharField(max_length=12,required=False) - group_name = forms.CharField(max_length=80,required=False) - primary_area = forms.ModelChoiceField(queryset=Group.objects.filter(type='area',state='active'),required=False) - type = forms.ModelChoiceField(queryset=GroupTypeName.objects.all(),required=False) - meeting_scheduled = forms.CharField(widget=forms.Select(choices=SEARCH_MEETING_CHOICES),required=False) - state = forms.ModelChoiceField(queryset=GroupStateName.objects.exclude(slug__in=('dormant','unknown')),required=False) diff --git a/ietf/secr/groups/tests.py b/ietf/secr/groups/tests.py deleted file mode 100644 index f4772abcfa..0000000000 --- a/ietf/secr/groups/tests.py +++ /dev/null @@ -1,69 +0,0 @@ -# Copyright The IETF Trust 2013-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -from django.urls import reverse -from ietf.utils.test_utils import TestCase -from ietf.group.models import Group -from ietf.secr.groups.forms import get_parent_group_choices -from ietf.group.factories import GroupFactory, RoleFactory -from ietf.meeting.factories import MeetingFactory -from ietf.person.factories import PersonFactory -import debug # pyflakes:ignore - -class GroupsTest(TestCase): - def test_get_parent_group_choices(self): - GroupFactory(type_id='area') - choices = get_parent_group_choices() - area = Group.objects.filter(type='area',state='active').first() - # This is opaque. Can it be rewritten to be more self-documenting? - self.assertEqual(choices[0][1][0][0],area.id) - - # ------- Test Search -------- # - def test_search(self): - "Test Search" - MeetingFactory(type_id='ietf') - group = GroupFactory() - url = reverse('ietf.secr.groups.views.search') - post_data = {'group_acronym':group.acronym,'submit':'Search'} - self.client.login(username="secretary", password="secretary+password") - response = self.client.post(url,post_data,follow=True) - self.assertContains(response, group.acronym) - - # ------- Test View -------- # - def test_view(self): - MeetingFactory(type_id='ietf') - group = GroupFactory() - url = reverse('ietf.secr.groups.views.view', kwargs={'acronym':group.acronym}) - self.client.login(username="secretary", password="secretary+password") - response = self.client.get(url) - self.assertEqual(response.status_code, 200) - - - # ------- Test People -------- # - def test_people_delete(self): - role = RoleFactory(name_id='member') - group = role.group - id = role.id - url = reverse('ietf.secr.groups.views.delete_role', kwargs={'acronym':group.acronym,'id':role.id}) - target = reverse('ietf.secr.groups.views.people', kwargs={'acronym':group.acronym}) - self.client.login(username="secretary", password="secretary+password") - response = self.client.get(url) - self.assertEqual(response.status_code, 200) - response = self.client.post(url, {'post':'yes'}) - self.assertRedirects(response, target) - self.assertFalse(group.role_set.filter(id=id)) - - def test_people_add(self): - person = PersonFactory() - group = GroupFactory() - url = reverse('ietf.secr.groups.views.people', kwargs={'acronym':group.acronym}) - post_data = {'group_acronym':group.acronym, - 'name':'chair', - 'person':'Joe Smith - (%s)' % person.id, - 'email':person.email_set.all()[0].address, - 'submit':'Add'} - self.client.login(username="secretary", password="secretary+password") - response = self.client.post(url,post_data,follow=True) - self.assertRedirects(response, url) - self.assertContains(response, 'added successfully') diff --git a/ietf/secr/groups/urls.py b/ietf/secr/groups/urls.py deleted file mode 100644 index 60d3566cab..0000000000 --- a/ietf/secr/groups/urls.py +++ /dev/null @@ -1,14 +0,0 @@ -from django.conf import settings - -from ietf.secr.groups import views -from ietf.utils.urls import url - -urlpatterns = [ - url(r'^$', views.search), - url(r'^blue-dot-report/$', views.blue_dot), - #(r'^ajax/get_ads/$', views.get_ads), - url(r'^%(acronym)s/$' % settings.URL_REGEXPS, views.view), - url(r'^%(acronym)s/delete/(?P\d{1,6})/$' % settings.URL_REGEXPS, views.delete_role), - url(r'^%(acronym)s/charter/$' % settings.URL_REGEXPS, views.charter), - url(r'^%(acronym)s/people/$' % settings.URL_REGEXPS, views.people), -] diff --git a/ietf/secr/groups/views.py b/ietf/secr/groups/views.py deleted file mode 100644 index 48541efe31..0000000000 --- a/ietf/secr/groups/views.py +++ /dev/null @@ -1,301 +0,0 @@ -from django.contrib import messages -from django.conf import settings -from django.shortcuts import render, get_object_or_404, redirect - -from ietf.group.models import Group, GroupEvent, Role -from ietf.group.utils import save_group_in_history, get_charter_text -from ietf.ietfauth.utils import role_required -from ietf.person.models import Person -from ietf.secr.groups.forms import RoleForm, SearchForm -from ietf.secr.utils.meeting import get_current_meeting -from ietf.liaisons.views import contacts_from_roles - -# ------------------------------------------------- -# Helper Functions -# ------------------------------------------------- - -def add_legacy_fields(group): - ''' - This function takes a Group object as input and adds legacy attributes: - start_date,proposed_date,concluded_date,meeting_scheduled - ''' - # it's possible there could be multiple records of a certain type in which case - # we just return the latest record - query = GroupEvent.objects.filter(group=group, type="changed_state").order_by('time') - proposed = query.filter(changestategroupevent__state="proposed") - meeting = get_current_meeting() - - if proposed: - group.proposed_date = proposed[0].time - active = query.filter(changestategroupevent__state="active") - if active: - group.start_date = active[0].time - concluded = query.filter(changestategroupevent__state="conclude") - if concluded: - group.concluded_date = concluded[0].time - - if group.session_set.filter(meeting__number=meeting.number): - group.meeting_scheduled = 'YES' - else: - group.meeting_scheduled = 'NO' - - group.chairs = group.role_set.filter(name="chair") - group.techadvisors = group.role_set.filter(name="techadv") - group.editors = group.role_set.filter(name="editor") - group.secretaries = group.role_set.filter(name="secr") - # Note: liaison_contacts is now a dict instead of a model instance with fields. In - # templates, the dict can still be accessed using '.contacts' and .cc_contacts', though. - group.liaison_contacts = dict( - contacts=contacts_from_roles(group.role_set.filter(name='liaison_contact')), - cc_contacts=contacts_from_roles(group.role_set.filter(name='liaison_cc_contact')), - ) - - #fill_in_charter_info(group) - -#-------------------------------------------------- -# AJAX Functions -# ------------------------------------------------- -''' -def get_ads(request): - """ AJAX function which takes a URL parameter, "area" and returns the area directors - in the form of a list of dictionaries with "id" and "value" keys(in json format). - Used to populate select options. - """ - - results=[] - area = request.GET.get('area','') - qs = AreaDirector.objects.filter(area=area) - for item in qs: - d = {'id': item.id, 'value': item.person.first_name + ' ' + item.person.last_name} - results.append(d) - - return HttpResponse(json.dumps(results), content_type='application/javascript') -''' -# ------------------------------------------------- -# Standard View Functions -# ------------------------------------------------- - - - -@role_required('Secretariat') -def blue_dot(request): - ''' - This is a report view. It returns a text/plain listing of working group chairs. - ''' - people = Person.objects.filter(role__name__slug='chair', - role__group__type='wg', - role__group__state__slug__in=('active','bof','proposed')).distinct() - chairs = [] - for person in people: - parts = person.name_parts() - groups = [ r.group.acronym for r in person.role_set.filter(name__slug='chair', - group__type='wg', - group__state__slug__in=('active','bof','proposed')) ] - entry = {'name':'%s, %s' % (parts[3], parts[1]), - 'groups': ', '.join(groups)} - chairs.append(entry) - - # sort the list - sorted_chairs = sorted(chairs, key = lambda a: a['name']) - - return render(request, 'groups/blue_dot_report.txt', { 'chairs':sorted_chairs }, - content_type="text/plain; charset=%s"%settings.DEFAULT_CHARSET, - ) - -@role_required('Secretariat') -def charter(request, acronym): - """ - View Group Charter - - **Templates:** - - * ``groups/charter.html`` - - **Template Variables:** - - * group, charter_text - - """ - - group = get_object_or_404(Group, acronym=acronym) - # TODO: get_charter_text() should be updated to return None - if group.charter: - charter_text = get_charter_text(group) - else: - charter_text = '' - - return render(request, 'groups/charter.html', { - 'group': group, - 'charter_text': charter_text}, - ) - -@role_required('Secretariat') -def delete_role(request, acronym, id): - """ - Handle deleting roles for groups (chair, editor, advisor, secretary) - - **Templates:** - - * none - - Redirects to people page on success. - - """ - group = get_object_or_404(Group, acronym=acronym) - role = get_object_or_404(Role, id=id) - - if request.method == 'POST' and request.POST['post'] == 'yes': - # save group - save_group_in_history(group) - - role.delete() - messages.success(request, 'The entry was deleted successfully') - return redirect('ietf.secr.groups.views.people', acronym=acronym) - - return render(request, 'confirm_delete.html', {'object': role}) - - -@role_required('Secretariat') -def people(request, acronym): - """ - Edit Group Roles (Chairs, Secretary, etc) - - **Templates:** - - * ``groups/people.html`` - - **Template Variables:** - - * form, group - - """ - - group = get_object_or_404(Group, acronym=acronym) - - if request.method == 'POST': - # we need to pass group for form validation - form = RoleForm(request.POST,group=group) - if form.is_valid(): - name = form.cleaned_data['name'] - person = form.cleaned_data['person'] - email = form.cleaned_data['email'] - - # save group - save_group_in_history(group) - - Role.objects.create(name=name, - person=person, - email=email, - group=group) - - if not email.origin or email.origin == person.user.username: - email.origin = "role: %s %s" % (group.acronym, name.slug) - email.save() - - messages.success(request, 'New %s added successfully!' % name) - return redirect('ietf.secr.groups.views.people', acronym=group.acronym) - else: - form = RoleForm(initial={'name':'chair', 'group_acronym':group.acronym}, group=group) - - return render(request, 'groups/people.html', { - 'form':form, - 'group':group}, - ) - -@role_required('Secretariat') -def search(request): - """ - Search IETF Groups - - **Templates:** - - * ``groups/search.html`` - - **Template Variables:** - - * form, results - - """ - results = [] - if request.method == 'POST': - form = SearchForm(request.POST) - - if form.is_valid(): - kwargs = {} - group_acronym = form.cleaned_data['group_acronym'] - group_name = form.cleaned_data['group_name'] - primary_area = form.cleaned_data['primary_area'] - meeting_scheduled = form.cleaned_data['meeting_scheduled'] - state = form.cleaned_data['state'] - type = form.cleaned_data['type'] - meeting = get_current_meeting() - - # construct search query - if group_acronym: - kwargs['acronym__istartswith'] = group_acronym - if group_name: - kwargs['name__istartswith'] = group_name - if primary_area: - kwargs['parent'] = primary_area - if state: - kwargs['state'] = state - if type: - kwargs['type'] = type - #else: - # kwargs['type__in'] = ('wg','rg','ietf','ag','sdo','team') - - if meeting_scheduled == 'YES': - kwargs['session__meeting__number'] = meeting.number - # perform query - if kwargs: - if meeting_scheduled == 'NO': - qs = Group.objects.filter(**kwargs).exclude(session__meeting__number=meeting.number).distinct() - else: - qs = Group.objects.filter(**kwargs).distinct() - else: - qs = Group.objects.all() - results = qs.order_by('acronym') - - # if there's just one result go straight to view - if len(results) == 1: - return redirect('ietf.secr.groups.views.view', acronym=results[0].acronym) - - # process GET argument to support link from area app - elif 'primary_area' in request.GET: - area = request.GET.get('primary_area','') - results = Group.objects.filter(parent__id=area,type='wg',state__in=('bof','active','proposed')).order_by('name') - form = SearchForm({'primary_area':area,'state':'','type':'wg'}) - else: - form = SearchForm(initial={'state':'active'}) - - # loop through results and tack on meeting_scheduled because it is no longer an - # attribute of the meeting model - for result in results: - add_legacy_fields(result) - - return render(request, 'groups/search.html', { - 'results': results, - 'form': form}, - ) - -@role_required('Secretariat') -def view(request, acronym): - """ - View IETF Group details - - **Templates:** - - * ``groups/view.html`` - - **Template Variables:** - - * group - - """ - - group = get_object_or_404(Group, acronym=acronym) - - add_legacy_fields(group) - - return render(request, 'groups/view.html', { 'group': group } ) - diff --git a/ietf/secr/meetings/blue_sheets.py b/ietf/secr/meetings/blue_sheets.py deleted file mode 100644 index 6175541212..0000000000 --- a/ietf/secr/meetings/blue_sheets.py +++ /dev/null @@ -1,96 +0,0 @@ -# Copyright The IETF Trust 2013-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -import io - -from django.conf import settings -from django.utils.encoding import force_bytes - -r''' -RTF quick reference (from Word2007RTFSpec9.doc): -\fs24 : sets the font size to 24 half points -\header : header on all pages -\headerf : header on first page only -\pard : resets any previous paragraph formatting -\plain : resets any previous character formatting -\qr : right-aligned -\tqc : centered tab -\tqr : flush-right tab -\tx : tab position in twips (1440/inch) from the left margin -\nowidctlpar : no window/orphan control -\widctlpar : window/orphan control -''' - -def create_blue_sheets(meeting, groups): - file = io.open(settings.SECR_BLUE_SHEET_PATH, 'wb') - - header = b'''{\\rtf1\\ansi\\ansicpg1252\\uc1 \\deff0\\deflang1033\\deflangfe1033 - {\\fonttbl{\\f0\\froman\\fcharset0\\fprq2{\\*\\panose 02020603050405020304}Times New Roman;}} - {\\colortbl;\\red0\\green0\\blue0;\\red0\\green0\\blue255;\\red0\\green255\\blue255;\\red0\\green255\\blue0; -\\red255\\green0\\blue255;\\red255\\green0\\blue0;\\red255\\green255\\blue0;\\red255\\green255\\blue255; -\\red0\\green0\\blue128;\\red0\\green128\\blue128;\\red0\\green128\\blue0;\\red128\\green0\\blue128; -\\red128\\green0\\blue0;\\red128\\green128\\blue0;\\red128\\green128\\blue128; -\\red192\\green192\\blue192;} - \\widowctrl\\ftnbj\\aenddoc\\hyphcaps0\\formshade\\viewkind1\\viewscale100\\pgbrdrhead\\pgbrdrfoot - \\fet0\\sectd \\pgnrestart\\linex0\\endnhere\\titlepg\\sectdefaultcl''' - - file.write(header) - - for group in groups: - group_header = b''' {\\header \\pard\\plain \\s15\\nowidctlpar\\widctlpar\\tqc\\tx4320\\tqr\\tx8640\\adjustright \\fs20\\cgrid - { Mailing List: %s \\tab\\tab Meeting # %s %s (%s) \\par } - \\pard \\s15\\nowidctlpar\\widctlpar\\tqc\\tx4320\\tqr\\tx8640\\adjustright - {\\b\\fs24 - \\par - \\par \\tab The NOTE WELL statement applies to this meeting. Participants acknowledge that these attendance records will be made available to the public. - \\par - \\par NAME ORGANIZATION - \\par \\tab - \\par }} - {\\footer \\pard\\plain \\s16\\qc\\nowidctlpar\\widctlpar\\tqc\\tx4320\\tqr\\tx8640\\adjustright \\fs20\\cgrid {\\cs17 Page } - {\\field{\\*\\fldinst {\\cs17 PAGE }}} - { \\par }} - {\\headerf \\pard\\plain \\s15\\qr\\nowidctlpar\\widctlpar\\tqc\\tx4320\\tqr\\tx8640\\adjustright \\fs20\\cgrid - {\\b\\fs24 Meeting # %s %s (%s) \\par }} - {\\footerf \\pard\\plain \\s16\\qc\\nowidctlpar\\widctlpar\\tqc\\tx4320\\tqr\\tx8640\\adjustright \\fs20\\cgrid - {Page 1 \\par }} - \\pard\\plain \\qc\\nowidctlpar\\widctlpar\\adjustright \\fs20\\cgrid - {\\b\\fs32 %s IETF Working Group Roster \\par } - \\pard \\nowidctlpar\\widctlpar\\adjustright - {\\fs28 \\par Working Group Session: %s \\par \\par } -{\\b \\fs24 Mailing List: %s \\tx5300\\tab Date: ___________________ \\par \\par Chairperson:_________________________________________________________ \\par \\par } - {\\tab \\tab } -{\\par \\tab The NOTE WELL statement applies to this meeting. Participants acknowledge that these attendance records will be made available to the public. \\par -\\par\\b NAME ORGANIZATION -\\par } - \\pard \\fi-90\\li90\\nowidctlpar\\widctlpar\\adjustright - {\\fs16 -''' % (force_bytes(group.list_email), - force_bytes(meeting.number), - force_bytes(group.acronym), - force_bytes(group.type), - force_bytes(meeting.number), - force_bytes(group.acronym), - force_bytes(group.type), - force_bytes(meeting.number), - force_bytes(group.name), - force_bytes(group.list_email), - ) - - file.write(group_header) - for x in range(1,117): - line = b'''\\par %s._________________________________________________ \\tab _____________________________________________________ - \\par - ''' % force_bytes(x) - file.write(line) - - footer = b'''} -\\pard \\nowidctlpar\\widctlpar\\adjustright -{\\fs16 \\sect } -\\sectd \\pgnrestart\\linex0\\endnhere\\titlepg\\sectdefaultcl -''' - file.write(footer) - - file.write(b'\n}') - file.close() diff --git a/ietf/secr/meetings/forms.py b/ietf/secr/meetings/forms.py index 92c7514184..1be22a9d6c 100644 --- a/ietf/secr/meetings/forms.py +++ b/ietf/secr/meetings/forms.py @@ -238,17 +238,8 @@ def clean_group(self): raise forms.ValidationError("ERROR: can't change group after materials have been uploaded") return group -class UploadBlueSheetForm(forms.Form): - file = forms.FileField(help_text='example: bluesheets-84-ancp-01.pdf') - - def clean_file(self): - file = self.cleaned_data['file'] - if not re.match(r'bluesheets-\d+',file.name): - raise forms.ValidationError('Incorrect filename format') - return file class RegularSessionEditForm(forms.ModelForm): class Meta: model = Session fields = ['agenda_note'] - diff --git a/ietf/secr/proceedings/__init__.py b/ietf/secr/meetings/migrations/__init__.py similarity index 100% rename from ietf/secr/proceedings/__init__.py rename to ietf/secr/meetings/migrations/__init__.py diff --git a/ietf/secr/meetings/tests.py b/ietf/secr/meetings/tests.py index bf052abe42..08c792ce1e 100644 --- a/ietf/secr/meetings/tests.py +++ b/ietf/secr/meetings/tests.py @@ -3,16 +3,11 @@ import datetime -import os -import shutil -from pathlib import Path from pyquery import PyQuery -from io import StringIO import debug # pyflakes:ignore -from django.conf import settings from django.urls import reverse from django.utils import timezone @@ -29,24 +24,6 @@ class SecrMeetingTestCase(TestCase): settings_temp_path_overrides = TestCase.settings_temp_path_overrides + ['AGENDA_PATH'] - def setUp(self): - super().setUp() - self.bluesheet_dir = self.tempdir('bluesheet') - self.bluesheet_path = os.path.join(self.bluesheet_dir,'blue_sheet.rtf') - self.saved_secr_blue_sheet_path = settings.SECR_BLUE_SHEET_PATH - settings.SECR_BLUE_SHEET_PATH = self.bluesheet_path - - # n.b., the bluesheet upload relies on SECR_PROCEEDINGS_DIR being the same - # as AGENDA_PATH. This is probably a bug, but may not be worth fixing if - # the secr app is on the way out. - self.saved_secr_proceedings_dir = settings.SECR_PROCEEDINGS_DIR - settings.SECR_PROCEEDINGS_DIR = settings.AGENDA_PATH - - def tearDown(self): - settings.SECR_PROCEEDINGS_DIR = self.saved_secr_proceedings_dir - settings.SECR_BLUE_SHEET_PATH = self.saved_secr_blue_sheet_path - shutil.rmtree(self.bluesheet_dir) - super().tearDown() def test_main(self): "Main Test" @@ -105,6 +82,10 @@ def test_add_meeting(self): [cn.slug for cn in new_meeting.group_conflict_types.all()], post_data['group_conflict_types'], ) + self.assertEqual( + new_meeting.session_request_lock_message, + "Session requests for this meeting have not yet opened.", + ) def test_add_meeting_default_conflict_types(self): """Add meeting should default to same conflict types as previous meeting""" @@ -170,34 +151,6 @@ def test_edit_meeting(self): [cn.slug for cn in meeting.group_conflict_types.all()], post_data['group_conflict_types'], ) - - def test_blue_sheets_upload(self): - "Test Bluesheets" - meeting = make_meeting_test_data() - (Path(settings.SECR_PROCEEDINGS_DIR) / str(meeting.number) / 'bluesheets').mkdir(parents=True) - - url = reverse('ietf.secr.meetings.views.blue_sheet',kwargs={'meeting_id':meeting.number}) - self.client.login(username="secretary", password="secretary+password") - response = self.client.get(url) - self.assertEqual(response.status_code, 200) - - # test upload - group = Group.objects.filter(type='wg',state='active').first() - file = StringIO('dummy bluesheet') - file.name = "bluesheets-%s-%s.pdf" % (meeting.number,group.acronym) - files = {'file':file} - response = self.client.post(url, files) - self.assertEqual(response.status_code, 302) - path = os.path.join(settings.SECR_PROCEEDINGS_DIR,str(meeting.number),'bluesheets') - self.assertEqual(len(os.listdir(path)),1) - - def test_blue_sheets_generate(self): - meeting = make_meeting_test_data() - url = reverse('ietf.secr.meetings.views.blue_sheet_generate',kwargs={'meeting_id':meeting.number}) - self.client.login(username="secretary", password="secretary+password") - response = self.client.post(url) - self.assertEqual(response.status_code, 302) - self.assertTrue(os.path.exists(self.bluesheet_path)) def test_notifications(self): "Test Notifications" @@ -392,6 +345,7 @@ def test_meetings_misc_session_edit(self): 'remote_instructions': 'http://webex.com/foobar', }) self.assertRedirects(response, redirect_url) + session = Session.objects.get(pk=session.pk) # get a clean instance to avoid cache problems timeslot = session.official_timeslotassignment().timeslot self.assertEqual(timeslot.time, new_time) @@ -445,4 +399,4 @@ def test_get_times(self): times = get_times(meeting,day) values = [ x[0] for x in times ] self.assertTrue(times) - self.assertTrue(timeslot.time.strftime('%H%M') in values) \ No newline at end of file + self.assertTrue(timeslot.time.strftime('%H%M') in values) diff --git a/ietf/secr/meetings/urls.py b/ietf/secr/meetings/urls.py index 96c61d47b4..cc51bb0584 100644 --- a/ietf/secr/meetings/urls.py +++ b/ietf/secr/meetings/urls.py @@ -7,10 +7,7 @@ url(r'^$', views.main), url(r'^add/$', views.add), # url(r'^ajax/get-times/(?P\d{1,6})/(?P\d)/$', views.ajax_get_times), # Not in use - url(r'^blue_sheet/$', views.blue_sheet_redirect), url(r'^(?P\d{1,6})/$', views.view), - url(r'^(?P\d{1,6})/blue_sheet/$', views.blue_sheet), - url(r'^(?P\d{1,6})/blue_sheet/generate/$', views.blue_sheet_generate), url(r'^(?P\d{1,6})/edit/$', views.edit_meeting), url(r'^(?P\d{1,6})/notifications/$', views.notifications), url(r'^(?P\d{1,6})/(?P[A-Za-z0-9_\-]+)/$', views.rooms), diff --git a/ietf/secr/meetings/views.py b/ietf/secr/meetings/views.py index bd0c8efa10..1f6f2f3297 100644 --- a/ietf/secr/meetings/views.py +++ b/ietf/secr/meetings/views.py @@ -1,9 +1,7 @@ -# Copyright The IETF Trust 2007-2019, All Rights Reserved +# Copyright The IETF Trust 2007-2025, All Rights Reserved # -*- coding: utf-8 -*- import datetime -import os -import time from django.conf import settings from django.contrib import messages @@ -19,19 +17,18 @@ from ietf.ietfauth.utils import role_required from ietf.utils.mail import send_mail from ietf.meeting.forms import duration_string -from ietf.meeting.helpers import get_meeting, make_materials_directories, populate_important_dates +from ietf.meeting.helpers import make_materials_directories, populate_important_dates from ietf.meeting.models import Meeting, Session, Room, TimeSlot, SchedTimeSessAssignment, Schedule, SchedulingEvent -from ietf.meeting.utils import add_event_info_to_session_qs, handle_upload_file +from ietf.meeting.utils import add_event_info_to_session_qs +from ietf.meeting.views_session_request import get_initial_session from ietf.name.models import SessionStatusName from ietf.group.models import Group, GroupEvent -from ietf.secr.meetings.blue_sheets import create_blue_sheets from ietf.secr.meetings.forms import ( BaseMeetingRoomFormSet, MeetingModelForm, MeetingSelectForm, MeetingRoomForm, MiscSessionForm, TimeSlotForm, RegularSessionEditForm, - UploadBlueSheetForm, MeetingRoomOptionsForm ) -from ietf.secr.sreq.views import get_initial_session + MeetingRoomOptionsForm ) from ietf.secr.utils.meeting import get_session, get_timeslot from ietf.mailtrigger.utils import gather_address_lists -from ietf.utils.timezone import date_today, make_aware +from ietf.utils.timezone import make_aware # prep for agenda changes @@ -129,7 +126,7 @@ def is_combined(session,meeting,schedule=None): def send_notifications(meeting, groups, person): ''' Send session scheduled email notifications for each group in groups. Person is the - user who initiated this action, request.uesr.get_profile(). + user who initiated this action, request.user.get_profile(). ''' now = timezone.now() for group in groups: @@ -150,7 +147,7 @@ def send_notifications(meeting, groups, person): t = d['timeslot'] dur = s.requested_duration.seconds/60 items[i]['duration'] = "%d:%02d" % (dur//60, dur%60) - items[i]['period'] = '%s-%s' % (t.time.strftime('%H%M'),(t.time + t.duration).strftime('%H%M')) + items[i]['period'] = f"{t.local_start_time().strftime('%H%M')}-{t.local_end_time().strftime('%H%M')} {t.tz()}" # send email first_event = SchedulingEvent.objects.filter(session=sessions[0]).select_related('by').order_by('time', 'id').first() @@ -226,9 +223,8 @@ def add(request): ) meeting.schedule = schedule - # we want to carry session request lock status over from previous meeting - previous_meeting = get_meeting( int(meeting.number) - 1 ) - meeting.session_request_lock_message = previous_meeting.session_request_lock_message + # Create meeting with session requests locked + meeting.session_request_lock_message = "Session requests for this meeting have not yet opened." meeting.save() populate_important_dates(meeting) @@ -256,72 +252,6 @@ def add(request): 'form': form}, ) -@role_required('Secretariat') -def blue_sheet(request, meeting_id): - ''' - Blue Sheet view. The user can generate blue sheets or upload scanned bluesheets - ''' - meeting = get_object_or_404(Meeting, number=meeting_id) - url = settings.SECR_BLUE_SHEET_URL - blank_sheets_path = settings.SECR_BLUE_SHEET_PATH - try: - last_run = time.ctime(os.stat(blank_sheets_path).st_ctime) - except OSError: - last_run = None - uploaded_sheets_path = os.path.join(settings.SECR_PROCEEDINGS_DIR,meeting.number,'bluesheets') - uploaded_files = sorted(os.listdir(uploaded_sheets_path)) - - if request.method == 'POST': - form = UploadBlueSheetForm(request.POST,request.FILES) - if form.is_valid(): - file = request.FILES['file'] - save_error = handle_upload_file(file,file.name,meeting,'bluesheets') - if save_error: - form.add_error(None, save_error) - else: - messages.success(request, 'File Uploaded') - return redirect('ietf.secr.meetings.views.blue_sheet', meeting_id=meeting.number) - else: - form = UploadBlueSheetForm() - - return render(request, 'meetings/blue_sheet.html', { - 'meeting': meeting, - 'url': url, - 'form': form, - 'last_run': last_run, - 'uploaded_files': uploaded_files}, - ) - -@role_required('Secretariat') -def blue_sheet_generate(request, meeting_id): - ''' - Generate bluesheets - ''' - meeting = get_object_or_404(Meeting, number=meeting_id) - - if request.method == "POST": - groups = Group.objects.filter( - type__in=['wg','rg','ag','rag','program'], - session__timeslotassignments__schedule__in=[meeting.schedule, meeting.schedule.base if meeting.schedule else None]).order_by('acronym') - create_blue_sheets(meeting, groups) - - messages.success(request, 'Blue Sheets generated') - return redirect('ietf.secr.meetings.views.blue_sheet', meeting_id=meeting.number) - -@role_required('Secretariat') -def blue_sheet_redirect(request): - ''' - This is the generic blue sheet URL. It gets the next IETF meeting and redirects - to the meeting specific URL. - ''' - today = date_today() - qs = Meeting.objects.filter(date__gt=today,type='ietf').order_by('date') - if qs: - meeting = qs[0] - else: - meeting = Meeting.objects.filter(type='ietf').order_by('-date')[0] - return redirect('ietf.secr.meetings.views.blue_sheet', meeting_id=meeting.number) - @role_required('Secretariat') def edit_meeting(request, meeting_id): ''' diff --git a/ietf/secr/middleware/dbquery.py b/ietf/secr/middleware/dbquery.py deleted file mode 100644 index 34c96bd12b..0000000000 --- a/ietf/secr/middleware/dbquery.py +++ /dev/null @@ -1,35 +0,0 @@ -# Copyright The IETF Trust 2014-2019, All Rights Reserved -#import logging - -from django.db import connection -from django.utils.log import getLogger # type: ignore - - -logger = getLogger(__name__) -#logger.setLevel(logging.DEBUG) -#logger.addHandler(logging.FileHandler(settings.SECR_LOG_FILE)) - -class QueryCountDebugMiddleware(object): - """ - This middleware will log the number of queries run - and the total time taken for each request (with a - status code of 200). It does not currently support - multi-db setups. - """ - def process_response(self, request, response): - #assert False, request.path - logger.debug('called middleware. %s:%s' % (request.path,len(connection.queries))) - if response.status_code == 200: - total_time = 0 - #for query in connection.queries: - # query_time = query.get('time') - # if query_time is None: - # django-debug-toolbar monkeypatches the connection - # cursor wrapper and adds extra information in each - # item in connection.queries. The query time is stored - # under the key "duration" rather than "time" and is - # in milliseconds, not seconds. - # query_time = query.get('duration', 0) / 1000 - # total_time += float(query_time) - logger.debug('%s: %s queries run, total %s seconds' % (request.path,len(connection.queries), total_time)) - return response diff --git a/ietf/secr/proceedings/forms.py b/ietf/secr/proceedings/forms.py deleted file mode 100644 index 26dd419d66..0000000000 --- a/ietf/secr/proceedings/forms.py +++ /dev/null @@ -1,44 +0,0 @@ -# Copyright The IETF Trust 2007-2019, All Rights Reserved - -from django import forms - -from ietf.doc.models import Document -from ietf.meeting.models import Session -from ietf.meeting.utils import add_event_info_to_session_qs - - -# --------------------------------------------- -# Globals -# --------------------------------------------- - -VALID_SLIDE_EXTENSIONS = ('.doc','.docx','.pdf','.ppt','.pptx','.txt','.zip') -VALID_MINUTES_EXTENSIONS = ('.txt','.html','.htm','.pdf') -VALID_AGENDA_EXTENSIONS = ('.txt','.html','.htm') -VALID_BLUESHEET_EXTENSIONS = ('.pdf','.jpg','.jpeg') - -#---------------------------------------------------------- -# Forms -#---------------------------------------------------------- - -class RecordingForm(forms.Form): - external_url = forms.URLField(label='Url') - session = forms.ModelChoiceField(queryset=Session.objects) - session.widget.attrs['class'] = "select2-field" - session.widget.attrs['data-minimum-input-length'] = 0 - - def __init__(self, *args, **kwargs): - self.meeting = kwargs.pop('meeting') - super(RecordingForm, self).__init__(*args,**kwargs) - self.fields['session'].queryset = add_event_info_to_session_qs( - Session.objects.filter(meeting=self.meeting, type__in=['regular','plenary','other']) - ).filter(current_status='sched').order_by('group__acronym') - -class RecordingEditForm(forms.ModelForm): - class Meta: - model = Document - fields = ['external_url'] - - def __init__(self, *args, **kwargs): - super(RecordingEditForm, self).__init__(*args, **kwargs) - self.fields['external_url'].label='Url' - diff --git a/ietf/secr/proceedings/migrations/0001_initial.py b/ietf/secr/proceedings/migrations/0001_initial.py deleted file mode 100644 index 2aee67d708..0000000000 --- a/ietf/secr/proceedings/migrations/0001_initial.py +++ /dev/null @@ -1,28 +0,0 @@ -# Copyright The IETF Trust 2018-2020, All Rights Reserved -# -*- coding: utf-8 -*- -# Generated by Django 1.11.10 on 2018-02-20 10:52 - - -from django.db import migrations - - -class Migration(migrations.Migration): - - initial = True - - dependencies = [ - ('meeting', '0001_initial'), - ] - - operations = [ - migrations.CreateModel( - name='InterimMeeting', - fields=[ - ], - options={ - 'proxy': True, - 'indexes': [], - }, - bases=('meeting.meeting',), - ), - ] diff --git a/ietf/secr/proceedings/models.py b/ietf/secr/proceedings/models.py deleted file mode 100644 index 75563429e5..0000000000 --- a/ietf/secr/proceedings/models.py +++ /dev/null @@ -1,62 +0,0 @@ -# Copyright The IETF Trust 2013-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -import os - -from django.conf import settings -from django.db import models - -from ietf.meeting.models import Meeting - - -class InterimManager(models.Manager): - '''A custom manager to limit objects to type=interim''' - def get_queryset(self): - return super(InterimManager, self).get_queryset().filter(type='interim') - -class InterimMeeting(Meeting): - ''' - This class is a proxy of Meeting. It's purpose is to provide extra methods that are - useful for an interim meeting, to help in templates. Most information is derived from - the session associated with this meeting. We are assuming there is only one. - ''' - class Meta: - proxy = True - - objects = InterimManager() - - def group(self): - return self.session_set.all()[0].group - - def agenda(self): # pylint: disable=method-hidden - session = self.session_set.all()[0] - agendas = session.materials.exclude(states__slug='deleted').filter(type='agenda') - if agendas: - return agendas[0] - else: - return None - - def minutes(self): - session = self.session_set.all()[0] - minutes = session.materials.exclude(states__slug='deleted').filter(type='minutes') - if minutes: - return minutes[0] - else: - return None - - def get_proceedings_path(self, group=None): - return os.path.join(self.get_materials_path(),'proceedings.html') - - def get_proceedings_url(self, group=None): - ''' - If the proceedings file doesn't exist return empty string. For use in templates. - ''' - if os.path.exists(self.get_proceedings_path()): - url = "%sproceedings/%s/proceedings.html" % ( - settings.IETF_HOST_URL, - self.number) - return url - else: - return '' - diff --git a/ietf/secr/proceedings/new_tables.sql b/ietf/secr/proceedings/new_tables.sql deleted file mode 100644 index 43d590f3d8..0000000000 --- a/ietf/secr/proceedings/new_tables.sql +++ /dev/null @@ -1,50 +0,0 @@ -CREATE TABLE `interim_slides` ( - `id` integer AUTO_INCREMENT NOT NULL PRIMARY KEY, - `meeting_num` integer NOT NULL, - `group_acronym_id` integer, - `slide_num` integer, - `slide_type_id` integer NOT NULL, - `slide_name` varchar(255) NOT NULL, - `irtf` integer NOT NULL, - `interim` bool NOT NULL, - `order_num` integer, - `in_q` integer -) -; -CREATE TABLE `interim_minutes` ( - `id` integer AUTO_INCREMENT NOT NULL PRIMARY KEY, - `meeting_num` integer NOT NULL, - `group_acronym_id` integer NOT NULL, - `filename` varchar(255) NOT NULL, - `irtf` integer NOT NULL, - `interim` bool NOT NULL -) -; -CREATE TABLE `interim_agenda` ( - `id` integer AUTO_INCREMENT NOT NULL PRIMARY KEY, - `meeting_num` integer NOT NULL, - `group_acronym_id` integer NOT NULL, - `filename` varchar(255) NOT NULL, - `irtf` integer NOT NULL, - `interim` bool NOT NULL -) -; -CREATE TABLE `interim_meetings` ( - `meeting_num` integer NOT NULL PRIMARY KEY AUTO_INCREMENT, - `start_date` date , - `end_date` date , - `city` varchar(255) , - `state` varchar(255) , - `country` varchar(255) , - `time_zone` integer, - `ack` longtext , - `agenda_html` longtext , - `agenda_text` longtext , - `future_meeting` longtext , - `overview1` longtext , - `overview2` longtext , - `group_acronym_id` integer -) -; -alter table interim_meetings auto_increment=201; - diff --git a/ietf/secr/proceedings/proc_utils.py b/ietf/secr/proceedings/proc_utils.py deleted file mode 100644 index 43d063d07f..0000000000 --- a/ietf/secr/proceedings/proc_utils.py +++ /dev/null @@ -1,309 +0,0 @@ -# Copyright The IETF Trust 2013-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -''' -proc_utils.py - -This module contains all the functions for generating static proceedings pages -''' -import datetime -import os -import pytz -import re -import subprocess -from urllib.parse import urlencode - -import debug # pyflakes:ignore - -from django.conf import settings -from django.core.exceptions import ObjectDoesNotExist - -from ietf.doc.models import Document, DocAlias, DocEvent, NewRevisionDocEvent, State -from ietf.group.models import Group -from ietf.meeting.models import Meeting, SessionPresentation, TimeSlot, SchedTimeSessAssignment, Session -from ietf.person.models import Person -from ietf.utils.log import log -from ietf.utils.mail import send_mail -from ietf.utils.timezone import make_aware - -AUDIO_FILE_RE = re.compile(r'ietf(?P[\d]+)-(?P.*)-(?P
    - - {% if form.non_field_errors %}{{ form.non_field_errors }}{% endif %} - {% for field in form.visible_fields %} - - - - - {% endfor %} - -
    {{ field.label_tag }}{% if field.field.required %} *{% endif %}{{ field.errors }}{{ field }}{% if field.help_text %}
    {{ field.help_text }}{% endif %}
    -
    -
      -
    • -
    • -
    -
    - - - - -{% endblock %} diff --git a/ietf/secr/templates/areas/list.html b/ietf/secr/templates/areas/list.html deleted file mode 100644 index a0ed1ae4a3..0000000000 --- a/ietf/secr/templates/areas/list.html +++ /dev/null @@ -1,42 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles %} -{% block title %}Areas{% endblock %} - -{% block extrahead %}{{ block.super }} - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Areas -{% endblock %} - -{% block content %} - -
    -

    Areas

    - - - - - - - - - - {% for item in results %} - - - - - - {% endfor %} - -
    NameAcronymStatus
    {{ item.name }}{{ item.acronym }}{{ item.state }}
    - -
    -
      -
    -
    -
    - -{% endblock %} diff --git a/ietf/secr/templates/areas/people.html b/ietf/secr/templates/areas/people.html deleted file mode 100644 index 7eeb77c593..0000000000 --- a/ietf/secr/templates/areas/people.html +++ /dev/null @@ -1,62 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles %} -{% block title %}Areas - People{% endblock %} - -{% block extrahead %}{{ block.super }} - - - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Areas - » {{ area.acronym }} - » People -{% endblock %} - -{% block content %} - -
    -

    Area Directors ({{ area.acronym }})

    - - {% for director in directors %} - {% csrf_token %} - - - - - {% endif %} - - - - {% endfor %} -
    {{ director.person.name }}{% if director.name.slug == "ad" %} - Voting Enabled - {% else %} -
    - - - -
    -
      -
    • -
    -
    -
    - -{% endblock %} \ No newline at end of file diff --git a/ietf/secr/templates/areas/view.html b/ietf/secr/templates/areas/view.html deleted file mode 100644 index 3b95988682..0000000000 --- a/ietf/secr/templates/areas/view.html +++ /dev/null @@ -1,51 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles %} -{% block title %}Areas - View{% endblock %} - -{% block extrahead %}{{ block.super }} - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Areas - » {{ area.acronym }} -{% endblock %} - -{% block content %} - -
    -

    Area - View

    - - - - - - - - - - -
    Area Acronym:{{ area.acronym }}
    Area Name:{{ area.name }}
    Status:{{ area.state }}
    Start Date:{{ area.start_date|date:"Y-m-d" }}
    Concluded Date:{{ area.concluded_date|date:"Y-m-d" }}
    Last Modified Date:{{ area.time|date:"Y-m-d" }}
    Comments:{{ area.comments}}
    - - - -
    -
      - -
    • -
    • -
    • -
    -
    -
    - -{% endblock %} diff --git a/ietf/secr/templates/base_secr.html b/ietf/secr/templates/base_secr.html index 47b893f043..18d77e47ba 100644 --- a/ietf/secr/templates/base_secr.html +++ b/ietf/secr/templates/base_secr.html @@ -1,5 +1,5 @@ -{% load staticfiles %} +{% load static %} diff --git a/ietf/secr/templates/base_secr_bootstrap.html b/ietf/secr/templates/base_secr_bootstrap.html index 2eee566a12..a326346847 100644 --- a/ietf/secr/templates/base_secr_bootstrap.html +++ b/ietf/secr/templates/base_secr_bootstrap.html @@ -1,5 +1,5 @@ -{% load staticfiles %} +{% load static %} diff --git a/ietf/secr/templates/base_site.html b/ietf/secr/templates/base_site.html index d369a40ec1..5e3ddc62d8 100644 --- a/ietf/secr/templates/base_site.html +++ b/ietf/secr/templates/base_site.html @@ -1,7 +1,7 @@ {% extends "base_secr.html" %} {% load i18n %} {% load ietf_filters %} -{% load staticfiles %} +{% load static %} {% block title %}{{ title }}{% if user|has_role:"Secretariat" %} Secretariat Dashboard {% else %} IETF Dashboard {% endif %}{% endblock %} diff --git a/ietf/secr/templates/base_site_bootstrap.html b/ietf/secr/templates/base_site_bootstrap.html index c1c2fdac6b..1653b26b85 100644 --- a/ietf/secr/templates/base_site_bootstrap.html +++ b/ietf/secr/templates/base_site_bootstrap.html @@ -1,7 +1,7 @@ {% extends "base_secr_bootstrap.html" %} {% load i18n %} {% load ietf_filters %} -{% load staticfiles %} +{% load static %} {% block title %}{{ title }}{% if user|has_role:"Secretariat" %} Secretariat Dashboard {% else %} WG Chair Dashboard {% endif %}{% endblock %} diff --git a/ietf/secr/templates/confirm_cancel.html b/ietf/secr/templates/confirm_cancel.html index 6bae631a7e..541c82863f 100644 --- a/ietf/secr/templates/confirm_cancel.html +++ b/ietf/secr/templates/confirm_cancel.html @@ -1,5 +1,5 @@ {% extends "base_site.html" %} -{% load staticfiles %} +{% load static %} {% block title %}Confirm Cancel{% endblock %} diff --git a/ietf/secr/templates/confirm_delete.html b/ietf/secr/templates/confirm_delete.html index ccfc7b1c2f..3f8fd19c8f 100644 --- a/ietf/secr/templates/confirm_delete.html +++ b/ietf/secr/templates/confirm_delete.html @@ -1,5 +1,5 @@ {% extends "base_site.html" %} -{% load staticfiles %} +{% load static %} {% block title %}Confirm Delete{% endblock %} diff --git a/ietf/secr/templates/console/main.html b/ietf/secr/templates/console/main.html deleted file mode 100644 index 5aadefc558..0000000000 --- a/ietf/secr/templates/console/main.html +++ /dev/null @@ -1,22 +0,0 @@ -{% extends "base_site.html" %} - -{% block title %}Console{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Console -{% endblock %} - -{% block content %} - -
    -

    Console

    - - - - - -
    Latest DocEvent{{ latest_docevent }}
    - -
    - -{% endblock %} diff --git a/ietf/secr/templates/groups/blue_dot_report.txt b/ietf/secr/templates/groups/blue_dot_report.txt deleted file mode 100644 index 7516b01dad..0000000000 --- a/ietf/secr/templates/groups/blue_dot_report.txt +++ /dev/null @@ -1,6 +0,0 @@ -BLUE DOT REPORT - -NAMES ROSTER BADGE --------------------------------------------------------------------------- -{% for chair in chairs %}{{ chair.name|safe|stringformat:"-33s" }}{{ chair.groups|stringformat:"-36s" }}BLUE -{% endfor %} \ No newline at end of file diff --git a/ietf/secr/templates/groups/charter.html b/ietf/secr/templates/groups/charter.html deleted file mode 100644 index 84649f4fde..0000000000 --- a/ietf/secr/templates/groups/charter.html +++ /dev/null @@ -1,25 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles %} - -{% block title %}Groups - Charter{% endblock %} - -{% block extrahead %}{{ block.super }} - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Groups - » {{ group.acronym }} - » Charter -{% endblock %} - -{% block content %} - -
    -

    Groups - Charter

    -
    -    {% if charter_txt %}{{ charter_text }}{% else %}Charter not found.{% endif %}
    -    
    -
    - -{% endblock %} diff --git a/ietf/secr/templates/groups/people.html b/ietf/secr/templates/groups/people.html deleted file mode 100644 index 3c207d3dbf..0000000000 --- a/ietf/secr/templates/groups/people.html +++ /dev/null @@ -1,72 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles widget_tweaks %} - -{% block title %}Groups - People{% endblock %} - -{% block extrahead %}{{ block.super }} - - - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Groups - » {{ group.acronym }} - » People -{% endblock %} - -{% block content %} - -
    -

    People

    - - - - - - - - - - {% if group.role_set.all %} - - {% for role in group.role_set.all %} - - - - - - - {% endfor %} - - {% endif %} -
    RoleNameEmailAction
    {{ role.name }}{{ role.person }}{{ role.email }}Delete
    - - - -
    -
      -
    • -
    -
    - -
    - -{% endblock %} \ No newline at end of file diff --git a/ietf/secr/templates/groups/search.html b/ietf/secr/templates/groups/search.html deleted file mode 100644 index fa7d3d2154..0000000000 --- a/ietf/secr/templates/groups/search.html +++ /dev/null @@ -1,39 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles %} - -{% block title %}Groups - Search{% endblock %} - -{% block extrahead %}{{ block.super }} - - - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Groups -{% endblock %} - -{% block content %} - -
    -

    Groups - Search

    -
    {% csrf_token %} - - - - {{ form.as_table }} - -
    - - {% include "includes/buttons_search.html" %} - -
    - - -
    - -{% endblock %} \ No newline at end of file diff --git a/ietf/secr/templates/groups/view.html b/ietf/secr/templates/groups/view.html deleted file mode 100644 index a0509b0630..0000000000 --- a/ietf/secr/templates/groups/view.html +++ /dev/null @@ -1,123 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles %} - -{% block title %}Groups - View{% endblock %} - -{% block extrahead %}{{ block.super }} - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Groups - » {{ group.acronym }} -{% endblock %} - -{% block content %} - -
    -
    -

    Groups - View

    - - - - - - - - - - {% comment %} - Here we need to check that group.area_director and group.area_director.area are defined before referencing. - Otherwise the template would raise errors if the group area director record didn't exist or - in the case of Area Director = TBD, the area field is NULL - {% endcomment %} - - - - - - - - - - {% if group.liaison_contacts %} - - {% endif %} - {% if group.features.has_chartering_process %} - - {% else %} - - {% endif %} - - - -
    Group Acronym:{{ group.acronym }}
    Group Name:{{ group.name }}
    Status:{{ group.state }}
    Type:{{ group.type }}
    Proposed Date:{{ group.proposed_date|date:"Y-m-d" }}
    Start Date:{{ group.start_date|date:"Y-m-d" }}
    Concluded Date:{{ group.concluded_date|date:"Y-m-d" }}
    Primary Area:{% if not group.parent %}(No Data){% else %} - {{ group.parent }} - {% endif %} -
    Primary Area Director:{% if group.ad_role %} - {{ group.ad_role.person }} - {% endif %} -
    Meeting Scheduled:{{ group.meeting_scheduled}}
    Email Address:{{ group.list_email }}
    Email Subscription:{{ group.list_subscribe }}
    Email Archive:{{ group.list_archive }}
    Default Liaison Contacts:{{ group.liaison_contacts.contacts }}
    Charter:View Charter
    Description:{{ group.description }}
    Comments:{{ group.comments }}
    Last Modified Date:{{ group.time }}
    - - -
    -
    - - - - - - - - -
    - -
    -
      -
    • -
    • - {% comment %} -
    • -
    • - {% endcomment %} -
    -
    -
    - -{% endblock %} \ No newline at end of file diff --git a/ietf/secr/templates/includes/activities.html b/ietf/secr/templates/includes/activities.html deleted file mode 100644 index 3e79c9aed4..0000000000 --- a/ietf/secr/templates/includes/activities.html +++ /dev/null @@ -1,23 +0,0 @@ -

    Activies Log

    - diff --git a/ietf/secr/templates/includes/buttons_next_cancel.html b/ietf/secr/templates/includes/buttons_next_cancel.html deleted file mode 100644 index 95d25f55bc..0000000000 --- a/ietf/secr/templates/includes/buttons_next_cancel.html +++ /dev/null @@ -1,6 +0,0 @@ -
    -
      -
    • -
    • -
    -
    diff --git a/ietf/secr/templates/includes/buttons_submit_cancel.html b/ietf/secr/templates/includes/buttons_submit_cancel.html deleted file mode 100644 index df40c98255..0000000000 --- a/ietf/secr/templates/includes/buttons_submit_cancel.html +++ /dev/null @@ -1,6 +0,0 @@ -
    -
      -
    • -
    • -
    -
    diff --git a/ietf/secr/templates/includes/group_search_results.html b/ietf/secr/templates/includes/group_search_results.html deleted file mode 100644 index fb8b2b06e3..0000000000 --- a/ietf/secr/templates/includes/group_search_results.html +++ /dev/null @@ -1,24 +0,0 @@ - - - - - - - - - - - - {% for item in results %} - - - - - - - - {% empty %} - - {% endfor %} - -
    Group NameGroup AcronymStatusTypeMeeting Scheduled
    {{item.name}}{{item.acronym}}{{item.state}}{{item.type}}{{item.meeting_scheduled}}
    No Results
    diff --git a/ietf/secr/templates/includes/proceedings_functions.html b/ietf/secr/templates/includes/proceedings_functions.html deleted file mode 100644 index 33e11a1489..0000000000 --- a/ietf/secr/templates/includes/proceedings_functions.html +++ /dev/null @@ -1,13 +0,0 @@ - -

    Use this to process meeting materials files that have been converted to PDF and uploaded to the server.

    -
      -
    • - -   {{ ppt_count }} PowerPoint files waiting to be converted -
    • -
    -

    Use this to input session recording information.

    -
      -
    • -
    • -
    diff --git a/ietf/secr/templates/includes/session_info.txt b/ietf/secr/templates/includes/session_info.txt index eea4a5f174..bffc13e3ef 100644 --- a/ietf/secr/templates/includes/session_info.txt +++ b/ietf/secr/templates/includes/session_info.txt @@ -14,7 +14,7 @@ Conflicts to Avoid: {% if session.adjacent_with_wg %} Adjacent with WG: {{ session.adjacent_with_wg }}{% endif %} {% if session.timeranges_display %} Can't meet: {{ session.timeranges_display|join:", " }}{% endif %} -People who must be present: +Participants who must be present: {% for person in session.bethere %} {{ person.ascii_name }} {% endfor %} Resources Requested: diff --git a/ietf/secr/templates/includes/sessions_footer.html b/ietf/secr/templates/includes/sessions_footer.html deleted file mode 100755 index a41a8b8db3..0000000000 --- a/ietf/secr/templates/includes/sessions_footer.html +++ /dev/null @@ -1,6 +0,0 @@ - \ No newline at end of file diff --git a/ietf/secr/templates/includes/sessions_request_form.html b/ietf/secr/templates/includes/sessions_request_form.html deleted file mode 100755 index 6d43447102..0000000000 --- a/ietf/secr/templates/includes/sessions_request_form.html +++ /dev/null @@ -1,133 +0,0 @@ -* Required Field -
    {% csrf_token %} - {{ form.session_forms.management_form }} - {% if form.non_field_errors %} - {{ form.non_field_errors }} - {% endif %} - - - - - - {% if group.features.acts_like_wg %} - - {% if not is_virtual %} - - {% endif %} - - {% else %}{# else not group.features.acts_like_wg #} - {% for session_form in form.session_forms %} - - {% endfor %} - {% endif %} - - - - - - - - - - {% if not is_virtual %} - - - - - - - - - - - - - - - - - - - - - - - {% endif %} - - - - - - -
    Working Group Name:{{ group.name }} ({{ group.acronym }})
    Area Name:{% if group.parent %}{{ group.parent.name }} ({{ group.parent.acronym }}){% endif %}
    Number of Sessions:*{{ form.num_session.errors }}{{ form.num_session }}
    Session 1:*{% include 'meeting/session_details_form.html' with form=form.session_forms.0 only %}
    Session 2:*{% include 'meeting/session_details_form.html' with form=form.session_forms.1 only %}
    Time between two sessions:{{ form.session_time_relation.errors }}{{ form.session_time_relation }}
    Additional Session Request:{{ form.third_session }} Check this box to request an additional session.
    - Additional slot may be available after agenda scheduling has closed and with the approval of an Area Director.
    -
    - Third Session: - {% include 'meeting/session_details_form.html' with form=form.session_forms.2 only %} -
    -
    Session {{ forloop.counter }}:*{% include 'meeting/session_details_form.html' with form=session_form only %}
    Number of Attendees:{% if not is_virtual %}*{% endif %}{{ form.attendees.errors }}{{ form.attendees }}
    People who must be present: - {{ form.bethere.errors }} - {{ form.bethere }} -

    - You should not include the Area Directors; they will be added automatically. -

    -
    Conflicts to Avoid: - - - - - - - {% for cname, cfield, cselector in form.wg_constraint_fields %} - - {% if forloop.first %}{% endif %} - - - - {% empty %}{# shown if there are no constraint fields #} - - {% endfor %} - {% if form.inactive_wg_constraints %} - {% for cname, value, field in form.inactive_wg_constraints %} - - {% if forloop.first %} - - {% endif %} - - - - {% endfor %} - {% endif %} - - - - - -
    Other WGs that included {{ group.name }} in their conflict lists:{{ session_conflicts.inbound|default:"None" }}
    WG Sessions:
    You may select multiple WGs within each category
    {{ cname|title }}{{ cselector }} -
    - {{ cfield.errors }}{{ cfield }} -
    No constraints are enabled for this meeting.
    - Disabled for this meeting - {{ cname|title }}
    {{ field }} {{ field.label }}
    BOF Sessions:If the sessions can not be found in the fields above, please enter free form requests in the Special Requests field below.
    -
    Resources requested: - {{ form.resources.errors }} {{ form.resources }} -
    Times during which this WG can not meet:
    Please explain any selections in Special Requests below.
    {{ form.timeranges.errors }}{{ form.timeranges }}
    - Plan session adjacent with another WG:
    - (Immediately before or after another WG, no break in between, in the same room.) -
    {{ form.adjacent_with_wg.errors }}{{ form.adjacent_with_wg }}
    - Joint session with:
    - (To request one session for multiple WGs together.) -
    {{ form.joint_with_groups_selector }} -
    - {{ form.joint_with_groups.errors }}{{ form.joint_with_groups }} -
    - Of the sessions requested by this WG, the joint session, if applicable, is: - {{ form.joint_for_session.errors }}{{ form.joint_for_session }}
    Special Requests:
     
    i.e. restrictions on meeting times / days, etc.
    (limit 200 characters)
    {{ form.comments.errors }}{{ form.comments }}
    - -
    -
      -
    • -
    • -
    -
    -
    \ No newline at end of file diff --git a/ietf/secr/templates/includes/sessions_request_view.html b/ietf/secr/templates/includes/sessions_request_view.html deleted file mode 100644 index 3854651d38..0000000000 --- a/ietf/secr/templates/includes/sessions_request_view.html +++ /dev/null @@ -1,63 +0,0 @@ -{% load ams_filters %} - - - - - - {% if form %} - {% include 'includes/sessions_request_view_formset.html' with formset=form.session_forms group=group session=session only %} - {% else %} - {% include 'includes/sessions_request_view_session_set.html' with session_set=sessions group=group session=session only %} - {% endif %} - - - - - - - - - - {% if not is_virtual %} - - - - - {% endif %} - - - - - - - - - {% if not is_virtual %} - - - - - - - - - {% endif %} - - -
    Working Group Name:{{ group.name }} ({{ group.acronym }})
    Area Name:{{ group.parent }}
    Number of Sessions Requested:{% if session.third_session %}3{% else %}{{ session.num_session }}{% endif %}
    Number of Attendees:{{ session.attendees }}
    Conflicts to Avoid: - {% if session_conflicts.outbound %} - - - {% for conflict in session_conflicts.outbound %} - - {% endfor %} - -
    {{ conflict.name|title }}: {{ conflict.groups }}
    - {% else %}None{% endif %} -
    Other WGs that included {{ group }} in their conflict list:{% if session_conflicts.inbound %}{{ session_conflicts.inbound }}{% else %}None so far{% endif %}
    Resources requested:{% if session.resources %}
      {% for resource in session.resources %}
    • {{ resource.desc }}
    • {% endfor %}
    {% else %}None so far{% endif %}
    People who must be present:{% if session.bethere %}
      {% for person in session.bethere %}
    • {{ person }}
    • {% endfor %}
    {% else %}None{% endif %}
    Can not meet on:{% if session.timeranges_display %}{{ session.timeranges_display|join:', ' }}{% else %}No constraints{% endif %}
    Adjacent with WG:{{ session.adjacent_with_wg|default:'No preference' }}
    Joint session: - {% if session.joint_with_groups %} - {{ session.joint_for_session_display }} with: {{ session.joint_with_groups }} - {% else %} - Not a joint session - {% endif %} -
    Special Requests:{{ session.comments }}
    diff --git a/ietf/secr/templates/includes/sessions_request_view_formset.html b/ietf/secr/templates/includes/sessions_request_view_formset.html deleted file mode 100644 index ff502dea30..0000000000 --- a/ietf/secr/templates/includes/sessions_request_view_formset.html +++ /dev/null @@ -1,30 +0,0 @@ -{% load ams_filters %}{# keep this in sync with sessions_request_view_session_set.html #} -{% for sess_form in formset %}{% if sess_form.cleaned_data and not sess_form.cleaned_data.DELETE %} - - Session {{ forloop.counter }}: - -
    -
    Length
    -
    {{ sess_form.cleaned_data.requested_duration.total_seconds|display_duration }}
    - {% if sess_form.cleaned_data.name %} -
    Name
    -
    {{ sess_form.cleaned_data.name }}
    {% endif %} - {% if sess_form.cleaned_data.purpose.slug != 'regular' %} -
    Purpose
    -
    - {{ sess_form.cleaned_data.purpose }} - {% if sess_form.cleaned_data.purpose.timeslot_types|length > 1 %}({{ sess_form.cleaned_data.type }} - ){% endif %} -
    - {% endif %} -
    - - - {% if group.features.acts_like_wg and forloop.counter == 2 and not is_virtual %} - - Time between sessions: - {% if session.session_time_relation_display %}{{ session.session_time_relation_display }}{% else %}No - preference{% endif %} - - {% endif %} -{% endif %}{% endfor %} \ No newline at end of file diff --git a/ietf/secr/templates/includes/sessions_request_view_session_set.html b/ietf/secr/templates/includes/sessions_request_view_session_set.html deleted file mode 100644 index 1f953ae3a7..0000000000 --- a/ietf/secr/templates/includes/sessions_request_view_session_set.html +++ /dev/null @@ -1,30 +0,0 @@ -{% load ams_filters %}{# keep this in sync with sessions_request_view_formset.html #} -{% for sess in session_set %} - - Session {{ forloop.counter }}: - -
    -
    Length
    -
    {{ sess.requested_duration.total_seconds|display_duration }}
    - {% if sess.name %} -
    Name
    -
    {{ sess.name }}
    {% endif %} - {% if sess.purpose.slug != 'regular' %} -
    Purpose
    -
    - {{ sess.purpose }} - {% if sess.purpose.timeslot_types|length > 1 %}({{ sess.type }} - ){% endif %} -
    - {% endif %} -
    - - - {% if group.features.acts_like_wg and forloop.counter == 2 and not is_virtual %} - - Time between sessions: - {% if session.session_time_relation_display %}{{ session.session_time_relation_display }}{% else %}No - preference{% endif %} - - {% endif %} -{% endfor %} \ No newline at end of file diff --git a/ietf/secr/templates/includes/upload_footer.html b/ietf/secr/templates/includes/upload_footer.html deleted file mode 100755 index cc7858e1f7..0000000000 --- a/ietf/secr/templates/includes/upload_footer.html +++ /dev/null @@ -1,6 +0,0 @@ - diff --git a/ietf/secr/templates/index.html b/ietf/secr/templates/index.html new file mode 100644 index 0000000000..9ea7021279 --- /dev/null +++ b/ietf/secr/templates/index.html @@ -0,0 +1,31 @@ +{# Copyright The IETF Trust 2007-2025, All Rights Reserved #} +{% extends "base.html" %} +{% load static %} +{% load ietf_filters %} +{% block title %}Secretariat Dashboard{% endblock %} +{% block content %} +

    Secretariat Dashboard

    +
    + {% if user|has_role:"Secretariat" %} +

    IESG

    + + +

    IDs and WGs Process

    + + +

    Meetings and Proceedings

    + + {% else %} + + {% endif %} +
    +{% endblock %} \ No newline at end of file diff --git a/ietf/secr/templates/main.html b/ietf/secr/templates/main.html deleted file mode 100644 index dcda4718de..0000000000 --- a/ietf/secr/templates/main.html +++ /dev/null @@ -1,81 +0,0 @@ -{% extends "base_site.html" %} -{% load ietf_filters %} - -{% block content %} -
    - - {% if user|has_role:"Secretariat" %} - - - - - - - - - - - - - - - {% else %} - - - - - - - - - - - - - - - {% endif %} - -
    -{% endblock %} \ No newline at end of file diff --git a/ietf/secr/templates/meetings/add.html b/ietf/secr/templates/meetings/add.html index 5a78255265..b2cc2617dc 100644 --- a/ietf/secr/templates/meetings/add.html +++ b/ietf/secr/templates/meetings/add.html @@ -1,5 +1,5 @@ {% extends "base_site.html" %} -{% load staticfiles %} +{% load static %} {% block title %}Meetings - Add{% endblock %} diff --git a/ietf/secr/templates/meetings/base_rooms_times.html b/ietf/secr/templates/meetings/base_rooms_times.html index dc08e9eb50..263418fabf 100644 --- a/ietf/secr/templates/meetings/base_rooms_times.html +++ b/ietf/secr/templates/meetings/base_rooms_times.html @@ -1,11 +1,9 @@ {% extends "base_site_bootstrap.html" %} -{% load staticfiles %} +{% load static %} {% block title %}Meetings{% endblock %} {% block extrahead %}{{ block.super }} - - {% endblock %} diff --git a/ietf/secr/templates/meetings/blue_sheet.html b/ietf/secr/templates/meetings/blue_sheet.html deleted file mode 100644 index d67efd9f63..0000000000 --- a/ietf/secr/templates/meetings/blue_sheet.html +++ /dev/null @@ -1,48 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles %} - -{% block title %}Meetings - Blue Sheet{% endblock %} - -{% block extrahead %}{{ block.super }} - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Meetings - » {{ meeting.number }} - » Blue Sheets -{% endblock %} - -{% block content %} - -
    -

    IETF {{ meeting.number }} - Blue Sheet

    - -

    Use this to generate blue sheets for meeting sessions.

    -
      -
    • -
      {% csrf_token %} - -
      -   Last run: - {% if last_run %} - {{ last_run }} - {% else %} - Never - {% endif %} -
    • -
    -

    Use this to download the blue sheets from the server.

    -
      -
    • - -
    • -
    - -
    -

    - Use the session details page for a group to upload scanned bluesheets. The session details pages for a group can be reached from the meeting's materials page. -

    -
    - -{% endblock %} diff --git a/ietf/secr/templates/meetings/edit_meeting.html b/ietf/secr/templates/meetings/edit_meeting.html index 773536e654..474373dbee 100644 --- a/ietf/secr/templates/meetings/edit_meeting.html +++ b/ietf/secr/templates/meetings/edit_meeting.html @@ -1,5 +1,5 @@ {% extends "base_site.html" %} -{% load staticfiles %} +{% load static %} {% block title %}Meetings - Edit{% endblock %} diff --git a/ietf/secr/templates/meetings/main.html b/ietf/secr/templates/meetings/main.html index 90c3802892..ff110dd978 100755 --- a/ietf/secr/templates/meetings/main.html +++ b/ietf/secr/templates/meetings/main.html @@ -1,5 +1,5 @@ {% extends "base_site.html" %} -{% load staticfiles %} +{% load static %} {% block title %}Meetings{% endblock %} diff --git a/ietf/secr/templates/meetings/notifications.html b/ietf/secr/templates/meetings/notifications.html index bf7099577e..dbe66ff283 100644 --- a/ietf/secr/templates/meetings/notifications.html +++ b/ietf/secr/templates/meetings/notifications.html @@ -1,5 +1,5 @@ {% extends "base_site.html" %} -{% load staticfiles %} +{% load static %} {% block title %}Meetings{% endblock %} diff --git a/ietf/secr/templates/meetings/regular_session_edit.html b/ietf/secr/templates/meetings/regular_session_edit.html index fbfba4f967..9993858be1 100644 --- a/ietf/secr/templates/meetings/regular_session_edit.html +++ b/ietf/secr/templates/meetings/regular_session_edit.html @@ -1,5 +1,5 @@ {% extends "base_site.html" %} -{% load staticfiles tz %} +{% load static tz %} {% block title %}Meetings{% endblock %} diff --git a/ietf/secr/templates/meetings/view.html b/ietf/secr/templates/meetings/view.html index d552d38dca..89bd8f7e03 100644 --- a/ietf/secr/templates/meetings/view.html +++ b/ietf/secr/templates/meetings/view.html @@ -1,5 +1,5 @@ {% extends "base_site.html" %} -{% load staticfiles %} +{% load static %} {% block title %}Meetings{% endblock %} @@ -37,7 +37,6 @@

    IETF {{ meeting.number }} - View

    • -
    • diff --git a/ietf/secr/templates/proceedings/audio_import_warning.txt b/ietf/secr/templates/proceedings/audio_import_warning.txt deleted file mode 100644 index 322a43f5fd..0000000000 --- a/ietf/secr/templates/proceedings/audio_import_warning.txt +++ /dev/null @@ -1,9 +0,0 @@ - -WARNING: - -After the last meeting session audio file import there are {{ unmatched_files|length }} -file(s) that were not matched to a timeslot. - -{% for file in unmatched_files %}{{ file }} -{% endfor %} - diff --git a/ietf/secr/templates/proceedings/interim_directory.html b/ietf/secr/templates/proceedings/interim_directory.html deleted file mode 100644 index 66d0b9714c..0000000000 --- a/ietf/secr/templates/proceedings/interim_directory.html +++ /dev/null @@ -1,36 +0,0 @@ -{% extends "base_site.html" %} - -{% block content %} - -

      Interim Meeting Proceedings

      - - - - - - - - {% for meeting in meetings %} - - - - {% if meeting.schedule %} - - {% else %} - - {% endif %} - {% if meeting.minutes %} - - {% else %} - - {% endif %} - {% if meeting.get_proceedings_url %} - - {% else %} - - {% endif %} - - {% endfor %} -
      DateGroup
      {{ meeting.date }}{{ meeting.group.acronym }}AgendaAgendaMinutesMinutesProceedingsProceedings
      - -{% endblock %} \ No newline at end of file diff --git a/ietf/secr/templates/proceedings/main.html b/ietf/secr/templates/proceedings/main.html deleted file mode 100644 index a9689430a9..0000000000 --- a/ietf/secr/templates/proceedings/main.html +++ /dev/null @@ -1,96 +0,0 @@ -{% extends "base_site.html" %} -{% load ietf_filters %} -{% load staticfiles %} -{% block title %}Proceeding manager{% endblock %} - -{% block extrahead %}{{ block.super }} - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Proceedings -{% endblock %} - -{% block content %} - -
      -

      Proceedings

      -
      - - - - - - - {% if meetings %} - - {% for meeting in meetings %} - - - - {% endfor %} - - {% endif %} -
      IETF Meeting
      - {{ meeting.number }} -
      - {% if user|has_role:"Secretariat" %} -
      -
        -
      • -
      -
      - {% endif %} -
      - -
      -
      - - - - - - - {% if interim_meetings %} - - {% for meeting in interim_meetings %} - - - - - - {% endfor %} - - {% endif %} -
      Interim Meeting
      {{ meeting.group.acronym }}{{ meeting.date }}
      -
      -
      -
        -
      • -
      -
      -
      - -
      - {% if not user|has_role:"Secretariat" %} -
      -
      -

      The list(s) above includes those meetings which you can upload materials for. Click on the meeting number or interim meeting date to continue.

      - {% endif %} - - -
      -
        -
      • -
      -
      -
      - -{% endblock %} - -{% block footer-extras %} - {% include "includes/upload_footer.html" %} -{% endblock %} -~ -~ -~ \ No newline at end of file diff --git a/ietf/secr/templates/proceedings/recording.html b/ietf/secr/templates/proceedings/recording.html deleted file mode 100755 index 96b73d65c7..0000000000 --- a/ietf/secr/templates/proceedings/recording.html +++ /dev/null @@ -1,121 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles tz %} - -{% block title %}Proceedings{% endblock %} - -{% block extrastyle %}{{ block.super }} - - -{% endblock %} - -{% block extrahead %}{{ block.super }} - - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - {% if meeting.type_id == "interim" %} - » Proceedings - » {{ meeting }} - {% else %} - » Proceedings - » {{ meeting.number }} - » Recording - {% endif %} -{% endblock %} - -{% block content %} - -
      - -

      Recording Metadata

      -
      {% csrf_token %} - - - - {{ form.as_table }} - -
      - -
      -
        -
      • -
      • -
      -
      - -
      - - - - {% if unmatched_recordings %} - - {% endif %} - -
      - - -{% endblock %} - -{% block footer-extras %} - {% include "includes/upload_footer.html" %} -{% endblock %} diff --git a/ietf/secr/templates/proceedings/recording_edit.html b/ietf/secr/templates/proceedings/recording_edit.html deleted file mode 100755 index 4a167db4e4..0000000000 --- a/ietf/secr/templates/proceedings/recording_edit.html +++ /dev/null @@ -1,41 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles %} -{% block title %}Edit Recording{% endblock %} - -{% block extrahead %}{{ block.super }} - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - {% if meeting.type_id == "interim" %} - » Proceedings - » {{ meeting }} - » Recording - » {{ recording.name }} - {% else %} - » Proceedings - » {{ meeting.number }} - » Recording - » {{ recording.name }} - {% endif %} -{% endblock %} - -{% block content %} - -
      -

      Recording Metadata for Group: {{ form.instance.group.acronym }} | Session: {{ form.instance.session_set.first.official_scheduledsession.timeslot.time }}

      -

      Edit Recording Metadata:

      -
      {% csrf_token %} - - - - {{ form.as_table }} - -
      - - {% include "includes/buttons_save_cancel.html" %} - -
      -
      - -{% endblock %} diff --git a/ietf/secr/templates/proceedings/report_id_activity.txt b/ietf/secr/templates/proceedings/report_id_activity.txt deleted file mode 100644 index 6f2aaf097a..0000000000 --- a/ietf/secr/templates/proceedings/report_id_activity.txt +++ /dev/null @@ -1,21 +0,0 @@ -IETF Activity since last IETF Meeting --------------------- - -IETF Activity since last IETF Meeting ({{ meeting.city }}) - -{{ new|stringformat:"3s" }} New I-Ds ({{ updated }} of which were updated, some ({{ updated_more }}) more than once) -{{ total_updated|stringformat:"3s" }} I-Ds were updated (Some more than once) -{{ last_call|stringformat:"3s" }} I-Ds Last Called -{{ approved|stringformat:"3s" }} I-Ds approved for publication - -In the final 4 weeks before meeting - -{{ ff_new_count|stringformat:"3s" }} New I-Ds were received - {{ ff_new_percent }} of total newbies since last meeting -{{ ff_update_count|stringformat:"3s" }} I-Ds were updated - {{ ff_update_percent }} of total updated since last meeting - - Week1 0 % - Week2 0 % - Week3 0 % - Week4 0 % - -The IESG Secretary. diff --git a/ietf/secr/templates/proceedings/report_progress_report.txt b/ietf/secr/templates/proceedings/report_progress_report.txt deleted file mode 100644 index 8c3d87548e..0000000000 --- a/ietf/secr/templates/proceedings/report_progress_report.txt +++ /dev/null @@ -1,48 +0,0 @@ -{% load ams_filters %} - IETF Activity since last IETF Meeting - {{ start_date }} to {{ end_date }} - - 1) {{ action_events.count }} IESG Protocol and Document Actions this period -{% for event in action_events %} - {{ event.doc.title }} ({{ event.doc.intended_std_level }}) -{% endfor %} - - 2) {{ lc_events.count }} IESG Last Calls issued to the IETF this period -{% for event in lc_events %} - {{ event.doc.title }} - {{ event.doc.file_tag|safe }} ({{ event.doc.intended_std_level }}) -{% endfor %} - - 3) {{ new_groups.count }} New Working Group(s) formed this period - {% for group in new_groups %} - {{ group }} ({{ group.acronym }}) - {% endfor %} - - 4) {{ concluded_groups.count }} Working Group(s) concluded this period - {% for group in concluded_groups %} - {{ group }} ({{ group.acronym }}) - {% endfor %} - - 5) {{ new_docs|length }} new or revised Internet-Drafts this period - - (o - Revised Internet-Draft; + - New Internet-Draft) - - WG I-D Title - ------- ------------------------------------------ - {% for doc in new_docs %} - ({{ doc.group.acronym|stringformat:"8s" }}) {% if doc.rev == "00" %} + {% else %} o {% endif %}{{ doc.title }} - {{ doc.file_tag|safe }} - {% endfor %} - - 6) {{ rfcs.count }} RFC(s) produced this period - - S - Standard; PS - Proposed Standard; DS - Draft Standard; - B - Best Current Practices; E - Experimental; I - Informational - - RFC Stat WG Published Title -------- -- ---------- ---------- ----------------------------------------- -{% for event in rfcs %} -{{ event.doc.canonical_name|upper }} {{ event.doc.intended_std_level.name|abbr_status|stringformat:"2s" }} ({{ event.doc.group.acronym|stringformat:"8s" }}) {{ event.time|date:"M d" }} {{ event.doc.title }} -{% endfor %} - - {{ counts.std }} Standards Track; {{ counts.bcp }} BCP; {{ counts.exp }} Experimental; {{ counts.inf }} Informational diff --git a/ietf/secr/templates/proceedings/select.html b/ietf/secr/templates/proceedings/select.html deleted file mode 100755 index e2ac7e50bf..0000000000 --- a/ietf/secr/templates/proceedings/select.html +++ /dev/null @@ -1,39 +0,0 @@ -{% extends "base_site.html" %} -{% load ietf_filters %} -{% load staticfiles %} -{% block title %}Proceedings{% endblock %} - -{% block extrahead %}{{ block.super }} - -{% endblock %} - - -{% block breadcrumbs %}{{ block.super }} - » Proceedings - » {{ meeting.number }} -{% endblock %} - -{% block instructions %} - Instructions -{% endblock %} - -{% block content %} - -
      - - - -
      - -{% endblock %} - -{% block footer-extras %} - {% include "includes/upload_footer.html" %} -{% endblock %} diff --git a/ietf/secr/templates/proceedings/status.html b/ietf/secr/templates/proceedings/status.html deleted file mode 100644 index f4b1160c55..0000000000 --- a/ietf/secr/templates/proceedings/status.html +++ /dev/null @@ -1,47 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles %} -{% block title %}Proceedings - Status{% endblock %} - -{% block extrahead %}{{ block.super }} - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Proceedings - » {{ meeting.meeting_num }} - » Status -{% endblock %} - -{% block content %} - - -
      • Changing one of the status below will result in changing the status of Proceedings {{ meeting.meeting_num }}.
      - -
      -

      IETF {{ meeting.meeting_num }}

      - - {% csrf_token %} - - - {% if not proceeding.frozen %} - - - - {% endif %} - {% if proceeding.frozen %} - - - - {% endif %} - - -
      Active Proceeding
      Frozen Proceeding
      - -
      -
        -
      • -
      -
      -
      - -{% endblock %} diff --git a/ietf/secr/templates/proceedings/view.html b/ietf/secr/templates/proceedings/view.html deleted file mode 100644 index c21d5c94b0..0000000000 --- a/ietf/secr/templates/proceedings/view.html +++ /dev/null @@ -1,59 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles %} -{% block title %}Proceedings - View{% endblock %} - -{% block extrahead %}{{ block.super }} - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Proceedings - » {{ meeting.number }} -{% endblock %} - -{% block content %} - -
      -

      IETF {{meeting.number}} Meeting - View

      - {% if meeting.frozen == 1 %} -
      • THIS IS A FROZEN PROCEEDING
      - {% endif %} - - - - - - - -
      Meeting Start Date: {{ meeting.date }}
      Meeting End Date: {{ meeting.end_date }}
      Meeting City: {{ meeting.city }}
      Meeting Country: {{ meeting.country }}
      - - - -
      - {% if meeting.frozen == 0 %} -
        -
      • -
      • -
      • -
      - {% endif %} - {% if meeting.frozen == 1 %} -
        -
      • -
      - {% endif %} -
      -
      - -{% endblock %} \ No newline at end of file diff --git a/ietf/secr/templates/proceedings/wait.html b/ietf/secr/templates/proceedings/wait.html deleted file mode 100644 index 9590f78541..0000000000 --- a/ietf/secr/templates/proceedings/wait.html +++ /dev/null @@ -1,30 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles %} -{% block title %}Proceeding manager{% endblock %} - -{% block extrahead %}{{ block.super }} - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Proceedings -{% endblock %} - -{% block content %} - -
      -

      Proceedings

      -
      -

      {{ message }}

      - loading... -
      - -{% endblock %} - -{% block footer-extras %} - {% include "includes/upload_footer.html" %} -{% endblock %} -~ -~ -~ - diff --git a/ietf/secr/templates/roles/main.html b/ietf/secr/templates/roles/main.html deleted file mode 100755 index 88be7cdccf..0000000000 --- a/ietf/secr/templates/roles/main.html +++ /dev/null @@ -1,84 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles %} - -{% block title %}Roles{% endblock %} - -{% block extrahead %}{{ block.super }} - - - - - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Roles -{% endblock %} -{% block instructions %} - Instructions -{% endblock %} - -{% block content %} - -
      -
      {% csrf_token %} -

      Role Tool

      - -
      - - - - -
      - -
      -
        -
      • -
      -
      - -
      - -{% endblock %} \ No newline at end of file diff --git a/ietf/secr/templates/roles/roles.html b/ietf/secr/templates/roles/roles.html deleted file mode 100644 index 615452a486..0000000000 --- a/ietf/secr/templates/roles/roles.html +++ /dev/null @@ -1,20 +0,0 @@ - - - - - - - - - - - {% for role in roles %} - - - - - - - {% endfor %} - -
      RoleNameEmailAction
      {{ role.name }}{{ role.person }}{{ role.email }}Delete
      diff --git a/ietf/secr/templates/rolodex/add.html b/ietf/secr/templates/rolodex/add.html index 272b844fa3..5adb738f2b 100644 --- a/ietf/secr/templates/rolodex/add.html +++ b/ietf/secr/templates/rolodex/add.html @@ -1,5 +1,5 @@ {% extends "base_site.html" %} -{% load staticfiles %} +{% load static %} {% block title %}Rolodex - Add{% endblock %} diff --git a/ietf/secr/templates/rolodex/edit.html b/ietf/secr/templates/rolodex/edit.html index 28a125f104..ed4c0f97e2 100644 --- a/ietf/secr/templates/rolodex/edit.html +++ b/ietf/secr/templates/rolodex/edit.html @@ -1,5 +1,5 @@ {% extends "base_site.html" %} -{% load staticfiles %} +{% load static %} {% block title %}Rolodex - Edit{% endblock %} diff --git a/ietf/secr/templates/rolodex/search.html b/ietf/secr/templates/rolodex/search.html index 8994cfabd1..065b0463f8 100644 --- a/ietf/secr/templates/rolodex/search.html +++ b/ietf/secr/templates/rolodex/search.html @@ -1,5 +1,5 @@ {% extends "base_site.html" %} -{% load staticfiles %} +{% load static %} {% block title %}Rolodex - Search{% endblock %} diff --git a/ietf/secr/templates/rolodex/view.html b/ietf/secr/templates/rolodex/view.html index 738ab3361a..d1a78cfaa5 100644 --- a/ietf/secr/templates/rolodex/view.html +++ b/ietf/secr/templates/rolodex/view.html @@ -44,9 +44,9 @@

      Roles

      {{ role.name }} {% if role.group.type.slug == "area" %} - {{ role.group.acronym }}{% if role.group.state.slug == "conclude" %} (concluded){% endif %} + {{ role.group.acronym }}{% if role.group.state.slug == "conclude" %} (concluded){% endif %} {% else %} - {{ role.group.acronym }}{% if role.group.state.slug == "conclude" %} (concluded){% endif %} + {{ role.group.acronym }}{% if role.group.state.slug == "conclude" %} (concluded){% endif %} {% endif %} {{ role.email }} diff --git a/ietf/secr/templates/sreq/confirm.html b/ietf/secr/templates/sreq/confirm.html deleted file mode 100755 index 4bf26dd858..0000000000 --- a/ietf/secr/templates/sreq/confirm.html +++ /dev/null @@ -1,57 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles %} - -{% block title %}Sessions - Confirm{% endblock %} - -{% block extrastyle %} - -{% endblock %} - -{% block extrahead %}{{ block.super }} - - {{ form.media }} -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Sessions - » New - » Session Request Confirmation -{% endblock %} - -{% block content %} - -
      -

      Sessions - Confirm

      - - {% include "includes/sessions_request_view.html" %} - - {% if group.features.acts_like_wg and form.session_forms.forms_to_keep|length > 2 %} -
      -

      - - Note: Your request for a third session must be approved by an area director before - being submitted to agenda@ietf.org. Click "Submit" below to email an approval - request to the area directors. - -

      -
      - {% endif %} - -
      - {% csrf_token %} - {{ form }} - {{ form.session_forms.management_form }} - {% for sf in form.session_forms %} - {% include 'meeting/session_details_form.html' with form=sf hidden=True only %} - {% endfor %} - {% include "includes/buttons_submit_cancel.html" %} -
      - -
      - -{% endblock %} \ No newline at end of file diff --git a/ietf/secr/templates/sreq/edit.html b/ietf/secr/templates/sreq/edit.html deleted file mode 100755 index b0bfbc1e0c..0000000000 --- a/ietf/secr/templates/sreq/edit.html +++ /dev/null @@ -1,39 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles %} -{% block title %}Sessions - Edit{% endblock %} - -{% block extrahead %}{{ block.super }} - - - {{ form.media }} - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Sessions - » {{ group.acronym }} - » Edit -{% endblock %} - -{% block instructions %} - Instructions -{% endblock %} - -{% block content %} -
      -

      IETF {{ meeting.number }}: Edit Session Request

      - -
      -{% endblock %} - -{% block footer-extras %} - {% include "includes/sessions_footer.html" %} -{% endblock %} \ No newline at end of file diff --git a/ietf/secr/templates/sreq/locked.html b/ietf/secr/templates/sreq/locked.html deleted file mode 100755 index 5f619f37cb..0000000000 --- a/ietf/secr/templates/sreq/locked.html +++ /dev/null @@ -1,30 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles %} - -{% block title %}Sessions{% endblock %} - -{% block extrahead %}{{ block.super }} - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Sessions (Locked) -{% endblock %} - -{% block content %} -

      » View list of timeslot requests

      -
      -

      Sessions - Status

      - -

      {{ message }}

      - -
      -
        -
      • -
      -
      - - -
      - -{% endblock %} \ No newline at end of file diff --git a/ietf/secr/templates/sreq/main.html b/ietf/secr/templates/sreq/main.html deleted file mode 100755 index bdb33bb77d..0000000000 --- a/ietf/secr/templates/sreq/main.html +++ /dev/null @@ -1,65 +0,0 @@ -{% extends "base_site.html" %} -{% load ietf_filters %} -{% load staticfiles %} - -{% block title %}Sessions{% endblock %} - -{% block extrahead %}{{ block.super }} - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Sessions -{% endblock %} -{% block instructions %} - Instructions -{% endblock %} - -{% block content %} -

      » View list of timeslot requests

      -
      -

      - Sessions Request Tool: IETF {{ meeting.number }} - {% if user|has_role:"Secretariat" %} - {% if is_locked %} - Tool Status: Locked - {% else %} - Tool Status: Unlocked - {% endif %} - {% endif %} -

      - -
      - -
      - -{% endblock %} - -{% block footer-extras %} - {% include "includes/sessions_footer.html" %} -{% endblock %} \ No newline at end of file diff --git a/ietf/secr/templates/sreq/new.html b/ietf/secr/templates/sreq/new.html deleted file mode 100755 index 2c6afb5576..0000000000 --- a/ietf/secr/templates/sreq/new.html +++ /dev/null @@ -1,43 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles %} - -{% block title %}Sessions- New{% endblock %} - -{% block extrahead %}{{ block.super }} - - - {{ form.media }} - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Sessions - » New Session Request -{% endblock %} - -{% block instructions %} - Instructions -{% endblock %} - -{% block content %} -
      -

      IETF {{ meeting.number }}: New Session Request

      - - {% include "includes/sessions_request_form.html" %} - -
      - -{% endblock %} - -{% block footer-extras %} - {% include "includes/sessions_footer.html" %} -{% endblock %} \ No newline at end of file diff --git a/ietf/secr/templates/sreq/not_meeting_notification.txt b/ietf/secr/templates/sreq/not_meeting_notification.txt deleted file mode 100644 index 1120f8480c..0000000000 --- a/ietf/secr/templates/sreq/not_meeting_notification.txt +++ /dev/null @@ -1,7 +0,0 @@ -{% load ams_filters %} - -{{ login|smart_login }} {{ group.acronym }} working group, indicated that the {{ group.acronym }} working group does not plan to hold a session at IETF {{ meeting.number }}. - -This message was generated and sent by the IETF Meeting Session Request Tool. - - diff --git a/ietf/secr/templates/sreq/session_approval_notification.txt b/ietf/secr/templates/sreq/session_approval_notification.txt deleted file mode 100644 index 7bb63aa3fa..0000000000 --- a/ietf/secr/templates/sreq/session_approval_notification.txt +++ /dev/null @@ -1,15 +0,0 @@ -Dear {{ group.parent }} Director(s): - -{{ header }} meeting session request has just been -submitted by {{ requester }}. -The third session requires your approval. - -To approve the session go to the session request view here: -{{ settings.IDTRACKER_BASE_URL }}{% url "ietf.secr.sreq.views.view" acronym=group.acronym %} -and click "Approve Third Session". - -Regards, - -The IETF Secretariat. - -{% include "includes/session_info.txt" %} diff --git a/ietf/secr/templates/sreq/session_cancel_notification.txt b/ietf/secr/templates/sreq/session_cancel_notification.txt deleted file mode 100644 index 3e6dd43f69..0000000000 --- a/ietf/secr/templates/sreq/session_cancel_notification.txt +++ /dev/null @@ -1,4 +0,0 @@ -{% load ams_filters %} - -A request to cancel a meeting session has just been submitted by {{ requester }}. - diff --git a/ietf/secr/templates/sreq/session_request_notification.txt b/ietf/secr/templates/sreq/session_request_notification.txt deleted file mode 100644 index db20060406..0000000000 --- a/ietf/secr/templates/sreq/session_request_notification.txt +++ /dev/null @@ -1,5 +0,0 @@ -{% load ams_filters %} - -{{ header }} meeting session request has just been submitted by {{ requester }}. - -{% include "includes/session_info.txt" %} diff --git a/ietf/secr/templates/sreq/tool_status.html b/ietf/secr/templates/sreq/tool_status.html deleted file mode 100755 index cf5131c226..0000000000 --- a/ietf/secr/templates/sreq/tool_status.html +++ /dev/null @@ -1,42 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles %} - -{% block title %}Sessions{% endblock %} - -{% block extrahead %}{{ block.super }} - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Sessions - » Session Status -{% endblock %} - -{% block content %} - -
      -

      Sessions - Status

      -

      Enter the message that you would like displayed to the WG Chair when this tool is locked.

      -
      {% csrf_token %} - - - - {{ form.as_table }} - -
      -
      -
        - {% if is_locked %} -
      • - {% else %} -
      • - {% endif %} -
      • -
      -
      - -
      - -
      - -{% endblock %} diff --git a/ietf/secr/templates/sreq/view.html b/ietf/secr/templates/sreq/view.html deleted file mode 100644 index 8cdefe194b..0000000000 --- a/ietf/secr/templates/sreq/view.html +++ /dev/null @@ -1,46 +0,0 @@ -{% extends "base_site.html" %} -{% load staticfiles %} - -{% block title %}Sessions - View{% endblock %} - -{% block extrahead %}{{ block.super }} - -{% endblock %} - -{% block breadcrumbs %}{{ block.super }} - » Sessions - » {{ group.acronym }} -{% endblock %} - -{% block instructions %} - Instructions -{% endblock %} - -{% block content %} - -
      -

      Sessions - View (meeting: {{ meeting.number }})

      - - {% include "includes/sessions_request_view.html" %} - -
      - - {% include "includes/activities.html" %} - -
      -
        -
      • - {% if show_approve_button %} -
      • - {% endif %} -
      • -
      • -
      -
      -
      - -{% endblock %} - -{% block footer-extras %} - {% include "includes/sessions_footer.html" %} -{% endblock %} diff --git a/ietf/secr/templates/telechat/base_telechat.html b/ietf/secr/templates/telechat/base_telechat.html index 73d42ea71e..1c8feaff6f 100644 --- a/ietf/secr/templates/telechat/base_telechat.html +++ b/ietf/secr/templates/telechat/base_telechat.html @@ -1,5 +1,5 @@ {% extends "base_site.html" %} -{% load staticfiles %} +{% load static %} {% block title %}Telechat{% endblock %} diff --git a/ietf/secr/templates/telechat/doc.html b/ietf/secr/templates/telechat/doc.html index 9d37db4cb0..6727e157f5 100644 --- a/ietf/secr/templates/telechat/doc.html +++ b/ietf/secr/templates/telechat/doc.html @@ -85,13 +85,13 @@

      Ballot Writeup

      {% if downrefs %}

      Downward References

      {% for ref in downrefs %} -

      Add {{ref.target.document.canonical_name}} - ({{ref.target.document.std_level}} - {{ref.target.document.stream.desc}}) +

      Add {{ref.target.name}} + ({{ref.target.std_level}} - {{ref.target.stream.desc}} stream) to downref registry.
      - {% if not ref.target.document.std_level %} + {% if not ref.target.std_level %} +++ Warning: The standards level has not been set yet!!!
      {% endif %} - {% if not ref.target.document.stream %} + {% if not ref.target.stream %} +++ Warning: document stream has not been set yet!!!
      {% endif %} {% endfor %}

      diff --git a/ietf/secr/templates/telechat/group.html b/ietf/secr/templates/telechat/group.html index 890c451e83..4e04f0e16e 100644 --- a/ietf/secr/templates/telechat/group.html +++ b/ietf/secr/templates/telechat/group.html @@ -3,7 +3,7 @@ Does anyone have an objection to the creation of this working group being sent for EXTERNAL REVIEW?

      External Review APPROVED; "The Secretariat will send a Working Group Review announcement with a copy to new-work and place it back on the agenda for the next telechat."

      External Review NOT APPROVED; -
      +
      The Secretariat will wait for instructions from
      The IESG decides the document needs more time in INTERNAL REVIEW. The Secretariat will put it back on the agenda for the next teleconference in the same category.
      The IESG has made changes since the charter was seen in INTERNAL REVIEW, and decides to send it back to INTERNAL REVIEW the charter again. diff --git a/ietf/secr/urls.py b/ietf/secr/urls.py index 196f139b2e..ab21046654 100644 --- a/ietf/secr/urls.py +++ b/ietf/secr/urls.py @@ -1,16 +1,22 @@ -from django.conf.urls import url, include +# Copyright The IETF Trust 2025, All Rights Reserved + +from django.conf import settings +from django.urls import re_path, include from django.views.generic import TemplateView +from django.views.generic.base import RedirectView urlpatterns = [ - url(r'^$', TemplateView.as_view(template_name='main.html')), - url(r'^announcement/', include('ietf.secr.announcement.urls')), - url(r'^areas/', include('ietf.secr.areas.urls')), - url(r'^console/', include('ietf.secr.console.urls')), - url(r'^groups/', include('ietf.secr.groups.urls')), - url(r'^meetings/', include('ietf.secr.meetings.urls')), - url(r'^proceedings/', include('ietf.secr.proceedings.urls')), - url(r'^roles/', include('ietf.secr.roles.urls')), - url(r'^rolodex/', include('ietf.secr.rolodex.urls')), - url(r'^sreq/', include('ietf.secr.sreq.urls')), - url(r'^telechat/', include('ietf.secr.telechat.urls')), + re_path(r'^$', TemplateView.as_view(template_name='index.html'), name='ietf.secr'), + re_path(r'^announcement/', include('ietf.secr.announcement.urls')), + re_path(r'^meetings/', include('ietf.secr.meetings.urls')), + re_path(r'^rolodex/', include('ietf.secr.rolodex.urls')), + # remove these redirects after 125 + re_path(r'^sreq/$', RedirectView.as_view(url='/meeting/session/request/', permanent=True)), + re_path(r'^sreq/%(acronym)s/$' % settings.URL_REGEXPS, RedirectView.as_view(url='/meeting/session/request/%(acronym)s/view/', permanent=True)), + re_path(r'^sreq/%(acronym)s/edit/$' % settings.URL_REGEXPS, RedirectView.as_view(url='/meeting/session/request/%(acronym)s/edit/', permanent=True)), + re_path(r'^sreq/%(acronym)s/new/$' % settings.URL_REGEXPS, RedirectView.as_view(url='/meeting/session/request/%(acronym)s/new/', permanent=True)), + re_path(r'^sreq/(?P[A-Za-z0-9_\-\+]+)/%(acronym)s/view/$' % settings.URL_REGEXPS, RedirectView.as_view(url='/meeting/%(num)s/session/request/%(acronym)s/view/', permanent=True)), + re_path(r'^sreq/(?P[A-Za-z0-9_\-\+]+)/%(acronym)s/edit/$' % settings.URL_REGEXPS, RedirectView.as_view(url='/meeting/%(num)s/session/request/%(acronym)s/edit/', permanent=True)), + # --------------------------------- + re_path(r'^telechat/', include('ietf.secr.telechat.urls')), ] diff --git a/ietf/secr/utils/decorators.py b/ietf/secr/utils/decorators.py index f635bc7ece..5887c3c9cc 100644 --- a/ietf/secr/utils/decorators.py +++ b/ietf/secr/utils/decorators.py @@ -1,12 +1,12 @@ # Copyright The IETF Trust 2013-2020, All Rights Reserved from functools import wraps +from urllib.parse import quote as urlquote from django.conf import settings from django.contrib.auth import REDIRECT_FIELD_NAME from django.core.exceptions import ObjectDoesNotExist from django.http import HttpResponseRedirect from django.shortcuts import render, get_object_or_404 -from django.utils.http import urlquote from ietf.ietfauth.utils import has_role from ietf.doc.models import Document diff --git a/ietf/secr/utils/document.py b/ietf/secr/utils/document.py index 0a34512a17..361bf836df 100644 --- a/ietf/secr/utils/document.py +++ b/ietf/secr/utils/document.py @@ -13,15 +13,6 @@ def get_full_path(doc): return None return os.path.join(doc.get_file_path(), doc.uploaded_filename) -def get_rfc_num(doc): - qs = doc.docalias.filter(name__startswith='rfc') - return qs[0].name[3:] if qs else None - -def is_draft(doc): - if doc.docalias.filter(name__startswith='rfc'): - return False - else: - return True def get_start_date(doc): ''' diff --git a/ietf/secr/utils/group.py b/ietf/secr/utils/group.py deleted file mode 100644 index a4c1c0f98a..0000000000 --- a/ietf/secr/utils/group.py +++ /dev/null @@ -1,74 +0,0 @@ -# Copyright The IETF Trust 2013-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -# Python imports -import io -import os - -# Django imports -from django.conf import settings -from django.core.exceptions import ObjectDoesNotExist - -# Datatracker imports -from ietf.group.models import Group -from ietf.ietfauth.utils import has_role - - - - -def current_nomcom(): - qs = Group.objects.filter(acronym__startswith='nomcom',state__slug="active").order_by('-time') - if qs.count(): - return qs[0] - else: - return None - -def get_charter_text(group): - ''' - Takes a group object and returns the text or the group's charter as a string - ''' - charter = group.charter - path = os.path.join(settings.CHARTER_PATH, '%s-%s.txt' % (charter.canonical_name(), charter.rev)) - f = io.open(path,'r') - text = f.read() - f.close() - - return text - -def get_my_groups(user,conclude=False): - ''' - Takes a Django user object (from request) - Returns a list of groups the user has access to. Rules are as follows - secretariat - has access to all groups - area director - has access to all groups in their area - wg chair or secretary - has access to their own group - chair of irtf has access to all irtf groups - - If user=None than all groups are returned. - concluded=True means include concluded groups. Need this to upload materials for groups - after they've been concluded. it happens. - ''' - my_groups = set() - states = ['bof','proposed','active'] - if conclude: - states.extend(['conclude','bof-conc']) - - all_groups = Group.objects.filter(type__features__has_meetings=True, state__in=states).order_by('acronym') - if user == None or has_role(user,'Secretariat'): - return all_groups - - try: - person = user.person - except ObjectDoesNotExist: - return list() - - for group in all_groups: - if group.role_set.filter(person=person,name__in=('chair','secr','ad')): - my_groups.add(group) - continue - if group.parent and group.parent.role_set.filter(person=person,name__in=('ad','chair')): - my_groups.add(group) - continue - - return list(my_groups) diff --git a/ietf/settings.py b/ietf/settings.py index 144f321cc7..3aa45a453c 100644 --- a/ietf/settings.py +++ b/ietf/settings.py @@ -1,4 +1,4 @@ -# Copyright The IETF Trust 2007-2022, All Rights Reserved +# Copyright The IETF Trust 2007-2026, All Rights Reserved # -*- coding: utf-8 -*- @@ -9,24 +9,42 @@ import os import sys import datetime +import pathlib import warnings +from hashlib import sha384 from typing import Any, Dict, List, Tuple # pyflakes:ignore +from django.http import UnreadablePostError +# DeprecationWarnings are suppressed by default, enable them warnings.simplefilter("always", DeprecationWarning) -warnings.filterwarnings("ignore", message="'urllib3\[secure\]' extra is deprecated") -warnings.filterwarnings("ignore", message="The logout\(\) view is superseded by") + +# Warnings that must be resolved for Django 5.x +warnings.filterwarnings("ignore", "Log out via GET requests is deprecated") # caused by oidc_provider +warnings.filterwarnings("ignore", message="The django.utils.timezone.utc alias is deprecated.", module="oidc_provider") +warnings.filterwarnings("ignore", message="The django.utils.datetime_safe module is deprecated.", module="tastypie") +warnings.filterwarnings("ignore", message="The USE_DEPRECATED_PYTZ setting,") # https://github.com/ietf-tools/datatracker/issues/5635 +warnings.filterwarnings("ignore", message="The is_dst argument to make_aware\\(\\)") # caused by django-filters when USE_DEPRECATED_PYTZ is true +warnings.filterwarnings("ignore", message="The USE_L10N setting is deprecated.") # https://github.com/ietf-tools/datatracker/issues/5648 +warnings.filterwarnings("ignore", message="django.contrib.auth.hashers.CryptPasswordHasher is deprecated.") # https://github.com/ietf-tools/datatracker/issues/5663 + +# Other DeprecationWarnings +warnings.filterwarnings("ignore", message="pkg_resources is deprecated as an API", module="pyang.plugin") warnings.filterwarnings("ignore", message="Report.file_reporters will no longer be available in Coverage.py 4.2", module="coverage.report") -warnings.filterwarnings("ignore", message="{% load staticfiles %} is deprecated") -warnings.filterwarnings("ignore", message="Using or importing the ABCs from 'collections' instead of from 'collections.abc' is deprecated", module="bleach") -warnings.filterwarnings("ignore", message="HTTPResponse.getheader\(\) is deprecated", module='selenium.webdriver') -try: - import syslog - syslog.openlog(str("datatracker"), syslog.LOG_PID, syslog.LOG_USER) -except ImportError: - pass +warnings.filterwarnings("ignore", message="currentThread\\(\\) is deprecated", module="coverage.pytracer") +warnings.filterwarnings("ignore", message="co_lnotab is deprecated", module="coverage.parser") +warnings.filterwarnings("ignore", message="datetime.datetime.utcnow\\(\\) is deprecated", module="botocore.auth") +warnings.filterwarnings("ignore", message="datetime.datetime.utcnow\\(\\) is deprecated", module="oic.utils.time_util") +warnings.filterwarnings("ignore", message="datetime.datetime.utcfromtimestamp\\(\\) is deprecated", module="oic.utils.time_util") +warnings.filterwarnings("ignore", message="datetime.datetime.utcfromtimestamp\\(\\) is deprecated", module="pytz.tzinfo") +warnings.filterwarnings("ignore", message="'instantiateVariableFont' is deprecated", module="weasyprint") -BASE_DIR = os.path.dirname(os.path.abspath(__file__)) -sys.path.append(os.path.abspath(BASE_DIR + "/..")) + +base_path = pathlib.Path(__file__).resolve().parent +BASE_DIR = str(base_path) + +project_path = base_path.parent +PROJECT_DIR = str(project_path) +sys.path.append(PROJECT_DIR) from ietf import __version__ import debug @@ -44,17 +62,12 @@ # Domain name of the IETF IETF_DOMAIN = 'ietf.org' +# Overriden in settings_local ADMINS = [ -# ('Henrik Levkowetz', 'henrik@levkowetz.com'), - ('Robert Sparks', 'rjsparks@nostrum.com'), -# ('Ole Laursen', 'olau@iola.dk'), - ('Ryan Cross', 'rcross@amsl.com'), - ('Glen Barney', 'glen@amsl.com'), - ('Maddy Conner', 'maddy@amsl.com'), - ('Kesara Rathnayaka', 'krathnayake@ietf.org'), + ('Tools Help', 'tools-help@ietf.org'), ] # type: List[Tuple[str, str]] -BUG_REPORT_EMAIL = "datatracker-project@ietf.org" +BUG_REPORT_EMAIL = "tools-help@ietf.org" PASSWORD_HASHERS = [ 'django.contrib.auth.hashers.Argon2PasswordHasher', @@ -64,6 +77,26 @@ 'django.contrib.auth.hashers.CryptPasswordHasher', ] + +PASSWORD_POLICY_MIN_LENGTH = 12 +PASSWORD_POLICY_ENFORCE_AT_LOGIN = False # should turn this on for prod + +AUTH_PASSWORD_VALIDATORS = [ + { + "NAME": "django.contrib.auth.password_validation.MinimumLengthValidator", + "OPTIONS": { + "min_length": PASSWORD_POLICY_MIN_LENGTH, + } + }, + { + "NAME": "ietf.ietfauth.password_validation.StrongPasswordValidator", + }, +] +# In dev environments, settings_local overrides the password validators. Save +# a handle to the original value so settings_test can restore it so tests match +# production. +ORIG_AUTH_PASSWORD_VALIDATORS = AUTH_PASSWORD_VALIDATORS + ALLOWED_HOSTS = [".ietf.org", ".ietf.org.", "209.208.19.216", "4.31.198.44", "127.0.0.1", "localhost", ] # Server name of the tools server @@ -80,21 +113,13 @@ DATABASES = { 'default': { - 'NAME': 'ietf_utf8', - 'ENGINE': 'django.db.backends.mysql', + 'NAME': 'datatracker', + 'ENGINE': 'django.db.backends.postgresql_psycopg2', 'USER': 'ietf', - #'PASSWORD': 'ietf', - 'OPTIONS': { - 'sql_mode': 'STRICT_TRANS_TABLES', - 'init_command': 'SET storage_engine=MyISAM; SET names "utf8"' - }, + #'PASSWORD': 'somepassword', }, } -DATABASE_TEST_OPTIONS = { - # Comment this out if your database doesn't support InnoDB - 'init_command': 'SET storage_engine=InnoDB', -} # Local time zone for this installation. Choices can be found here: # http://www.postgresql.org/docs/8.1/static/datetime-keywords.html#DATETIME-TIMEZONE-SET-TABLE @@ -114,7 +139,27 @@ # to load the internationalization machinery. USE_I18N = False +# Django 4.0 changed the default setting of USE_L10N to True. The setting +# is deprecated and will be removed in Django 5.0. +USE_L10N = False + USE_TZ = True +USE_DEPRECATED_PYTZ = True # supported until Django 5 + +# The DjangoDivFormRenderer is a transitional class that opts in to defaulting to the div.html +# template for formsets. This will become the default behavior in Django 5.0. This configuration +# can be removed at that point. +# See https://docs.djangoproject.com/en/4.2/releases/4.1/#forms +FORM_RENDERER = "django.forms.renderers.DjangoDivFormRenderer" + +# Default primary key field type to use for models that don’t have a field with primary_key=True. +# In the future (relative to 4.2), the default will become 'django.db.models.BigAutoField.' +DEFAULT_AUTO_FIELD = 'django.db.models.AutoField' + +# OIDC configuration +_SITE_URL = os.environ.get("OIDC_SITE_URL", None) +if _SITE_URL is not None: + SITE_URL = _SITE_URL if SERVER_MODE == 'production': MEDIA_ROOT = '/a/www/www6s/lib/dt/media/' @@ -150,7 +195,6 @@ # Absolute path to the directory static files should be collected to. # Example: "/var/www/example.com/static/" - SERVE_CDN_PHOTOS = True SERVE_CDN_FILES_LOCALLY_IN_DEV_MODE = True @@ -160,7 +204,7 @@ STATIC_URL = "/static/" STATIC_ROOT = os.path.abspath(BASE_DIR + "/../static/") else: - STATIC_URL = "https://www.ietf.org/lib/dt/%s/"%__version__ + STATIC_URL = "https://static.ietf.org/dt/%s/"%__version__ STATIC_ROOT = "/a/www/www6s/lib/dt/%s/"%__version__ # List of finder classes that know how to find static files in @@ -170,165 +214,145 @@ 'django.contrib.staticfiles.finders.AppDirectoriesFinder', ) +# Client-side static.ietf.org URL +STATIC_IETF_ORG = "https://static.ietf.org" +# Server-side static.ietf.org URL (used in pdfized) +STATIC_IETF_ORG_INTERNAL = STATIC_IETF_ORG + +ENABLE_BLOBSTORAGE = True + +# "standard" retry mode is used, which does exponential backoff with a base factor of 2 +# and a cap of 20. +BLOBSTORAGE_MAX_ATTEMPTS = 5 # boto3 default is 3 (for "standard" retry mode) +BLOBSTORAGE_CONNECT_TIMEOUT = 10 # seconds; boto3 default is 60 +BLOBSTORAGE_READ_TIMEOUT = 10 # seconds; boto3 default is 60 + +# Caching for agenda data in seconds +AGENDA_CACHE_TIMEOUT_DEFAULT = 8 * 24 * 60 * 60 # 8 days +AGENDA_CACHE_TIMEOUT_CURRENT_MEETING = 6 * 60 # 6 minutes + WSGI_APPLICATION = "ietf.wsgi.application" -AUTHENTICATION_BACKENDS = ( 'django.contrib.auth.backends.ModelBackend', ) +AUTHENTICATION_BACKENDS = ( 'ietf.ietfauth.backends.CaseInsensitiveModelBackend', ) -FILE_UPLOAD_PERMISSIONS = 0o644 +FILE_UPLOAD_PERMISSIONS = 0o644 -# ------------------------------------------------------------------------ -# Django/Python Logging Framework Modifications +FIRST_V3_RFC = 8650 -# Filter out "Invalid HTTP_HOST" emails -# Based on http://www.tiwoc.de/blog/2013/03/django-prevent-email-notification-on-suspiciousoperation/ -from django.core.exceptions import SuspiciousOperation -def skip_suspicious_operations(record): - if record.exc_info: - exc_value = record.exc_info[1] - if isinstance(exc_value, SuspiciousOperation): - return False - return True -# Filter out UreadablePostError: -from django.http import UnreadablePostError +# +# Logging config +# + +# Callback to filter out UnreadablePostError: def skip_unreadable_post(record): if record.exc_info: - exc_type, exc_value = record.exc_info[:2] # pylint: disable=unused-variable + exc_type, exc_value = record.exc_info[:2] # pylint: disable=unused-variable if isinstance(exc_value, UnreadablePostError): return False return True -# Copied from DEFAULT_LOGGING as of Django 1.10.5 on 22 Feb 2017, and modified -# to incorporate html logging, invalid http_host filtering, and more. -# Changes from the default has comments. - -# The Python logging flow is as follows: -# (see https://docs.python.org/2.7/howto/logging.html#logging-flow) -# -# Init: get a Logger: logger = logging.getLogger(name) -# -# Logging call, e.g. logger.error(level, msg, *args, exc_info=(...), extra={...}) -# --> Logger (discard if level too low for this logger) -# (create log record from level, msg, args, exc_info, extra) -# --> Filters (discard if any filter attach to logger rejects record) -# --> Handlers (discard if level too low for handler) -# --> Filters (discard if any filter attached to handler rejects record) -# --> Formatter (format log record and emit) -# - LOGGING = { - 'version': 1, - 'disable_existing_loggers': False, - # - 'loggers': { - 'django': { - 'handlers': ['debug_console', 'mail_admins'], - 'level': 'INFO', + "version": 1, + "disable_existing_loggers": False, + "loggers": { + "celery": { + "handlers": ["console"], + "level": "INFO", }, - 'django.request': { - 'handlers': ['debug_console'], - 'level': 'ERROR', + "datatracker": { + "handlers": ["console"], + "level": "INFO", }, - 'django.server': { - 'handlers': ['django.server'], - 'level': 'INFO', + "django": { + "handlers": ["console", "mail_admins"], + "level": "INFO", }, - 'django.security': { - 'handlers': ['debug_console', ], - 'level': 'INFO', + "django.request": {"level": "ERROR"}, # only log 5xx, ignore 4xx + "django.security": { + # SuspiciousOperation errors - log to console only + "handlers": ["console"], + "propagate": False, # no further handling please }, - 'oidc_provider': { - 'handlers': ['debug_console', ], - 'level': 'DEBUG', - }, - }, - # - # No logger filters - # - 'handlers': { - 'console': { - 'level': 'DEBUG', - 'class': 'logging.StreamHandler', - 'formatter': 'plain', + "django.server": { + # Only used by Django's runserver development server + "handlers": ["django.server"], + "level": "INFO", }, - 'syslog': { - 'level': 'DEBUG', - 'class': 'logging.handlers.SysLogHandler', - 'facility': 'user', - 'formatter': 'plain', - 'address': '/dev/log', + "oidc_provider": { + "handlers": ["console"], + "level": "DEBUG", }, - 'debug_console': { - # Active only when DEBUG=True - 'level': 'DEBUG', - 'filters': ['require_debug_true'], - 'class': 'logging.StreamHandler', - 'formatter': 'plain', + }, + "handlers": { + "console": { + "level": "DEBUG", + "class": "logging.StreamHandler", + "formatter": "plain", }, - 'django.server': { - 'level': 'INFO', - 'class': 'logging.StreamHandler', - 'formatter': 'django.server', + "debug_console": { + "level": "DEBUG", + "filters": ["require_debug_true"], + "class": "logging.StreamHandler", + "formatter": "plain", }, - 'mail_admins': { - 'level': 'ERROR', - 'filters': [ - 'require_debug_false', - 'skip_suspicious_operations', # custom - 'skip_unreadable_posts', # custom + "django.server": { + "level": "INFO", + "class": "logging.StreamHandler", + "formatter": "django.server", + }, + "mail_admins": { + "level": "ERROR", + "filters": [ + "require_debug_false", + "skip_unreadable_posts", ], - 'class': 'django.utils.log.AdminEmailHandler', - 'include_html': True, # non-default - } + "class": "django.utils.log.AdminEmailHandler", + "include_html": True, + }, }, - # # All these are used by handlers - 'filters': { - 'require_debug_false': { - '()': 'django.utils.log.RequireDebugFalse', - }, - 'require_debug_true': { - '()': 'django.utils.log.RequireDebugTrue', + "filters": { + "require_debug_false": { + "()": "django.utils.log.RequireDebugFalse", }, - # custom filter, function defined above: - 'skip_suspicious_operations': { - '()': 'django.utils.log.CallbackFilter', - 'callback': skip_suspicious_operations, + "require_debug_true": { + "()": "django.utils.log.RequireDebugTrue", }, # custom filter, function defined above: - 'skip_unreadable_posts': { - '()': 'django.utils.log.CallbackFilter', - 'callback': skip_unreadable_post, + "skip_unreadable_posts": { + "()": "django.utils.log.CallbackFilter", + "callback": skip_unreadable_post, }, }, - # And finally the formatters - 'formatters': { - 'django.server': { - '()': 'django.utils.log.ServerFormatter', - 'format': '[%(server_time)s] %(message)s', + "formatters": { + "django.server": { + "()": "django.utils.log.ServerFormatter", + "format": "[%(server_time)s] %(message)s", + }, + "plain": { + "style": "{", + "format": "{levelname}: {name}:{lineno}: {message}", }, - 'plain': { - 'style': '{', - 'format': '{levelname}: {name}:{lineno}: {message}', + "json": { + "class": "ietf.utils.jsonlogger.DatatrackerJsonFormatter", + "style": "{", + "format": ( + "{asctime}{levelname}{message}{name}{pathname}{lineno}{funcName}" + "{process}{status_code}" + ), }, }, } -# This should be overridden by settings_local for any logger where debug (or -# other) custom log settings are wanted. Use "ietf/manage.py showloggers -l" -# to show registered loggers. The content here should match the levels above -# and is shown as an example: -UTILS_LOGGER_LEVELS: Dict[str, str] = { -# 'django': 'INFO', -# 'django.server': 'INFO', -} - -# End logging -# ------------------------------------------------------------------------ - X_FRAME_OPTIONS = 'SAMEORIGIN' -CSRF_TRUSTED_ORIGINS = ['ietf.org', '*.ietf.org', 'meetecho.com', '*.meetecho.com', 'gather.town', '*.gather.town', ] +CSRF_TRUSTED_ORIGINS = [ + "https://ietf.org", + "https://*.ietf.org", + 'https://meetecho.com', + 'https://*.meetecho.com', +] CSRF_COOKIE_SAMESITE = 'None' CSRF_COOKIE_SECURE = True @@ -338,11 +362,7 @@ def skip_unreadable_post(record): SESSION_COOKIE_SECURE = True SESSION_EXPIRE_AT_BROWSER_CLOSE = False -# We want to use the JSON serialisation, as it's safer -- but there is /secr/ -# code which stashes objects in the session that can't be JSON serialized. -# Switch when that code is rewritten. -#SESSION_SERIALIZER = "django.contrib.sessions.serializers.JSONSerializer" -SESSION_SERIALIZER = 'django.contrib.sessions.serializers.PickleSerializer' +SESSION_SERIALIZER = "django.contrib.sessions.serializers.JSONSerializer" SESSION_ENGINE = "django.contrib.sessions.backends.cache" SESSION_SAVE_EVERY_REQUEST = True SESSION_CACHE_ALIAS = 'sessions' @@ -358,6 +378,7 @@ def skip_unreadable_post(record): ], 'OPTIONS': { 'context_processors': [ + 'ietf.context_processors.traceparent_id', 'django.contrib.auth.context_processors.auth', 'django.template.context_processors.debug', # makes 'sql_queries' available in templates 'django.template.context_processors.i18n', @@ -372,6 +393,7 @@ def skip_unreadable_post(record): 'ietf.context_processors.settings_info', 'ietf.secr.context_processors.secr_revision_info', 'ietf.context_processors.rfcdiff_base_url', + 'ietf.context_processors.timezone_now', ], 'loaders': [ ('django.template.loaders.cached.Loader', ( @@ -389,42 +411,45 @@ def skip_unreadable_post(record): MIDDLEWARE = [ - 'django.middleware.csrf.CsrfViewMiddleware', - 'corsheaders.middleware.CorsMiddleware', # see docs on CORS_REPLACE_HTTPS_REFERER before using it - 'django.middleware.common.CommonMiddleware', - 'django.contrib.sessions.middleware.SessionMiddleware', - 'django.contrib.auth.middleware.AuthenticationMiddleware', - 'django.contrib.messages.middleware.MessageMiddleware', - 'django.middleware.http.ConditionalGetMiddleware', - 'simple_history.middleware.HistoryRequestMiddleware', + "ietf.middleware.add_otel_traceparent_header", + "django.middleware.csrf.CsrfViewMiddleware", + "corsheaders.middleware.CorsMiddleware", # see docs on CORS_REPLACE_HTTPS_REFERER before using it + "django.middleware.common.CommonMiddleware", + "django.contrib.sessions.middleware.SessionMiddleware", + "django.contrib.auth.middleware.AuthenticationMiddleware", + "django.contrib.messages.middleware.MessageMiddleware", + "ietf.middleware.is_authenticated_header_middleware", + "django.middleware.http.ConditionalGetMiddleware", + "simple_history.middleware.HistoryRequestMiddleware", # comment in this to get logging of SQL insert and update statements: - #'ietf.middleware.sql_log_middleware', - 'ietf.middleware.SMTPExceptionMiddleware', - 'ietf.middleware.Utf8ExceptionMiddleware', - 'ietf.middleware.redirect_trailing_period_middleware', - 'django_referrer_policy.middleware.ReferrerPolicyMiddleware', - 'django.middleware.clickjacking.XFrameOptionsMiddleware', - 'django.middleware.security.SecurityMiddleware', - # 'csp.middleware.CSPMiddleware', - 'ietf.middleware.unicode_nfkc_normalization_middleware', + #"ietf.middleware.sql_log_middleware", + "ietf.middleware.SMTPExceptionMiddleware", + "ietf.middleware.Utf8ExceptionMiddleware", + "ietf.middleware.redirect_trailing_period_middleware", + "django.middleware.clickjacking.XFrameOptionsMiddleware", + "django.middleware.security.SecurityMiddleware", + "ietf.middleware.unicode_nfkc_normalization_middleware", ] ROOT_URLCONF = 'ietf.urls' -DJANGO_VITE_ASSETS_PATH = os.path.join(BASE_DIR, 'static/dist-neue') +# Configure django_vite +DJANGO_VITE: dict = {"default": {}} if DEBUG: - DJANGO_VITE_MANIFEST_PATH = os.path.join(BASE_DIR, 'static/dist-neue/manifest.json') + DJANGO_VITE["default"]["manifest_path"] = os.path.join( + BASE_DIR, 'static/dist-neue/manifest.json' + ) # Additional locations of static files (in addition to each app's static/ dir) STATICFILES_DIRS = ( - DJANGO_VITE_ASSETS_PATH, + os.path.join(BASE_DIR, "static/dist-neue"), # for django_vite os.path.join(BASE_DIR, 'static/dist'), os.path.join(BASE_DIR, 'secr/static/dist'), ) INSTALLED_APPS = [ # Django apps - 'django.contrib.admin', + 'ietf.admin', # replaces django.contrib.admin 'django.contrib.admindocs', 'django.contrib.auth', 'django.contrib.contenttypes', @@ -439,16 +464,21 @@ def skip_unreadable_post(record): 'django_vite', 'django_bootstrap5', 'django_celery_beat', + 'django_celery_results', 'corsheaders', 'django_markup', - 'django_password_strength', - 'form_utils', + 'django_filters', 'oidc_provider', + 'drf_spectacular', + 'drf_standardized_errors', + 'rest_framework', + 'rangefilter', 'simple_history', 'tastypie', 'widget_tweaks', # IETF apps 'ietf.api', + 'ietf.blobdb', 'ietf.community', 'ietf.dbtemplate', 'ietf.doc', @@ -469,18 +499,14 @@ def skip_unreadable_post(record): 'ietf.release', 'ietf.review', 'ietf.stats', + 'ietf.status', 'ietf.submit', 'ietf.sync', 'ietf.utils', # IETF Secretariat apps 'ietf.secr.announcement', - 'ietf.secr.areas', - 'ietf.secr.groups', 'ietf.secr.meetings', - 'ietf.secr.proceedings', - 'ietf.secr.roles', 'ietf.secr.rolodex', - 'ietf.secr.sreq', 'ietf.secr.telechat', ] @@ -520,8 +546,6 @@ def skip_unreadable_post(record): CORS_ALLOW_METHODS = ( 'GET', 'OPTIONS', ) CORS_URLS_REGEX = r'^(/api/.*|.*\.json|.*/json/?)$' -# Setting for django_referrer_policy.middleware.ReferrerPolicyMiddleware -REFERRER_POLICY = 'strict-origin-when-cross-origin' # django.middleware.security.SecurityMiddleware SECURE_BROWSER_XSS_FILTER = True @@ -532,6 +556,9 @@ def skip_unreadable_post(record): #SECURE_REDIRECT_EXEMPT #SECURE_SSL_HOST #SECURE_SSL_REDIRECT = True +# Relax the COOP policy to allow Meetecho authentication pop-up +SECURE_CROSS_ORIGIN_OPENER_POLICY = "unsafe-none" +SECURE_REFERRER_POLICY = "strict-origin-when-cross-origin" # Override this in your settings_local with the IP addresses relevant for you: INTERNAL_IPS = ( @@ -540,15 +567,83 @@ def skip_unreadable_post(record): '::1', ) +# django-rest-framework configuration +REST_FRAMEWORK = { + "DEFAULT_AUTHENTICATION_CLASSES": [ + "ietf.api.authentication.ApiKeyAuthentication", + "rest_framework.authentication.SessionAuthentication", + ], + "DEFAULT_PERMISSION_CLASSES": [ + "ietf.api.permissions.HasApiKey", + ], + "DEFAULT_RENDERER_CLASSES": [ + "rest_framework.renderers.JSONRenderer", + ], + "DEFAULT_PARSER_CLASSES": [ + "rest_framework.parsers.JSONParser", + ], + "DEFAULT_SCHEMA_CLASS": "drf_standardized_errors.openapi.AutoSchema", + "EXCEPTION_HANDLER": "drf_standardized_errors.handler.exception_handler", +} + +# DRF OpenApi schema settings +SPECTACULAR_SETTINGS = { + "TITLE": "Datatracker API", + "DESCRIPTION": "Datatracker API", + "VERSION": "1.0.0", + "SCHEMA_PATH_PREFIX": "/api/", + "COMPONENT_SPLIT_REQUEST": True, + "COMPONENT_NO_READ_ONLY_REQUIRED": True, + "SERVERS": [ + {"url": "http://localhost:8000", "description": "local dev server"}, + {"url": "https://datatracker.ietf.org", "description": "production server"}, + ], + # The following settings are needed for drf-standardized-errors + "ENUM_NAME_OVERRIDES": { + "ValidationErrorEnum": "drf_standardized_errors.openapi_serializers.ValidationErrorEnum.choices", + "ClientErrorEnum": "drf_standardized_errors.openapi_serializers.ClientErrorEnum.choices", + "ServerErrorEnum": "drf_standardized_errors.openapi_serializers.ServerErrorEnum.choices", + "ErrorCode401Enum": "drf_standardized_errors.openapi_serializers.ErrorCode401Enum.choices", + "ErrorCode403Enum": "drf_standardized_errors.openapi_serializers.ErrorCode403Enum.choices", + "ErrorCode404Enum": "drf_standardized_errors.openapi_serializers.ErrorCode404Enum.choices", + "ErrorCode405Enum": "drf_standardized_errors.openapi_serializers.ErrorCode405Enum.choices", + "ErrorCode406Enum": "drf_standardized_errors.openapi_serializers.ErrorCode406Enum.choices", + "ErrorCode415Enum": "drf_standardized_errors.openapi_serializers.ErrorCode415Enum.choices", + "ErrorCode429Enum": "drf_standardized_errors.openapi_serializers.ErrorCode429Enum.choices", + "ErrorCode500Enum": "drf_standardized_errors.openapi_serializers.ErrorCode500Enum.choices", + }, + "POSTPROCESSING_HOOKS": ["drf_standardized_errors.openapi_hooks.postprocess_schema_enums"], +} + +# DRF Standardized Errors settings +DRF_STANDARDIZED_ERRORS = { + # enable the standardized errors when DEBUG=True for unhandled exceptions. + # By default, this is set to False so you're able to view the traceback in + # the terminal and get more information about the exception. + "ENABLE_IN_DEBUG_FOR_UNHANDLED_EXCEPTIONS": False, + # ONLY the responses that correspond to these status codes will appear + # in the API schema. + "ALLOWED_ERROR_STATUS_CODES": [ + "400", + # "401", + # "403", + "404", + # "405", + # "406", + # "415", + # "429", + # "500", + ], + +} + # no slash at end IDTRACKER_BASE_URL = "https://datatracker.ietf.org" RFCDIFF_BASE_URL = "https://author-tools.ietf.org/iddiff" IDNITS_BASE_URL = "https://author-tools.ietf.org/api/idnits" +IDNITS3_BASE_URL = "https://author-tools.ietf.org/idnits3/results" IDNITS_SERVICE_URL = "https://author-tools.ietf.org/idnits" -# Content security policy configuration (django-csp) -CSP_DEFAULT_SRC = ("'self'", "'unsafe-inline'", f"data: {IDTRACKER_BASE_URL} https://www.ietf.org/ https://analytics.ietf.org/ https://fonts.googleapis.com/") - # The name of the method to use to invoke the test suite TEST_RUNNER = 'ietf.utils.test_runner.IetfTestRunner' @@ -558,8 +653,6 @@ def skip_unreadable_post(record): TEST_DIFF_FAILURE_DIR = "/tmp/test/failure/" -TEST_GHOSTDRIVER_LOG_PATH = "ghostdriver.log" - # These are regexes TEST_URL_COVERAGE_EXCLUDE = [ r"^\^admin/", @@ -586,9 +679,10 @@ def skip_unreadable_post(record): "ietf/utils/test_runner.py", "ietf/name/generate_fixtures.py", "ietf/review/import_from_review_tool.py", - "ietf/stats/backfill_data.py", "ietf/utils/patch.py", "ietf/utils/test_data.py", + "ietf/utils/jstest.py", + "ietf/utils/coverage.py", ] # These are code line regex patterns @@ -602,12 +696,15 @@ def skip_unreadable_post(record): ] # These are filename globs. They are used by test_parse_templates() and -# get_template_paths() +# get_template_paths(). Globs are applied via pathlib.Path().match, using +# the path to the template from the project root. TEST_TEMPLATE_IGNORE = [ - ".*", # dot-files - "*~", # tilde temp-files - "#*", # files beginning with a hashmark - "500.html" # isn't loaded by regular loader, but checked by test_500_page() + ".*", # dot-files + "*~", # tilde temp-files + "#*", # files beginning with a hashmark + "500.html", # isn't loaded by regular loader, but checked by test_500_page() + "ietf/templates/admin/meeting/RegistrationTicket/change_list.html", + "ietf/templates/admin/meeting/Registration/change_list.html", ] TEST_COVERAGE_MAIN_FILE = os.path.join(BASE_DIR, "../release-coverage.json") @@ -615,8 +712,8 @@ def skip_unreadable_post(record): TEST_CODE_COVERAGE_CHECKER = None if SERVER_MODE != 'production': - import coverage - TEST_CODE_COVERAGE_CHECKER = coverage.Coverage(source=[ BASE_DIR ], cover_pylib=False, omit=TEST_CODE_COVERAGE_EXCLUDE_FILES) + from ietf.utils.coverage import CoverageManager + TEST_CODE_COVERAGE_CHECKER = CoverageManager() TEST_CODE_COVERAGE_REPORT_PATH = "coverage/" TEST_CODE_COVERAGE_REPORT_URL = os.path.join(STATIC_URL, TEST_CODE_COVERAGE_REPORT_PATH, "index.html") @@ -643,6 +740,7 @@ def skip_unreadable_post(record): "acronym": r"(?P[-a-z0-9]+)", "bofreq": r"(?Pbofreq-[-a-z0-9]+)", "charter": r"(?Pcharter-[-a-z0-9]+)", + "statement": r"(?Pstatement-[-a-z0-9]+)", "date": r"(?P\d{4}-\d{2}-\d{2})", "name": r"(?P[A-Za-z0-9._+-]+?)", "document": r"(?P[a-z][-a-z0-9]+)", # regular document names @@ -651,40 +749,98 @@ def skip_unreadable_post(record): "schedule_name": r"(?P[A-Za-z0-9-:_]+)", } +STORAGES: dict[str, Any] = { + "default": {"BACKEND": "django.core.files.storage.FileSystemStorage"}, + "staticfiles": {"BACKEND": "django.contrib.staticfiles.storage.StaticFilesStorage"}, +} + +# Storages for artifacts stored as blobs +ARTIFACT_STORAGE_NAMES: list[str] = [ + "active-draft", + "agenda", + "bibxml-ids", + "bluesheets", + "bofreq", + "charter", + "chatlog", + "conflrev", + "draft", + "floorplan", + "indexes", + "liai-att", + "meetinghostlogo", + "minutes", + "narrativeminutes", + "photo", + "polls", + "procmaterials", + "review", + "rfc", + "slides", + "staging", + "statchg", + "statement", +] +for storagename in ARTIFACT_STORAGE_NAMES: + STORAGES[storagename] = { + "BACKEND": "ietf.doc.storage.StoredObjectBlobdbStorage", + "OPTIONS": {"bucket_name": storagename}, + } + +# Buckets / doc types of meeting materials the CF worker is allowed to serve. This +# differs from the list in Session.meeting_related() by the omission of "recording" +MATERIALS_TYPES_SERVED_BY_WORKER = [ + "agenda", + "bluesheets", + "chatlog", + "minutes", + "narrativeminutes", + "polls", + "procmaterials", + "slides", +] + +# Other storages +STORAGES["red_bucket"] = { + "BACKEND": "django.core.files.storage.InMemoryStorage", + "OPTIONS": {"location": "red_bucket"}, +} + # Override this in settings_local.py if needed # *_PATH variables ends with a slash/ . -#DOCUMENT_PATH_PATTERN = '/a/www/ietf-ftp/{doc.type_id}/' DOCUMENT_PATH_PATTERN = '/a/ietfdata/doc/{doc.type_id}/' INTERNET_DRAFT_PATH = '/a/ietfdata/doc/draft/repository' INTERNET_DRAFT_PDF_PATH = '/a/www/ietf-datatracker/pdf/' RFC_PATH = '/a/www/ietf-ftp/rfc/' CHARTER_PATH = '/a/ietfdata/doc/charter/' +CHARTER_COPY_PATH = '/a/www/ietf-ftp/ietf' # copy 1wg-charters files here if set +CHARTER_COPY_OTHER_PATH = '/a/ftp/ietf' +CHARTER_COPY_THIRD_PATH = '/a/ftp/charter' +GROUP_SUMMARY_PATH = '/a/www/ietf-ftp/ietf' BOFREQ_PATH = '/a/ietfdata/doc/bofreq/' CONFLICT_REVIEW_PATH = '/a/ietfdata/doc/conflict-review' STATUS_CHANGE_PATH = '/a/ietfdata/doc/status-change' AGENDA_PATH = '/a/www/www6s/proceedings/' MEETINGHOST_LOGO_PATH = AGENDA_PATH # put these in the same place as other proceedings files -IPR_DOCUMENT_PATH = '/a/www/ietf-ftp/ietf/IPR/' -IESG_TASK_FILE = '/a/www/www6/iesg/internal/task.txt' -IESG_ROLL_CALL_FILE = '/a/www/www6/iesg/internal/rollcall.txt' -IESG_ROLL_CALL_URL = 'https://www6.ietf.org/iesg/internal/rollcall.txt' -IESG_MINUTES_FILE = '/a/www/www6/iesg/internal/minutes.txt' -IESG_MINUTES_URL = 'https://www6.ietf.org/iesg/internal/minutes.txt' -IESG_WG_EVALUATION_DIR = "/a/www/www6/iesg/evaluation" # Move drafts to this directory when they expire INTERNET_DRAFT_ARCHIVE_DIR = '/a/ietfdata/doc/draft/collection/draft-archive/' -# The following directory contains linked copies of all drafts, but don't -# write anything to this directory -- its content is maintained by ghostlinkd: +# The following directory contains copies of all drafts - it used to be +# a set of hardlinks maintained by ghostlinkd, but is now explicitly written to INTERNET_ALL_DRAFTS_ARCHIVE_DIR = '/a/ietfdata/doc/draft/archive' MEETING_RECORDINGS_DIR = '/a/www/audio' DERIVED_DIR = '/a/ietfdata/derived' +FTP_DIR = '/a/ftp' +ALL_ID_DOWNLOAD_DIR = '/a/www/www6s/download' +NFS_METRICS_TMP_DIR = '/a/tmp' DOCUMENT_FORMAT_ALLOWLIST = ["txt", "ps", "pdf", "xml", "html", ] # Mailing list info URL for lists hosted on the IETF servers -MAILING_LIST_INFO_URL = "https://www.ietf.org/mailman/listinfo/%(list_addr)s" +MAILING_LIST_INFO_URL = "https://mailman3.%(domain)s/mailman3/lists/%(list_addr)s.%(domain)s" MAILING_LIST_ARCHIVE_URL = "https://mailarchive.ietf.org" +MAILING_LIST_ARCHIVE_SEARCH_URL = "https://mailarchive.ietf.org/api/v1/message/search/" +MAILING_LIST_ARCHIVE_API_KEY = "changeme" # Liaison Statement Tool settings (one is used in DOC_HREFS below) LIAISON_UNIVERSAL_FROM = 'Liaison Statement Management Tool ' @@ -696,7 +852,7 @@ def skip_unreadable_post(record): DOC_HREFS = { "charter": "https://www.ietf.org/charter/{doc.name}-{doc.rev}.txt", "draft": "https://www.ietf.org/archive/id/{doc.name}-{doc.rev}.txt", - "rfc": "https://www.rfc-editor.org/rfc/rfc{doc.rfcnum}.txt", + "rfc": "https://www.rfc-editor.org/rfc/rfc{doc.rfc_number}.txt", "slides": "https://www.ietf.org/slides/{doc.name}-{doc.rev}", "procmaterials": "https://www.ietf.org/procmaterials/{doc.name}-{doc.rev}", "conflrev": "https://www.ietf.org/cr/{doc.name}-{doc.rev}.txt", @@ -716,44 +872,6 @@ def skip_unreadable_post(record): CACHE_MIDDLEWARE_SECONDS = 300 CACHE_MIDDLEWARE_KEY_PREFIX = '' -# The default with no CACHES setting is 'django.core.cache.backends.locmem.LocMemCache' -# This setting is possibly overridden further down, after the import of settings_local -CACHES = { - 'default': { - 'BACKEND': 'django.core.cache.backends.memcached.MemcachedCache', - 'LOCATION': '127.0.0.1:11211', - 'VERSION': __version__, - 'KEY_PREFIX': 'ietf:dt', - }, - 'sessions': { - 'BACKEND': 'django.core.cache.backends.memcached.MemcachedCache', - 'LOCATION': '127.0.0.1:11211', - # No release-specific VERSION setting. - 'KEY_PREFIX': 'ietf:dt', - }, - 'htmlized': { - 'BACKEND': 'django.core.cache.backends.filebased.FileBasedCache', - 'LOCATION': '/a/cache/datatracker/htmlized', - 'OPTIONS': { - 'MAX_ENTRIES': 100000, # 100,000 - }, - }, - 'pdfized': { - 'BACKEND': 'django.core.cache.backends.filebased.FileBasedCache', - 'LOCATION': '/a/cache/datatracker/pdfized', - 'OPTIONS': { - 'MAX_ENTRIES': 100000, # 100,000 - }, - }, - 'slowpages': { - 'BACKEND': 'django.core.cache.backends.filebased.FileBasedCache', - 'LOCATION': '/a/cache/datatracker/slowpages', - 'OPTIONS': { - 'MAX_ENTRIES': 5000, - }, - }, -} - HTMLIZER_VERSION = 1 HTMLIZER_URL_PREFIX = "/doc/html" HTMLIZER_CACHE_TIME = 60*60*24*14 # 14 days @@ -766,16 +884,14 @@ def skip_unreadable_post(record): SESSION_REQUEST_FROM_EMAIL = 'IETF Meeting Session Request Tool ' SECRETARIAT_SUPPORT_EMAIL = "support@ietf.org" -SECRETARIAT_ACTION_EMAIL = "ietf-action@ietf.org" -SECRETARIAT_INFO_EMAIL = "ietf-info@ietf.org" +SECRETARIAT_ACTION_EMAIL = SECRETARIAT_SUPPORT_EMAIL +SECRETARIAT_INFO_EMAIL = SECRETARIAT_SUPPORT_EMAIL # Put real password in settings_local.py IANA_SYNC_PASSWORD = "secret" IANA_SYNC_CHANGES_URL = "https://datatracker.iana.org:4443/data-tracker/changes" IANA_SYNC_PROTOCOLS_URL = "https://www.iana.org/protocols/" -RFC_TEXT_RSYNC_SOURCE="ftp.rfc-editor.org::rfcs-text-only" - RFC_EDITOR_SYNC_PASSWORD="secret" RFC_EDITOR_SYNC_NOTIFICATION_URL = "https://www.rfc-editor.org/parser/parser.php" RFC_EDITOR_GROUP_NOTIFICATION_EMAIL = "webmaster@rfc-editor.org" @@ -783,17 +899,18 @@ def skip_unreadable_post(record): RFC_EDITOR_QUEUE_URL = "https://www.rfc-editor.org/queue2.xml" RFC_EDITOR_INDEX_URL = "https://www.rfc-editor.org/rfc/rfc-index.xml" RFC_EDITOR_ERRATA_JSON_URL = "https://www.rfc-editor.org/errata.json" -RFC_EDITOR_ERRATA_URL = "https://www.rfc-editor.org/errata_search.php?rfc={rfc_number}" RFC_EDITOR_INLINE_ERRATA_URL = "https://www.rfc-editor.org/rfc/inline-errata/rfc{rfc_number}.html" +RFC_EDITOR_ERRATA_BASE_URL = "https://www.rfc-editor.org/errata/" RFC_EDITOR_INFO_BASE_URL = "https://www.rfc-editor.org/info/" + # NomCom Tool settings ROLODEX_URL = "" NOMCOM_PUBLIC_KEYS_DIR = '/a/www/nomcom/public_keys/' NOMCOM_FROM_EMAIL = 'nomcom-chair-{year}@ietf.org' OPENSSL_COMMAND = '/usr/bin/openssl' DAYS_TO_EXPIRE_NOMINATION_LINK = '' -NOMINEE_FEEDBACK_TYPES = ['comment', 'questio', 'nomina'] +NOMINEE_FEEDBACK_TYPES = ['comment', 'questio', 'nomina', 'obe'] # SlideSubmission settings SLIDE_STAGING_PATH = '/a/www/www6s/staging/' @@ -848,7 +965,8 @@ def skip_unreadable_post(record): # Max time to allow for validation before a submission is subject to cancellation IDSUBMIT_MAX_VALIDATION_TIME = datetime.timedelta(minutes=20) -IDSUBMIT_MANUAL_STAGING_DIR = '/tmp/' +# Age at which a submission expires if not posted +IDSUBMIT_EXPIRATION_AGE = datetime.timedelta(days=14) IDSUBMIT_FILE_TYPES = ( 'txt', @@ -888,6 +1006,7 @@ def skip_unreadable_post(record): MEETING_DOC_LOCAL_HREFS = { "agenda": "/meeting/{meeting.number}/materials/{doc.name}-{doc.rev}", "minutes": "/meeting/{meeting.number}/materials/{doc.name}-{doc.rev}", + "narrativeminutes": "/meeting/{meeting.number}/materials/{doc.name}-{doc.rev}", "slides": "/meeting/{meeting.number}/materials/{doc.name}-{doc.rev}", "chatlog": "/meeting/{meeting.number}/materials/{doc.name}-{doc.rev}", "polls": "/meeting/{meeting.number}/materials/{doc.name}-{doc.rev}", @@ -899,6 +1018,7 @@ def skip_unreadable_post(record): MEETING_DOC_CDN_HREFS = { "agenda": "https://www.ietf.org/proceedings/{meeting.number}/agenda/{doc.name}-{doc.rev}", "minutes": "https://www.ietf.org/proceedings/{meeting.number}/minutes/{doc.name}-{doc.rev}", + "narrativeminutes": "https://www.ietf.org/proceedings/{meeting.number}/narrative-minutes/{doc.name}-{doc.rev}", "slides": "https://www.ietf.org/proceedings/{meeting.number}/slides/{doc.name}-{doc.rev}", "recording": "{doc.external_url}", "bluesheets": "https://www.ietf.org/proceedings/{meeting.number}/bluesheets/{doc.uploaded_filename}", @@ -910,6 +1030,7 @@ def skip_unreadable_post(record): MEETING_DOC_OLD_HREFS = { "agenda": "/meeting/{meeting.number}/materials/{doc.name}", "minutes": "/meeting/{meeting.number}/materials/{doc.name}", + "narrativeminutes" : "/meeting/{meeting.number}/materials/{doc.name}", "slides": "/meeting/{meeting.number}/materials/{doc.name}", "recording": "{doc.external_url}", "bluesheets": "https://www.ietf.org/proceedings/{meeting.number}/bluesheets/{doc.uploaded_filename}", @@ -919,6 +1040,7 @@ def skip_unreadable_post(record): MEETING_DOC_GREFS = { "agenda": "/meeting/{meeting.number}/materials/{doc.name}", "minutes": "/meeting/{meeting.number}/materials/{doc.name}", + "narrativeminutes": "/meeting/{meeting.number}/materials/{doc.name}", "slides": "/meeting/{meeting.number}/materials/{doc.name}", "recording": "{doc.external_url}", "bluesheets": "https://www.ietf.org/proceedings/{meeting.number}/bluesheets/{doc.uploaded_filename}", @@ -932,6 +1054,7 @@ def skip_unreadable_post(record): MEETING_VALID_UPLOAD_EXTENSIONS = { 'agenda': ['.txt','.html','.htm', '.md', ], 'minutes': ['.txt','.html','.htm', '.md', '.pdf', ], + 'narrativeminutes': ['.txt','.html','.htm', '.md', '.pdf', ], 'slides': ['.doc','.docx','.pdf','.ppt','.pptx','.txt', ], # Note the removal of .zip 'bluesheets': ['.pdf', '.txt', ], 'procmaterials':['.pdf', ], @@ -941,6 +1064,7 @@ def skip_unreadable_post(record): MEETING_VALID_UPLOAD_MIME_TYPES = { 'agenda': ['text/plain', 'text/html', 'text/markdown', 'text/x-markdown', ], 'minutes': ['text/plain', 'text/html', 'application/pdf', 'text/markdown', 'text/x-markdown', ], + 'narrativeminutes': ['text/plain', 'text/html', 'application/pdf', 'text/markdown', 'text/x-markdown', ], 'slides': [], 'bluesheets': ['application/pdf', 'text/plain', ], 'procmaterials':['application/pdf', ], @@ -995,32 +1119,35 @@ def skip_unreadable_post(record): # ============================================================================== -RSYNC_BINARY = '/usr/bin/rsync' YANGLINT_BINARY = '/usr/bin/yanglint' DE_GFM_BINARY = '/usr/bin/de-gfm.ruby2.5' # Account settings DAYS_TO_EXPIRE_REGISTRATION_LINK = 3 MINUTES_TO_EXPIRE_RESET_PASSWORD_LINK = 60 -HTPASSWD_COMMAND = "/usr/bin/htpasswd" -HTPASSWD_FILE = "/www/htpasswd" # Generation of pdf files GHOSTSCRIPT_COMMAND = "/usr/bin/gs" -# Generation of bibxml files (currently only for internet drafts) +# Generation of bibxml files (currently only for Internet-Drafts) BIBXML_BASE_PATH = '/a/ietfdata/derived/bibxml' # Timezone files for iCalendar TZDATA_ICS_PATH = BASE_DIR + '/../vzic/zoneinfo/' -SECR_BLUE_SHEET_PATH = '/a/www/ietf-datatracker/documents/blue_sheet.rtf' -SECR_BLUE_SHEET_URL = IDTRACKER_BASE_URL + '/documents/blue_sheet.rtf' -SECR_INTERIM_LISTING_DIR = '/a/www/www6/meeting/interim' -SECR_MAX_UPLOAD_SIZE = 40960000 -SECR_PROCEEDINGS_DIR = '/a/www/www6s/proceedings/' -SECR_PPT2PDF_COMMAND = ['/usr/bin/soffice','--headless','--convert-to','pdf:writer_globaldocument_pdf_Export','--outdir'] -STATS_REGISTRATION_ATTENDEES_JSON_URL = 'https://registration.ietf.org/{number}/attendees/' +DATATRACKER_MAX_UPLOAD_SIZE = 40960000 +PPT2PDF_COMMAND = [ + "/usr/bin/soffice", + "--headless", # no GUI + "--safe-mode", # use a new libreoffice profile every time (ensures no reliance on accumulated profile config) + "--norestore", # don't attempt to restore files after a previous crash (ensures that one crash won't block future conversions until UI intervention) + "--convert-to", "pdf:writer_globaldocument_pdf_Export", + "--outdir" +] + +REGISTRATION_PARTICIPANTS_API_URL = 'https://registration.ietf.org/api/v1/participants-dt/' +REGISTRATION_PARTICIPANTS_API_KEY = 'changeme' + PROCEEDINGS_VERSION_CHANGES = [ 0, # version 1 97, # version 2: meeting 97 and later (was number was NEW_PROCEEDINGS_START) @@ -1040,7 +1167,6 @@ def skip_unreadable_post(record): # CHAT_ARCHIVE_URL_PATTERN = 'https://www.ietf.org/jabber/logs/{chat_room_name}?C=M;O=D' PYFLAKES_DEFAULT_ARGS= ["ietf", ] -VULTURE_DEFAULT_ARGS= ["ietf", ] # Automatic Scheduling # @@ -1087,16 +1213,6 @@ def skip_unreadable_post(record): TEST_DATA_DIR = os.path.abspath(BASE_DIR + "/../test/data") -# Path to the email alias lists. Used by ietf.utils.aliases -DRAFT_ALIASES_PATH = os.path.join(TEST_DATA_DIR, "draft-aliases") -DRAFT_VIRTUAL_PATH = os.path.join(TEST_DATA_DIR, "draft-virtual") -DRAFT_VIRTUAL_DOMAIN = "virtual.ietf.org" - -GROUP_ALIASES_PATH = os.path.join(TEST_DATA_DIR, "group-aliases") -GROUP_VIRTUAL_PATH = os.path.join(TEST_DATA_DIR, "group-virtual") -GROUP_VIRTUAL_DOMAIN = "virtual.ietf.org" - -POSTCONFIRM_PATH = "/a/postconfirm/wrapper" USER_PREFERENCE_DEFAULTS = { "expires_soon" : "14", @@ -1112,21 +1228,23 @@ def skip_unreadable_post(record): "@ietf.org$", ] +# Configuration for django-markup MARKUP_SETTINGS = { 'restructuredtext': { 'settings_overrides': { + 'report_level': 3, # error (3) or severe (4) only 'initial_header_level': 3, 'doctitle_xform': False, 'footnote_references': 'superscript', 'trim_footnote_reference_space': True, 'default_reference_context': 'view', + 'raw_enabled': False, # critical for security + 'file_insertion_enabled': False, # critical for security 'link_base': '' } } } -MAILMAN_LIB_DIR = '/usr/lib/mailman' - # This is the number of seconds required between subscribing to an ietf # mailing list and datatracker account creation being accepted LIST_ACCOUNT_DELAY = 60*60*25 # 25 hours @@ -1135,14 +1253,13 @@ def skip_unreadable_post(record): SILENCED_SYSTEM_CHECKS = [ "fields.W342", # Setting unique=True on a ForeignKey has the same effect as using a OneToOneField. + "fields.W905", # django.contrib.postgres.fields.CICharField is deprecated. (see https://github.com/ietf-tools/datatracker/issues/5660) ] CHECKS_LIBRARY_PATCHES_TO_APPLY = [ 'patch/change-oidc-provider-field-sizes-228.patch', 'patch/fix-oidc-access-token-post.patch', 'patch/fix-jwkest-jwt-logging.patch', - 'patch/fix-django-password-strength-kwargs.patch', - 'patch/add-django-http-cookie-value-none.patch', 'patch/django-cookie-delete-with-all-settings.patch', 'patch/tastypie-django22-fielderror-response.patch', ] @@ -1185,6 +1302,19 @@ def skip_unreadable_post(record): CELERY_BROKER_URL = 'amqp://mq/' CELERY_BEAT_SCHEDULER = 'django_celery_beat.schedulers:DatabaseScheduler' CELERY_BEAT_SYNC_EVERY = 1 # update DB after every event +CELERY_BEAT_CRON_STARTING_DEADLINE = 1800 # seconds after a missed deadline before abandoning a cron task +CELERY_BROKER_CONNECTION_RETRY_ON_STARTUP = True # the default, but setting it squelches a warning +# Use a result backend so we can chain tasks. This uses the rpc backend, see +# https://docs.celeryq.dev/en/stable/userguide/tasks.html#rpc-result-backend-rabbitmq-qpid +# Results can be retrieved only once and only by the caller of the task. Results will be +# lost if the message broker restarts. +CELERY_RESULT_BACKEND = 'django-cache' # use a Django cache for results +CELERY_CACHE_BACKEND = 'celery-results' # which Django cache to use +CELERY_RESULT_EXPIRES = datetime.timedelta(minutes=5) # how long are results valid? (Default is 1 day) +CELERY_TASK_IGNORE_RESULT = True # ignore results unless specifically enabled for a task +CELERY_TASK_ROUTES = { + "ietf.blobdb.tasks.pybob_the_blob_replicator_task": {"queue": "blobdb"} +} # Meetecho API setup: Uncomment this and provide real credentials to enable # Meetecho conference creation for interim session requests @@ -1194,8 +1324,22 @@ def skip_unreadable_post(record): # 'client_id': 'datatracker', # 'client_secret': 'some secret', # 'request_timeout': 3.01, # python-requests doc recommend slightly > a multiple of 3 seconds +# # How many minutes before/after session to enable slide update API. Defaults to 15. Set to None to disable, +# # or < 0 to _always_ send updates (useful for debugging) +# 'slides_notify_time': 15, +# 'debug': False, # if True, API calls will be echoed as debug instead of sent (only works for slides for now) # } +# Meetecho URLs - instantiate with url.format(session=some_session) +MEETECHO_ONSITE_TOOL_URL = "https://meetings.conf.meetecho.com/onsite{session.meeting.number}/?session={session.pk}" +MEETECHO_VIDEO_STREAM_URL = "https://meetings.conf.meetecho.com/ietf{session.meeting.number}/?session={session.pk}" +MEETECHO_AUDIO_STREAM_URL = "https://mp3.conf.meetecho.com/ietf{session.meeting.number}/{session.pk}.m3u" +MEETECHO_SESSION_RECORDING_URL = "https://meetecho-player.ietf.org/playout/?session={session_label}" + +# Errata system api configuration +# settings should provide +# ERRATA_METADATA_NOTIFICATION_URL +# ERRATA_METADATA_NOTIFICATION_API_KEY # Put the production SECRET_KEY in settings_local.py, and also any other # sensitive or site-specific changes. DO NOT commit settings_local.py to svn. @@ -1216,6 +1360,143 @@ def skip_unreadable_post(record): MIDDLEWARE += DEV_MIDDLEWARE TEMPLATES[0]['OPTIONS']['context_processors'] += DEV_TEMPLATE_CONTEXT_PROCESSORS +if "CACHES" not in locals(): + if SERVER_MODE == "production": + MEMCACHED_HOST = os.environ.get("MEMCACHED_SERVICE_HOST", "127.0.0.1") + MEMCACHED_PORT = os.environ.get("MEMCACHED_SERVICE_PORT", "11211") + CACHES = { + "default": { + "BACKEND": "ietf.utils.cache.LenientMemcacheCache", + "LOCATION": f"{MEMCACHED_HOST}:{MEMCACHED_PORT}", + "VERSION": __version__, + "KEY_PREFIX": "ietf:dt", + # Key function is default except with sha384-encoded key + "KEY_FUNCTION": lambda key, key_prefix, version: ( + f"{key_prefix}:{version}:{sha384(str(key).encode('utf8')).hexdigest()}" + ), + }, + "agenda": { + "BACKEND": "ietf.utils.cache.LenientMemcacheCache", + "LOCATION": f"{MEMCACHED_HOST}:{MEMCACHED_PORT}", + # No release-specific VERSION setting. + "KEY_PREFIX": "ietf:dt:agenda", + # Key function is default except with sha384-encoded key + "KEY_FUNCTION": lambda key, key_prefix, version: ( + f"{key_prefix}:{version}:{sha384(str(key).encode('utf8')).hexdigest()}" + ), + }, + "proceedings": { + "BACKEND": "ietf.utils.cache.LenientMemcacheCache", + "LOCATION": f"{MEMCACHED_HOST}:{MEMCACHED_PORT}", + # No release-specific VERSION setting. + "KEY_PREFIX": "ietf:dt:proceedings", + # Key function is default except with sha384-encoded key + "KEY_FUNCTION": lambda key, key_prefix, version: ( + f"{key_prefix}:{version}:{sha384(str(key).encode('utf8')).hexdigest()}" + ), + }, + "sessions": { + "BACKEND": "ietf.utils.cache.LenientMemcacheCache", + "LOCATION": f"{MEMCACHED_HOST}:{MEMCACHED_PORT}", + # No release-specific VERSION setting. + "KEY_PREFIX": "ietf:dt", + }, + "htmlized": { + "BACKEND": "django.core.cache.backends.filebased.FileBasedCache", + "LOCATION": "/a/cache/datatracker/htmlized", + "OPTIONS": { + "MAX_ENTRIES": 100000, # 100,000 + }, + }, + "pdfized": { + "BACKEND": "django.core.cache.backends.filebased.FileBasedCache", + "LOCATION": "/a/cache/datatracker/pdfized", + "OPTIONS": { + "MAX_ENTRIES": 100000, # 100,000 + }, + }, + "slowpages": { + "BACKEND": "django.core.cache.backends.filebased.FileBasedCache", + "LOCATION": "/a/cache/datatracker/slowpages", + "OPTIONS": { + "MAX_ENTRIES": 5000, + }, + }, + "celery-results": { + "BACKEND": "django.core.cache.backends.memcached.PyMemcacheCache", + "LOCATION": f"{MEMCACHED_HOST}:{MEMCACHED_PORT}", + "KEY_PREFIX": "ietf:celery", + }, + } + else: + CACHES = { + "default": { + "BACKEND": "django.core.cache.backends.dummy.DummyCache", + #'BACKEND': 'ietf.utils.cache.LenientMemcacheCache', + #'LOCATION': '127.0.0.1:11211', + #'BACKEND': 'django.core.cache.backends.filebased.FileBasedCache', + "VERSION": __version__, + "KEY_PREFIX": "ietf:dt", + }, + "agenda": { + "BACKEND": "django.core.cache.backends.dummy.DummyCache", + # "BACKEND": "ietf.utils.cache.LenientMemcacheCache", + # "LOCATION": "127.0.0.1:11211", + # No release-specific VERSION setting. + "KEY_PREFIX": "ietf:dt:agenda", + # Key function is default except with sha384-encoded key + "KEY_FUNCTION": lambda key, key_prefix, version: ( + f"{key_prefix}:{version}:{sha384(str(key).encode('utf8')).hexdigest()}" + ), + }, + "proceedings": { + "BACKEND": "django.core.cache.backends.dummy.DummyCache", + # "BACKEND": "ietf.utils.cache.LenientMemcacheCache", + # "LOCATION": "127.0.0.1:11211", + # No release-specific VERSION setting. + "KEY_PREFIX": "ietf:dt:proceedings", + # Key function is default except with sha384-encoded key + "KEY_FUNCTION": lambda key, key_prefix, version: ( + f"{key_prefix}:{version}:{sha384(str(key).encode('utf8')).hexdigest()}" + ), + }, + "sessions": { + "BACKEND": "django.core.cache.backends.locmem.LocMemCache", + }, + "htmlized": { + "BACKEND": "django.core.cache.backends.dummy.DummyCache", + #'BACKEND': 'django.core.cache.backends.filebased.FileBasedCache', + "LOCATION": "/var/cache/datatracker/htmlized", + "OPTIONS": { + "MAX_ENTRIES": 1000, + }, + }, + "pdfized": { + "BACKEND": "django.core.cache.backends.dummy.DummyCache", + #'BACKEND': 'django.core.cache.backends.filebased.FileBasedCache', + "LOCATION": "/var/cache/datatracker/pdfized", + "OPTIONS": { + "MAX_ENTRIES": 1000, + }, + }, + "slowpages": { + "BACKEND": "django.core.cache.backends.dummy.DummyCache", + #'BACKEND': 'django.core.cache.backends.filebased.FileBasedCache', + "LOCATION": "/var/cache/datatracker/", + "OPTIONS": { + "MAX_ENTRIES": 5000, + }, + }, + "celery-results": { + "BACKEND": "django.core.cache.backends.memcached.PyMemcacheCache", + "LOCATION": "app:11211", + "KEY_PREFIX": "ietf:celery", + }, + } + +PUBLISH_IPR_STATES = ['posted', 'removed', 'removed_objfalse'] + +ADVERTISE_VERSIONS = ["markdown", "pyang", "rfc2html", "xml2rfc"] # We provide a secret key only for test and development modes. It's # absolutely vital that django fails to start in production mode unless a @@ -1226,44 +1507,6 @@ def skip_unreadable_post(record): loaders = TEMPLATES[0]['OPTIONS']['loaders'] loaders = tuple(l for e in loaders for l in (e[1] if isinstance(e, tuple) and "cached.Loader" in e[0] else (e,))) TEMPLATES[0]['OPTIONS']['loaders'] = loaders - - CACHES = { - 'default': { - 'BACKEND': 'django.core.cache.backends.dummy.DummyCache', - #'BACKEND': 'django.core.cache.backends.memcached.MemcachedCache', - #'LOCATION': '127.0.0.1:11211', - #'BACKEND': 'django.core.cache.backends.filebased.FileBasedCache', - 'VERSION': __version__, - 'KEY_PREFIX': 'ietf:dt', - }, - 'sessions': { - 'BACKEND': 'django.core.cache.backends.locmem.LocMemCache', - }, - 'htmlized': { - 'BACKEND': 'django.core.cache.backends.dummy.DummyCache', - #'BACKEND': 'django.core.cache.backends.filebased.FileBasedCache', - 'LOCATION': '/var/cache/datatracker/htmlized', - 'OPTIONS': { - 'MAX_ENTRIES': 1000, - }, - }, - 'pdfized': { - 'BACKEND': 'django.core.cache.backends.dummy.DummyCache', - #'BACKEND': 'django.core.cache.backends.filebased.FileBasedCache', - 'LOCATION': '/var/cache/datatracker/pdfized', - 'OPTIONS': { - 'MAX_ENTRIES': 1000, - }, - }, - 'slowpages': { - 'BACKEND': 'django.core.cache.backends.dummy.DummyCache', - #'BACKEND': 'django.core.cache.backends.filebased.FileBasedCache', - 'LOCATION': '/var/cache/datatracker/', - 'OPTIONS': { - 'MAX_ENTRIES': 5000, - }, - }, - } SESSION_ENGINE = "django.contrib.sessions.backends.db" if 'SECRET_KEY' not in locals(): @@ -1272,17 +1515,28 @@ def skip_unreadable_post(record): NOMCOM_APP_SECRET = b'\x9b\xdas1\xec\xd5\xa0SI~\xcb\xd4\xf5t\x99\xc4i\xd7\x9f\x0b\xa9\xe8\xfeY\x80$\x1e\x12tN:\x84' ALLOWED_HOSTS = ['*',] - + try: # see https://github.com/omarish/django-cprofile-middleware - import django_cprofile_middleware # pyflakes:ignore - MIDDLEWARE = MIDDLEWARE + ['django_cprofile_middleware.middleware.ProfilerMiddleware', ] + import django_cprofile_middleware # pyflakes:ignore + + MIDDLEWARE = MIDDLEWARE + [ + "django_cprofile_middleware.middleware.ProfilerMiddleware", + ] + DJANGO_CPROFILE_MIDDLEWARE_REQUIRE_STAFF = ( + False # Do not use this setting for a public site! + ) except ImportError: pass # Cannot have this set to True if we're using http: from the dev-server: CSRF_COOKIE_SECURE = False CSRF_COOKIE_SAMESITE = 'Lax' + CSRF_TRUSTED_ORIGINS += ['http://localhost:8000', 'http://127.0.0.1:8000', 'http://[::1]:8000'] SESSION_COOKIE_SECURE = False SESSION_COOKIE_SAMESITE = 'Lax' - + + +YOUTUBE_DOMAINS = ['www.youtube.com', 'youtube.com', 'youtu.be', 'm.youtube.com', 'youtube-nocookie.com', 'www.youtube-nocookie.com'] + +IETF_DOI_PREFIX = "10.17487" diff --git a/ietf/settings_sqlitetest.py b/ietf/settings_sqlitetest.py deleted file mode 100644 index 784e8dea64..0000000000 --- a/ietf/settings_sqlitetest.py +++ /dev/null @@ -1,58 +0,0 @@ -# Copyright The IETF Trust 2010-2020, All Rights Reserved -# -*- coding: utf-8 -*- - - -# Standard settings except we use SQLite and skip migrations, this is -# useful for speeding up tests that depend on the test database, try -# for instance: -# -# ./manage.py test --settings=settings_sqlitetest doc.ChangeStateTestCase -# - -import os -from ietf.settings import * # pyflakes:ignore -from ietf.settings import TEST_CODE_COVERAGE_CHECKER, BASE_DIR, PHOTOS_DIRNAME -import debug # pyflakes:ignore -debug.debug = True - -# Use a different hostname, to catch hardcoded values -IDTRACKER_BASE_URL = "https://sqlitetest.ietf.org" - -# Workaround to avoid spending minutes stepping through the migrations in -# every test run. The result of this is to use the 'syncdb' way of creating -# the test database instead of doing it through the migrations. Taken from -# https://gist.github.com/NotSqrt/5f3c76cd15e40ef62d09 - -class DisableMigrations(object): - - def __contains__(self, item): - return True - - def __getitem__(self, item): - return None - -MIGRATION_MODULES = DisableMigrations() - - -DATABASES = { - 'default': { - 'NAME': 'test.db', - 'ENGINE': 'django.db.backends.sqlite3', - }, - } - -if TEST_CODE_COVERAGE_CHECKER and not TEST_CODE_COVERAGE_CHECKER._started: # pyflakes:ignore - TEST_CODE_COVERAGE_CHECKER.start() # pyflakes:ignore - -NOMCOM_PUBLIC_KEYS_DIR=os.path.abspath("tmp-nomcom-public-keys-dir") - -MEDIA_ROOT = os.path.join(os.path.dirname(BASE_DIR), 'test/media/') # pyflakes:ignore -MEDIA_URL = '/test/media/' -PHOTOS_DIR = MEDIA_ROOT + PHOTOS_DIRNAME # pyflakes:ignore - -# Undo any developer-dependent middleware when running the tests -MIDDLEWARE = [ c for c in MIDDLEWARE if not c in DEV_MIDDLEWARE ] # pyflakes:ignore - -TEMPLATES[0]['OPTIONS']['context_processors'] = [ p for p in TEMPLATES[0]['OPTIONS']['context_processors'] if not p in DEV_TEMPLATE_CONTEXT_PROCESSORS ] # pyflakes:ignore - -REQUEST_PROFILE_STORE_ANONYMOUS_SESSIONS = False diff --git a/ietf/settings_test.py b/ietf/settings_test.py new file mode 100755 index 0000000000..e7ebc13eb2 --- /dev/null +++ b/ietf/settings_test.py @@ -0,0 +1,126 @@ +# Copyright The IETF Trust 2010-2023, All Rights Reserved +# -*- coding: utf-8 -*- + + +# Standard settings except we use Postgres and skip migrations, this is +# useful for speeding up tests that depend on the test database, try +# for instance: +# +# ./manage.py test --settings=settings_test doc.ChangeStateTestCase +# + +import atexit +import os +import shutil +import tempfile +from ietf.settings import * # pyflakes:ignore +from ietf.settings import ORIG_AUTH_PASSWORD_VALIDATORS, STORAGES +import debug # pyflakes:ignore +debug.debug = True + +# Use a different hostname, to catch hardcoded values +IDTRACKER_BASE_URL = "https://postgrestest.ietf.org" + +# Workaround to avoid spending minutes stepping through the migrations in +# every test run. The result of this is to use the 'syncdb' way of creating +# the test database instead of doing it through the migrations. Taken from +# https://gist.github.com/NotSqrt/5f3c76cd15e40ef62d09 + +class DisableMigrations(object): + + def __contains__(self, item): + return True + + def __getitem__(self, item): + return None + +MIGRATION_MODULES = DisableMigrations() + + +DATABASES = { + 'default': { + 'HOST': 'db', + 'PORT': '5432', + 'NAME': 'test.db', + 'ENGINE': 'django.db.backends.postgresql', + 'USER': 'django', + 'PASSWORD': 'RkTkDPFnKpko', + }, + } + +# test with a single DB - do not use a DB router +BLOBDB_DATABASE = "default" +DATABASE_ROUTERS = [] # type: ignore + +if TEST_CODE_COVERAGE_CHECKER: # pyflakes:ignore + TEST_CODE_COVERAGE_CHECKER.start() # pyflakes:ignore + +def tempdir_with_cleanup(**kwargs): + """Utility to create a temporary dir and arrange cleanup""" + _dir = tempfile.mkdtemp(**kwargs) + atexit.register(shutil.rmtree, _dir) + return _dir + + +NOMCOM_PUBLIC_KEYS_DIR = tempdir_with_cleanup(suffix="-nomcom-public-keys-dir") + +MEDIA_ROOT = tempdir_with_cleanup(suffix="-media") +PHOTOS_DIRNAME = "photo" +PHOTOS_DIR = os.path.join(MEDIA_ROOT, PHOTOS_DIRNAME) +os.mkdir(PHOTOS_DIR) + +# Undo any developer-dependent middleware when running the tests +MIDDLEWARE = [ c for c in MIDDLEWARE if not c in DEV_MIDDLEWARE ] # pyflakes:ignore + +TEMPLATES[0]['OPTIONS']['context_processors'] = [ p for p in TEMPLATES[0]['OPTIONS']['context_processors'] if not p in DEV_TEMPLATE_CONTEXT_PROCESSORS ] # pyflakes:ignore + +REQUEST_PROFILE_STORE_ANONYMOUS_SESSIONS = False + +# Override loggers with a safer set in case things go to the log during testing. Specifically, +# make sure there are no syslog loggers that might send things to a real syslog. +LOGGING["loggers"] = { # pyflakes:ignore + 'django': { + 'handlers': ['debug_console'], + 'level': 'INFO', + }, + 'django.request': { + 'handlers': ['debug_console'], + 'level': 'ERROR', + }, + 'django.server': { + 'handlers': ['django.server'], + 'level': 'INFO', + }, + 'django.security': { + 'handlers': ['debug_console', ], + 'level': 'INFO', + }, + 'oidc_provider': { + 'handlers': ['debug_console', ], + 'level': 'DEBUG', + }, + 'datatracker': { + 'handlers': ['debug_console'], + 'level': 'INFO', + }, + 'celery': { + 'handlers': ['debug_console'], + 'level': 'INFO', + }, +} + +# Restore AUTH_PASSWORD_VALIDATORS if they were reset in settings_local +try: + AUTH_PASSWORD_VALIDATORS = ORIG_AUTH_PASSWORD_VALIDATORS +except NameError: + pass + +# Use InMemoryStorage for red bucket and r2-rfc storages +STORAGES["red_bucket"] = { + "BACKEND": "django.core.files.storage.InMemoryStorage", + "OPTIONS": {"location": "red_bucket"}, +} +STORAGES["r2-rfc"] = { + "BACKEND": "django.core.files.storage.InMemoryStorage", + "OPTIONS": {"location": "r2-rfc"}, +} diff --git a/ietf/settings_testcrawl.py b/ietf/settings_testcrawl.py index a1b5ce8946..edb978757a 100644 --- a/ietf/settings_testcrawl.py +++ b/ietf/settings_testcrawl.py @@ -27,9 +27,14 @@ 'MAX_ENTRIES': 10000, }, }, + 'agenda': { + 'BACKEND': 'django.core.cache.backends.locmem.LocMemCache', + }, + 'proceedings': { + 'BACKEND': 'django.core.cache.backends.locmem.LocMemCache', + }, 'sessions': { 'BACKEND': 'django.core.cache.backends.locmem.LocMemCache', - # No version-specific VERSION setting. }, 'htmlized': { 'BACKEND': 'django.core.cache.backends.dummy.DummyCache', diff --git a/ietf/static/css/custom-bs-import.scss b/ietf/static/css/custom-bs-import.scss new file mode 100644 index 0000000000..644efdcf10 --- /dev/null +++ b/ietf/static/css/custom-bs-import.scss @@ -0,0 +1,58 @@ +@import "bootstrap/scss/functions"; + +// Enable negative margin classes. +$enable-negative-margins: true; + +// Don't add carets to dropdowns by default. +// $enable-caret: false; + +$popover-max-width: 100%; + +// Override default fonts + +$font-family-sans-serif: "Inter", +system-ui, +-apple-system, +"Segoe UI", +Roboto, +"Helvetica Neue", +"Noto Sans", +"Liberation Sans", +Arial, +sans-serif, +"Apple Color Emoji", +"Segoe UI Emoji", +"Segoe UI Symbol", +"Noto Color Emoji"; +$font-family-monospace: "Noto Sans Mono", +SFMono-Regular, +Menlo, +Monaco, +Consolas, +"Liberation Mono", +"Courier New", +monospace; + +// Enable color modes +$color-mode-type: data; + +@import "bootstrap/scss/variables"; +@import "bootstrap/scss/variables-dark"; + +$h1-font-size: $font-size-base * 2.2; +$h2-font-size: $font-size-base * 1.8; +$h3-font-size: $font-size-base * 1.6; +$h4-font-size: $font-size-base * 1.4; +$h5-font-size: $font-size-base * 1.2; +$h6-font-size: $font-size-base; + +// Default is gray-800, which is the same as the range slider background. +$light-bg-subtle-dark: mix($gray-800, $black); + +@import "bootstrap/scss/maps"; +@import "bootstrap/scss/mixins"; +@import "bootstrap/scss/utilities"; +@import "bootstrap/scss/root"; + + + diff --git a/ietf/static/css/datepicker.scss b/ietf/static/css/datepicker.scss index c25b790e0b..b193ccda3a 100644 --- a/ietf/static/css/datepicker.scss +++ b/ietf/static/css/datepicker.scss @@ -1,683 +1,32 @@ -/*! - * Datepicker for Bootstrap v1.9.0 (https://github.com/uxsolutions/bootstrap-datepicker) - * - * Licensed under the Apache License v2.0 (http://www.apache.org/licenses/LICENSE-2.0) - */ +@import "custom-bs-import"; - .datepicker { - border-radius: 4px; - direction: ltr; - } - .datepicker-inline { - width: 220px; - } - .datepicker-rtl { - direction: rtl; - } - .datepicker-rtl.dropdown-menu { - left: auto; - } - .datepicker-rtl table tr td span { - float: right; - } - .datepicker-dropdown { - top: 0; - left: 0; - padding: 4px; - } - .datepicker-dropdown:before { - content: ''; - display: inline-block; - border-left: 7px solid transparent; - border-right: 7px solid transparent; - border-bottom: 7px solid rgba(0, 0, 0, 0.15); - border-top: 0; - border-bottom-color: rgba(0, 0, 0, 0.2); - position: absolute; - } - .datepicker-dropdown:after { - content: ''; - display: inline-block; - border-left: 6px solid transparent; - border-right: 6px solid transparent; - border-bottom: 6px solid #fff; - border-top: 0; - position: absolute; - } - .datepicker-dropdown.datepicker-orient-left:before { - left: 6px; - } - .datepicker-dropdown.datepicker-orient-left:after { - left: 7px; - } - .datepicker-dropdown.datepicker-orient-right:before { - right: 6px; - } - .datepicker-dropdown.datepicker-orient-right:after { - right: 7px; - } - .datepicker-dropdown.datepicker-orient-bottom:before { - top: -7px; - } - .datepicker-dropdown.datepicker-orient-bottom:after { - top: -6px; - } - .datepicker-dropdown.datepicker-orient-top:before { - bottom: -7px; - border-bottom: 0; - border-top: 7px solid rgba(0, 0, 0, 0.15); - } - .datepicker-dropdown.datepicker-orient-top:after { - bottom: -6px; - border-bottom: 0; - border-top: 6px solid #fff; - } - .datepicker table { - margin: 0; - -webkit-touch-callout: none; - -webkit-user-select: none; - -khtml-user-select: none; - -moz-user-select: none; - -ms-user-select: none; - user-select: none; - } - .datepicker table tr td, - .datepicker table tr th { - text-align: center; - width: 30px; - height: 30px; - border-radius: 4px; - border: none; - } - .table-striped .datepicker table tr td, - .table-striped .datepicker table tr th { - background-color: transparent; - } - .datepicker table tr td.old, - .datepicker table tr td.new { - color: #777777; - } - .datepicker table tr td.day:hover, - .datepicker table tr td.focused { - background: #eeeeee; - cursor: pointer; - } - .datepicker table tr td.disabled, - .datepicker table tr td.disabled:hover { - background: none; - color: #777777; - cursor: default; - } - .datepicker table tr td.highlighted { - color: #000; - background-color: #d9edf7; - border-color: #85c5e5; - border-radius: 0; - } - .datepicker table tr td.highlighted:focus, - .datepicker table tr td.highlighted.focus { - color: #000; - background-color: #afd9ee; - border-color: #298fc2; - } - .datepicker table tr td.highlighted:hover { - color: #000; - background-color: #afd9ee; - border-color: #52addb; - } - .datepicker table tr td.highlighted:active, - .datepicker table tr td.highlighted.active { - color: #000; - background-color: #afd9ee; - border-color: #52addb; - } - .datepicker table tr td.highlighted:active:hover, - .datepicker table tr td.highlighted.active:hover, - .datepicker table tr td.highlighted:active:focus, - .datepicker table tr td.highlighted.active:focus, - .datepicker table tr td.highlighted:active.focus, - .datepicker table tr td.highlighted.active.focus { - color: #000; - background-color: #91cbe8; - border-color: #298fc2; - } - .datepicker table tr td.highlighted.disabled:hover, - .datepicker table tr td.highlighted[disabled]:hover, - fieldset[disabled] .datepicker table tr td.highlighted:hover, - .datepicker table tr td.highlighted.disabled:focus, - .datepicker table tr td.highlighted[disabled]:focus, - fieldset[disabled] .datepicker table tr td.highlighted:focus, - .datepicker table tr td.highlighted.disabled.focus, - .datepicker table tr td.highlighted[disabled].focus, - fieldset[disabled] .datepicker table tr td.highlighted.focus { - background-color: #d9edf7; - border-color: #85c5e5; - } - .datepicker table tr td.highlighted.focused { - background: #afd9ee; - } - .datepicker table tr td.highlighted.disabled, - .datepicker table tr td.highlighted.disabled:active { - background: #d9edf7; - color: #777777; - } - .datepicker table tr td.today { - color: #000; - background-color: #ffdb99; - border-color: #ffb733; - } - .datepicker table tr td.today:focus, - .datepicker table tr td.today.focus { - color: #000; - background-color: #ffc966; - border-color: #b37400; - } - .datepicker table tr td.today:hover { - color: #000; - background-color: #ffc966; - border-color: #f59e00; - } - .datepicker table tr td.today:active, - .datepicker table tr td.today.active { - color: #000; - background-color: #ffc966; - border-color: #f59e00; - } - .datepicker table tr td.today:active:hover, - .datepicker table tr td.today.active:hover, - .datepicker table tr td.today:active:focus, - .datepicker table tr td.today.active:focus, - .datepicker table tr td.today:active.focus, - .datepicker table tr td.today.active.focus { - color: #000; - background-color: #ffbc42; - border-color: #b37400; - } - .datepicker table tr td.today.disabled:hover, - .datepicker table tr td.today[disabled]:hover, - fieldset[disabled] .datepicker table tr td.today:hover, - .datepicker table tr td.today.disabled:focus, - .datepicker table tr td.today[disabled]:focus, - fieldset[disabled] .datepicker table tr td.today:focus, - .datepicker table tr td.today.disabled.focus, - .datepicker table tr td.today[disabled].focus, - fieldset[disabled] .datepicker table tr td.today.focus { - background-color: #ffdb99; - border-color: #ffb733; - } - .datepicker table tr td.today.focused { - background: #ffc966; - } - .datepicker table tr td.today.disabled, - .datepicker table tr td.today.disabled:active { - background: #ffdb99; - color: #777777; - } - .datepicker table tr td.range { - color: #000; - background-color: #eeeeee; - border-color: #bbbbbb; - border-radius: 0; - } - .datepicker table tr td.range:focus, - .datepicker table tr td.range.focus { - color: #000; - background-color: #d5d5d5; - border-color: #7c7c7c; - } - .datepicker table tr td.range:hover { - color: #000; - background-color: #d5d5d5; - border-color: #9d9d9d; - } - .datepicker table tr td.range:active, - .datepicker table tr td.range.active { - color: #000; - background-color: #d5d5d5; - border-color: #9d9d9d; - } - .datepicker table tr td.range:active:hover, - .datepicker table tr td.range.active:hover, - .datepicker table tr td.range:active:focus, - .datepicker table tr td.range.active:focus, - .datepicker table tr td.range:active.focus, - .datepicker table tr td.range.active.focus { - color: #000; - background-color: #c3c3c3; - border-color: #7c7c7c; - } - .datepicker table tr td.range.disabled:hover, - .datepicker table tr td.range[disabled]:hover, - fieldset[disabled] .datepicker table tr td.range:hover, - .datepicker table tr td.range.disabled:focus, - .datepicker table tr td.range[disabled]:focus, - fieldset[disabled] .datepicker table tr td.range:focus, - .datepicker table tr td.range.disabled.focus, - .datepicker table tr td.range[disabled].focus, - fieldset[disabled] .datepicker table tr td.range.focus { - background-color: #eeeeee; - border-color: #bbbbbb; - } - .datepicker table tr td.range.focused { - background: #d5d5d5; - } - .datepicker table tr td.range.disabled, - .datepicker table tr td.range.disabled:active { - background: #eeeeee; - color: #777777; - } - .datepicker table tr td.range.highlighted { - color: #000; - background-color: #e4eef3; - border-color: #9dc1d3; - } - .datepicker table tr td.range.highlighted:focus, - .datepicker table tr td.range.highlighted.focus { - color: #000; - background-color: #c1d7e3; - border-color: #4b88a6; - } - .datepicker table tr td.range.highlighted:hover { - color: #000; - background-color: #c1d7e3; - border-color: #73a6c0; - } - .datepicker table tr td.range.highlighted:active, - .datepicker table tr td.range.highlighted.active { - color: #000; - background-color: #c1d7e3; - border-color: #73a6c0; - } - .datepicker table tr td.range.highlighted:active:hover, - .datepicker table tr td.range.highlighted.active:hover, - .datepicker table tr td.range.highlighted:active:focus, - .datepicker table tr td.range.highlighted.active:focus, - .datepicker table tr td.range.highlighted:active.focus, - .datepicker table tr td.range.highlighted.active.focus { - color: #000; - background-color: #a8c8d8; - border-color: #4b88a6; - } - .datepicker table tr td.range.highlighted.disabled:hover, - .datepicker table tr td.range.highlighted[disabled]:hover, - fieldset[disabled] .datepicker table tr td.range.highlighted:hover, - .datepicker table tr td.range.highlighted.disabled:focus, - .datepicker table tr td.range.highlighted[disabled]:focus, - fieldset[disabled] .datepicker table tr td.range.highlighted:focus, - .datepicker table tr td.range.highlighted.disabled.focus, - .datepicker table tr td.range.highlighted[disabled].focus, - fieldset[disabled] .datepicker table tr td.range.highlighted.focus { - background-color: #e4eef3; - border-color: #9dc1d3; - } - .datepicker table tr td.range.highlighted.focused { - background: #c1d7e3; - } - .datepicker table tr td.range.highlighted.disabled, - .datepicker table tr td.range.highlighted.disabled:active { - background: #e4eef3; - color: #777777; - } - .datepicker table tr td.range.today { - color: #000; - background-color: #f7ca77; - border-color: #f1a417; - } - .datepicker table tr td.range.today:focus, - .datepicker table tr td.range.today.focus { - color: #000; - background-color: #f4b747; - border-color: #815608; - } - .datepicker table tr td.range.today:hover { - color: #000; - background-color: #f4b747; - border-color: #bf800c; - } - .datepicker table tr td.range.today:active, - .datepicker table tr td.range.today.active { - color: #000; - background-color: #f4b747; - border-color: #bf800c; - } - .datepicker table tr td.range.today:active:hover, - .datepicker table tr td.range.today.active:hover, - .datepicker table tr td.range.today:active:focus, - .datepicker table tr td.range.today.active:focus, - .datepicker table tr td.range.today:active.focus, - .datepicker table tr td.range.today.active.focus { - color: #000; - background-color: #f2aa25; - border-color: #815608; - } - .datepicker table tr td.range.today.disabled:hover, - .datepicker table tr td.range.today[disabled]:hover, - fieldset[disabled] .datepicker table tr td.range.today:hover, - .datepicker table tr td.range.today.disabled:focus, - .datepicker table tr td.range.today[disabled]:focus, - fieldset[disabled] .datepicker table tr td.range.today:focus, - .datepicker table tr td.range.today.disabled.focus, - .datepicker table tr td.range.today[disabled].focus, - fieldset[disabled] .datepicker table tr td.range.today.focus { - background-color: #f7ca77; - border-color: #f1a417; - } - .datepicker table tr td.range.today.disabled, - .datepicker table tr td.range.today.disabled:active { - background: #f7ca77; - color: #777777; - } - .datepicker table tr td.selected, - .datepicker table tr td.selected.highlighted { - color: #fff; - background-color: #777777; - border-color: #555555; - text-shadow: 0 -1px 0 rgba(0, 0, 0, 0.25); - } - .datepicker table tr td.selected:focus, - .datepicker table tr td.selected.highlighted:focus, - .datepicker table tr td.selected.focus, - .datepicker table tr td.selected.highlighted.focus { - color: #fff; - background-color: #5e5e5e; - border-color: #161616; - } - .datepicker table tr td.selected:hover, - .datepicker table tr td.selected.highlighted:hover { - color: #fff; - background-color: #5e5e5e; - border-color: #373737; - } - .datepicker table tr td.selected:active, - .datepicker table tr td.selected.highlighted:active, - .datepicker table tr td.selected.active, - .datepicker table tr td.selected.highlighted.active { - color: #fff; - background-color: #5e5e5e; - border-color: #373737; - } - .datepicker table tr td.selected:active:hover, - .datepicker table tr td.selected.highlighted:active:hover, - .datepicker table tr td.selected.active:hover, - .datepicker table tr td.selected.highlighted.active:hover, - .datepicker table tr td.selected:active:focus, - .datepicker table tr td.selected.highlighted:active:focus, - .datepicker table tr td.selected.active:focus, - .datepicker table tr td.selected.highlighted.active:focus, - .datepicker table tr td.selected:active.focus, - .datepicker table tr td.selected.highlighted:active.focus, - .datepicker table tr td.selected.active.focus, - .datepicker table tr td.selected.highlighted.active.focus { - color: #fff; - background-color: #4c4c4c; - border-color: #161616; - } - .datepicker table tr td.selected.disabled:hover, - .datepicker table tr td.selected.highlighted.disabled:hover, - .datepicker table tr td.selected[disabled]:hover, - .datepicker table tr td.selected.highlighted[disabled]:hover, - fieldset[disabled] .datepicker table tr td.selected:hover, - fieldset[disabled] .datepicker table tr td.selected.highlighted:hover, - .datepicker table tr td.selected.disabled:focus, - .datepicker table tr td.selected.highlighted.disabled:focus, - .datepicker table tr td.selected[disabled]:focus, - .datepicker table tr td.selected.highlighted[disabled]:focus, - fieldset[disabled] .datepicker table tr td.selected:focus, - fieldset[disabled] .datepicker table tr td.selected.highlighted:focus, - .datepicker table tr td.selected.disabled.focus, - .datepicker table tr td.selected.highlighted.disabled.focus, - .datepicker table tr td.selected[disabled].focus, - .datepicker table tr td.selected.highlighted[disabled].focus, - fieldset[disabled] .datepicker table tr td.selected.focus, - fieldset[disabled] .datepicker table tr td.selected.highlighted.focus { - background-color: #777777; - border-color: #555555; - } - .datepicker table tr td.active, - .datepicker table tr td.active.highlighted { - color: #fff; - background-color: #337ab7; - border-color: #2e6da4; - text-shadow: 0 -1px 0 rgba(0, 0, 0, 0.25); - } - .datepicker table tr td.active:focus, - .datepicker table tr td.active.highlighted:focus, - .datepicker table tr td.active.focus, - .datepicker table tr td.active.highlighted.focus { - color: #fff; - background-color: #286090; - border-color: #122b40; - } - .datepicker table tr td.active:hover, - .datepicker table tr td.active.highlighted:hover { - color: #fff; - background-color: #286090; - border-color: #204d74; - } - .datepicker table tr td.active:active, - .datepicker table tr td.active.highlighted:active, - .datepicker table tr td.active.active, - .datepicker table tr td.active.highlighted.active { - color: #fff; - background-color: #286090; - border-color: #204d74; - } - .datepicker table tr td.active:active:hover, - .datepicker table tr td.active.highlighted:active:hover, - .datepicker table tr td.active.active:hover, - .datepicker table tr td.active.highlighted.active:hover, - .datepicker table tr td.active:active:focus, - .datepicker table tr td.active.highlighted:active:focus, - .datepicker table tr td.active.active:focus, - .datepicker table tr td.active.highlighted.active:focus, - .datepicker table tr td.active:active.focus, - .datepicker table tr td.active.highlighted:active.focus, - .datepicker table tr td.active.active.focus, - .datepicker table tr td.active.highlighted.active.focus { - color: #fff; - background-color: #204d74; - border-color: #122b40; - } - .datepicker table tr td.active.disabled:hover, - .datepicker table tr td.active.highlighted.disabled:hover, - .datepicker table tr td.active[disabled]:hover, - .datepicker table tr td.active.highlighted[disabled]:hover, - fieldset[disabled] .datepicker table tr td.active:hover, - fieldset[disabled] .datepicker table tr td.active.highlighted:hover, - .datepicker table tr td.active.disabled:focus, - .datepicker table tr td.active.highlighted.disabled:focus, - .datepicker table tr td.active[disabled]:focus, - .datepicker table tr td.active.highlighted[disabled]:focus, - fieldset[disabled] .datepicker table tr td.active:focus, - fieldset[disabled] .datepicker table tr td.active.highlighted:focus, - .datepicker table tr td.active.disabled.focus, - .datepicker table tr td.active.highlighted.disabled.focus, - .datepicker table tr td.active[disabled].focus, - .datepicker table tr td.active.highlighted[disabled].focus, - fieldset[disabled] .datepicker table tr td.active.focus, - fieldset[disabled] .datepicker table tr td.active.highlighted.focus { - background-color: #337ab7; - border-color: #2e6da4; - } - .datepicker table tr td span { - display: block; - width: 23%; - height: 54px; - line-height: 54px; - float: left; - margin: 1%; - cursor: pointer; - border-radius: 4px; - } - .datepicker table tr td span:hover, - .datepicker table tr td span.focused { - background: #eeeeee; - } - .datepicker table tr td span.disabled, - .datepicker table tr td span.disabled:hover { - background: none; - color: #777777; - cursor: default; - } - .datepicker table tr td span.active, - .datepicker table tr td span.active:hover, - .datepicker table tr td span.active.disabled, - .datepicker table tr td span.active.disabled:hover { - color: #fff; - background-color: #337ab7; - border-color: #2e6da4; - text-shadow: 0 -1px 0 rgba(0, 0, 0, 0.25); - } - .datepicker table tr td span.active:focus, - .datepicker table tr td span.active:hover:focus, - .datepicker table tr td span.active.disabled:focus, - .datepicker table tr td span.active.disabled:hover:focus, - .datepicker table tr td span.active.focus, - .datepicker table tr td span.active:hover.focus, - .datepicker table tr td span.active.disabled.focus, - .datepicker table tr td span.active.disabled:hover.focus { - color: #fff; - background-color: #286090; - border-color: #122b40; - } - .datepicker table tr td span.active:hover, - .datepicker table tr td span.active:hover:hover, - .datepicker table tr td span.active.disabled:hover, - .datepicker table tr td span.active.disabled:hover:hover { - color: #fff; - background-color: #286090; - border-color: #204d74; - } - .datepicker table tr td span.active:active, - .datepicker table tr td span.active:hover:active, - .datepicker table tr td span.active.disabled:active, - .datepicker table tr td span.active.disabled:hover:active, - .datepicker table tr td span.active.active, - .datepicker table tr td span.active:hover.active, - .datepicker table tr td span.active.disabled.active, - .datepicker table tr td span.active.disabled:hover.active { - color: #fff; - background-color: #286090; - border-color: #204d74; - } - .datepicker table tr td span.active:active:hover, - .datepicker table tr td span.active:hover:active:hover, - .datepicker table tr td span.active.disabled:active:hover, - .datepicker table tr td span.active.disabled:hover:active:hover, - .datepicker table tr td span.active.active:hover, - .datepicker table tr td span.active:hover.active:hover, - .datepicker table tr td span.active.disabled.active:hover, - .datepicker table tr td span.active.disabled:hover.active:hover, - .datepicker table tr td span.active:active:focus, - .datepicker table tr td span.active:hover:active:focus, - .datepicker table tr td span.active.disabled:active:focus, - .datepicker table tr td span.active.disabled:hover:active:focus, - .datepicker table tr td span.active.active:focus, - .datepicker table tr td span.active:hover.active:focus, - .datepicker table tr td span.active.disabled.active:focus, - .datepicker table tr td span.active.disabled:hover.active:focus, - .datepicker table tr td span.active:active.focus, - .datepicker table tr td span.active:hover:active.focus, - .datepicker table tr td span.active.disabled:active.focus, - .datepicker table tr td span.active.disabled:hover:active.focus, - .datepicker table tr td span.active.active.focus, - .datepicker table tr td span.active:hover.active.focus, - .datepicker table tr td span.active.disabled.active.focus, - .datepicker table tr td span.active.disabled:hover.active.focus { - color: #fff; - background-color: #204d74; - border-color: #122b40; - } - .datepicker table tr td span.active.disabled:hover, - .datepicker table tr td span.active:hover.disabled:hover, - .datepicker table tr td span.active.disabled.disabled:hover, - .datepicker table tr td span.active.disabled:hover.disabled:hover, - .datepicker table tr td span.active[disabled]:hover, - .datepicker table tr td span.active:hover[disabled]:hover, - .datepicker table tr td span.active.disabled[disabled]:hover, - .datepicker table tr td span.active.disabled:hover[disabled]:hover, - fieldset[disabled] .datepicker table tr td span.active:hover, - fieldset[disabled] .datepicker table tr td span.active:hover:hover, - fieldset[disabled] .datepicker table tr td span.active.disabled:hover, - fieldset[disabled] .datepicker table tr td span.active.disabled:hover:hover, - .datepicker table tr td span.active.disabled:focus, - .datepicker table tr td span.active:hover.disabled:focus, - .datepicker table tr td span.active.disabled.disabled:focus, - .datepicker table tr td span.active.disabled:hover.disabled:focus, - .datepicker table tr td span.active[disabled]:focus, - .datepicker table tr td span.active:hover[disabled]:focus, - .datepicker table tr td span.active.disabled[disabled]:focus, - .datepicker table tr td span.active.disabled:hover[disabled]:focus, - fieldset[disabled] .datepicker table tr td span.active:focus, - fieldset[disabled] .datepicker table tr td span.active:hover:focus, - fieldset[disabled] .datepicker table tr td span.active.disabled:focus, - fieldset[disabled] .datepicker table tr td span.active.disabled:hover:focus, - .datepicker table tr td span.active.disabled.focus, - .datepicker table tr td span.active:hover.disabled.focus, - .datepicker table tr td span.active.disabled.disabled.focus, - .datepicker table tr td span.active.disabled:hover.disabled.focus, - .datepicker table tr td span.active[disabled].focus, - .datepicker table tr td span.active:hover[disabled].focus, - .datepicker table tr td span.active.disabled[disabled].focus, - .datepicker table tr td span.active.disabled:hover[disabled].focus, - fieldset[disabled] .datepicker table tr td span.active.focus, - fieldset[disabled] .datepicker table tr td span.active:hover.focus, - fieldset[disabled] .datepicker table tr td span.active.disabled.focus, - fieldset[disabled] .datepicker table tr td span.active.disabled:hover.focus { - background-color: #337ab7; - border-color: #2e6da4; - } - .datepicker table tr td span.old, - .datepicker table tr td span.new { - color: #777777; - } - .datepicker .datepicker-switch { - width: 145px; - } - .datepicker .datepicker-switch, - .datepicker .prev, - .datepicker .next, - .datepicker tfoot tr th { - cursor: pointer; - } - .datepicker .datepicker-switch:hover, - .datepicker .prev:hover, - .datepicker .next:hover, - .datepicker tfoot tr th:hover { - background: #eeeeee; - } - .datepicker .prev.disabled, - .datepicker .next.disabled { - visibility: hidden; - } - .datepicker .cw { - font-size: 10px; - width: 12px; - padding: 0 2px 0 5px; - vertical-align: middle; - } - .input-group.date .input-group-addon { - cursor: pointer; - } - .input-daterange { - width: 100%; - } - .input-daterange input { - text-align: center; - } - .input-daterange input:first-child { - border-radius: 3px 0 0 3px; - } - .input-daterange input:last-child { - border-radius: 0 3px 3px 0; - } - .input-daterange .input-group-addon { - width: auto; - min-width: 16px; - padding: 4px 5px; - line-height: 1.42857143; - border-width: 1px 0; - margin-left: -5px; - margin-right: -5px; - } - /*# sourceMappingURL=bootstrap-datepicker3.css.map */ +// FIXME: color.scale doesn't seem to work with CSS variables, so avoid those:` +$dp-cell-focus-background-color: $dropdown-link-hover-bg !default; + +@import "vanillajs-datepicker/sass/datepicker-bs5"; + +[data-bs-theme="dark"] .datepicker-picker { + .datepicker-header, + .datepicker-controls .btn, + .datepicker-main, + .datepicker-footer { + background-color: $gray-800; + } + + .datepicker-cell:hover { + background-color: $gray-700; + } + + .datepicker-cell.day.focused { + background-color: $gray-600; + } + + .datepicker-cell.day.selected.focused { + background-color: $blue; + } + + .datepicker-controls .btn:hover { + background-color:$gray-700; + color: $gray-400; + } +} diff --git a/ietf/static/css/document_html.scss b/ietf/static/css/document_html.scss index 42e0af1a64..47ef8d64b4 100644 --- a/ietf/static/css/document_html.scss +++ b/ietf/static/css/document_html.scss @@ -1,19 +1,9 @@ @use "sass:map"; -// FIXME: It's not clear why these three variables remain unset by bs5, but just -// set them to placeholder values so the CSS embedded in the HTML validates. -$btn-font-family: inherit !default; -$nav-link-font-weight: inherit !default; -$tooltip-margin: inherit !default; - -@import "bootstrap/scss/functions"; -@import "bootstrap/scss/variables"; -@import "bootstrap/scss/maps"; -@import "bootstrap/scss/mixins"; -@import "bootstrap/scss/utilities"; -@import "bootstrap/scss/root"; +@import "custom-bs-import"; // Layout & components +// Only import what we need: @import "bootstrap/scss/reboot"; @import "bootstrap/scss/type"; // @import "bootstrap/scss/images"; @@ -23,7 +13,7 @@ $tooltip-margin: inherit !default; @import "bootstrap/scss/forms"; @import "bootstrap/scss/buttons"; @import "bootstrap/scss/transitions"; -// @import "bootstrap/scss/dropdown"; +@import "bootstrap/scss/dropdown"; @import "bootstrap/scss/button-group"; @import "bootstrap/scss/nav"; @import "bootstrap/scss/navbar"; @@ -63,12 +53,42 @@ $tooltip-margin: inherit !default; scrollbar-width: none; } -.sidebar-toggle[aria-expanded="true"] { +.sidebar-toggle[aria-expanded="true"] .sidebar-shown { display: none; } -.sidebar-toggle[aria-expanded="false"] { - display: inherit; +.sidebar-toggle[aria-expanded="false"] .sidebar-collapsed { + display: none; +} + +.sidebar-toolbar { + z-index: 1; +} + +// Toggle classes for dark/light modes +[data-bs-theme="dark"] { + .d-dm-none { + display: none; + } + + .d-lm-none { + display: initial; + } +} + +[data-bs-theme="light"] { + .d-dm-none { + display: initial; + } + + .d-lm-none { + display: none; + } +} + +// Show theme toggler checkbox +.dropdown-menu .active .bi { + display: block !important; } @media screen { @@ -108,28 +128,15 @@ $tooltip-margin: inherit !default; } @media screen { - @include media-breakpoint-only(xs) { - font-size: min(7pt, var(--doc-ptsize-max)); - } - @include media-breakpoint-up(sm) { - font-size: min(9.5pt, var(--doc-ptsize-max)); + // the viewport-width ("vw") constants are magic; they seem to work for + // many monospace fonts, but may need tweaking + @include media-breakpoint-up(xs) { + font-size: min(2.2vw, var(--doc-ptsize-max)); } @include media-breakpoint-up(md) { - font-size: min(9.5pt, var(--doc-ptsize-max)); - } - - @include media-breakpoint-up(lg) { - font-size: min(11pt, var(--doc-ptsize-max)); - } - - @include media-breakpoint-up(xl) { - font-size: min(13pt, var(--doc-ptsize-max)); - } - - @include media-breakpoint-up(xxl) { - font-size: min(16pt, var(--doc-ptsize-max)); + font-size: min(1.6vw, var(--doc-ptsize-max)); } .grey, @@ -151,6 +158,7 @@ $tooltip-margin: inherit !default; pre, code { font-size: 1em; + overflow: visible; } pre { @@ -166,6 +174,27 @@ $tooltip-margin: inherit !default; .rfcmarkup { + // A lot of plaintext documents seem to have line lengths >72ch. + // To handle that, we calculate with 80ch here and adjust some of the + // font sizes down accordingly. + pre { + width: 80ch; + white-space: pre-wrap; + } + + @media screen { + + // the viewport-width ("vw") constants are magic; they seem to work for + // many monospace fonts, but may need tweaking + @include media-breakpoint-up(xs) { + font-size: min(2vw, var(--doc-ptsize-max)); + } + + @include media-breakpoint-up(md) { + font-size: min(1.5vw, var(--doc-ptsize-max)); + } + } + h1, h2, h3, @@ -291,6 +320,11 @@ tbody.meta tr { background-color: $danger; } +.badge-generic { + color: white; + background-color: $danger; +} + #toc-nav { width: inherit; overscroll-behavior-y: none; // Prevent overscrolling from scrolling the main content @@ -322,11 +356,13 @@ tbody.meta tr { page-break-inside: avoid; } + /* a:link, a:visited { // color: inherit; // text-decoration: none; } +*/ .newpage { page-break-before: always !important; diff --git a/ietf/static/css/document_html_txt.scss b/ietf/static/css/document_html_txt.scss index 7016dba3d4..a5991056c9 100644 --- a/ietf/static/css/document_html_txt.scss +++ b/ietf/static/css/document_html_txt.scss @@ -1,5 +1,7 @@ +// Based on https://github.com/martinthomson/rfc-txt-html/blob/main/txt.css + :root { - --line: 1.2em; + --line: 1.3em; --block: 0 0 0 3ch; --paragraph: var(--line) 0 var(--line) 3ch; } @@ -36,38 +38,38 @@ // font-size: 1em; line-height: var(--line); width: 72ch; - // margin: var(--line) 3ch; margin-top: var(--line); margin-right: 3ch; margin-bottom: var(--line); margin-left: 3ch; + h1, h2, h3, h4, h5 { font-weight: bold; font-size: inherit; line-height: inherit; - // margin: var(--line) 0; - @include margin-line; + @include margin-line; // margin: var(--line) 0; } .section-number { margin-right: 1ch; } p { - // margin: var(--paragraph); - @include margin-paragraph; + margin: 0; +} +section > p, section > div > p { + @include margin-paragraph; // margin: var(--paragraph); } -aside, ol, dl { - // margin: var(--block); - @include margin-block; +aside { + @include margin-block; // margin: var(--block); } figure { margin: 0; } blockquote { - // margin: var(--paragraph); - @include margin-paragraph; - border-left: 2px solid darkgrey; + @include margin-paragraph; // margin: var(--paragraph); + padding-left: calc(2ch - 2px); + border-left: 2px solid var(--bs-border-color); } /* Header junk */ @@ -90,6 +92,7 @@ blockquote { } #identifiers dd { margin: 0; + margin-left: 0 !important; /* grr */ width: 47ch; /* HAXX: this gets around the lack of text-content-trim support */ display: flex; @@ -122,7 +125,6 @@ blockquote { } #identifiers dd.obsoletes { padding-left: 10ch; - width: 37ch; } #identifiers dd.updates::before { content: "Updates:"; @@ -130,7 +132,6 @@ blockquote { } #identifiers dd.updates { padding-left: 8ch; - width: 39ch; } #identifiers dd:is(.updates, .obsoletes) a { margin: 0 0 0 1ch; @@ -164,6 +165,7 @@ blockquote { #identifiers dd.authors .author { display: inline-block; margin: 0 2ch 0 1ch; + min-height: calc(2 * var(--line)); } #identifiers dd.authors .author:last-of-type { margin-right: 0; @@ -201,7 +203,7 @@ blockquote { #title { clear: left; text-align: center; - margin-top: 2em; + margin-top: calc(2 * var(--line)); } #rfcnum { display: none; @@ -239,59 +241,88 @@ ol { } ul { margin: 0 0 0 4ch; +} +ul.ulEmpty { + list-style: none; +} +ul:not(.ulEmpty) { list-style-type: '*'; } -ul ul { +ul ul:not(.ulEmpty) { list-style-type: '-'; } :is(ul, ol) ul { margin-left: 1ch; } -ul ul ul { +ul ul ul:not(.ulEmpty) { list-style-type: 'o'; } li { - // margin: var(--line) 0; - @include margin-line; + @include margin-line; // margin: var(--line) 0; padding: 0 0 0 2ch; } -.compact li { +dl { margin: 0; } -:is(li, dd) p:first-child { - margin: 0; +section > dl, section > div > dl { + @include margin-block; // margin: var(--block); +} +dl, dt { + clear: left; } dt { float: left; - clear: left; + font-weight: bold; margin: 0 2ch 0 0; break-after: avoid; } dd { - // margin: var(--paragraph); - @include margin-paragraph; + @include margin-paragraph; // margin: var(--paragraph); + margin-left: 3ch !important; /* override attribute added by xml2rfc */ + padding: 0; break-before: avoid; } -dl.compact :is(dt, dd) { - margin-top: 0; +dl.dlNewline > dd { + clear: left; +} +dl.olPercent > dt { + min-width: 4ch; +} +dl.olPercent > dd { + margin-left: 6ch !important; /* as above */ +} +dl > dd > dl { + margin-top: var(--line); margin-bottom: 0; } dl.references dt { margin-right: 1ch; } dl.references dd { - margin-left: 11ch; + margin-left: 11ch !important; /* grr */ } :is(dd, span).break { display: none; } +:is(li, dd, blockquote, aside) :is(p, ol, ul:not(.toc), dl):not(:first-child) { + margin-top: var(--line); +} +:is(li, dd, blockquote, aside) .break:first-child + :is(p, ol, ul, dl) { + margin-top: 0; +} +:is(ol:is(.compact, .olCompact), ul:is(.compact, .ulCompact)) > li, +:is(dl.compact, .dlCompact) > :is(dt, dd) { + margin-top: 0; + margin-bottom: 0; +} /* Figures, tables */ pre { - // margin: var(--line) 0; - @include margin-line; + @include margin-line; // margin: var(--line) 0; } div:is(.artwork, .sourcecode) { + margin-top: var(--line); + margin-bottom: var(--line); display: flex; flex-wrap: nowrap; align-items: end; @@ -313,7 +344,7 @@ div:is(.artwork, .sourcecode) pre { flex: 0 0 content; margin: 0; max-width: 72ch; - overflow: auto; + overflow: auto clip; } div:is(.artwork, .sourcecode) .pilcrow { flex: 0 0 1ch; @@ -344,7 +375,7 @@ thead, tfoot { border-bottom-style: double; } td, th { - border: 1px solid black; + border: 1px solid var(--bs-border-color); // padding: var(--half-line) 1ch; padding-top: var(--half-line); padding-right: 1ch; @@ -362,8 +393,8 @@ td, th { } /* Links */ -a.selfRef, a.pilcrow { - color: black; +a.selfRef, a.pilcrow, .iref + a.internal { + color: inherit; text-decoration: none; } a.relref, a.xref { @@ -392,17 +423,14 @@ sup, sub { } /* Authors */ -address { +address, address.vcard { font-style: normal; - // margin: 2em 0 var(--line) 3ch; - margin-top: 2em; + // margin: var(--line) 0 var(--line) 3ch + margin-top: var(--line); margin-right: 0; margin-bottom: var(--line); margin-left: 3ch; } -h2 + address { - margin-top: 1em; -} address .tel, address .email { // margin: var(--line) 0 0; margin-top: var(--line); @@ -413,4 +441,28 @@ address .tel, address .email { address .tel + .email { margin: 0; } + +/* haxx */ +section > p, section > dl.references > dd { + /* Really long lines can wrap when all else fails. + * This won't affect
       or , or cases where soft-wrapping occurs.
      +   * Mostly this exists so that long URLs wrap properly in Safari, which
      +   * doesn't break words at '/' like other browsers. */
      +  overflow-wrap: break-word;
      +}
      +
      +/* From https://github.com/martinthomson/rfc-css/blob/main/rfc.css */
      +/* SVG Trick: a prefix match works because only black and white are allowed */
      +svg :is([stroke="black"], [stroke^="#000"]) {
      +    stroke: var(--bs-body-color);
      +}
      +svg :is([stroke="white"], [stroke^="#fff"]) {
      +    stroke: var(--bs-body-bg);
      +}
      +svg :is([fill="black"], [fill^="#000"], :not([fill])) {
      +    fill: var(--bs-body-color);
      +}
      +svg :is([fill="white"], [fill^="#fff"]) {
      +    fill: var(--bs-body-bg);
      +}
       }
      diff --git a/ietf/static/css/edit-meeting-schedule.scss b/ietf/static/css/edit-meeting-schedule.scss
      deleted file mode 100644
      index e69de29bb2..0000000000
      diff --git a/ietf/static/css/highcharts.scss b/ietf/static/css/highcharts.scss
      new file mode 100644
      index 0000000000..d2f5d5e0e7
      --- /dev/null
      +++ b/ietf/static/css/highcharts.scss
      @@ -0,0 +1,6 @@
      +@import "npm:highcharts/css/highcharts.css";
      +@import "custom-bs-import";
      +
      +.highcharts-container {
      +    font-family: $font-family-sans-serif;
      +}
      diff --git a/ietf/static/css/ietf.scss b/ietf/static/css/ietf.scss
      index 9cfb610b79..6695c57b13 100644
      --- a/ietf/static/css/ietf.scss
      +++ b/ietf/static/css/ietf.scss
      @@ -1,32 +1,9 @@
       @use "sass:map";
       
      -@import "bootstrap/scss/functions";
      -
      -// Enable negative margin classes.
      -$enable-negative-margins: true;
      -
      -// Don't add carets to dropdowns by default.
      -// $enable-caret: false;
      -
      -$popover-max-width: 100%;
      -
      -// Only import what we need:
      -
      -@import "bootstrap/scss/variables";
      -
      -$h1-font-size: $font-size-base * 2.2;
      -$h2-font-size: $font-size-base * 1.8;
      -$h3-font-size: $font-size-base * 1.6;
      -$h4-font-size: $font-size-base * 1.4;
      -$h5-font-size: $font-size-base * 1.2;
      -$h6-font-size: $font-size-base;
      -
      -@import "bootstrap/scss/maps";
      -@import "bootstrap/scss/mixins";
      -@import "bootstrap/scss/utilities";
      -@import "bootstrap/scss/root";
      +@import "custom-bs-import";
       
       // Layout & components
      +// Only import what we need:
       @import "bootstrap/scss/reboot";
       @import "bootstrap/scss/type";
       @import "bootstrap/scss/images";
      @@ -70,7 +47,7 @@ url("npm:bootstrap-icons/font/fonts/bootstrap-icons.woff") format("woff");
       @import "bootstrap-icons/font/bootstrap-icons";
       
       // Leave room for fixed-top navbar...
      -body {
      +body.navbar-offset {
           padding-top: 60px;
       }
       
      @@ -79,6 +56,55 @@ html {
           scroll-padding-top: 60px;
       }
       
      +// Toggle classes for dark/light modes
      +[data-bs-theme="dark"] {
      +    .d-dm-none {
      +        display: none;
      +    }
      +
      +    .d-lm-none {
      +        display: initial;
      +    }
      +}
      +
      +[data-bs-theme="light"] {
      +    .d-dm-none {
      +        display: initial;
      +    }
      +
      +    .d-lm-none {
      +        display: none;
      +    }
      +}
      +
      +// Make submenus open on hover.
      +@include media-breakpoint-up(lg) {
      +    .dropdown-menu>li>ul {
      +        display: none;
      +    }
      +
      +    .dropdown-menu>li:hover>ul {
      +        display: block;
      +    }
      +
      +}
      +
      +@include media-breakpoint-up(md) {
      +    .leftmenu .nav>li>ul {
      +        display: none;
      +    }
      +
      +    .leftmenu .nav>li:hover>ul {
      +        display: block;
      +    }
      +}
      +
      +:is(.dropdown-menu, .leftmenu .nav) .dropdown-menu {
      +    top: 0;
      +    left: 100%;
      +    right: auto;
      +}
      +
       // Make textareas in forms use a monospace font
       textarea.form-control {
           font-family: $font-family-code;
      @@ -97,6 +123,11 @@ pre {
           width: 60px;
       }
       
      +// Style preformatted alert messages better.
      +.preformatted {
      +    white-space: pre-line;
      +}
      +
       .leftmenu {
           width: 13em;
       
      @@ -122,8 +153,8 @@ pre {
               --#{$prefix}dropdown-divider-bg: #{$dropdown-divider-bg};
               --#{$prefix}dropdown-divider-margin-y: #{$dropdown-divider-margin-y};
               --#{$prefix}dropdown-box-shadow: #{$dropdown-box-shadow};
      -        --#{$prefix}dropdown-link-color: #{$dropdown-link-color};
      -        --#{$prefix}dropdown-link-hover-color: #{$dropdown-link-hover-color};
      +        --#{$prefix}dropdown-link-color: #{$nav-link-color};
      +        --#{$prefix}dropdown-link-hover-color: #{$nav-link-hover-color};
               --#{$prefix}dropdown-link-hover-bg: #{$dropdown-link-hover-bg};
               --#{$prefix}dropdown-link-active-color: #{$dropdown-link-active-color};
               --#{$prefix}dropdown-link-active-bg: #{$dropdown-link-active-bg};
      @@ -198,16 +229,20 @@ th,
       .group-menu .dropdown-menu {
           height: auto;
           width: auto;
      -    max-height: 35em;
      +    max-height: 95vh;
           overflow-x: hidden;
           overflow-y: auto;
       }
       
       // Helper to constrain the size of the main logo
       .ietflogo {
      -    width: 75%;
      +    width: 100%;
           max-width: 300px;
       }
      +.ietflogo > img {
      +    min-width: 100px;
      +    width: 100%;
      +}
       
       // Make revision numbers pagination items fixed-width
       .revision-list {
      @@ -220,6 +255,12 @@ th,
           }
       }
       
      +.charter.revision-list {
      +    .page-item {
      +        width: auto;
      +    }
      +}
      +
       // Style the photo cards
       .photo {
           width: 12em;
      @@ -256,13 +297,13 @@ th,
       }
       
       // Styles for d3.js graphical SVG timelines
      -#timeline {
      +#doc-timeline {
           font-size: small;
       
           .axis path,
           .axis line {
               fill: none;
      -        stroke: black;
      +        stroke: var(--bs-body-color);
           }
       
           .axis.y path,
      @@ -279,7 +320,7 @@ th,
           }
       
           .bar text {
      -        fill: black;
      +        fill: var(--bs-body-color);
               dominant-baseline: central;
               pointer-events: none;
           }
      @@ -318,7 +359,7 @@ th,
       }
       
       .ballot-icon table .my {
      -    border: 2 * $table-border-width solid #000;
      +    border: calc(2 * $table-border-width) solid var(--bs-emphasis-color);
       }
       
       // See https://getbootstrap.com/docs/5.1/customize/color/#all-colors
      @@ -430,36 +471,64 @@ td.position-recuse {
       }
       
       td.position-norecord {
      -    background-color: $white; // $color-norecord;
      +    background-color: transparent;
       }
       
       td.position-empty {
           border: none !important;
       }
       
      -tr.position-moretime-row,
      -tr.position-notready-row,
      -tr.position-discuss-row,
      -tr.position-block-row {
      -    background-color: tint-color($color-discuss, 85%);
      -}
      +[data-bs-theme="light"] {
       
      -tr.position-yes-row {
      -    background-color: tint-color($color-yes, 75%);
      -}
      +    tr.position-moretime-row,
      +    tr.position-notready-row,
      +    tr.position-discuss-row,
      +    tr.position-block-row {
      +        background-color: tint-color($color-discuss, 85%);
      +    }
       
      -tr.position-noobj-row {
      -    background-color: tint-color($color-noobj, 50%);
      -}
      +    tr.position-yes-row {
      +        background-color: tint-color($color-yes, 75%);
      +    }
       
      -tr.position-abstain-row {
      -    background-color: tint-color($color-abstain, 85%);
      -}
      +    tr.position-noobj-row {
      +        background-color: tint-color($color-noobj, 50%);
      +    }
       
      -tr.position-recuse-row {
      -    background-color: tint-color($color-recuse, 85%);
      +    tr.position-abstain-row {
      +        background-color: tint-color($color-abstain, 85%);
      +    }
      +
      +    tr.position-recuse-row {
      +        background-color: tint-color($color-recuse, 85%);
      +    }
       }
       
      +[data-bs-theme="dark"] {
      +
      +    tr.position-moretime-row,
      +    tr.position-notready-row,
      +    tr.position-discuss-row,
      +    tr.position-block-row {
      +        background-color: shade-color($color-discuss, 65%);
      +    }
      +
      +    tr.position-yes-row {
      +        background-color: shade-color($color-yes, 65%);
      +    }
      +
      +    tr.position-noobj-row {
      +        background-color: shade-color($color-noobj, 65%);
      +    }
      +
      +    tr.position-abstain-row {
      +        background-color: shade-color($color-abstain, 65%);
      +    }
      +
      +    tr.position-recuse-row {
      +        background-color: shade-color($color-recuse, 65%);
      +    }
      +}
       
       /* === Edit Meeting Schedule ====================================== */
       
      @@ -497,7 +566,7 @@ tr.position-recuse-row {
       }
       
       .edit-meeting-schedule .edit-grid .day-label .swap-days:hover {
      -    color: #666;
      +    color: var(--bs-secondary-color);
       }
       
       .edit-meeting-schedule #swap-days-modal .modal-body label {
      @@ -529,15 +598,15 @@ tr.position-recuse-row {
       }
       
       .edit-meeting-schedule .edit-grid .time-header .time-label.would-violate-hint {
      -    background-color: #ffe0e0;
      -    outline: #ffe0e0 solid 0.4em;
      +    background-color: var(--bs-danger-bg-subtle);
      +    outline: var(--bs-danger-bg-subtle) solid 0.4em;
       }
       
       .edit-meeting-schedule .edit-grid .time-header .time-label span {
           display: inline-block;
           width: 100%;
           text-align: center;
      -    color: #444444;
      +    color: var(--bs-secondary-color);
       }
       
       .edit-meeting-schedule .edit-grid .timeslots {
      @@ -549,7 +618,7 @@ tr.position-recuse-row {
       .edit-meeting-schedule .edit-grid .timeslot {
           position: relative;
           display: inline-block;
      -    background-color: #f4f4f4;
      +    background-color: var(--bs-secondary-bg);
           height: 100%;
           overflow: hidden;
       }
      @@ -562,7 +631,7 @@ tr.position-recuse-row {
           width: 100%;
           align-items: center;
           justify-content: center;
      -    color: #999;
      +    color: var(--bs-tertiary-color);
       }
       
       .edit-meeting-schedule .edit-grid .timeslot .drop-target {
      @@ -579,22 +648,22 @@ tr.position-recuse-row {
       }
       
       .edit-meeting-schedule .edit-grid .timeslot.overfull {
      -    border-right: 0.3em dashed #f55000;
      +    border-right: 0.3em dashed var(--bs-danger);
           /* cut-off illusion */
       }
       
       .edit-meeting-schedule .edit-grid .timeslot.would-violate-hint {
      -    background-color: #ffe0e0;
      -    outline: #ffe0e0 solid 0.4em;
      +    background-color: var(--bs-danger-bg-subtle);
      +    outline: var(--bs-danger-bg-subtle) solid 0.4em;
       }
       
       .edit-meeting-schedule .edit-grid .timeslot.would-violate-hint.dropping {
      -    background-color: #ccb3b3;
      +    background-color: var(--bs-danger);
       }
       
       .edit-meeting-schedule .constraints .encircled,
       .edit-meeting-schedule .formatted-constraints .encircled {
      -    border: 1px solid #000;
      +    border: 1px solid var( --bs-body-color);
           border-radius: 1em;
           padding: 0 0.3em;
           text-align: center;
      @@ -607,7 +676,7 @@ tr.position-recuse-row {
       
       /* sessions */
       .edit-meeting-schedule .session {
      -    background-color: #fff;
      +    background-color: var(--bs-body-bg);
           margin: 0.2em;
           padding-right: 0.2em;
           padding-left: 0.5em;
      @@ -619,15 +688,15 @@ tr.position-recuse-row {
       
       .edit-meeting-schedule .session.selected {
           cursor: grabbing;
      -    outline: #0000ff solid 0.2em;
      -    /* blue, width matches margin on .session */
      +    outline: var(--bs-primary) solid 0.2em;
      +    /* width matches margin on .session */
           z-index: 2;
           /* render above timeslot outlines */
       }
       
       .edit-meeting-schedule .session.other-session-selected {
      -    outline: #00008b solid 0.2em;
      -    /* darkblue, width matches margin on .session */
      +    outline: 0.3em solid var(--bs-info);
      +    box-shadow: 0 0 1em var(--bs-info);
           z-index: 2;
           /* render above timeslot outlines */
       }
      @@ -638,7 +707,7 @@ tr.position-recuse-row {
       
       .edit-meeting-schedule .session.readonly {
           cursor: default;
      -    background-color: #ddd;
      +    background-color: var(--bs-dark-bg-subtle);
       }
       
       .edit-meeting-schedule .session.hidden-parent * {
      @@ -652,13 +721,12 @@ tr.position-recuse-row {
       }
       
       .edit-meeting-schedule .session.highlight {
      -    outline-color: #ff8c00;
      -    /* darkorange */
      -    background-color: #f3f3f3;
      +    outline-color: var(--bs-warning);
      +    background-color: var(--bs-light);
       }
       
       .edit-meeting-schedule .session.would-violate-hint {
      -    outline: 0.3em solid #F55000;
      +    outline: 0.3em solid var(--bs-danger);
           z-index: 1;
           /* raise up so the outline is not overdrawn */
       }
      @@ -680,6 +748,7 @@ tr.position-recuse-row {
       
       .edit-meeting-schedule .edit-grid,
       .edit-meeting-schedule .session {
      +    // Removing this font-family style causes selenium tests to fail :-(
           font-family: arial, helvetica, sans-serif;
           font-size: 11px;
       }
      @@ -751,9 +820,9 @@ tr.position-recuse-row {
           bottom: 0;
           left: 0;
           width: 100%;
      -    border-top: 0.2em solid #ccc;
      +    border-top: 0.2em solid var(--bs-border-color);
           margin-bottom: 2em;
      -    background-color: #fff;
      +    background-color: var(--bs-body-bg);
           opacity: 0.95;
           z-index: 5;
           /* raise above edit-grid items */
      @@ -768,7 +837,7 @@ tr.position-recuse-row {
           min-height: 4em;
           max-height: 13em;
           overflow-y: auto;
      -    background-color: #f4f4f4;
      +    background-color: var(--bs-secondary-bg);
       }
       
       .edit-meeting-schedule .unassigned-sessions.dropping {
      @@ -825,7 +894,7 @@ tr.position-recuse-row {
           font-weight: normal;
           margin-right: 1em;
           padding: 0 1em;
      -    border: 0.1em solid #eee;
      +    border: 0.1em solid var(--bs-border-color);
           cursor: pointer;
       }
       
      @@ -891,7 +960,7 @@ tr.position-recuse-row {
       }
       
       .edit-meeting-timeslots-and-misc-sessions .room-row {
      -    border-bottom: 1px solid #ccc;
      +    border-bottom: 1px solid var(--bs-border-color);
           // height: 20px;
           display: flex;
           cursor: pointer;
      @@ -911,13 +980,13 @@ tr.position-recuse-row {
       }
       
       .edit-meeting-timeslots-and-misc-sessions .timeline.hover {
      -    background: radial-gradient(#999 1px, transparent 1px);
      +    background: radial-gradient(var(--bs-tertiary-color) 1px, transparent 1px);
           background-size: 20px 20px;
       }
       
       .edit-meeting-timeslots-and-misc-sessions .timeline.selected.hover,
       .edit-meeting-timeslots-and-misc-sessions .timeline.selected {
      -    background: radial-gradient(#999 2px, transparent 2px);
      +    background: radial-gradient(var(--bs-tertiary-color) 2px, transparent 2px);
           background-size: 20px 20px;
       }
       
      @@ -933,8 +1002,8 @@ tr.position-recuse-row {
           white-space: nowrap;
           cursor: pointer;
           padding-left: 0.2em;
      -    border-left: 1px solid #999;
      -    border-right: 1px solid #999;
      +    border-left: 1px solid var(--bs-border-color);
      +    border-right: 1px solid var(--bs-border-color);
       }
       
       .edit-meeting-timeslots-and-misc-sessions .timeslot:hover {
      @@ -954,10 +1023,10 @@ tr.position-recuse-row {
           bottom: 0;
           left: 0;
           width: 100%;
      -    border-top: 0.2em solid #ccc;
      +    border-top: 0.2em solid var(--bs-border-color);
           padding-top: 0.2em;
           margin-bottom: 2em;
      -    background-color: #fff;
      +    background-color: var(--bs-body-bg);
           opacity: 0.95;
       }
       
      @@ -1005,7 +1074,7 @@ tr.position-recuse-row {
       }
       
       .timeslot-edit .tstable div.timeslot {
      -    border: #000000 solid 1px;
      +    border: var(--bs-body-color) solid 1px;
           border-radius: 0.5em;
           padding: 0.5em;
       }
      @@ -1035,7 +1104,7 @@ tr.position-recuse-row {
       }
       
       .timeslot-edit .tstable .tstype_unavail {
      -    background-color: #666;
      +    background-color: var(--bs-secondary-color);
       }
       
       .timeslot-edit .official-use-warning {
      @@ -1118,3 +1187,49 @@ tr.position-recuse-row {
               }
           }
       }
      +
      +blockquote {
      +    padding-left: 1rem;
      +    border-left: solid 1px var(--bs-body-color);
      +}
      +
      +iframe.status {
      +    background-color:transparent;
      +    border:none;
      +    width:100%;
      +    height:3.5em;
      +}
      +
      +.overflow-shadows {
      +    transition: box-shadow 0.5s;
      +}
      +
      +.overflow-shadows--both {
      +    box-shadow: inset 0px 21px 18px -20px var(--bs-body-color),
      +                inset 0px -21px 18px -20px var(--bs-body-color);
      +}
      +
      +.overflow-shadows--top-only {
      +    box-shadow: inset 0px 21px 18px -20px var(--bs-body-color);
      +}
      +
      +.overflow-shadows--bottom-only {
      +    box-shadow: inset 0px -21px 18px -20px var(--bs-body-color);
      +}
      +
      +#navbar-doc-search-wrapper {
      +    position: relative;
      +}
      +
      +#navbar-doc-search-results {
      +    max-height: 400px;
      +    overflow-y: auto;
      +    min-width: auto;
      +    left: 0;
      +    right: 0;
      +
      +    .dropdown-item {
      +        white-space: normal;
      +        overflow-wrap: break-word;
      +    }
      +}
      diff --git a/ietf/static/css/list.scss b/ietf/static/css/list.scss
      index d52fc879a7..595bf360d5 100644
      --- a/ietf/static/css/list.scss
      +++ b/ietf/static/css/list.scss
      @@ -1,6 +1,4 @@
      -// Import bootstrap helpers
      -@import "bootstrap/scss/functions";
      -@import "bootstrap/scss/variables";
      +@import "custom-bs-import";
       
       table .sort {
           cursor: pointer;
      diff --git a/ietf/static/css/select2.scss b/ietf/static/css/select2.scss
      index 44824a358a..c8e3da7adc 100644
      --- a/ietf/static/css/select2.scss
      +++ b/ietf/static/css/select2.scss
      @@ -1,5 +1,7 @@
      -@import "bootstrap/scss/functions";
      -@import "bootstrap/scss/variables";
      -@import "bootstrap/scss/mixins";
      +@import "custom-bs-import";
      +
      +// FIXME: bs-5.3.0 workaround from https://github.com/apalfrey/select2-bootstrap-5-theme/issues/75#issuecomment-1573265695
      +$s2bs5-border-color: $border-color;
      +
       @import "select2/src/scss/core";
       @import "select2-bootstrap-5-theme/src/include-all";
      diff --git a/ietf/static/images/arrow-ani.webp b/ietf/static/images/arrow-ani.webp
      deleted file mode 100644
      index 379407db15..0000000000
      Binary files a/ietf/static/images/arrow-ani.webp and /dev/null differ
      diff --git a/ietf/static/images/iab-logo-white.svg b/ietf/static/images/iab-logo-white.svg
      new file mode 100644
      index 0000000000..264b7bb842
      --- /dev/null
      +++ b/ietf/static/images/iab-logo-white.svg
      @@ -0,0 +1,64 @@
      +
      +
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +
      diff --git a/ietf/static/images/ietf-logo-nor-white.svg b/ietf/static/images/ietf-logo-nor-white.svg
      index 004e58af80..42c033600f 100644
      --- a/ietf/static/images/ietf-logo-nor-white.svg
      +++ b/ietf/static/images/ietf-logo-nor-white.svg
      @@ -1,26 +1,136 @@
      -
      -	
      -		
      -	
      -	
      -	
      -	
      -	
      -	
      -	
      -	
      -	
      -	
      -	
      -	
      -	
      -	
      -	
      -	
      -	
      -	
      -	
      -	
      -	
      +
      +
      +  
      +  
      +    
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
       
      diff --git a/ietf/static/images/ietf-logo-white.svg b/ietf/static/images/ietf-logo-white.svg
      new file mode 100644
      index 0000000000..2417f917ce
      --- /dev/null
      +++ b/ietf/static/images/ietf-logo-white.svg
      @@ -0,0 +1,146 @@
      +
      +
      +  
      +  
      +    
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +  
      +
      diff --git a/ietf/static/images/irtf-logo-white.svg b/ietf/static/images/irtf-logo-white.svg
      new file mode 100644
      index 0000000000..a67412581e
      --- /dev/null
      +++ b/ietf/static/images/irtf-logo-white.svg
      @@ -0,0 +1,65 @@
      +
      +
      +  
      +  
      +  R
      +  
      +  
      +  
      +  
      +  
      +
      diff --git a/ietf/static/images/irtf-logo.svg b/ietf/static/images/irtf-logo.svg
      index be64890b25..10b2a96816 100644
      --- a/ietf/static/images/irtf-logo.svg
      +++ b/ietf/static/images/irtf-logo.svg
      @@ -6,7 +6,7 @@
          version="1.1"
          id="svg303"
          sodipodi:docname="irtf-logo.svg"
      -   inkscape:version="1.2.1 (9c6d41e4, 2022-07-14)"
      +   inkscape:version="1.2.2 (b0a84865, 2022-12-01)"
          xmlns:inkscape="http://www.inkscape.org/namespaces/inkscape"
          xmlns:sodipodi="http://sodipodi.sourceforge.net/DTD/sodipodi-0.dtd"
          xmlns="http://www.w3.org/2000/svg"
      @@ -24,12 +24,12 @@
            inkscape:deskcolor="#d1d1d1"
            showgrid="false"
            inkscape:zoom="1.0373737"
      -     inkscape:cx="304.1334"
      -     inkscape:cy="446.80136"
      +     inkscape:cx="303.65142"
      +     inkscape:cy="447.76535"
            inkscape:window-width="1797"
            inkscape:window-height="1083"
            inkscape:window-x="951"
      -     inkscape:window-y="445"
      +     inkscape:window-y="417"
            inkscape:window-maximized="0"
            inkscape:current-layer="svg303" />
         
           R
            {
      +    const form = document.getElementById('delete_recordings_form')
      +    const dialog = document.getElementById('delete_confirm_dialog')
      +    const dialog_link = document.getElementById('delete_confirm_link')
      +    const dialog_submit = document.getElementById('delete_confirm_submit')
      +    const dialog_cancel = document.getElementById('delete_confirm_cancel')
      +
      +    dialog.style.maxWidth = '30vw'
      +
      +    form.addEventListener('submit', (e) => {
      +        e.preventDefault()
      +        dialog_submit.value = e.submitter.value
      +        const recording_link = e.submitter.closest('tr').querySelector('a')
      +        dialog_link.setAttribute('href', recording_link.getAttribute('href'))
      +        dialog_link.textContent = recording_link.textContent
      +        dialog.showModal()
      +    })
      +
      +    dialog_cancel.addEventListener('click', (e) => {
      +        e.preventDefault()
      +        dialog.close()
      +    })
      +
      +    document.addEventListener('keydown', (e) => {
      +        if (dialog.open && e.key === 'Escape') {
      +            dialog.close()
      +        }
      +    })
      +})
      diff --git a/ietf/static/js/agenda_personalize.js b/ietf/static/js/agenda_personalize.js
      deleted file mode 100644
      index edb2787f3c..0000000000
      --- a/ietf/static/js/agenda_personalize.js
      +++ /dev/null
      @@ -1,41 +0,0 @@
      -// Copyright The IETF Trust 2021, All Rights Reserved
      -
      -/**
      - * Agenda personalization JS methods
      - *
      - * Requires agenda_timezone.js and timezone.js be included.
      - */
      -'use strict';
      -
      -/**
      - * Update the checkbox state to match the filter parameters
      - */
      -function updateAgendaCheckboxes(filter_params) {
      -    var selection_inputs = document.getElementsByName('selected-sessions');
      -    selection_inputs.forEach((inp) => {
      -        const item_keywords = inp.dataset.filterKeywords.toLowerCase()
      -            .split(',');
      -        if (
      -            agenda_filter.keyword_match(item_keywords, filter_params.show) &&
      -            !agenda_filter.keyword_match(item_keywords, filter_params.hide)
      -        ) {
      -            inp.checked = true;
      -        } else {
      -            inp.checked = false;
      -        }
      -    });
      -}
      -
      -window.handleFilterParamUpdate = function (filter_params) {
      -    updateAgendaCheckboxes(filter_params);
      -};
      -
      -window.handleTableClick = function (event) {
      -    if (event.target.name === 'selected-sessions') {
      -        // hide the tooltip after clicking on a checkbox
      -        const jqElt = jQuery(event.target);
      -        if (jqElt.tooltip) {
      -            jqElt.tooltip('hide');
      -        }
      -    }
      -};
      \ No newline at end of file
      diff --git a/ietf/static/js/agenda_timezone.js b/ietf/static/js/agenda_timezone.js
      deleted file mode 100644
      index 187bea0e88..0000000000
      --- a/ietf/static/js/agenda_timezone.js
      +++ /dev/null
      @@ -1,306 +0,0 @@
      -// Copyright The IETF Trust 2021, All Rights Reserved
      -/*
      - Timezone support specific to the agenda page
      -
      - To properly handle timezones other than local, needs a method to retrieve
      - the current timezone. Set this by passing a method taking no parameters and
      - returning the current timezone to the set_current_tz_cb() method.
      - This should be done before calling anything else in the file.
      - */
      -(function() {
      -    'use strict';
      -
      -    const local_timezone = moment.tz.guess();
      -
      -    // get_current_tz_cb must be overwritten using set_current_tz_cb
      -    let get_current_tz_cb = function() {
      -        throw new Error('Tried to get current timezone before callback registered. Use set_current_tz_cb().');
      -    };
      -
      -    // Initialize moments
      -    function initialize_moments() {
      -        const times = $('.time');
      -        $.each(times, function (i, item) {
      -            item.start_ts = moment.unix(this.getAttribute("data-start-time"))
      -                .utc();
      -            item.end_ts = moment.unix(this.getAttribute("data-end-time"))
      -                .utc();
      -            if (this.hasAttribute("data-weekday")) {
      -                item.format = 2;
      -            } else {
      -                item.format = 1;
      -            }
      -            if (this.hasAttribute("format")) {
      -                item.format = +this.getAttribute("format");
      -            }
      -        });
      -        const things_with_slots = $('[data-slot-start-ts]');
      -        $.each(things_with_slots, function (i, item) {
      -            item.slot_start_ts = moment.unix(this.getAttribute("data-slot-start-ts"))
      -                .utc();
      -            item.slot_end_ts = moment.unix(this.getAttribute("data-slot-end-ts"))
      -                .utc();
      -        });
      -    }
      -
      -    function format_time(t, tz, fmt) {
      -        let out;
      -        const mtz = window.meeting_timezone || "UTC";
      -        switch (fmt) {
      -        case 0:
      -            out = t.tz(tz)
      -                .format('dddd, ') + '' +
      -                t.tz(tz)
      -                    .format('MMMM Do YYYY, ') + '' +
      -                t.tz(tz)
      -                    .format('HH:mm') + '' +
      -                t.tz(tz)
      -                    .format(' Z z') + '';
      -            break;
      -        case 1:
      -            // Note, this code does not work if the meeting crosses the
      -            // year boundary.
      -            out = t.tz(tz)
      -                .format("HH:mm");
      -            if (+t.tz(tz)
      -                .dayOfYear() < +t.tz(mtz)
      -                .dayOfYear()) {
      -                out = out + " (-1)";
      -            } else if (+t.tz(tz)
      -                .dayOfYear() > +t.tz(mtz)
      -                .dayOfYear()) {
      -                out = out + " (+1)";
      -            }
      -            break;
      -        case 2:
      -            out = t.tz(mtz)
      -                .format("dddd, ")
      -                .toUpperCase() +
      -                t.tz(tz)
      -                    .format("HH:mm");
      -            if (+t.tz(tz)
      -                .dayOfYear() < +t.tz(mtz)
      -                .dayOfYear()) {
      -                out = out + " (-1)";
      -            } else if (+t.tz(tz)
      -                .dayOfYear() > +t.tz(mtz)
      -                .dayOfYear()) {
      -                out = out + " (+1)";
      -            }
      -            break;
      -        case 3:
      -            out = t.utc()
      -                .format("YYYY-MM-DD");
      -            break;
      -        case 4:
      -            out = t.tz(tz)
      -                .format("YYYY-MM-DD HH:mm");
      -            break;
      -        case 5:
      -            out = t.tz(tz)
      -                .format("HH:mm");
      -            break;
      -        }
      -        return out;
      -    }
      -
      -    // Format tooltip notice
      -    function format_tooltip_notice(start, end) {
      -        let notice = "";
      -
      -        if (end.isBefore()) {
      -            notice = "Event ended " + end.fromNow();
      -        } else if (start.isAfter()) {
      -            notice = "Event will start " + start.fromNow();
      -        } else {
      -            notice = "Event started " + start.fromNow() + " and will end " +
      -                end.fromNow();
      -        }
      -        return '' + notice + '';
      -    }
      -
      -    // Format tooltip table
      -    function format_tooltip_table(start, end) {
      -        const current_timezone = get_current_tz_cb();
      -        let out = '
      '; - if (window.meeting_timezone !== "") { - out += ''; - } - out += ''; - if (current_timezone !== 'UTC') { - out += ''; - } - out += ''; - out += '
      Session startSession end
      Meeting timezone' + - format_time(start, window.meeting_timezone, 0) + '' + - format_time(end, window.meeting_timezone, 0) + '
      Local timezone' + - format_time(start, local_timezone, 0) + '' + - format_time(end, local_timezone, 0) + '
      Selected Timezone' + - format_time(start, current_timezone, 0) + '' + - format_time(end, current_timezone, 0) + '
      UTC' + - format_time(start, 'UTC', 0) + '' + - format_time(end, 'UTC', 0) + '
      ' + format_tooltip_notice(start, end) + ''; - return out; - } - - // Format tooltip for item - function format_tooltip(start, end) { - return '
      ' + - format_tooltip_table(start, end) + - '
      '; - } - - // Add tooltips - function add_tooltips() { - $('.time') - .each(function () { - const tooltip = $(format_tooltip(this.start_ts, this.end_ts)); - tooltip[0].start_ts = this.start_ts; - tooltip[0].end_ts = this.end_ts; - tooltip[0].ustart_ts = moment(this.start_ts) - .add(-2, 'hours'); - tooltip[0].uend_ts = moment(this.end_ts) - .add(2, 'hours'); - $(this) - .closest("th, td") - .attr("data-bs-toggle", "popover") - .attr("data-bs-content", $(tooltip) - .html()) - .popover({ - html: true, - sanitize: false, - trigger: "hover" - }); - }); - } - - // Update times on the agenda based on the selected timezone - function update_times(newtz) { - $('.current-tz') - .html(newtz.replaceAll("_", " ").replaceAll("/", " / ")); - $('.time') - .each(function () { - if (this.format === 4) { - const tz = this.start_ts.tz(newtz).format(" z"); - const start_doy = this.start_ts.tz(newtz).dayOfYear(); - const end_doy = this.end_ts.tz(newtz).dayOfYear(); - if (start_doy === end_doy) { - $(this) - .html(format_time(this.start_ts, newtz, this.format) + - '
      -' + format_time(this.end_ts, newtz, 5) + tz); - } else { - $(this) - .html(format_time(this.start_ts, newtz, this.format) + - '
      -' + - format_time(this.end_ts, newtz, this.format) + tz); - } - } else { - $(this) - .html(format_time(this.start_ts, newtz, this.format) + '
      -' + - format_time(this.end_ts, newtz, this.format)); - } - }); - update_tooltips_all(); - update_clock(); - } - - // Update hrefs in anchor tags with the "now-link" class. Mark the target with the "current-session" class. - function update_now_link(agenda_rows, ongoing_rows, later_rows) { - agenda_rows.removeClass('current-session'); - const links_to_update = $('a.now-link'); - if (ongoing_rows.length > 0) { - // sessions are ongoing - find those with the latest start time and mark the first of them as "now" - const last_start_time = ongoing_rows[ongoing_rows.length - 1].slot_start_ts; - for (let ii=0; ii < ongoing_rows.length; ii++) { - const dt = ongoing_rows[ii].slot_start_ts.diff(last_start_time, 'seconds'); - if (Math.abs(dt) < 1) { - $(ongoing_rows[ii]).addClass('current-session'); - links_to_update.attr('href', '#' + ongoing_rows[ii].id); - break; - } - } - } else if (later_rows.length > 0) { - // There were no ongoing sessions, look for the next one to start and mark as current - $(later_rows[0]).addClass('current-session'); - links_to_update.attr('href', '#' + later_rows[0].id); - } else { - // No sessions in the future - meeting has apparently ended - links_to_update.attr('href', '#'); - links_to_update.addClass('disabled'); // mark link - } - } - - function update_ongoing_sessions() { - const agenda_rows = $('[data-slot-start-ts]'); - const now_moment = moment(); - const ongoing_rows = agenda_rows.filter(function () { - return now_moment.isBetween(this.slot_start_ts, this.slot_end_ts); - }); - const later_rows = agenda_rows.filter(function() { return now_moment.isBefore(this.slot_start_ts); }); - // Highlight ongoing based on the current time - agenda_rows.removeClass("table-warning"); - ongoing_rows.addClass("table-warning"); - update_now_link(agenda_rows, ongoing_rows, later_rows); // update any "now-link" anchors - } - - // Update tooltips - function update_tooltips() { - const tooltips = $('.timetooltiptext'); - tooltips.filter(function () { - return moment() - .isBetween(this.ustart_ts, this.uend_ts); - }) - .each(function () { - $(this) - .html(format_tooltip_table(this.start_ts, this.end_ts)); - }); - } - - // Update all tooltips - function update_tooltips_all() { - const tooltips = $('.timetooltiptext'); - tooltips.each(function () { - $(this) - .html(format_tooltip_table(this.start_ts, this.end_ts)); - }); - } - - // Update clock - function update_clock() { - $('span.current-time') - .html(format_time(moment(), get_current_tz_cb(), 0)); - } - - function urlParam(name) { - const results = new RegExp('[\?&]' + name + '=([^&#]*)') - .exec(window.location.href); - if (results === null) { - return null; - } else { - return results[1] || 0; - } - } - - function init_timers(speedup) { - speedup = speedup || 1; - const fast_timer = 60000 / (speedup > 600 ? 600 : speedup); - update_clock(); - update_ongoing_sessions(); - setInterval(function () { update_clock(); }, fast_timer); - setInterval(function () { update_ongoing_sessions(); }, fast_timer); - setInterval(function () { update_tooltips(); }, fast_timer); - setInterval(function () { update_tooltips_all(); }, 3600000 / speedup); - } - - /***** make public interface available on window *****/ - window.initialize_moments = initialize_moments; - window.add_tooltips = add_tooltips; - window.update_times = update_times; - window.urlParam = urlParam; - window.init_timers = init_timers; - - // set method used to find current time zone - window.set_current_tz_cb = function (fn) { - get_current_tz_cb = fn; - }; -})(); diff --git a/ietf/static/js/announcement.js b/ietf/static/js/announcement.js new file mode 100644 index 0000000000..95465120fa --- /dev/null +++ b/ietf/static/js/announcement.js @@ -0,0 +1,57 @@ +const announcementApp = (function() { + 'use strict'; + return { + // functions for Announcement + checkToField: function() { + document.documentElement.scrollTop = 0; // For most browsers + const toField = document.getElementById('id_to'); + const toCustomInput = document.getElementById('id_to_custom'); + const toCustomDiv = toCustomInput.closest('div.row'); + + if (toField.value === 'Other...') { + toCustomDiv.style.display = 'flex'; // Show the custom field + } else { + toCustomDiv.style.display = 'none'; // Hide the custom field + toCustomInput.value = ''; // Optionally clear the input value if hidden + } + } + }; +})(); + +// Extra care is required to ensure the back button +// works properly for the optional to_custom field. +// Take the case when a user selects "Other..." for +// "To" field. The "To custom" field appears and they +// enter a new address there. +// In Chrome, when the form is submitted and then the user +// uses the back button (or browser back), the page loads +// from bfcache then the javascript DOMContentLoaded event +// handler is run, hiding the empty to_custom field, THEN the +// browser autofills the form fields. Because to_submit +// is now hidden it does not get a value. This is a very +// bad experience for the user because the to_custom field +// was unexpectedly cleared and hidden. If they notice this +// they would need to know to first select another "To" +// option, then select "Other..." again just to get the +// to_custom field visible so they can re-enter the custom +// address. +// The solution is to use setTimeout to run checkToField +// after a short delay, giving the browser time to autofill +// the form fields before it checks to see if the to_custom +// field is empty and hides it. + +document.addEventListener('DOMContentLoaded', function() { + // Run the visibility check after allowing cache to populate values + setTimeout(announcementApp.checkToField, 300); + + const toField = document.getElementById('id_to'); + toField.addEventListener('change', announcementApp.checkToField); +}); + +// Handle back/forward navigation with pageshow +window.addEventListener('pageshow', function(event) { + if (event.persisted) { + // Then apply visibility logic after cache restoration + setTimeout(announcementApp.checkToField, 300); + } +}); \ No newline at end of file diff --git a/ietf/static/js/attendees-chart.js b/ietf/static/js/attendees-chart.js new file mode 100644 index 0000000000..fed3b1289c --- /dev/null +++ b/ietf/static/js/attendees-chart.js @@ -0,0 +1,58 @@ +(function () { + var raw = document.getElementById('attendees-chart-data'); + if (!raw) return; + var chartData = JSON.parse(raw.textContent); + var chart = null; + var currentBreakdown = 'type'; + + // Override the global transparent background set by highcharts.js so the + // export menu and fullscreen view use the page background color. + var container = document.getElementById('attendees-pie-chart'); + var bodyBg = getComputedStyle(document.body).backgroundColor; + container.style.setProperty('--highcharts-background-color', bodyBg); + + function renderChart(breakdown) { + var seriesData = chartData[breakdown].map(function (item) { + return { name: item[0], y: item[1] }; + }); + if (chart) chart.destroy(); + chart = Highcharts.chart(container, { + chart: { type: 'pie', height: 400 }, + title: { text: null }, + tooltip: { pointFormat: '{point.name}: {point.y} ({point.percentage:.1f}%)' }, + plotOptions: { + pie: { + dataLabels: { + enabled: true, + format: '{point.name}
      {point.y} ({point.percentage:.1f}%)', + }, + showInLegend: false, + } + }, + series: [{ name: 'Attendees', data: seriesData }], + }); + } + + var modal = document.getElementById('attendees-chart-modal'); + + // Render (or re-render) the chart each time the modal becomes fully visible, + // so Highcharts can measure the container dimensions correctly. + modal.addEventListener('shown.bs.modal', function () { + renderChart(currentBreakdown); + }); + + // Release the chart when the modal closes to avoid stale renders. + modal.addEventListener('hidden.bs.modal', function () { + if (chart) { + chart.destroy(); + chart = null; + } + }); + + document.querySelectorAll('[name="attendees-breakdown"]').forEach(function (radio) { + radio.addEventListener('change', function () { + currentBreakdown = this.value; + renderChart(currentBreakdown); + }); + }); +})(); diff --git a/ietf/static/js/complete-review.js b/ietf/static/js/complete-review.js index a359dac237..3a58ba9700 100644 --- a/ietf/static/js/complete-review.js +++ b/ietf/static/js/complete-review.js @@ -24,6 +24,8 @@ $(document) .before(mailArchiveSearchTemplate); var mailArchiveSearch = form.find(".mail-archive-search"); + const isReviewer = mailArchiveSearch.data('isReviewer'); + const searchMailArchiveUrl = mailArchiveSearch.data('searchMailArchiveUrl'); var retrievingData = null; @@ -190,4 +192,4 @@ $(document) form.find("[name=review_submission][value=link]") .trigger("click"); } - }); \ No newline at end of file + }); diff --git a/ietf/static/js/custom_striped.js b/ietf/static/js/custom_striped.js new file mode 100644 index 0000000000..480ad7cf82 --- /dev/null +++ b/ietf/static/js/custom_striped.js @@ -0,0 +1,16 @@ +// Copyright The IETF Trust 2025, All Rights Reserved + +document.addEventListener('DOMContentLoaded', () => { + // add stripes + const firstRow = document.querySelector('.custom-stripe .row') + if (firstRow) { + const parent = firstRow.parentElement; + const allRows = Array.from(parent.children).filter(child => child.classList.contains('row')) + allRows.forEach((row, index) => { + row.classList.remove('bg-light') + if (index % 2 === 1) { + row.classList.add('bg-light') + } + }) + } +}) diff --git a/ietf/static/js/datepicker.js b/ietf/static/js/datepicker.js index 1fe6d0314b..43d80acb5f 100644 --- a/ietf/static/js/datepicker.js +++ b/ietf/static/js/datepicker.js @@ -1,2039 +1,54 @@ -/*! - * Datepicker for Bootstrap v1.9.0 (https://github.com/uxsolutions/bootstrap-datepicker) - * - * Licensed under the Apache License v2.0 (http://www.apache.org/licenses/LICENSE-2.0) - */ - -(function(factory){ - if (typeof define === 'function' && define.amd) { - define(['jquery'], factory); - } else if (typeof exports === 'object') { - factory(require('jquery')); - } else { - factory(jQuery); - } -}(function($, undefined){ - function UTCDate(){ - return new Date(Date.UTC.apply(Date, arguments)); - } - function UTCToday(){ - var today = new Date(); - return UTCDate(today.getFullYear(), today.getMonth(), today.getDate()); - } - function isUTCEquals(date1, date2) { - return ( - date1.getUTCFullYear() === date2.getUTCFullYear() && - date1.getUTCMonth() === date2.getUTCMonth() && - date1.getUTCDate() === date2.getUTCDate() - ); - } - function alias(method, deprecationMsg){ - return function(){ - if (deprecationMsg !== undefined) { - $.fn.datepicker.deprecated(deprecationMsg); - } - - return this[method].apply(this, arguments); - }; - } - function isValidDate(d) { - return d && !isNaN(d.getTime()); - } - - var DateArray = (function(){ - var extras = { - get: function(i){ - return this.slice(i)[0]; - }, - contains: function(d){ - // Array.indexOf is not cross-browser; - // $.inArray doesn't work with Dates - var val = d && d.valueOf(); - for (var i=0, l=this.length; i < l; i++) - // Use date arithmetic to allow dates with different times to match - if (0 <= this[i].valueOf() - val && this[i].valueOf() - val < 1000*60*60*24) - return i; - return -1; - }, - remove: function(i){ - this.splice(i,1); - }, - replace: function(new_array){ - if (!new_array) - return; - if (!Array.isArray(new_array)) - new_array = [new_array]; - this.clear(); - this.push.apply(this, new_array); - }, - clear: function(){ - this.length = 0; - }, - copy: function(){ - var a = new DateArray(); - a.replace(this); - return a; - } - }; - - return function(){ - var a = []; - a.push.apply(a, arguments); - $.extend(a, extras); - return a; - }; - })(); - - - // Picker object - - var Datepicker = function(element, options){ - $.data(element, 'datepicker', this); - - this._events = []; - this._secondaryEvents = []; - - this._process_options(options); - - this.dates = new DateArray(); - this.viewDate = this.o.defaultViewDate; - this.focusDate = null; - - this.element = $(element); - this.isInput = this.element.is('input'); - this.inputField = this.isInput ? this.element : this.element.find('input'); - this.component = this.element.hasClass('date') ? this.element.find('.add-on, .input-group-addon, .input-group-append, .input-group-prepend, .btn') : false; - if (this.component && this.component.length === 0) - this.component = false; - this.isInline = !this.component && this.element.is('div'); - - this.picker = $(DPGlobal.template); - - // Checking templates and inserting - if (this._check_template(this.o.templates.leftArrow)) { - this.picker.find('.prev').html(this.o.templates.leftArrow); - } - - if (this._check_template(this.o.templates.rightArrow)) { - this.picker.find('.next').html(this.o.templates.rightArrow); - } - - this._buildEvents(); - this._attachEvents(); - - if (this.isInline){ - this.picker.addClass('datepicker-inline').appendTo(this.element); - } - else { - this.picker.addClass('datepicker-dropdown dropdown-menu'); - } - - if (this.o.rtl){ - this.picker.addClass('datepicker-rtl'); - } - - if (this.o.calendarWeeks) { - this.picker.find('.datepicker-days .datepicker-switch, thead .datepicker-title, tfoot .today, tfoot .clear') - .attr('colspan', function(i, val){ - return Number(val) + 1; - }); - } - - this._process_options({ - startDate: this._o.startDate, - endDate: this._o.endDate, - daysOfWeekDisabled: this.o.daysOfWeekDisabled, - daysOfWeekHighlighted: this.o.daysOfWeekHighlighted, - datesDisabled: this.o.datesDisabled - }); - - this._allow_update = false; - this.setViewMode(this.o.startView); - this._allow_update = true; - - this.fillDow(); - this.fillMonths(); - - this.update(); - - if (this.isInline){ - this.show(); - } - }; - - Datepicker.prototype = { - constructor: Datepicker, - - _resolveViewName: function(view){ - $.each(DPGlobal.viewModes, function(i, viewMode){ - if (view === i || $.inArray(view, viewMode.names) !== -1){ - view = i; - return false; - } - }); - - return view; - }, - - _resolveDaysOfWeek: function(daysOfWeek){ - if (!Array.isArray(daysOfWeek)) - daysOfWeek = daysOfWeek.split(/[,\s]*/); - return $.map(daysOfWeek, Number); - }, - - _check_template: function(tmp){ - try { - // If empty - if (tmp === undefined || tmp === "") { - return false; - } - // If no html, everything ok - if ((tmp.match(/[<>]/g) || []).length <= 0) { - return true; - } - // Checking if html is fine - var jDom = $(tmp); - return jDom.length > 0; - } - catch (ex) { - return false; - } - }, - - _process_options: function(opts){ - // Store raw options for reference - this._o = $.extend({}, this._o, opts); - // Processed options - var o = this.o = $.extend({}, this._o); - - // Check if "de-DE" style date is available, if not language should - // fallback to 2 letter code eg "de" - var lang = o.language; - if (!dates[lang]){ - lang = lang.split('-')[0]; - if (!dates[lang]) - lang = defaults.language; - } - o.language = lang; - - // Retrieve view index from any aliases - o.startView = this._resolveViewName(o.startView); - o.minViewMode = this._resolveViewName(o.minViewMode); - o.maxViewMode = this._resolveViewName(o.maxViewMode); - - // Check view is between min and max - o.startView = Math.max(this.o.minViewMode, Math.min(this.o.maxViewMode, o.startView)); - - // true, false, or Number > 0 - if (o.multidate !== true){ - o.multidate = Number(o.multidate) || false; - if (o.multidate !== false) - o.multidate = Math.max(0, o.multidate); - } - o.multidateSeparator = String(o.multidateSeparator); - - o.weekStart %= 7; - o.weekEnd = (o.weekStart + 6) % 7; - - var format = DPGlobal.parseFormat(o.format); - if (o.startDate !== -Infinity){ - if (!!o.startDate){ - if (o.startDate instanceof Date) - o.startDate = this._local_to_utc(this._zero_time(o.startDate)); - else - o.startDate = DPGlobal.parseDate(o.startDate, format, o.language, o.assumeNearbyYear); - } - else { - o.startDate = -Infinity; - } - } - if (o.endDate !== Infinity){ - if (!!o.endDate){ - if (o.endDate instanceof Date) - o.endDate = this._local_to_utc(this._zero_time(o.endDate)); - else - o.endDate = DPGlobal.parseDate(o.endDate, format, o.language, o.assumeNearbyYear); - } - else { - o.endDate = Infinity; - } - } - - o.daysOfWeekDisabled = this._resolveDaysOfWeek(o.daysOfWeekDisabled||[]); - o.daysOfWeekHighlighted = this._resolveDaysOfWeek(o.daysOfWeekHighlighted||[]); - - o.datesDisabled = o.datesDisabled||[]; - if (!Array.isArray(o.datesDisabled)) { - o.datesDisabled = o.datesDisabled.split(','); - } - o.datesDisabled = $.map(o.datesDisabled, function(d){ - return DPGlobal.parseDate(d, format, o.language, o.assumeNearbyYear); - }); - - var plc = String(o.orientation).toLowerCase().split(/\s+/g), - _plc = o.orientation.toLowerCase(); - plc = $.grep(plc, function(word){ - return /^auto|left|right|top|bottom$/.test(word); - }); - o.orientation = {x: 'auto', y: 'auto'}; - if (!_plc || _plc === 'auto') - ; // no action - else if (plc.length === 1){ - switch (plc[0]){ - case 'top': - case 'bottom': - o.orientation.y = plc[0]; - break; - case 'left': - case 'right': - o.orientation.x = plc[0]; - break; - } - } - else { - _plc = $.grep(plc, function(word){ - return /^left|right$/.test(word); - }); - o.orientation.x = _plc[0] || 'auto'; - - _plc = $.grep(plc, function(word){ - return /^top|bottom$/.test(word); - }); - o.orientation.y = _plc[0] || 'auto'; - } - if (o.defaultViewDate instanceof Date || typeof o.defaultViewDate === 'string') { - o.defaultViewDate = DPGlobal.parseDate(o.defaultViewDate, format, o.language, o.assumeNearbyYear); - } else if (o.defaultViewDate) { - var year = o.defaultViewDate.year || new Date().getFullYear(); - var month = o.defaultViewDate.month || 0; - var day = o.defaultViewDate.day || 1; - o.defaultViewDate = UTCDate(year, month, day); - } else { - o.defaultViewDate = UTCToday(); - } - }, - _applyEvents: function(evs){ - for (var i=0, el, ch, ev; i < evs.length; i++){ - el = evs[i][0]; - if (evs[i].length === 2){ - ch = undefined; - ev = evs[i][1]; - } else if (evs[i].length === 3){ - ch = evs[i][1]; - ev = evs[i][2]; - } - el.on(ev, ch); - } - }, - _unapplyEvents: function(evs){ - for (var i=0, el, ev, ch; i < evs.length; i++){ - el = evs[i][0]; - if (evs[i].length === 2){ - ch = undefined; - ev = evs[i][1]; - } else if (evs[i].length === 3){ - ch = evs[i][1]; - ev = evs[i][2]; - } - el.off(ev, ch); - } - }, - _buildEvents: function(){ - var events = { - keyup: $.proxy(function(e){ - if ($.inArray(e.keyCode, [27, 37, 39, 38, 40, 32, 13, 9]) === -1) - this.update(); - }, this), - keydown: $.proxy(this.keydown, this), - paste: $.proxy(this.paste, this) +import { + Datepicker +} from 'vanillajs-datepicker'; + +global.enable_datepicker = function (el) { + // we need to translate from bootstrap-datepicker options to + // vanillajs-datepicker options + const view_mode = { + day: 0, + days: 0, + month: 1, + months: 1, + year: 2, + years: 2, + decade: 3, + decades: 3 + }; + + let options = { + buttonClass: "btn" + }; + if (el.dataset.dateFormat) { + options = { ...options, + format: el.dataset.dateFormat + }; + if (!el.dataset.dateFormat.includes("dd")) { + options = { ...options, + pickLevel: 1 }; - - if (this.o.showOnFocus === true) { - events.focus = $.proxy(this.show, this); - } - - if (this.isInput) { // single input - this._events = [ - [this.element, events] - ]; - } - // component: input + button - else if (this.component && this.inputField.length) { - this._events = [ - // For components that are not readonly, allow keyboard nav - [this.inputField, events], - [this.component, { - click: $.proxy(this.show, this) - }] - ]; - } - else { - this._events = [ - [this.element, { - click: $.proxy(this.show, this), - keydown: $.proxy(this.keydown, this) - }] - ]; - } - this._events.push( - // Component: listen for blur on element descendants - [this.element, '*', { - blur: $.proxy(function(e){ - this._focused_from = e.target; - }, this) - }], - // Input: listen for blur on element - [this.element, { - blur: $.proxy(function(e){ - this._focused_from = e.target; - }, this) - }] - ); - - if (this.o.immediateUpdates) { - // Trigger input updates immediately on changed year/month - this._events.push([this.element, { - 'changeYear changeMonth': $.proxy(function(e){ - this.update(e.date); - }, this) - }]); - } - - this._secondaryEvents = [ - [this.picker, { - click: $.proxy(this.click, this) - }], - [this.picker, '.prev, .next', { - click: $.proxy(this.navArrowsClick, this) - }], - [this.picker, '.day:not(.disabled)', { - click: $.proxy(this.dayCellClick, this) - }], - [$(window), { - resize: $.proxy(this.place, this) - }], - [$(document), { - 'mousedown touchstart': $.proxy(function(e){ - // Clicked outside the datepicker, hide it - if (!( - this.element.is(e.target) || - this.element.find(e.target).length || - this.picker.is(e.target) || - this.picker.find(e.target).length || - this.isInline - )){ - this.hide(); - } - }, this) - }] - ]; - }, - _attachEvents: function(){ - this._detachEvents(); - this._applyEvents(this._events); - }, - _detachEvents: function(){ - this._unapplyEvents(this._events); - }, - _attachSecondaryEvents: function(){ - this._detachSecondaryEvents(); - this._applyEvents(this._secondaryEvents); - }, - _detachSecondaryEvents: function(){ - this._unapplyEvents(this._secondaryEvents); - }, - _trigger: function(event, altdate){ - var date = altdate || this.dates.get(-1), - local_date = this._utc_to_local(date); - - this.element.trigger({ - type: event, - date: local_date, - viewMode: this.viewMode, - dates: $.map(this.dates, this._utc_to_local), - format: $.proxy(function(ix, format){ - if (arguments.length === 0){ - ix = this.dates.length - 1; - format = this.o.format; - } else if (typeof ix === 'string'){ - format = ix; - ix = this.dates.length - 1; - } - format = format || this.o.format; - var date = this.dates.get(ix); - return DPGlobal.formatDate(date, format, this.o.language); - }, this) - }); - }, - - show: function(){ - if (this.inputField.is(':disabled') || (this.inputField.prop('readonly') && this.o.enableOnReadonly === false)) - return; - if (!this.isInline) - this.picker.appendTo(this.o.container); - this.place(); - this.picker.show(); - this._attachSecondaryEvents(); - this._trigger('show'); - if ((window.navigator.msMaxTouchPoints || 'ontouchstart' in document) && this.o.disableTouchKeyboard) { - $(this.element).blur(); - } - return this; - }, - - hide: function(){ - if (this.isInline || !this.picker.is(':visible')) - return this; - this.focusDate = null; - this.picker.hide().detach(); - this._detachSecondaryEvents(); - this.setViewMode(this.o.startView); - - if (this.o.forceParse && this.inputField.val()) - this.setValue(); - this._trigger('hide'); - return this; - }, - - destroy: function(){ - this.hide(); - this._detachEvents(); - this._detachSecondaryEvents(); - this.picker.remove(); - delete this.element.data().datepicker; - if (!this.isInput){ - delete this.element.data().date; - } - return this; - }, - - paste: function(e){ - var dateString; - if (e.originalEvent.clipboardData && e.originalEvent.clipboardData.types - && $.inArray('text/plain', e.originalEvent.clipboardData.types) !== -1) { - dateString = e.originalEvent.clipboardData.getData('text/plain'); - } else if (window.clipboardData) { - dateString = window.clipboardData.getData('Text'); - } else { - return; - } - this.setDate(dateString); - this.update(); - e.preventDefault(); - }, - - _utc_to_local: function(utc){ - if (!utc) { - return utc; - } - - var local = new Date(utc.getTime() + (utc.getTimezoneOffset() * 60000)); - - if (local.getTimezoneOffset() !== utc.getTimezoneOffset()) { - local = new Date(utc.getTime() + (local.getTimezoneOffset() * 60000)); - } - - return local; - }, - _local_to_utc: function(local){ - return local && new Date(local.getTime() - (local.getTimezoneOffset()*60000)); - }, - _zero_time: function(local){ - return local && new Date(local.getFullYear(), local.getMonth(), local.getDate()); - }, - _zero_utc_time: function(utc){ - return utc && UTCDate(utc.getUTCFullYear(), utc.getUTCMonth(), utc.getUTCDate()); - }, - - getDates: function(){ - return $.map(this.dates, this._utc_to_local); - }, - - getUTCDates: function(){ - return $.map(this.dates, function(d){ - return new Date(d); - }); - }, - - getDate: function(){ - return this._utc_to_local(this.getUTCDate()); - }, - - getUTCDate: function(){ - var selected_date = this.dates.get(-1); - if (selected_date !== undefined) { - return new Date(selected_date); - } else { - return null; - } - }, - - clearDates: function(){ - this.inputField.val(''); - this.update(); - this._trigger('changeDate'); - - if (this.o.autoclose) { - this.hide(); - } - }, - - setDates: function(){ - var args = Array.isArray(arguments[0]) ? arguments[0] : arguments; - this.update.apply(this, args); - this._trigger('changeDate'); - this.setValue(); - return this; - }, - - setUTCDates: function(){ - var args = Array.isArray(arguments[0]) ? arguments[0] : arguments; - this.setDates.apply(this, $.map(args, this._utc_to_local)); - return this; - }, - - setDate: alias('setDates'), - setUTCDate: alias('setUTCDates'), - remove: alias('destroy', 'Method `remove` is deprecated and will be removed in version 2.0. Use `destroy` instead'), - - setValue: function(){ - var formatted = this.getFormattedDate(); - this.inputField.val(formatted); - return this; - }, - - getFormattedDate: function(format){ - if (format === undefined) - format = this.o.format; - - var lang = this.o.language; - return $.map(this.dates, function(d){ - return DPGlobal.formatDate(d, format, lang); - }).join(this.o.multidateSeparator); - }, - - getStartDate: function(){ - return this.o.startDate; - }, - - setStartDate: function(startDate){ - this._process_options({startDate: startDate}); - this.update(); - this.updateNavArrows(); - return this; - }, - - getEndDate: function(){ - return this.o.endDate; - }, - - setEndDate: function(endDate){ - this._process_options({endDate: endDate}); - this.update(); - this.updateNavArrows(); - return this; - }, - - setDaysOfWeekDisabled: function(daysOfWeekDisabled){ - this._process_options({daysOfWeekDisabled: daysOfWeekDisabled}); - this.update(); - return this; - }, - - setDaysOfWeekHighlighted: function(daysOfWeekHighlighted){ - this._process_options({daysOfWeekHighlighted: daysOfWeekHighlighted}); - this.update(); - return this; - }, - - setDatesDisabled: function(datesDisabled){ - this._process_options({datesDisabled: datesDisabled}); - this.update(); - return this; - }, - - place: function(){ - if (this.isInline) - return this; - var calendarWidth = this.picker.outerWidth(), - calendarHeight = this.picker.outerHeight(), - visualPadding = 10, - container = $(this.o.container), - windowWidth = container.width(), - scrollTop = this.o.container === 'body' ? $(document).scrollTop() : container.scrollTop(), - appendOffset = container.offset(); - - var parentsZindex = [0]; - this.element.parents().each(function(){ - var itemZIndex = $(this).css('z-index'); - if (itemZIndex !== 'auto' && Number(itemZIndex) !== 0) parentsZindex.push(Number(itemZIndex)); - }); - var zIndex = Math.max.apply(Math, parentsZindex) + this.o.zIndexOffset; - var offset = this.component ? this.component.parent().offset() : this.element.offset(); - var height = this.component ? this.component.outerHeight(true) : this.element.outerHeight(false); - var width = this.component ? this.component.outerWidth(true) : this.element.outerWidth(false); - var left = offset.left - appendOffset.left; - var top = offset.top - appendOffset.top; - - if (this.o.container !== 'body') { - top += scrollTop; - } - - this.picker.removeClass( - 'datepicker-orient-top datepicker-orient-bottom '+ - 'datepicker-orient-right datepicker-orient-left' - ); - - if (this.o.orientation.x !== 'auto'){ - this.picker.addClass('datepicker-orient-' + this.o.orientation.x); - if (this.o.orientation.x === 'right') - left -= calendarWidth - width; - } - // auto x orientation is best-placement: if it crosses a window - // edge, fudge it sideways - else { - if (offset.left < 0) { - // component is outside the window on the left side. Move it into visible range - this.picker.addClass('datepicker-orient-left'); - left -= offset.left - visualPadding; - } else if (left + calendarWidth > windowWidth) { - // the calendar passes the widow right edge. Align it to component right side - this.picker.addClass('datepicker-orient-right'); - left += width - calendarWidth; - } else { - if (this.o.rtl) { - // Default to right - this.picker.addClass('datepicker-orient-right'); - } else { - // Default to left - this.picker.addClass('datepicker-orient-left'); - } - } - } - - // auto y orientation is best-situation: top or bottom, no fudging, - // decision based on which shows more of the calendar - var yorient = this.o.orientation.y, - top_overflow; - if (yorient === 'auto'){ - top_overflow = -scrollTop + top - calendarHeight; - yorient = top_overflow < 0 ? 'bottom' : 'top'; - } - - this.picker.addClass('datepicker-orient-' + yorient); - if (yorient === 'top') - top -= calendarHeight + parseInt(this.picker.css('padding-top')); - else - top += height; - - if (this.o.rtl) { - var right = windowWidth - (left + width); - this.picker.css({ - top: top, - right: right, - zIndex: zIndex - }); - } else { - this.picker.css({ - top: top, - left: left, - zIndex: zIndex - }); - } - return this; - }, - - _allow_update: true, - update: function(){ - if (!this._allow_update) - return this; - - var oldDates = this.dates.copy(), - dates = [], - fromArgs = false; - if (arguments.length){ - $.each(arguments, $.proxy(function(i, date){ - if (date instanceof Date) - date = this._local_to_utc(date); - dates.push(date); - }, this)); - fromArgs = true; - } else { - dates = this.isInput - ? this.element.val() - : this.element.data('date') || this.inputField.val(); - if (dates && this.o.multidate) - dates = dates.split(this.o.multidateSeparator); - else - dates = [dates]; - delete this.element.data().date; - } - - dates = $.map(dates, $.proxy(function(date){ - return DPGlobal.parseDate(date, this.o.format, this.o.language, this.o.assumeNearbyYear); - }, this)); - dates = $.grep(dates, $.proxy(function(date){ - return ( - !this.dateWithinRange(date) || - !date - ); - }, this), true); - this.dates.replace(dates); - - if (this.o.updateViewDate) { - if (this.dates.length) - this.viewDate = new Date(this.dates.get(-1)); - else if (this.viewDate < this.o.startDate) - this.viewDate = new Date(this.o.startDate); - else if (this.viewDate > this.o.endDate) - this.viewDate = new Date(this.o.endDate); - else - this.viewDate = this.o.defaultViewDate; - } - - if (fromArgs){ - // setting date by clicking - this.setValue(); - this.element.change(); - } - else if (this.dates.length){ - // setting date by typing - if (String(oldDates) !== String(this.dates) && fromArgs) { - this._trigger('changeDate'); - this.element.change(); - } - } - if (!this.dates.length && oldDates.length) { - this._trigger('clearDate'); - this.element.change(); - } - - this.fill(); - return this; - }, - - fillDow: function(){ - if (this.o.showWeekDays) { - var dowCnt = this.o.weekStart, - html = ''; - if (this.o.calendarWeeks){ - html += ' '; - } - while (dowCnt < this.o.weekStart + 7){ - html += ''+dates[this.o.language].daysMin[(dowCnt++)%7]+''; - } - html += ''; - this.picker.find('.datepicker-days thead').append(html); - } - }, - - fillMonths: function(){ - var localDate = this._utc_to_local(this.viewDate); - var html = ''; - var focused; - for (var i = 0; i < 12; i++){ - focused = localDate && localDate.getMonth() === i ? ' focused' : ''; - html += '' + dates[this.o.language].monthsShort[i] + ''; - } - this.picker.find('.datepicker-months td').html(html); - }, - - setRange: function(range){ - if (!range || !range.length) - delete this.range; - else - this.range = $.map(range, function(d){ - return d.valueOf(); - }); - this.fill(); - }, - - getClassNames: function(date){ - var cls = [], - year = this.viewDate.getUTCFullYear(), - month = this.viewDate.getUTCMonth(), - today = UTCToday(); - if (date.getUTCFullYear() < year || (date.getUTCFullYear() === year && date.getUTCMonth() < month)){ - cls.push('old'); - } else if (date.getUTCFullYear() > year || (date.getUTCFullYear() === year && date.getUTCMonth() > month)){ - cls.push('new'); - } - if (this.focusDate && date.valueOf() === this.focusDate.valueOf()) - cls.push('focused'); - // Compare internal UTC date with UTC today, not local today - if (this.o.todayHighlight && isUTCEquals(date, today)) { - cls.push('today'); - } - if (this.dates.contains(date) !== -1) - cls.push('active'); - if (!this.dateWithinRange(date)){ - cls.push('disabled'); - } - if (this.dateIsDisabled(date)){ - cls.push('disabled', 'disabled-date'); - } - if ($.inArray(date.getUTCDay(), this.o.daysOfWeekHighlighted) !== -1){ - cls.push('highlighted'); - } - - if (this.range){ - if (date > this.range[0] && date < this.range[this.range.length-1]){ - cls.push('range'); - } - if ($.inArray(date.valueOf(), this.range) !== -1){ - cls.push('selected'); - } - if (date.valueOf() === this.range[0]){ - cls.push('range-start'); - } - if (date.valueOf() === this.range[this.range.length-1]){ - cls.push('range-end'); } - } - return cls; - }, - - _fill_yearsView: function(selector, cssClass, factor, year, startYear, endYear, beforeFn){ - var html = ''; - var step = factor / 10; - var view = this.picker.find(selector); - var startVal = Math.floor(year / factor) * factor; - var endVal = startVal + step * 9; - var focusedVal = Math.floor(this.viewDate.getFullYear() / step) * step; - var selected = $.map(this.dates, function(d){ - return Math.floor(d.getUTCFullYear() / step) * step; - }); - - var classes, tooltip, before; - for (var currVal = startVal - step; currVal <= endVal + step; currVal += step) { - classes = [cssClass]; - tooltip = null; - - if (currVal === startVal - step) { - classes.push('old'); - } else if (currVal === endVal + step) { - classes.push('new'); - } - if ($.inArray(currVal, selected) !== -1) { - classes.push('active'); - } - if (currVal < startYear || currVal > endYear) { - classes.push('disabled'); - } - if (currVal === focusedVal) { - classes.push('focused'); - } - - if (beforeFn !== $.noop) { - before = beforeFn(new Date(currVal, 0, 1)); - if (before === undefined) { - before = {}; - } else if (typeof before === 'boolean') { - before = {enabled: before}; - } else if (typeof before === 'string') { - before = {classes: before}; - } - if (before.enabled === false) { - classes.push('disabled'); - } - if (before.classes) { - classes = classes.concat(before.classes.split(/\s+/)); - } - if (before.tooltip) { - tooltip = before.tooltip; - } - } - - html += '' + currVal + ''; - } - - view.find('.datepicker-switch').text(startVal + '-' + endVal); - view.find('td').html(html); - }, - - fill: function(){ - var d = new Date(this.viewDate), - year = d.getUTCFullYear(), - month = d.getUTCMonth(), - startYear = this.o.startDate !== -Infinity ? this.o.startDate.getUTCFullYear() : -Infinity, - startMonth = this.o.startDate !== -Infinity ? this.o.startDate.getUTCMonth() : -Infinity, - endYear = this.o.endDate !== Infinity ? this.o.endDate.getUTCFullYear() : Infinity, - endMonth = this.o.endDate !== Infinity ? this.o.endDate.getUTCMonth() : Infinity, - todaytxt = dates[this.o.language].today || dates['en'].today || '', - cleartxt = dates[this.o.language].clear || dates['en'].clear || '', - titleFormat = dates[this.o.language].titleFormat || dates['en'].titleFormat, - todayDate = UTCToday(), - titleBtnVisible = (this.o.todayBtn === true || this.o.todayBtn === 'linked') && todayDate >= this.o.startDate && todayDate <= this.o.endDate && !this.weekOfDateIsDisabled(todayDate), - tooltip, - before; - if (isNaN(year) || isNaN(month)) - return; - this.picker.find('.datepicker-days .datepicker-switch') - .text(DPGlobal.formatDate(d, titleFormat, this.o.language)); - this.picker.find('tfoot .today') - .text(todaytxt) - .css('display', titleBtnVisible ? 'table-cell' : 'none'); - this.picker.find('tfoot .clear') - .text(cleartxt) - .css('display', this.o.clearBtn === true ? 'table-cell' : 'none'); - this.picker.find('thead .datepicker-title') - .text(this.o.title) - .css('display', typeof this.o.title === 'string' && this.o.title !== '' ? 'table-cell' : 'none'); - this.updateNavArrows(); - this.fillMonths(); - var prevMonth = UTCDate(year, month, 0), - day = prevMonth.getUTCDate(); - prevMonth.setUTCDate(day - (prevMonth.getUTCDay() - this.o.weekStart + 7)%7); - var nextMonth = new Date(prevMonth); - if (prevMonth.getUTCFullYear() < 100){ - nextMonth.setUTCFullYear(prevMonth.getUTCFullYear()); - } - nextMonth.setUTCDate(nextMonth.getUTCDate() + 42); - nextMonth = nextMonth.valueOf(); - var html = []; - var weekDay, clsName; - while (prevMonth.valueOf() < nextMonth){ - weekDay = prevMonth.getUTCDay(); - if (weekDay === this.o.weekStart){ - html.push(''); - if (this.o.calendarWeeks){ - // ISO 8601: First week contains first thursday. - // ISO also states week starts on Monday, but we can be more abstract here. - var - // Start of current week: based on weekstart/current date - ws = new Date(+prevMonth + (this.o.weekStart - weekDay - 7) % 7 * 864e5), - // Thursday of this week - th = new Date(Number(ws) + (7 + 4 - ws.getUTCDay()) % 7 * 864e5), - // First Thursday of year, year from thursday - yth = new Date(Number(yth = UTCDate(th.getUTCFullYear(), 0, 1)) + (7 + 4 - yth.getUTCDay()) % 7 * 864e5), - // Calendar week: ms between thursdays, div ms per day, div 7 days - calWeek = (th - yth) / 864e5 / 7 + 1; - html.push(''+ calWeek +''); - } - } - clsName = this.getClassNames(prevMonth); - clsName.push('day'); - - var content = prevMonth.getUTCDate(); - - if (this.o.beforeShowDay !== $.noop){ - before = this.o.beforeShowDay(this._utc_to_local(prevMonth)); - if (before === undefined) - before = {}; - else if (typeof before === 'boolean') - before = {enabled: before}; - else if (typeof before === 'string') - before = {classes: before}; - if (before.enabled === false) - clsName.push('disabled'); - if (before.classes) - clsName = clsName.concat(before.classes.split(/\s+/)); - if (before.tooltip) - tooltip = before.tooltip; - if (before.content) - content = before.content; - } - - //Check if uniqueSort exists (supported by jquery >=1.12 and >=2.2) - //Fallback to unique function for older jquery versions - if (typeof $.uniqueSort === "function") { - clsName = $.uniqueSort(clsName); - } else { - clsName = $.unique(clsName); - } - - html.push('' + content + ''); - tooltip = null; - if (weekDay === this.o.weekEnd){ - html.push(''); - } - prevMonth.setUTCDate(prevMonth.getUTCDate() + 1); - } - this.picker.find('.datepicker-days tbody').html(html.join('')); - - var monthsTitle = dates[this.o.language].monthsTitle || dates['en'].monthsTitle || 'Months'; - var months = this.picker.find('.datepicker-months') - .find('.datepicker-switch') - .text(this.o.maxViewMode < 2 ? monthsTitle : year) - .end() - .find('tbody span').removeClass('active'); - - $.each(this.dates, function(i, d){ - if (d.getUTCFullYear() === year) - months.eq(d.getUTCMonth()).addClass('active'); - }); - - if (year < startYear || year > endYear){ - months.addClass('disabled'); - } - if (year === startYear){ - months.slice(0, startMonth).addClass('disabled'); - } - if (year === endYear){ - months.slice(endMonth+1).addClass('disabled'); - } - - if (this.o.beforeShowMonth !== $.noop){ - var that = this; - $.each(months, function(i, month){ - var moDate = new Date(year, i, 1); - var before = that.o.beforeShowMonth(moDate); - if (before === undefined) - before = {}; - else if (typeof before === 'boolean') - before = {enabled: before}; - else if (typeof before === 'string') - before = {classes: before}; - if (before.enabled === false && !$(month).hasClass('disabled')) - $(month).addClass('disabled'); - if (before.classes) - $(month).addClass(before.classes); - if (before.tooltip) - $(month).prop('title', before.tooltip); - }); - } - - // Generating decade/years picker - this._fill_yearsView( - '.datepicker-years', - 'year', - 10, - year, - startYear, - endYear, - this.o.beforeShowYear - ); - - // Generating century/decades picker - this._fill_yearsView( - '.datepicker-decades', - 'decade', - 100, - year, - startYear, - endYear, - this.o.beforeShowDecade - ); - - // Generating millennium/centuries picker - this._fill_yearsView( - '.datepicker-centuries', - 'century', - 1000, - year, - startYear, - endYear, - this.o.beforeShowCentury - ); - }, - - updateNavArrows: function(){ - if (!this._allow_update) - return; - - var d = new Date(this.viewDate), - year = d.getUTCFullYear(), - month = d.getUTCMonth(), - startYear = this.o.startDate !== -Infinity ? this.o.startDate.getUTCFullYear() : -Infinity, - startMonth = this.o.startDate !== -Infinity ? this.o.startDate.getUTCMonth() : -Infinity, - endYear = this.o.endDate !== Infinity ? this.o.endDate.getUTCFullYear() : Infinity, - endMonth = this.o.endDate !== Infinity ? this.o.endDate.getUTCMonth() : Infinity, - prevIsDisabled, - nextIsDisabled, - factor = 1; - switch (this.viewMode){ - case 4: - factor *= 10; - /* falls through */ - case 3: - factor *= 10; - /* falls through */ - case 2: - factor *= 10; - /* falls through */ - case 1: - prevIsDisabled = Math.floor(year / factor) * factor <= startYear; - nextIsDisabled = Math.floor(year / factor) * factor + factor > endYear; - break; - case 0: - prevIsDisabled = year <= startYear && month <= startMonth; - nextIsDisabled = year >= endYear && month >= endMonth; - break; - } - - this.picker.find('.prev').toggleClass('disabled', prevIsDisabled); - this.picker.find('.next').toggleClass('disabled', nextIsDisabled); - }, - - click: function(e){ - e.preventDefault(); - e.stopPropagation(); - - var target, dir, day, year, month; - target = $(e.target); - - // Clicked on the switch - if (target.hasClass('datepicker-switch') && this.viewMode !== this.o.maxViewMode){ - this.setViewMode(this.viewMode + 1); - } - - // Clicked on today button - if (target.hasClass('today') && !target.hasClass('day')){ - this.setViewMode(0); - this._setDate(UTCToday(), this.o.todayBtn === 'linked' ? null : 'view'); - } - - // Clicked on clear button - if (target.hasClass('clear')){ - this.clearDates(); - } - - if (!target.hasClass('disabled')){ - // Clicked on a month, year, decade, century - if (target.hasClass('month') - || target.hasClass('year') - || target.hasClass('decade') - || target.hasClass('century')) { - this.viewDate.setUTCDate(1); - - day = 1; - if (this.viewMode === 1){ - month = target.parent().find('span').index(target); - year = this.viewDate.getUTCFullYear(); - this.viewDate.setUTCMonth(month); - } else { - month = 0; - year = Number(target.text()); - this.viewDate.setUTCFullYear(year); - } - - this._trigger(DPGlobal.viewModes[this.viewMode - 1].e, this.viewDate); - - if (this.viewMode === this.o.minViewMode){ - this._setDate(UTCDate(year, month, day)); - } else { - this.setViewMode(this.viewMode - 1); - this.fill(); - } - } - } - - if (this.picker.is(':visible') && this._focused_from){ - this._focused_from.focus(); - } - delete this._focused_from; - }, - - dayCellClick: function(e){ - var $target = $(e.currentTarget); - var timestamp = $target.data('date'); - var date = new Date(timestamp); - - if (this.o.updateViewDate) { - if (date.getUTCFullYear() !== this.viewDate.getUTCFullYear()) { - this._trigger('changeYear', this.viewDate); - } - - if (date.getUTCMonth() !== this.viewDate.getUTCMonth()) { - this._trigger('changeMonth', this.viewDate); - } - } - this._setDate(date); - }, - - // Clicked on prev or next - navArrowsClick: function(e){ - var $target = $(e.currentTarget); - var dir = $target.hasClass('prev') ? -1 : 1; - if (this.viewMode !== 0){ - dir *= DPGlobal.viewModes[this.viewMode].navStep * 12; - } - this.viewDate = this.moveMonth(this.viewDate, dir); - this._trigger(DPGlobal.viewModes[this.viewMode].e, this.viewDate); - this.fill(); - }, - - _toggle_multidate: function(date){ - var ix = this.dates.contains(date); - if (!date){ - this.dates.clear(); - } - - if (ix !== -1){ - if (this.o.multidate === true || this.o.multidate > 1 || this.o.toggleActive){ - this.dates.remove(ix); - } - } else if (this.o.multidate === false) { - this.dates.clear(); - this.dates.push(date); - } - else { - this.dates.push(date); - } - - if (typeof this.o.multidate === 'number') - while (this.dates.length > this.o.multidate) - this.dates.remove(0); - }, - - _setDate: function(date, which){ - if (!which || which === 'date') - this._toggle_multidate(date && new Date(date)); - if ((!which && this.o.updateViewDate) || which === 'view') - this.viewDate = date && new Date(date); - - this.fill(); - this.setValue(); - if (!which || which !== 'view') { - this._trigger('changeDate'); - } - this.inputField.trigger('change'); - if (this.o.autoclose && (!which || which === 'date')){ - this.hide(); - } - }, - - moveDay: function(date, dir){ - var newDate = new Date(date); - newDate.setUTCDate(date.getUTCDate() + dir); - - return newDate; - }, - - moveWeek: function(date, dir){ - return this.moveDay(date, dir * 7); - }, - - moveMonth: function(date, dir){ - if (!isValidDate(date)) - return this.o.defaultViewDate; - if (!dir) - return date; - var new_date = new Date(date.valueOf()), - day = new_date.getUTCDate(), - month = new_date.getUTCMonth(), - mag = Math.abs(dir), - new_month, test; - dir = dir > 0 ? 1 : -1; - if (mag === 1){ - test = dir === -1 - // If going back one month, make sure month is not current month - // (eg, Mar 31 -> Feb 31 == Feb 28, not Mar 02) - ? function(){ - return new_date.getUTCMonth() === month; - } - // If going forward one month, make sure month is as expected - // (eg, Jan 31 -> Feb 31 == Feb 28, not Mar 02) - : function(){ - return new_date.getUTCMonth() !== new_month; - }; - new_month = month + dir; - new_date.setUTCMonth(new_month); - // Dec -> Jan (12) or Jan -> Dec (-1) -- limit expected date to 0-11 - new_month = (new_month + 12) % 12; - } - else { - // For magnitudes >1, move one month at a time... - for (var i=0; i < mag; i++) - // ...which might decrease the day (eg, Jan 31 to Feb 28, etc)... - new_date = this.moveMonth(new_date, dir); - // ...then reset the day, keeping it in the new month - new_month = new_date.getUTCMonth(); - new_date.setUTCDate(day); - test = function(){ - return new_month !== new_date.getUTCMonth(); - }; - } - // Common date-resetting loop -- if date is beyond end of month, make it - // end of month - while (test()){ - new_date.setUTCDate(--day); - new_date.setUTCMonth(new_month); - } - return new_date; - }, - - moveYear: function(date, dir){ - return this.moveMonth(date, dir*12); - }, - - moveAvailableDate: function(date, dir, fn){ - do { - date = this[fn](date, dir); - - if (!this.dateWithinRange(date)) - return false; - - fn = 'moveDay'; - } - while (this.dateIsDisabled(date)); - - return date; - }, - - weekOfDateIsDisabled: function(date){ - return $.inArray(date.getUTCDay(), this.o.daysOfWeekDisabled) !== -1; - }, - - dateIsDisabled: function(date){ - return ( - this.weekOfDateIsDisabled(date) || - $.grep(this.o.datesDisabled, function(d){ - return isUTCEquals(date, d); - }).length > 0 - ); - }, - - dateWithinRange: function(date){ - return date >= this.o.startDate && date <= this.o.endDate; - }, - - keydown: function(e){ - if (!this.picker.is(':visible')){ - if (e.keyCode === 40 || e.keyCode === 27) { // allow down to re-show picker - this.show(); - e.stopPropagation(); - } - return; - } - var dateChanged = false, - dir, newViewDate, - focusDate = this.focusDate || this.viewDate; - switch (e.keyCode){ - case 27: // escape - if (this.focusDate){ - this.focusDate = null; - this.viewDate = this.dates.get(-1) || this.viewDate; - this.fill(); - } - else - this.hide(); - e.preventDefault(); - e.stopPropagation(); - break; - case 37: // left - case 38: // up - case 39: // right - case 40: // down - if (!this.o.keyboardNavigation || this.o.daysOfWeekDisabled.length === 7) - break; - dir = e.keyCode === 37 || e.keyCode === 38 ? -1 : 1; - if (this.viewMode === 0) { - if (e.ctrlKey){ - newViewDate = this.moveAvailableDate(focusDate, dir, 'moveYear'); - - if (newViewDate) - this._trigger('changeYear', this.viewDate); - } else if (e.shiftKey){ - newViewDate = this.moveAvailableDate(focusDate, dir, 'moveMonth'); - - if (newViewDate) - this._trigger('changeMonth', this.viewDate); - } else if (e.keyCode === 37 || e.keyCode === 39){ - newViewDate = this.moveAvailableDate(focusDate, dir, 'moveDay'); - } else if (!this.weekOfDateIsDisabled(focusDate)){ - newViewDate = this.moveAvailableDate(focusDate, dir, 'moveWeek'); - } - } else if (this.viewMode === 1) { - if (e.keyCode === 38 || e.keyCode === 40) { - dir = dir * 4; - } - newViewDate = this.moveAvailableDate(focusDate, dir, 'moveMonth'); - } else if (this.viewMode === 2) { - if (e.keyCode === 38 || e.keyCode === 40) { - dir = dir * 4; - } - newViewDate = this.moveAvailableDate(focusDate, dir, 'moveYear'); - } - if (newViewDate){ - this.focusDate = this.viewDate = newViewDate; - this.setValue(); - this.fill(); - e.preventDefault(); - } - break; - case 13: // enter - if (!this.o.forceParse) - break; - focusDate = this.focusDate || this.dates.get(-1) || this.viewDate; - if (this.o.keyboardNavigation) { - this._toggle_multidate(focusDate); - dateChanged = true; - } - this.focusDate = null; - this.viewDate = this.dates.get(-1) || this.viewDate; - this.setValue(); - this.fill(); - if (this.picker.is(':visible')){ - e.preventDefault(); - e.stopPropagation(); - if (this.o.autoclose) - this.hide(); - } - break; - case 9: // tab - this.focusDate = null; - this.viewDate = this.dates.get(-1) || this.viewDate; - this.fill(); - this.hide(); - break; - } - if (dateChanged){ - if (this.dates.length) - this._trigger('changeDate'); - else - this._trigger('clearDate'); - this.inputField.trigger('change'); - } - }, - - setViewMode: function(viewMode){ - this.viewMode = viewMode; - this.picker - .children('div') - .hide() - .filter('.datepicker-' + DPGlobal.viewModes[this.viewMode].clsName) - .show(); - this.updateNavArrows(); - this._trigger('changeViewMode', new Date(this.viewDate)); - } - }; - - var DateRangePicker = function(element, options){ - $.data(element, 'datepicker', this); - this.element = $(element); - this.inputs = $.map(options.inputs, function(i){ - return i.jquery ? i[0] : i; - }); - delete options.inputs; - - this.keepEmptyValues = options.keepEmptyValues; - delete options.keepEmptyValues; - - datepickerPlugin.call($(this.inputs), options) - .on('changeDate', $.proxy(this.dateUpdated, this)); - - this.pickers = $.map(this.inputs, function(i){ - return $.data(i, 'datepicker'); - }); - this.updateDates(); - }; - DateRangePicker.prototype = { - updateDates: function(){ - this.dates = $.map(this.pickers, function(i){ - return i.getUTCDate(); - }); - this.updateRanges(); - }, - updateRanges: function(){ - var range = $.map(this.dates, function(d){ - return d.valueOf(); - }); - $.each(this.pickers, function(i, p){ - p.setRange(range); - }); - }, - clearDates: function(){ - $.each(this.pickers, function(i, p){ - p.clearDates(); - }); - }, - dateUpdated: function(e){ - // `this.updating` is a workaround for preventing infinite recursion - // between `changeDate` triggering and `setUTCDate` calling. Until - // there is a better mechanism. - if (this.updating) - return; - this.updating = true; - - var dp = $.data(e.target, 'datepicker'); - - if (dp === undefined) { - return; - } - - var new_date = dp.getUTCDate(), - keep_empty_values = this.keepEmptyValues, - i = $.inArray(e.target, this.inputs), - j = i - 1, - k = i + 1, - l = this.inputs.length; - if (i === -1) - return; - - $.each(this.pickers, function(i, p){ - if (!p.getUTCDate() && (p === dp || !keep_empty_values)) - p.setUTCDate(new_date); - }); - - if (new_date < this.dates[j]){ - // Date being moved earlier/left - while (j >= 0 && new_date < this.dates[j]){ - this.pickers[j--].setUTCDate(new_date); - } - } else if (new_date > this.dates[k]){ - // Date being moved later/right - while (k < l && new_date > this.dates[k]){ - this.pickers[k++].setUTCDate(new_date); - } - } - this.updateDates(); - - delete this.updating; - }, - destroy: function(){ - $.map(this.pickers, function(p){ p.destroy(); }); - $(this.inputs).off('changeDate', this.dateUpdated); - delete this.element.data().datepicker; - }, - remove: alias('destroy', 'Method `remove` is deprecated and will be removed in version 2.0. Use `destroy` instead') - }; - - function opts_from_el(el, prefix){ - // Derive options from element data-attrs - var data = $(el).data(), - out = {}, inkey, - replace = new RegExp('^' + prefix.toLowerCase() + '([A-Z])'); - prefix = new RegExp('^' + prefix.toLowerCase()); - function re_lower(_,a){ - return a.toLowerCase(); - } - for (var key in data) - if (prefix.test(key)){ - inkey = key.replace(replace, re_lower); - out[inkey] = data[key]; - } - return out; - } - - function opts_from_locale(lang){ - // Derive options from locale plugins - var out = {}; - // Check if "de-DE" style date is available, if not language should - // fallback to 2 letter code eg "de" - if (!dates[lang]){ - lang = lang.split('-')[0]; - if (!dates[lang]) - return; - } - var d = dates[lang]; - $.each(locale_opts, function(i,k){ - if (k in d) - out[k] = d[k]; - }); - return out; - } - - var old = $.fn.datepicker; - var datepickerPlugin = function(option){ - var args = Array.apply(null, arguments); - args.shift(); - var internal_return; - this.each(function(){ - var $this = $(this), - data = $this.data('datepicker'), - options = typeof option === 'object' && option; - if (!data){ - var elopts = opts_from_el(this, 'date'), - // Preliminary options - xopts = $.extend({}, defaults, elopts, options), - locopts = opts_from_locale(xopts.language), - // Options priority: js args, data-attrs, locales, defaults - opts = $.extend({}, defaults, locopts, elopts, options); - if ($this.hasClass('input-daterange') || opts.inputs){ - $.extend(opts, { - inputs: opts.inputs || $this.find('input').toArray() - }); - data = new DateRangePicker(this, opts); - } - else { - data = new Datepicker(this, opts); - } - $this.data('datepicker', data); - } - if (typeof option === 'string' && typeof data[option] === 'function'){ - internal_return = data[option].apply(data, args); - } - }); - - if ( - internal_return === undefined || - internal_return instanceof Datepicker || - internal_return instanceof DateRangePicker - ) - return this; - - if (this.length > 1) - throw new Error('Using only allowed for the collection of a single element (' + option + ' function)'); - else - return internal_return; - }; - $.fn.datepicker = datepickerPlugin; - - var defaults = $.fn.datepicker.defaults = { - assumeNearbyYear: false, - autoclose: false, - beforeShowDay: $.noop, - beforeShowMonth: $.noop, - beforeShowYear: $.noop, - beforeShowDecade: $.noop, - beforeShowCentury: $.noop, - calendarWeeks: false, - clearBtn: false, - toggleActive: false, - daysOfWeekDisabled: [], - daysOfWeekHighlighted: [], - datesDisabled: [], - endDate: Infinity, - forceParse: true, - format: 'mm/dd/yyyy', - keepEmptyValues: false, - keyboardNavigation: true, - language: 'en', - minViewMode: 0, - maxViewMode: 4, - multidate: false, - multidateSeparator: ',', - orientation: "auto", - rtl: false, - startDate: -Infinity, - startView: 0, - todayBtn: false, - todayHighlight: false, - updateViewDate: true, - weekStart: 0, - disableTouchKeyboard: false, - enableOnReadonly: true, - showOnFocus: true, - zIndexOffset: 10, - container: 'body', - immediateUpdates: false, - title: '', - templates: { - leftArrow: '«', - rightArrow: '»' - }, - showWeekDays: true - }; - var locale_opts = $.fn.datepicker.locale_opts = [ - 'format', - 'rtl', - 'weekStart' - ]; - $.fn.datepicker.Constructor = Datepicker; - var dates = $.fn.datepicker.dates = { - en: { - days: ["Sunday", "Monday", "Tuesday", "Wednesday", "Thursday", "Friday", "Saturday"], - daysShort: ["Sun", "Mon", "Tue", "Wed", "Thu", "Fri", "Sat"], - daysMin: ["Su", "Mo", "Tu", "We", "Th", "Fr", "Sa"], - months: ["January", "February", "March", "April", "May", "June", "July", "August", "September", "October", "November", "December"], - monthsShort: ["Jan", "Feb", "Mar", "Apr", "May", "Jun", "Jul", "Aug", "Sep", "Oct", "Nov", "Dec"], - today: "Today", - clear: "Clear", - titleFormat: "MM yyyy" - } - }; - - var DPGlobal = { - viewModes: [ - { - names: ['days', 'month'], - clsName: 'days', - e: 'changeMonth' - }, - { - names: ['months', 'year'], - clsName: 'months', - e: 'changeYear', - navStep: 1 - }, - { - names: ['years', 'decade'], - clsName: 'years', - e: 'changeDecade', - navStep: 10 - }, - { - names: ['decades', 'century'], - clsName: 'decades', - e: 'changeCentury', - navStep: 100 - }, - { - names: ['centuries', 'millennium'], - clsName: 'centuries', - e: 'changeMillennium', - navStep: 1000 - } - ], - validParts: /dd?|DD?|mm?|MM?|yy(?:yy)?/g, - nonpunctuation: /[^ -\/:-@\u5e74\u6708\u65e5\[-`{-~\t\n\r]+/g, - parseFormat: function(format){ - if (typeof format.toValue === 'function' && typeof format.toDisplay === 'function') - return format; - // IE treats \0 as a string end in inputs (truncating the value), - // so it's a bad format delimiter, anyway - var separators = format.replace(this.validParts, '\0').split('\0'), - parts = format.match(this.validParts); - if (!separators || !separators.length || !parts || parts.length === 0){ - throw new Error("Invalid date format."); - } - return {separators: separators, parts: parts}; - }, - parseDate: function(date, format, language, assumeNearby){ - if (!date) - return undefined; - if (date instanceof Date) - return date; - if (typeof format === 'string') - format = DPGlobal.parseFormat(format); - if (format.toValue) - return format.toValue(date, format, language); - var fn_map = { - d: 'moveDay', - m: 'moveMonth', - w: 'moveWeek', - y: 'moveYear' - }, - dateAliases = { - yesterday: '-1d', - today: '+0d', - tomorrow: '+1d' - }, - parts, part, dir, i, fn; - if (date in dateAliases){ - date = dateAliases[date]; - } - if (/^[\-+]\d+[dmwy]([\s,]+[\-+]\d+[dmwy])*$/i.test(date)){ - parts = date.match(/([\-+]\d+)([dmwy])/gi); - date = new Date(); - for (i=0; i < parts.length; i++){ - part = parts[i].match(/([\-+]\d+)([dmwy])/i); - dir = Number(part[1]); - fn = fn_map[part[2].toLowerCase()]; - date = Datepicker.prototype[fn](date, dir); - } - return Datepicker.prototype._zero_utc_time(date); - } - - parts = date && date.match(this.nonpunctuation) || []; - - function applyNearbyYear(year, threshold){ - if (threshold === true) - threshold = 10; - - // if year is 2 digits or less, than the user most likely is trying to get a recent century - if (year < 100){ - year += 2000; - // if the new year is more than threshold years in advance, use last century - if (year > ((new Date()).getFullYear()+threshold)){ - year -= 100; - } - } - - return year; - } - - var parsed = {}, - setters_order = ['yyyy', 'yy', 'M', 'MM', 'm', 'mm', 'd', 'dd'], - setters_map = { - yyyy: function(d,v){ - return d.setUTCFullYear(assumeNearby ? applyNearbyYear(v, assumeNearby) : v); - }, - m: function(d,v){ - if (isNaN(d)) - return d; - v -= 1; - while (v < 0) v += 12; - v %= 12; - d.setUTCMonth(v); - while (d.getUTCMonth() !== v) - d.setUTCDate(d.getUTCDate()-1); - return d; - }, - d: function(d,v){ - return d.setUTCDate(v); - } - }, - val, filtered; - setters_map['yy'] = setters_map['yyyy']; - setters_map['M'] = setters_map['MM'] = setters_map['mm'] = setters_map['m']; - setters_map['dd'] = setters_map['d']; - date = UTCToday(); - var fparts = format.parts.slice(); - // Remove noop parts - if (parts.length !== fparts.length){ - fparts = $(fparts).filter(function(i,p){ - return $.inArray(p, setters_order) !== -1; - }).toArray(); - } - // Process remainder - function match_part(){ - var m = this.slice(0, parts[i].length), - p = parts[i].slice(0, m.length); - return m.toLowerCase() === p.toLowerCase(); - } - if (parts.length === fparts.length){ - var cnt; - for (i=0, cnt = fparts.length; i < cnt; i++){ - val = parseInt(parts[i], 10); - part = fparts[i]; - if (isNaN(val)){ - switch (part){ - case 'MM': - filtered = $(dates[language].months).filter(match_part); - val = $.inArray(filtered[0], dates[language].months) + 1; - break; - case 'M': - filtered = $(dates[language].monthsShort).filter(match_part); - val = $.inArray(filtered[0], dates[language].monthsShort) + 1; - break; - } - } - parsed[part] = val; - } - var _date, s; - for (i=0; i < setters_order.length; i++){ - s = setters_order[i]; - if (s in parsed && !isNaN(parsed[s])){ - _date = new Date(date); - setters_map[s](_date, parsed[s]); - if (!isNaN(_date)) - date = _date; - } - } - } - return date; - }, - formatDate: function(date, format, language){ - if (!date) - return ''; - if (typeof format === 'string') - format = DPGlobal.parseFormat(format); - if (format.toDisplay) - return format.toDisplay(date, format, language); - var val = { - d: date.getUTCDate(), - D: dates[language].daysShort[date.getUTCDay()], - DD: dates[language].days[date.getUTCDay()], - m: date.getUTCMonth() + 1, - M: dates[language].monthsShort[date.getUTCMonth()], - MM: dates[language].months[date.getUTCMonth()], - yy: date.getUTCFullYear().toString().substring(2), - yyyy: date.getUTCFullYear() - }; - val.dd = (val.d < 10 ? '0' : '') + val.d; - val.mm = (val.m < 10 ? '0' : '') + val.m; - date = []; - var seps = $.extend([], format.separators); - for (var i=0, cnt = format.parts.length; i <= cnt; i++){ - if (seps.length) - date.push(seps.shift()); - date.push(val[format.parts[i]]); - } - return date.join(''); - }, - headTemplate: ''+ - ''+ - ''+ - ''+ - ''+ - ''+defaults.templates.leftArrow+''+ - ''+ - ''+defaults.templates.rightArrow+''+ - ''+ - '', - contTemplate: '', - footTemplate: ''+ - ''+ - ''+ - ''+ - ''+ - ''+ - ''+ - '' - }; - DPGlobal.template = '
      '+ - '
      '+ - ''+ - DPGlobal.headTemplate+ - ''+ - DPGlobal.footTemplate+ - '
      '+ - '
      '+ - '
      '+ - ''+ - DPGlobal.headTemplate+ - DPGlobal.contTemplate+ - DPGlobal.footTemplate+ - '
      '+ - '
      '+ - '
      '+ - ''+ - DPGlobal.headTemplate+ - DPGlobal.contTemplate+ - DPGlobal.footTemplate+ - '
      '+ - '
      '+ - '
      '+ - ''+ - DPGlobal.headTemplate+ - DPGlobal.contTemplate+ - DPGlobal.footTemplate+ - '
      '+ - '
      '+ - '
      '+ - ''+ - DPGlobal.headTemplate+ - DPGlobal.contTemplate+ - DPGlobal.footTemplate+ - '
      '+ - '
      '+ - '
      '; - - $.fn.datepicker.DPGlobal = DPGlobal; - - - /* DATEPICKER NO CONFLICT - * =================== */ - - $.fn.datepicker.noConflict = function(){ - $.fn.datepicker = old; - return this; - }; - - /* DATEPICKER VERSION - * =================== */ - $.fn.datepicker.version = '1.9.0'; - - $.fn.datepicker.deprecated = function(msg){ - var console = window.console; - if (console && console.warn) { - console.warn('DEPRECATED: ' + msg); - } - }; - - - /* DATEPICKER DATA-API - * ================== */ + } + if (el.dataset.dateMinViewMode && view_mode[el.dataset.dateMinViewMode]) { + options = { ...options, + startView: view_mode[el.dataset.dateMinViewMode] + }; + } + if (el.dataset.dateViewMode && view_mode[el.dataset.dateViewMode]) { + options = { ...options, + maxView: view_mode[el.dataset.dateViewMode] + }; + } + if (el.dataset.dateAutoclose) { + options = { ...options, + autohide: el.dataset.dateAutoclose + }; + } - $(document).on( - 'focus.datepicker.data-api click.datepicker.data-api', - '[data-provide="datepicker"]', - function(e){ - var $this = $(this); - if ($this.data('datepicker')) - return; - e.preventDefault(); - // component click requires us to explicitly show it - datepickerPlugin.call($this, 'show'); - } - ); - $(function(){ - datepickerPlugin.call($('[data-provide="datepicker-inline"]')); - }); + new Datepicker(el, options); +} -})); +document.addEventListener("DOMContentLoaded", function () { + const elems = document.querySelectorAll('[data-provide="datepicker"]'); + elems.forEach(el => enable_datepicker(el)); +}); diff --git a/ietf/static/js/document_html.js b/ietf/static/js/document_html.js index 9da32a0a7d..3e609f3965 100644 --- a/ietf/static/js/document_html.js +++ b/ietf/static/js/document_html.js @@ -42,20 +42,25 @@ document.addEventListener("DOMContentLoaded", function (event) { // Set up a nav pane const toc_pane = document.getElementById("toc-nav"); - populate_nav(toc_pane, - `#content h2, #content h3, #content h4, #content h5, #content h6 - #content .h1, #content .h2, #content .h3, #content .h4, #content .h5, #content .h6`, - ["py-0"]); + const headings = document.querySelectorAll(`#content :is(h2, h3, h4, h5, h6, .h2, .h3, .h4, .h5, .h6)`); + populate_nav(toc_pane, headings, ["py-0"]); // activate pref buttons selected by pref cookies or localStorage const in_localStorage = ["deftab", "reflinks"]; + const btn_pref = { + "sidebar": "on", + "deftab": "docinfo", + "htmlconf": "html", + "pagedeps": "reference", + "reflinks": "refsection" + }; document.querySelectorAll("#pref-tab-pane .btn-check") .forEach(btn => { const id = btn.id.replace("-radio", ""); const val = in_localStorage.includes(btn.name) ? localStorage.getItem(btn.name) : cookies.get(btn.name); - if (val == id) { + if (val === id || (val === null && btn_pref[btn.name] === id)) { btn.checked = true; } @@ -112,4 +117,83 @@ document.addEventListener("DOMContentLoaded", function (event) { } }); } + + // Rewrite these CSS properties so that the values are available for restyling. + document.querySelectorAll("svg [style]").forEach(el => { + // Push these CSS properties into their own attributes + const SVG_PRESENTATION_ATTRS = new Set([ + 'alignment-baseline', 'baseline-shift', 'clip', 'clip-path', 'clip-rule', + 'color', 'color-interpolation', 'color-interpolation-filters', + 'color-rendering', 'cursor', 'direction', 'display', 'dominant-baseline', + 'fill', 'fill-opacity', 'fill-rule', 'filter', 'flood-color', + 'flood-opacity', 'font-family', 'font-size', 'font-size-adjust', + 'font-stretch', 'font-style', 'font-variant', 'font-weight', + 'image-rendering', 'letter-spacing', 'lighting-color', 'marker-end', + 'marker-mid', 'marker-start', 'mask', 'opacity', 'overflow', 'paint-order', + 'pointer-events', 'shape-rendering', 'stop-color', 'stop-opacity', + 'stroke', 'stroke-dasharray', 'stroke-dashoffset', 'stroke-linecap', + 'stroke-linejoin', 'stroke-miterlimit', 'stroke-opacity', 'stroke-width', + 'text-anchor', 'text-decoration', 'text-rendering', 'unicode-bidi', + 'vector-effect', 'visibility', 'word-spacing', 'writing-mode', + ]); + + // Simple CSS splitter: respects quoted strings and parens so semicolons + // inside url(...) or "..." don't get treated as declaration boundaries. + function parseDeclarations(styleText) { + const decls = []; + let buf = ''; + let inStr = false; + let strChar = ''; + let escaped = false; + let depth = 0; + + for (const ch of styleText) { + if (inStr) { + if (escaped) { + escaped = false; + } else if (ch === '\\') { + escaped = true; + } else if (ch === strChar) { + inStr = false; + } + } else if (ch === '"' || ch === "'") { + inStr = true; + strChar = ch; + } else if (ch === '(') { + depth++; + } else if (ch === ')') { + depth--; + } else if (ch === ';' && depth === 0) { + const trimmed = buf.trim(); + if (trimmed) { + decls.push(trimmed); + } + buf = ''; + continue; + } + buf += ch; + } + const trimmed = buf.trim(); + if (trimmed) { + decls.push(trimmed); + } + return decls; + } + + const remainder = []; + for (const decl of parseDeclarations(el.getAttribute('style'))) { + const [prop, val] = decl.split(":", 2).map(v => v.trim()); + if (val && !/!important$/.test(val) && SVG_PRESENTATION_ATTRS.has(prop)) { + el.setAttribute(prop, val); + } else { + remainder.push(decl); + } + } + + if (remainder.length > 0) { + el.setAttribute('style', remainder.join('; ')); + } else { + el.removeAttribute('style'); + } + }); }); diff --git a/ietf/static/js/document_timeline.js b/ietf/static/js/document_timeline.js index babde0deeb..d8532c3623 100644 --- a/ietf/static/js/document_timeline.js +++ b/ietf/static/js/document_timeline.js @@ -86,7 +86,7 @@ function scale_x() { } function update_x_axis() { - d3.select("#timeline svg .x.axis") + d3.select("#doc-timeline svg .x.axis") .call(x_axis) .selectAll("text") .style("text-anchor", "end") @@ -96,7 +96,7 @@ function update_x_axis() { function update_timeline() { bar_y = {}; scale_x(); - var chart = d3.select("#timeline svg") + var chart = d3.select("#doc-timeline svg") .attr("width", width); // enter data (skip the last pseudo entry) var bar = chart.selectAll("g") @@ -111,12 +111,12 @@ function draw_timeline() { bar_height = parseFloat($("body") .css("line-height")); - var div = $("#timeline"); + var div = $("#doc-timeline"); div.addClass("my-3"); if (div.is(":empty")) { div.append(""); } - var chart = d3.select("#timeline svg") + var chart = d3.select("#doc-timeline svg") .attr("width", width); var defs = chart.append("defs"); @@ -249,7 +249,7 @@ d3.json("doc.json") published: expiration_date(data[data.length - 1]) }); - width = $("#timeline") + width = $("#doc-timeline") .width(); draw_timeline(); } @@ -258,11 +258,11 @@ d3.json("doc.json") $(window) .on({ resize: function () { - var g = $("#timeline svg"); + var g = $("#doc-timeline svg"); g.remove(); - width = $("#timeline") + width = $("#doc-timeline") .width(); - $("#timeline") + $("#doc-timeline") .append(g); update_timeline(); } diff --git a/ietf/static/js/draft-submit.js b/ietf/static/js/draft-submit.js index d3657a8377..38ac7eb263 100644 --- a/ietf/static/js/draft-submit.js +++ b/ietf/static/js/draft-submit.js @@ -1,69 +1,96 @@ -$(document) - .ready(function () { - // fill in submitter info when an author button is clicked - $("form.idsubmit button.author") - .on("click", function () { - var name = $(this) - .data("name"); - var email = $(this) - .data("email"); +$(function () { + // fill in submitter info when an author button is clicked + $("form.idsubmit button.author") + .on("click", function () { + var name = $(this) + .data("name"); + var email = $(this) + .data("email"); - $(this) - .parents("form") - .find("input[name=submitter-name]") - .val(name || ""); - $(this) - .parents("form") - .find("input[name=submitter-email]") - .val(email || ""); - }); + $(this) + .parents("form") + .find("input[name=submitter-name]") + .val(name || ""); + $(this) + .parents("form") + .find("input[name=submitter-email]") + .val(email || ""); + }); - $("form.idsubmit") - .on("submit", function () { - if (this.submittedAlready) - return false; - else { - this.submittedAlready = true; - return true; - } - }); + $("form.idsubmit") + .on("submit", function () { + if (this.submittedAlready) + return false; + else { + this.submittedAlready = true; + return true; + } + }); - $("form.idsubmit #add-author") - .on("click", function () { - // clone the last author block and make it empty - var cloner = $("#cloner"); - var next = cloner.clone(); - next.find('input:not([type=hidden])') - .val(''); + $("form.idsubmit #add-author") + .on("click", function () { + // clone the last author block and make it empty + var cloner = $("#cloner"); + var next = cloner.clone(); + next.find('input:not([type=hidden])') + .val(''); - // find the author number - var t = next.children('h3') - .text(); - var n = parseInt(t.replace(/\D/g, '')); + // find the author number + var t = next.children('h3') + .text(); + var n = parseInt(t.replace(/\D/g, '')); - // change the number in attributes and text - next.find('*') - .each(function () { - var e = this; - $.each(['id', 'for', 'name', 'value'], function (i, v) { - if ($(e) - .attr(v)) { - $(e) - .attr(v, $(e) - .attr(v) - .replace(n - 1, n)); - } - }); + // change the number in attributes and text + next.find('*') + .each(function () { + var e = this; + $.each(['id', 'for', 'name', 'value'], function (i, v) { + if ($(e) + .attr(v)) { + $(e) + .attr(v, $(e) + .attr(v) + .replace(n - 1, n)); + } }); + }); + + t = t.replace(n, n + 1); + next.children('h3') + .text(t); - t = t.replace(n, n + 1); - next.children('h3') - .text(t); + // move the cloner id to next and insert next into the DOM + cloner.removeAttr('id'); + next.attr('id', 'cloner'); + next.insertAfter(cloner); - // move the cloner id to next and insert next into the DOM - cloner.removeAttr('id'); - next.attr('id', 'cloner'); - next.insertAfter(cloner); + }); - }); - }); + // If draft is validating, poll until validation is complete, then reload the page + const submissionValidatingAlert = document.getElementById('submission-validating-alert'); + if (submissionValidatingAlert) { + let statusPollTimer; + const statusUrl = submissionValidatingAlert.dataset['submissionStatusUrl']; + let statusPollInterval = 2000; // ms + const maxPollInterval = 32000; // ms + + function checkStatus() { + if (statusPollInterval < maxPollInterval) { + statusPollInterval *= 2; + } + const xhr = new XMLHttpRequest(); + xhr.open("GET", statusUrl, true); + xhr.onload = (e) => { + if (xhr.response && xhr.response.state !== 'validating') { + location.reload(); + } else { + statusPollTimer = setTimeout(checkStatus, statusPollInterval); + } + }; + xhr.onerror = (e) => {statusPollTimer = setTimeout(checkStatus, statusPollInterval);}; + xhr.responseType = 'json'; + xhr.send(''); + } + statusPollTimer = setTimeout(checkStatus, statusPollInterval); + } +}); diff --git a/ietf/static/js/edit-meeting-schedule.js b/ietf/static/js/edit-meeting-schedule.js index cf0bd364cf..2a73a8c29d 100644 --- a/ietf/static/js/edit-meeting-schedule.js +++ b/ietf/static/js/edit-meeting-schedule.js @@ -50,6 +50,7 @@ $(function () { let sessionPurposeInputs = schedEditor.find('.session-purpose-toggles input'); let timeSlotGroupInputs = schedEditor.find("#timeslot-group-toggles-modal .modal-body .individual-timeslots input"); let sessionParentInputs = schedEditor.find(".session-parent-toggles input"); + let sessionParentToggleAll = schedEditor.find(".session-parent-toggles .session-parent-toggle-all") const classes_to_hide = '.hidden-timeslot-group,.hidden-timeslot-type'; // hack to work around lack of position sticky support in old browsers, see https://caniuse.com/#feat=css-sticky @@ -487,13 +488,13 @@ $(function () { // Disable a particular swap modal radio input let updateSwapRadios = function (labels, radios, disableValue, datePrecision) { - labels.removeClass('text-muted'); + labels.removeClass('text-body-secondary'); radios.prop('disabled', false); radios.prop('checked', false); // disable the input requested by value let disableInput = radios.filter('[value="' + disableValue + '"]'); if (disableInput) { - disableInput.parent().addClass('text-muted'); + disableInput.parent().addClass('text-body-secondary'); disableInput.prop('disabled', true); } if (officialSchedule) { @@ -502,7 +503,7 @@ $(function () { const past_radios = radios.filter( (_, radio) => parseISOTimestamp(radio.closest('*[data-start]').dataset.start).isSameOrBefore(now, datePrecision) ); - past_radios.parent().addClass('text-muted'); + past_radios.parent().addClass('text-body-secondary'); past_radios.prop('disabled', true); } return disableInput; // return the input that was specifically disabled, if any @@ -647,12 +648,9 @@ $(function () { function updateTimeSlotDurationViolations() { timeslots.each(function () { - let total = 0; - jQuery(this).find(".session").each(function () { - total += +jQuery(this).data("duration"); - }); - - jQuery(this).toggleClass("overfull", total > +jQuery(this).data("duration")); + const sessionsInSlot = Array.from(this.getElementsByClassName('session')); + const requiredDuration = Math.max(sessionsInSlot.map(elt => Number(elt.dataset.duration))); + this.classList.toggle('overfull', requiredDuration > Number(this.dataset.duration)); }); } @@ -772,6 +770,17 @@ $(function () { sessionParentInputs.on("click", updateSessionParentToggling); updateSessionParentToggling(); + // Toggle _all_ session parents + function toggleAllSessionParents() { + if (sessionParentInputs.filter(":checked").length < sessionParentInputs.length) { + sessionParentInputs.prop("checked", true); + } else { + sessionParentInputs.prop("checked", false); + } + updateSessionParentToggling(); + } + sessionParentToggleAll.on("click", toggleAllSessionParents); + // Toggling timeslot types function updateTimeSlotTypeToggling() { const checkedTypes = jQuery.map(timeSlotTypeInputs.filter(":checked"), elt => elt.value); @@ -862,10 +871,10 @@ $(function () { .not('.hidden') .length === 0) { purpose_input.setAttribute('disabled', 'disabled'); - purpose_input.closest('.session-purpose-toggle').classList.add('text-muted'); + purpose_input.closest('.session-purpose-toggle').classList.add('text-body-secondary'); } else { purpose_input.removeAttribute('disabled'); - purpose_input.closest('.session-purpose-toggle').classList.remove('text-muted'); + purpose_input.closest('.session-purpose-toggle').classList.remove('text-body-secondary'); } }); } @@ -1023,4 +1032,4 @@ $(function () { .on("mouseleave", ".other-session", function () { sessions.filter("#session" + this.dataset.othersessionid).removeClass("highlight"); }); -}); \ No newline at end of file +}); diff --git a/ietf/static/js/edit-milestones.js b/ietf/static/js/edit-milestones.js index 2a7e2ce2ae..2b64900d6c 100644 --- a/ietf/static/js/edit-milestones.js +++ b/ietf/static/js/edit-milestones.js @@ -134,6 +134,11 @@ $(document) window.setupSelect2Field($(this)); // from select2-field.js }); + new_edit_milestone.find("[data-provide='datepicker']") + .each(function () { + enable_datepicker($(this)[0]); // from datepicker.js + }); + if (!group_uses_milestone_dates) { setOrderControlValue(); } @@ -231,4 +236,4 @@ $(document) var el = document.getElementById('dragdropcontainer'); Sortable.create(el, options); } - }); \ No newline at end of file + }); diff --git a/ietf/static/js/fullcalendar.js b/ietf/static/js/fullcalendar.js index 0d58a24e70..dfdad730e9 100644 --- a/ietf/static/js/fullcalendar.js +++ b/ietf/static/js/fullcalendar.js @@ -1,5 +1,9 @@ import { Calendar } from '@fullcalendar/core'; import dayGridPlugin from '@fullcalendar/daygrid'; +import iCalendarPlugin from '@fullcalendar/icalendar'; +import bootstrap5Plugin from '@fullcalendar/bootstrap5'; global.FullCalendar = Calendar; -global.dayGridPlugin = dayGridPlugin; \ No newline at end of file +global.dayGridPlugin = dayGridPlugin; +global.iCalendarPlugin = iCalendarPlugin; +global.bootstrap5Plugin = bootstrap5Plugin; diff --git a/ietf/static/js/highcharts.js b/ietf/static/js/highcharts.js index f9b7aa6154..6c3b68051f 100644 --- a/ietf/static/js/highcharts.js +++ b/ietf/static/js/highcharts.js @@ -3,11 +3,113 @@ import Highcharts from "highcharts"; import Highcharts_Exporting from "highcharts/modules/exporting"; import Highcharts_Offline_Exporting from "highcharts/modules/offline-exporting"; import Highcharts_Export_Data from "highcharts/modules/export-data"; -import Highcharts_Accessibility from"highcharts/modules/accessibility"; +import Highcharts_Accessibility from "highcharts/modules/accessibility"; +import Highcharts_Sunburst from "highcharts/modules/sunburst"; + +document.documentElement.style.setProperty("--highcharts-background-color", "transparent"); Highcharts_Exporting(Highcharts); Highcharts_Offline_Exporting(Highcharts); Highcharts_Export_Data(Highcharts); Highcharts_Accessibility(Highcharts); +Highcharts_Sunburst(Highcharts); + +Highcharts.setOptions({ + chart: { + height: "100%", + styledMode: true, + }, + credits: { + enabled: false + }, +}); window.Highcharts = Highcharts; + +window.group_stats = function (url, chart_selector) { + $.getJSON(url, function (data) { + $(chart_selector) + .each(function (_, e) { + const dataset = e.dataset.dataset; + if (!dataset) { + console.log("dataset data attribute not set"); + return; + } + const area = e.dataset.area; + if (!area) { + console.log("area data attribute not set"); + return; + } + + const chart = Highcharts.chart(e, { + title: { + text: `${dataset == "docs" ? "Documents" : "Pages"} in ${area.toUpperCase()}` + }, + series: [{ + type: "sunburst", + data: [], + tooltip: { + pointFormatter: function () { + return `There ${this.value == 1 ? "is" : "are"} ${this.value} ${dataset == "docs" ? "documents" : "pages"} in ${this.name}.`; + } + }, + dataLabels: { + formatter() { + return this.point.active ? this.point.name : `(${this.point.name})`; + } + }, + allowDrillToNode: true, + cursor: 'pointer', + levels: [{ + level: 1, + color: "transparent", + levelSize: { + value: .5 + } + }, { + level: 2, + colorByPoint: true + }, { + level: 3, + colorVariation: { + key: "brightness", + to: 0.5 + } + }] + }], + }); + + // limit data to area if set and (for now) drop docs + const slice = data.filter(d => (area == "ietf" && d.grandparent == area) || d.parent == area || d.id == area) + .map((d) => { + return { + value: d[dataset], + id: d.id, + parent: d.parent, + grandparent: d.grandparent, + active: d.active, + }; + }) + .sort((a, b) => { + if (a.parent != b.parent) { + if (a.parent < b.parent) { + return -1; + } + if (a.parent > b.parent) { + return 1; + } + } else if (a.parent == area) { + if (a.id < b.id) { + return 1; + } + if (a.id > b.id) { + return -1; + } + return 0; + } + return b.value - a.value; + }); + chart.series[0].setData(slice); + }); + }); +} diff --git a/ietf/static/js/highstock.js b/ietf/static/js/highstock.js index e1965acb62..05b1250ed0 100644 --- a/ietf/static/js/highstock.js +++ b/ietf/static/js/highstock.js @@ -5,9 +5,20 @@ import Highcharts_Offline_Exporting from "highcharts/modules/offline-exporting"; import Highcharts_Export_Data from "highcharts/modules/export-data"; import Highcharts_Accessibility from"highcharts/modules/accessibility"; +document.documentElement.style.setProperty("--highcharts-background-color", "transparent"); + Highcharts_Exporting(Highcharts); Highcharts_Offline_Exporting(Highcharts); Highcharts_Export_Data(Highcharts); Highcharts_Accessibility(Highcharts); +Highcharts.setOptions({ + chart: { + styledMode: true, + }, + credits: { + enabled: false + }, +}); + window.Highcharts = Highcharts; diff --git a/ietf/static/js/ietf.js b/ietf/static/js/ietf.js index 17165bbfbd..09fa324e42 100644 --- a/ietf/static/js/ietf.js +++ b/ietf/static/js/ietf.js @@ -13,7 +13,7 @@ import "bootstrap/js/dist/scrollspy"; import "bootstrap/js/dist/tab"; // import "bootstrap/js/dist/toast"; import "bootstrap/js/dist/tooltip"; - +import { debounce } from 'lodash-es'; import jquery from "jquery"; window.$ = window.jQuery = jquery; @@ -57,7 +57,7 @@ $(document) var text = $(this) .text(); // insert some at strategic places - var newtext = text.replace(/([@._])/g, "$1"); + var newtext = text.replace(/(\S)([@._+])(\S)/g, "$1$2$3"); if (newtext === text) { return; } @@ -91,20 +91,57 @@ $(document) // }); }); -$(document) - .ready(function () { +function overflowShadows(el) { + function handleScroll(){ + const canScrollUp = el.scrollTop > 0 + const canScrollDown = el.offsetHeight + el.scrollTop < el.scrollHeight + el.classList.toggle("overflow-shadows--both", canScrollUp && canScrollDown) + el.classList.toggle("overflow-shadows--top-only", canScrollUp && !canScrollDown) + el.classList.toggle("overflow-shadows--bottom-only", !canScrollUp && canScrollDown) + } - function dropdown_hover(e) { - var navbar = $(this) - .closest(".navbar"); - if (navbar.length === 0 || navbar.find(".navbar-toggler") - .is(":hidden")) { - $(this) - .children(".dropdown-toggle") - .dropdown(e.type == "mouseenter" ? "show" : "hide"); - } + el.addEventListener("scroll", handleScroll, {passive: true}) + handleScroll() + + const observer = new IntersectionObserver(handleScroll) + observer.observe(el) // el won't have scrollTop etc when hidden, so we need to recalculate when it's revealed + + return () => { + el.removeEventListener("scroll", handleScroll) + observer.unobserve(el) + } +} + +function ensureDropdownOnscreen(elm) { + const handlePlacement = () => { + if(!(elm instanceof HTMLElement)) { + return } + const rect = elm.getBoundingClientRect() + const BUFFER_PX = 5 // additional distance from bottom of viewport + const existingStyleTop = parseInt(elm.style.top, 10) + const offscreenBy = Math.round(window.innerHeight - (rect.top + rect.height) - BUFFER_PX) + if(existingStyleTop === offscreenBy) { + console.log(`Already set top to ${offscreenBy}. Ignoring`) + // already set, nothing to do + return + } + if(offscreenBy < 0) { + elm.style.top = `${offscreenBy}px` + } + } + + const debouncedHandler = debounce(handlePlacement, 100) + + const observer = new MutationObserver(debouncedHandler) + observer.observe(elm, { + attributes: true + }) +} + +$(document) + .ready(function () { // load data for the menu $.ajax({ url: $(document.body) @@ -120,7 +157,7 @@ $(document) } attachTo.find(".dropdown-menu") .remove(); - var menu = ['

      diff --git a/ietf/templates/base.html b/ietf/templates/base.html index 1571ff5000..b0df04f30a 100644 --- a/ietf/templates/base.html +++ b/ietf/templates/base.html @@ -1,4 +1,4 @@ -{# Copyright The IETF Trust 2015-2022, All Rights Reserved #} +{# Copyright The IETF Trust 2015-2023, All Rights Reserved #} {% load analytical %} {% load ietf_filters static %} @@ -6,7 +6,7 @@ {% origin %} {% load django_bootstrap5 %} {% load django_vite %} - + {% analytical_head_top %} @@ -15,14 +15,13 @@ {% block title %}No title{% endblock %} - {% comment Halloween %} - - {% endcomment %} - - - + + + + {# load this in the head, to prevent flickering #} + @@ -33,27 +32,27 @@ {% analytical_head_bottom %} - {% analytical_body_top %} + {% include "base/status.html" %} Skip to main content -

    • {% for g in user|managed_review_groups %}
    • - {{ g.acronym }} reviews
    • {% endfor %} {% endif %} - {% if user|active_nomcoms %} + {% with user|active_nomcoms as nomcoms %}{% if nomcoms %} {% if flavor == 'top' %}
    • @@ -179,15 +190,15 @@
    • NomComs
    • - {% for g in user|active_nomcoms %} + {% for nomcom in nomcoms %}
    • - - {{ g.acronym|capfirst }} + + {{ nomcom|capfirst }}
    • {% endfor %} - {% endif %} + {% endif %}{% endwith %} {% endif %} {% if flavor == 'top' %}
    • @@ -197,29 +208,50 @@ RFC streams
    • - IAB
    • - IRTF
    • - ISE
    • - Editorial
    • + {% if flavor == 'top' %} +
    • +
    • + {% endif %} +
    • + Subseries +
    • +
    • + + STD + + + BCP + + + FYI + +
    • {% if flavor == 'top' %}
    @@ -241,47 +273,67 @@ {% endif %}
  • - Agenda
  • - Materials
  • - Floor plan
  • - Registration
  • - Important dates
  • - + Request a session
  • + {% if user|can_request_interim %} +
  • + + Request an interim meeting + +
  • + {% endif %}
  • - Session requests
  • + {% if user|matman_groups %} + {% if flavor == 'top' %}
  • {% endif %} +
  • Manage
  • + {% for g in user|matman_groups %} +
  • + + {{ g.acronym }} {{ g.type_id }} meetings + +
  • + {% endfor %} + {% endif %} {% if flavor == 'top' %} {% if flavor == 'top' %}
  • @@ -292,7 +344,7 @@
  • {% endif %}
  • - Upcoming meetings @@ -307,13 +359,13 @@
  • {% endif %}
  • - Past meetings
  • - Meeting proceedings @@ -339,56 +391,64 @@
  • {% endif %}
  • - IPR disclosures
  • - Liaison statements
  • + {% if user|has_role:"Secretariat,IAB,Liaison Manager,Liaison Coordinator" %} +
  • + + List of other SDO groups + +
  • + {% endif %}
  • - IESG agenda
  • - NomComs
  • - Downref registry
  • - Statistics
  • - - Tutorials - -
  • -
  • - API Help
  • - Release notes
  • +
  • + + System status + +
  • {% if flavor == 'top' %}
  • {% endif %}
  • - @@ -432,4 +492,4 @@ {% endif %} {% if flavor == 'top' %} {% include "base/menu_user.html" %} -{% endif %} \ No newline at end of file +{% endif %} diff --git a/ietf/templates/base/menu_user.html b/ietf/templates/base/menu_user.html index bb68855b6e..fd921638a4 100644 --- a/ietf/templates/base/menu_user.html +++ b/ietf/templates/base/menu_user.html @@ -1,4 +1,4 @@ -{# Copyright The IETF Trust 2015, All Rights Reserved #} +{# Copyright The IETF Trust 2015-2023, All Rights Reserved #} {% load origin %} {% origin %} {% load ietf_filters %} @@ -23,7 +23,7 @@ {% url 'django.contrib.auth.views.logout' as logout_url %} {% if request.get_full_path == logout_url %}
  • - Sign in @@ -32,14 +32,16 @@ {% else %} {% if user.is_authenticated %}
  • - - Sign out - +
    + {% csrf_token %} + +
  • - Account info @@ -47,35 +49,35 @@
  • {% if user and user.person %}
  • - Public profile page
  • {% endif %}
  • - Preferences
  • - API keys
  • - Change password
  • - Change username @@ -83,21 +85,21 @@
  • {% else %}
  • - + href="{% url 'ietf.ietfauth.views.login' %}?next={{ request.get_full_path|removeprefix:'/accounts/login/?next='|urlencode }}"> Sign in
  • - Password reset
  • - Preferences @@ -107,15 +109,46 @@ {% endif %} {% if not request.user.is_authenticated %}
  • - New account
  • {% endif %} +
  • + + List subscriptions + + +
  • {% if user|has_role:"Reviewer" %}
  • - My reviews @@ -125,31 +158,31 @@ {% if flavor == "top" %}
  • {% endif %}
  • AD dashboard
  • - My docs
  • - Next telechat
  • - Discusses
  • - Milestone review
  • - Last Call docs @@ -161,35 +194,29 @@ IETF secretariat
  • - Telechat dates
  • - Management items
  • - Milestones
  • - Sync discrepancies
  • -
  • - - Account allowlist - -
  • {% endif %} {% if user|has_role:"IANA" %} {% if flavor == "top" %} @@ -200,7 +227,7 @@ IANA
  • - Sync discrepancies @@ -215,7 +242,7 @@ RFC Editor
  • - Sync discrepancies diff --git a/ietf/templates/base/menu_wg.html b/ietf/templates/base/menu_wg.html index 3ab7ca399b..5ca1be8a4e 100644 --- a/ietf/templates/base/menu_wg.html +++ b/ietf/templates/base/menu_wg.html @@ -2,8 +2,11 @@ {% load origin %} {% origin %} {% for p in parents %} + {% if p.acronym == "iab" %} +
  • IESG
  • + {% endif%}
  • - {{ p.short_name }} diff --git a/ietf/templates/base/status.html b/ietf/templates/base/status.html new file mode 100644 index 0000000000..33e1abf699 --- /dev/null +++ b/ietf/templates/base/status.html @@ -0,0 +1,2 @@ + +
    \ No newline at end of file diff --git a/ietf/templates/community/atom.xml b/ietf/templates/community/atom.xml index 32e3b00292..01dcdfeee7 100644 --- a/ietf/templates/community/atom.xml +++ b/ietf/templates/community/atom.xml @@ -3,7 +3,7 @@ {{ title }} {{ subtitle }} {{ id }} - {{ updated|date:"Y-m-d\TH:i:sO" }} + {{ updated.isoformat }} @@ -17,11 +17,11 @@ - {{ entry.id }} + urn:datatracker-ietf-org:event:{{ entry.id }} - {{ entry.time|date:"Y-m-d\TH:i:sO" }} + {{ entry.time.isoformat }} - {{ entry.time|date:"Y-m-d\TH:i:sO" }} + {{ entry.time.isoformat }} {{ entry.by }} diff --git a/ietf/templates/community/list_menu.html b/ietf/templates/community/list_menu.html index d54695775f..009d01152d 100644 --- a/ietf/templates/community/list_menu.html +++ b/ietf/templates/community/list_menu.html @@ -3,18 +3,18 @@ {% if clist.pk != None %} + href="{% if clist.group %}{% url "ietf.community.views.subscription" acronym=clist.group.acronym %}{% else %}{% url "ietf.community.views.subscription" email_or_name=email_or_name %}{% endif %}"> {% if subscribed %} Change subscription @@ -24,7 +24,7 @@ {% endif %} + href="{% if clist.group %}{% url "ietf.community.views.export_to_csv" acronym=clist.group.acronym %}{% else %}{% url "ietf.community.views.export_to_csv" email_or_name=email_or_name %}{% endif %}"> Export as CSV - \ No newline at end of file + diff --git a/ietf/templates/community/notification_email.txt b/ietf/templates/community/notification_email.txt index d3a6c2214b..7c214de30a 100644 --- a/ietf/templates/community/notification_email.txt +++ b/ietf/templates/community/notification_email.txt @@ -12,6 +12,6 @@ Change by {{ event.by }} on {{ event.time }}: Best regards, - The Datatracker draft tracking service + The Datatracker Internet-Draft tracking service (for the IETF Secretariat) {% endautoescape %} diff --git a/ietf/templates/community/subscription.html b/ietf/templates/community/subscription.html index e8b562a47d..4b92a610e1 100644 --- a/ietf/templates/community/subscription.html +++ b/ietf/templates/community/subscription.html @@ -32,7 +32,7 @@

    Existing subscriptions

    {% endif %}

    Add new subscription

    -

    +

    The email addresses you can choose between are those registered in your profile.

    diff --git a/ietf/templates/community/untrack_document.html b/ietf/templates/community/untrack_document.html index 985dcad89b..fe94081a79 100644 --- a/ietf/templates/community/untrack_document.html +++ b/ietf/templates/community/untrack_document.html @@ -1,17 +1,14 @@ {# Copyright The IETF Trust 2015, All Rights Reserved #} +{% extends "base.html" %} {% load origin %} {% origin %} -{% load django_bootstrap5 %} {% block title %}Remove tracking of document {{ name }}{% endblock %} -{% bootstrap_messages %} -
    - {% csrf_token %} -

    - Remove {{ name }} from the list? -

    - -
    +{% block content %} +
    + {% csrf_token %} +

    + Remove {{ name }} from the list? +

    + +
    +{% endblock %} \ No newline at end of file diff --git a/ietf/templates/community/view_list.html b/ietf/templates/community/view_list.html index 17650beb10..a543eaf7cf 100644 --- a/ietf/templates/community/view_list.html +++ b/ietf/templates/community/view_list.html @@ -12,7 +12,7 @@

    {{ clist.long_name }}

    {% bootstrap_messages %} {% if can_manage_list %} + href="{% url "ietf.community.views.manage_list" email_or_name=email_or_name %}"> Manage list @@ -22,4 +22,4 @@

    {{ clist.long_name }}

    {% endblock %} {% block js %} -{% endblock %} \ No newline at end of file +{% endblock %} diff --git a/ietf/templates/cookies/settings.html b/ietf/templates/cookies/settings.html index 5ded2f3bd6..3515491c35 100644 --- a/ietf/templates/cookies/settings.html +++ b/ietf/templates/cookies/settings.html @@ -12,7 +12,7 @@

    User settings

    cookies disabled then you will not be able to change the settings (everything still continues to work by using default settings).

    -

    How many days is considered "new"?

    +

    How many days is considered "new"?

    This setting affects how many days are considered "new enough" to get the special highlighting in the documents table. The default setting is {{ defaults.new_enough }} days.

    @@ -60,7 +60,7 @@

    How many days is considered "new"?

    90 days

    -

    How many days is considered "soon"?

    +

    How many days is considered "soon"?

    This setting tells what is considered "soon" when showing documents that are going to be expire soon. The Default setting is {{ defaults.expires_soon }} days.

    @@ -108,7 +108,7 @@

    How many days is considered "soon"?

    90 days

    -

    Show full document text by default?

    +

    Show full document text by default?

    Show the full text immediately on the document page instead of only showing beginning of it. This defaults to {{ defaults.full_draft }}.

    @@ -128,7 +128,7 @@

    Show full document text by default?

    On

    -

    Show the left-hand menu?

    +

    Show the left-hand menu?

    Show the left-hand menu on all regular pages? This defaults to {{ defaults.left_menu }}.

    @@ -148,4 +148,24 @@

    Show the left-hand menu?

    On

    + + +

    Which color mode should be used?

    +

    + Use a light, dark or automatic (as indicated by OS) color mode. +

    +

    + + + Light + + + + Dark + + + + Auto + +

    {% endblock %} \ No newline at end of file diff --git a/ietf/templates/debug.html b/ietf/templates/debug.html index f7788e776c..c459359d3e 100644 --- a/ietf/templates/debug.html +++ b/ietf/templates/debug.html @@ -103,7 +103,7 @@
  • - {% endfor %} - - {% endif %}" - {% endif %} - {% with label.2 as up_is_good %} - {% if prev < count %} - class="bi bi-arrow-up-right-circle{% if count %}-fill{% endif %} {{ up_is_good|yesno:'text-success,text-danger,text-muted' }}" - {% elif prev > count %} - class="bi bi-arrow-down-right-circle{% if count %}-fill{% endif %} {{ up_is_good|yesno:'text-danger,text-success,text-muted' }}" - {% else %} - class="bi bi-arrow-right-circle text-muted" - {% endif %} - > - {% endwith %} - {% endif %} -{% endif %} \ No newline at end of file diff --git a/ietf/templates/doc/ad_list.html b/ietf/templates/doc/ad_list.html index 189754e8ac..cac709021e 100644 --- a/ietf/templates/doc/ad_list.html +++ b/ietf/templates/doc/ad_list.html @@ -3,42 +3,80 @@ {% load origin static %} {% load ietf_filters %} {% block pagehead %} - + + {% endblock %} -{% block title %}Area directors{% endblock %} +{% block morecss %} + table .border-bottom { border-bottom-color: var(--highcharts-neutral-color-80) !important; } + .highcharts-container .highcharts-axis-labels { + font-size: .7rem; + fill: var(--bs-body-color) + } + .highcharts-container .highcharts-graph { stroke-width: 2.5; } + .highcharts-container .highcharts-color-0 { + fill: var(--bs-body-color); + stroke: var(--bs-primary); + } + .highcharts-container .highcharts-data-label text { + font-size: 1rem; + font-weight: inherit; + fill: var(--bs-body-color) + } +{% endblock %} +{% block title %}IESG Dashboard{% endblock %} {% block content %} {% origin %} -

    Area Director Workload

    +

    IESG Dashboard

    {% if user|has_role:"Area Director,Secretariat" %}
    - {{ delta.days }}-day trend indicators + {{ delta }}-day trend graphs are only shown to logged-in Area Directors.
    {% endif %} - {% for group in workload %} -

    {{ group.group_type }} State Counts

    - +

    + Documents in IESG Processing + IESG view of Working Groups +

    + {% for dt in metadata %} +

    {{ dt.type.1 }} State Counts

    +
    - {% for g, desc, up_is_good in group.group_names %} - + {% endif %} + {% for state, state_name in dt.states %} + {% endfor %} - {% for ad, ad_data in group.counts %} + {% for ad in dt.ads %} - {% for label, count, prev, docs_delta in ad_data %} - + {% endif %} + {% for state, state_name in dt.states %} + {% endfor %} @@ -47,9 +85,19 @@

    {{ group.group_type }} Stat

    - {% for label, count, prev in group.sums %} - + {% endif %} + {% for state, state_name in dt.states %} + {% endfor %} @@ -69,4 +117,177 @@

    {{ group.group_type }} Stat }); }); + + {{ data|json_script:"data" }} + + + + + {% endblock %} \ No newline at end of file diff --git a/ietf/templates/doc/add_comment.html b/ietf/templates/doc/add_comment.html index 0783b2486c..211cb380db 100644 --- a/ietf/templates/doc/add_comment.html +++ b/ietf/templates/doc/add_comment.html @@ -2,13 +2,14 @@ {# Copyright The IETF Trust 2015, All Rights Reserved #} {% load origin %} {% load django_bootstrap5 %} -{% block title %}Add comment for {{ doc }}{% endblock %} +{% block title %}Add comment for +{% if review_req %} {{ review_req }} {% else %} {{ doc }} {% endif %} {% endblock %} {% block content %} {% origin %}

    Add comment
    - {{ doc }} + {% if review_req %} {{ review_req }} {% else %} {{ doc }} {% endif %}

    {% csrf_token %} @@ -17,7 +18,12 @@

    The comment will be added to the history trail.

    {% bootstrap_button button_type="submit" content="Submit" %} + {% if review_req %} + Back + {% else %} Back + {% endif %} {% endblock %} \ No newline at end of file diff --git a/ietf/templates/doc/add_sessionpresentation.html b/ietf/templates/doc/add_sessionpresentation.html index a5bcf7096a..0bd751a465 100644 --- a/ietf/templates/doc/add_sessionpresentation.html +++ b/ietf/templates/doc/add_sessionpresentation.html @@ -8,7 +8,7 @@

    Add document to session
    - {{ doc.name }} + {{ doc.name }}
    {{ doc.title }}

    diff --git a/ietf/templates/doc/badge/doc-badge-draft.html b/ietf/templates/doc/badge/doc-badge-draft.html new file mode 100644 index 0000000000..f7f66b6c5e --- /dev/null +++ b/ietf/templates/doc/badge/doc-badge-draft.html @@ -0,0 +1,16 @@ +{% load origin %} +{% load static %} +{% load ietf_filters %} +{% load person_filters %} +{% origin %} +{# Non-RFC #} + +{% if doc.became_rfc %} + This is an older version of an Internet-Draft that was ultimately published as {{doc.became_rfc.name|prettystdname}}. +{% elif snapshot and doc.rev != latest_rev %} + This is an older version of an Internet-Draft whose latest revision state is "{{ doc.doc.get_state }}". +{% else %} + {% if snapshot and doc.rev == latest_rev %}{{ doc.doc.get_state }}{% else %}{{ doc.get_state }}{% endif %} Internet-Draft + {% if submission %}({{ submission|safe }}){% endif %} + {% if resurrected_by %}- resurrect requested by {{ resurrected_by }}{% endif %} +{% endif %} \ No newline at end of file diff --git a/ietf/templates/doc/badge/doc-badge-rfc.html b/ietf/templates/doc/badge/doc-badge-rfc.html new file mode 100644 index 0000000000..780f14a54f --- /dev/null +++ b/ietf/templates/doc/badge/doc-badge-rfc.html @@ -0,0 +1,13 @@ +{% load origin %} +{% load static %} +{% load ietf_filters %} +{% load person_filters %} +{% origin %} + +RFC + {% if not document_html %} + - {{ doc.std_level }} + {% else %} + {{ doc.std_level }} + {% endif %} + diff --git a/ietf/templates/doc/ballot/approvaltext.html b/ietf/templates/doc/ballot/approvaltext.html index 1b228ba375..3cb632b8f8 100644 --- a/ietf/templates/doc/ballot/approvaltext.html +++ b/ietf/templates/doc/ballot/approvaltext.html @@ -9,7 +9,7 @@

    Approval announcement writeup
    - {{ doc }} + {{ doc }}

    {% csrf_token %} @@ -29,7 +29,7 @@

    href="{% url 'ietf.doc.views_ballot.approve_ballot' name=doc.name %}">Approve ballot {% endif %} + href="{% url "ietf.doc.views_doc.document_main" name=doc.name %}"> Back diff --git a/ietf/templates/doc/ballot/approve_ballot.html b/ietf/templates/doc/ballot/approve_ballot.html index a146acf4e9..30dd05fa43 100644 --- a/ietf/templates/doc/ballot/approve_ballot.html +++ b/ietf/templates/doc/ballot/approve_ballot.html @@ -8,7 +8,7 @@

    Approve ballot
    - {{ doc }} + {{ doc }}

    {% csrf_token %} @@ -21,7 +21,7 @@

    {% endif %} + href="{% url "ietf.doc.views_doc.document_main" name=doc.name %}"> Back diff --git a/ietf/templates/doc/ballot/approve_downrefs.html b/ietf/templates/doc/ballot/approve_downrefs.html index 67caf992c7..ad528c67bf 100644 --- a/ietf/templates/doc/ballot/approve_downrefs.html +++ b/ietf/templates/doc/ballot/approve_downrefs.html @@ -8,20 +8,20 @@

    Approve downward references
    - {{ doc }} + {{ doc }}

    The ballot for - {{ doc }} + {{ doc }} was just approved.

    {% if not downrefs_to_rfc %}

    No downward references for - {{ doc }} + {{ doc }}

    Back + href="{% url "ietf.doc.views_doc.document_main" name=doc.name %}">Back {% else %}

    Add downward references to RFCs to the DOWNREF registry, if they were identified in the IETF Last Call and approved by the Sponsoring Area Director. @@ -41,7 +41,7 @@

    {% csrf_token %} {% bootstrap_form approve_downrefs_form %} + href="{% url "ietf.doc.views_doc.document_main" name=doc.name %}"> Add no DOWNREF entries diff --git a/ietf/templates/doc/ballot/ballot_comment_mail.txt b/ietf/templates/doc/ballot/ballot_comment_mail.txt index add55c68ef..8fb709b7f5 100644 --- a/ietf/templates/doc/ballot/ballot_comment_mail.txt +++ b/ietf/templates/doc/ballot/ballot_comment_mail.txt @@ -5,7 +5,7 @@ When responding, please keep the subject line intact and reply to all email addresses included in the To and CC lines. (Feel free to cut this introductory paragraph, however.) -{% if doc.type_id == "draft" and doc.stream_id != "irtf" %} +{% if doc.type_id == "draft" and doc.stream_id == "ietf" %} Please refer to https://www.ietf.org/about/groups/iesg/statements/handling-ballot-positions/ for more information about how to handle DISCUSS and COMMENT positions. {% endif %} diff --git a/ietf/templates/doc/ballot/ballot_issued.html b/ietf/templates/doc/ballot/ballot_issued.html index 5549cd1189..dfa03896e9 100644 --- a/ietf/templates/doc/ballot/ballot_issued.html +++ b/ietf/templates/doc/ballot/ballot_issued.html @@ -7,11 +7,11 @@

    Ballot issued
    - {{ doc }} + {{ doc }}

    Ballot for - {{ doc }} + {{ doc }} has been sent out.

    {% if doc.telechat_date %} @@ -24,5 +24,5 @@

    {% endif %} Back + href="{% url "ietf.doc.views_doc.document_main" name=doc.name %}">Back {% endblock %} diff --git a/ietf/templates/doc/ballot/clear_ballot.html b/ietf/templates/doc/ballot/clear_ballot.html index d1f731ba70..09e7dfef1b 100644 --- a/ietf/templates/doc/ballot/clear_ballot.html +++ b/ietf/templates/doc/ballot/clear_ballot.html @@ -8,20 +8,20 @@

    Clear ballot
    - {{ doc }} + {{ doc }}

    {% csrf_token %}

    Clear the ballot for - {{ doc }}? + {{ doc }}?
    This will clear all ballot positions and discuss entries.

    + href="{% url "ietf.doc.views_doc.document_main" name=doc.name %}"> Back diff --git a/ietf/templates/doc/ballot/defer_ballot.html b/ietf/templates/doc/ballot/defer_ballot.html index 85f887c9ab..ae7099e9e1 100644 --- a/ietf/templates/doc/ballot/defer_ballot.html +++ b/ietf/templates/doc/ballot/defer_ballot.html @@ -8,20 +8,20 @@

    Defer ballot
    - {{ doc }} + {{ doc }}

    {% csrf_token %}

    Defer the ballot for - {{ doc }}? + {{ doc }}?
    The ballot will then be put on the IESG agenda of {{ telechat_date }}.

    + href="{% url "ietf.doc.views_doc.document_main" name=doc.name %}"> Back diff --git a/ietf/templates/doc/ballot/edit_position.html b/ietf/templates/doc/ballot/edit_position.html index 1b11fbb90a..b57e9a3652 100644 --- a/ietf/templates/doc/ballot/edit_position.html +++ b/ietf/templates/doc/ballot/edit_position.html @@ -8,7 +8,7 @@

    Change position for {{ balloter.plain_name }}
    - {{ doc }} + {{ doc }}

    {% if ballot.ballot_type.question %}
    @@ -20,43 +20,86 @@

    Ballot deferred by {{ ballot_deferred.by }} on {{ ballot_deferred.time|date:"Y-m-d" }}.

    {% endif %} +
    +
    + {% if form.errors or cc_select_form.errors or additional_cc_form.errors %} +
    + There were errors in the submitted form -- see below. Please correct these and resubmit. +
    + {% if form.errors %} +
    Position entry
    + {% bootstrap_form_errors form %} + {% endif %} + {% if cc_select_form.errors %} +
    CC selection
    + {% bootstrap_form_errors cc_select_form %} + {% endif %} + {% if additional_cc_form.errors %} +
    Additional Cc Addresses
    + {% bootstrap_form_errors additional_cc_form %} + {% endif %} + {% endif %}
    {% csrf_token %} {% for field in form %} {% if field.name == "discuss" %}
    {% endif %} {% bootstrap_field field %} {% if field.name == "discuss" and old_pos and old_pos.discuss_time %} -
    Last edited {{ old_pos.discuss_time }}
    +
    Last saved {{ old_pos.discuss_time }}
    {% elif field.name == "comment" and old_pos and old_pos.comment_time %} -
    Last edited {{ old_pos.comment_time }}
    +
    Last saved {{ old_pos.comment_time }}
    {% endif %} {% if field.name == "discuss" %}
    {% endif %} {% endfor %} + {% bootstrap_form cc_select_form %} + {% bootstrap_form additional_cc_form %}
    + - + {% if doc.type_id == "draft" or doc.type_id == "conflrev" %} {% if doc.stream.slug != "irtf" %} {% if ballot_deferred %} + name="Undefer" + value="Undefer">Undefer ballot {% else %} + name="Defer" + value="Defer">Defer ballot {% endif %} {% endif %} {% endif %} + href="{% url "ietf.doc.views_doc.document_main" name=doc.name %}"> Back
    - + + + {% endblock %} {% block js %} + + {% endblock %} \ No newline at end of file diff --git a/ietf/templates/doc/ballot/irsg_ballot_approve.html b/ietf/templates/doc/ballot/irsg_ballot_approve.html index 8b28f04c3a..9eb0cc80be 100644 --- a/ietf/templates/doc/ballot/irsg_ballot_approve.html +++ b/ietf/templates/doc/ballot/irsg_ballot_approve.html @@ -12,7 +12,7 @@

    Issue ballot
    - {{ doc }} + {{ doc }}

    {{ question }} diff --git a/ietf/templates/doc/ballot/irsg_ballot_close.html b/ietf/templates/doc/ballot/irsg_ballot_close.html index ab49ec79d7..22405df1c1 100644 --- a/ietf/templates/doc/ballot/irsg_ballot_close.html +++ b/ietf/templates/doc/ballot/irsg_ballot_close.html @@ -8,7 +8,7 @@

    Close ballot
    - {{ doc }} + {{ doc }}

    {{ question }} diff --git a/ietf/templates/doc/ballot/lastcalltext.html b/ietf/templates/doc/ballot/lastcalltext.html index 5b49f6b372..fe2b884c2b 100644 --- a/ietf/templates/doc/ballot/lastcalltext.html +++ b/ietf/templates/doc/ballot/lastcalltext.html @@ -9,7 +9,7 @@

    Last call text
    - {{ doc }} + {{ doc }}

    {% csrf_token %} @@ -39,7 +39,7 @@

    href="{% url 'ietf.doc.views_ballot.make_last_call' name=doc.name %}">Issue last call {% endif %} + href="{% url "ietf.doc.views_doc.document_main" name=doc.name %}"> Back diff --git a/ietf/templates/doc/ballot/rfceditornote.html b/ietf/templates/doc/ballot/rfceditornote.html index ff6ca7078c..8a6d57379d 100644 --- a/ietf/templates/doc/ballot/rfceditornote.html +++ b/ietf/templates/doc/ballot/rfceditornote.html @@ -8,7 +8,7 @@

    RFC Editor Note
    - {{ doc }} + {{ doc }}

    {% bootstrap_messages %}
    @@ -31,7 +31,7 @@

    Clear + href="{% url "ietf.doc.views_doc.document_main" name=doc.name %}"> Back diff --git a/ietf/templates/doc/ballot/rsab_ballot_approve.html b/ietf/templates/doc/ballot/rsab_ballot_approve.html new file mode 100644 index 0000000000..d607348722 --- /dev/null +++ b/ietf/templates/doc/ballot/rsab_ballot_approve.html @@ -0,0 +1,23 @@ +{% extends "base.html" %} +{# Copyright The IETF Trust 2022, All Rights Reserved #} +{% load origin %} +{% load static %} +{% load django_bootstrap5 %} +{% block title %}Issue ballot for {{ doc }}?{% endblock %} +{% block content %} + {% origin %} +

    + Issue ballot +
    + {{ doc }} +

    +

    + {{ question }} +

    +
    + {% csrf_token %} + {# curly percent bootstrap_form approval_text_form curly percent #} + + + +{% endblock %} diff --git a/ietf/templates/doc/ballot/rsab_ballot_close.html b/ietf/templates/doc/ballot/rsab_ballot_close.html new file mode 100644 index 0000000000..f6da3e052d --- /dev/null +++ b/ietf/templates/doc/ballot/rsab_ballot_close.html @@ -0,0 +1,22 @@ +{% extends "base.html" %} +{# Copyright The IETF Trust 2022, All Rights Reserved #} +{% load origin %} +{% load django_bootstrap5 %} +{% block title %}Close ballot for {{ doc }}{% endblock %} +{% block content %} + {% origin %} +

    + Close ballot +
    + {{ doc }} +

    +

    + {{ question }} +

    +
    + {% csrf_token %} + {# curly percent bootstrap_form approval_text_form curly percent #} + + + +{% endblock %} diff --git a/ietf/templates/doc/ballot/send_ballot_comment.html b/ietf/templates/doc/ballot/send_ballot_comment.html deleted file mode 100644 index 8f87bf51ad..0000000000 --- a/ietf/templates/doc/ballot/send_ballot_comment.html +++ /dev/null @@ -1,44 +0,0 @@ -{% extends "base.html" %} -{# Copyright The IETF Trust 2015, All Rights Reserved #} -{% load origin %} -{% load django_bootstrap5 %} -{% load ietf_filters %} -{% block title %}Send ballot position for {{ balloter }} on {{ doc }}{% endblock %} -{% block content %} - {% origin %} -

    - Send ballot position for {{ balloter }} -
    - {{ doc }} -

    -
    - {% csrf_token %} -
    - - -
    -
    - - -
    - {% bootstrap_form cc_select_form %} -
    - - -
    Separate email addresses with commas.
    -
    -
    - - -
    -
    -

    Body

    -
    {{ body|maybewordwrap }}
    -
    - - - Back - - -{% endblock %} diff --git a/ietf/templates/doc/ballot/undefer_ballot.html b/ietf/templates/doc/ballot/undefer_ballot.html index df2217fa2f..4e86698160 100644 --- a/ietf/templates/doc/ballot/undefer_ballot.html +++ b/ietf/templates/doc/ballot/undefer_ballot.html @@ -8,7 +8,7 @@

    Undefer ballot
    - {{ doc }} + {{ doc }}

    {% csrf_token %} @@ -19,7 +19,7 @@

    + href="{% url "ietf.doc.views_doc.document_main" name=doc.name %}"> Back diff --git a/ietf/templates/doc/ballot/writeupnotes.html b/ietf/templates/doc/ballot/writeupnotes.html index 481e00e134..8e985c15c7 100644 --- a/ietf/templates/doc/ballot/writeupnotes.html +++ b/ietf/templates/doc/ballot/writeupnotes.html @@ -8,18 +8,23 @@

    Ballot writeup and notes
    - {{ doc }} + {{ doc }}

    {% csrf_token %} {% bootstrap_form ballot_writeup_form %}
    - Technical summary, Working Group summary, document quality, personnel, IRTF note, IESG note, IANA note. This text will be appended to all announcements and messages to the IRTF or RFC Editor. - {% if ballot_issue_danger %} + Technical summary, Working Group summary, document quality, personnel, IANA note. This text will be appended to all announcements and messages to the IRTF or RFC Editor. + {% if warn_lc %}

    This document has not completed IETF Last Call. Please do not issue the ballot early without good reason.

    {% endif %} + {% if warn_unexpected_state %} +

    + This document is in an IESG state of "{{warn_unexpected_state}}". It would be unexpected to issue a ballot while in this state. +

    + {% endif %}
    + href="{% url "ietf.doc.views_doc.document_main" name=doc.name %}"> Back -{% endblock %} +{% endblock %} \ No newline at end of file diff --git a/ietf/templates/doc/ballot_popup.html b/ietf/templates/doc/ballot_popup.html index 815a4f3958..d2589cd54c 100644 --- a/ietf/templates/doc/ballot_popup.html +++ b/ietf/templates/doc/ballot_popup.html @@ -1,4 +1,4 @@ -{# Copyright The IETF Trust 2015, All Rights Reserved #} +{# Copyright The IETF Trust 2015-2022, All Rights Reserved #} {% load origin %} {% origin %} {% load ietf_filters %} @@ -24,10 +24,10 @@
    - - - - - - - - - - - {% for req in req_group.list %} + {% if req_group.grouper != "spam" or request.user|has_role:"Secretariat" %} +

    {{ req_group.grouper|capfirst }} BOF Requests

    +
    NameDateTitleResponsibleEditors
    + - - - - - + + + + + - {% endfor %} - -
    - {{ req.name }}-{{ req.rev }} - {{ req.latest_revision_event.time|date:"Y-m-d" }}{{ req.title }} - {% for person in req.responsible %} - {% person_link person %}{% if not forloop.last %},{% endif %} - {% endfor %} - - {% for person in req.editors %} - {% person_link person %}{% if not forloop.last %},{% endif %} - {% endfor %} - NameDateTitleResponsibleEditors
    + + + {% for req in req_group.list %} + + + {{ req.name }}-{{ req.rev }} + + {{ req.latest_revision_event.time|date:"Y-m-d" }} + {{ req.title }} + + {% for person in req.responsible %} + {% person_link person %}{% if not forloop.last %},{% endif %} + {% endfor %} + + + {% for person in req.editors %} + {% person_link person %}{% if not forloop.last %},{% endif %} + {% endfor %} + + + {% endfor %} + + + {% endif %} {% endfor %} {% endif %} {% endblock %} diff --git a/ietf/templates/doc/bofreq/bofreq_template.md b/ietf/templates/doc/bofreq/bofreq_template.md index c6269d7f4f..49c5e511a5 100644 --- a/ietf/templates/doc/bofreq/bofreq_template.md +++ b/ietf/templates/doc/bofreq/bofreq_template.md @@ -1,16 +1,15 @@ -# Name: Exact MPLS Edges (EXAMPLE) (There's an acronym for anything if you really want one ;-) +# Name: EXAct MPLs Edges (EXAMPLE) (There's an acronym for anything if you really want one ;-) ## Description Replace this with a few paragraphs describing the BOF request. Fill in the details below. Keep items in the order they appear here. ## Required Details -- Status: (not) WG Forming -- Responsible AD: name +- Status: "not WG Forming" or "WG forming" +- Responsible AD: name (or at least area(s) if you know) - BOF proponents: name , name (1-3 people - who are requesting and coordinating discussion for proposal) -- BOF chairs: TBD - Number of people expected to attend: 100 -- Length of session (1 or 2 hours): 2 hours +- Length of session (1 or usually 2 hours): 2 hours - Conflicts (whole Areas and/or WGs) - Chair Conflicts: TBD - Technology Overlap: TBD @@ -25,13 +24,13 @@ To allow evaluation of your proposal, please include the following items: - Open source projects (if any) implementing this work: ## Agenda - - Items, drafts, speakers, timing + - Items, Internet-Drafts, speakers, timing - Or a URL -## Links to the mailing list, draft charter if any, relevant Internet-Drafts, etc. +## Links to the mailing list, draft charter if any (for WG-forming BoF), relevant Internet-Drafts, etc. - Mailing List: https://www.ietf.org/mailman/listinfo/example - Draft charter: {{ settings.IDTRACKER_BASE_URL }}{% url 'ietf.doc.views_doc.document_main' name='charter-ietf-EXAMPLE' %} - - Relevant drafts: + - Relevant Internet-Drafts: - Use Cases: - {{ settings.IDTRACKER_BASE_URL }}{% url 'ietf.doc.views_doc.document_main' name='draft-blah-uses' %} - Solutions diff --git a/ietf/templates/doc/bofreq/change_editors.html b/ietf/templates/doc/bofreq/change_editors.html index 546ad7179b..0c30cdecb4 100644 --- a/ietf/templates/doc/bofreq/change_editors.html +++ b/ietf/templates/doc/bofreq/change_editors.html @@ -9,14 +9,14 @@

    Change editors
    - {{ titletext }} + {{ titletext }}

    {% csrf_token %} {% bootstrap_form form %} + href="{% url "ietf.doc.views_doc.document_main" name=doc.name %}"> Back
    diff --git a/ietf/templates/doc/bofreq/change_responsible.html b/ietf/templates/doc/bofreq/change_responsible.html index 79ca29f22b..8c51c6e1f4 100644 --- a/ietf/templates/doc/bofreq/change_responsible.html +++ b/ietf/templates/doc/bofreq/change_responsible.html @@ -9,14 +9,14 @@

    Change Responsible Leadership
    - {{ titletext }} + {{ titletext }}

    {% csrf_token %} {% bootstrap_form form %} + href="{% url "ietf.doc.views_doc.document_main" name=doc.name %}"> Back
    diff --git a/ietf/templates/doc/bofreq/new_bofreq.html b/ietf/templates/doc/bofreq/new_bofreq.html index 9506d9c8e6..cda6f73b90 100644 --- a/ietf/templates/doc/bofreq/new_bofreq.html +++ b/ietf/templates/doc/bofreq/new_bofreq.html @@ -1,20 +1,29 @@ {% extends "base.html" %} -{# Copyright The IETF Trust 2021, All Rights Reserved #} +{# Copyright The IETF Trust 2021-2026, All Rights Reserved #} {% load origin django_bootstrap5 static textfilters %} {% block title %}Start a new BOF Request{% endblock %} {% block content %} {% origin %}

    Start a new BOF Request

    - The IAB will also attempt to provide BoF Shepherds as described in their document on the subject only on request from the IESG. If you feel that your BoF would benefit from an IAB BoF Shepherd, please discuss this with your Area Director. + BoF proponents are strongly encouraged to review the following sources before submitting requests:

    +

    - Choose a short descriptive title for your request. Take time to choose a good initial title - it will be used to make the filename for your request's content. The title can be changed later, but the filename will not change. + The IAB will also attempt to provide BoF Shepherds as described in their document on the subject only on request from the IESG. + If you feel that your BoF would benefit from an IAB BoF Shepherd, please discuss this with your Area Director. +

    +

    + Choose a short descriptive title for your request. Take time to choose a good initial title - it will be used to make the filename for your request's content. + The title can be changed later, but the filename will not change.

    For example, a request with a title of "A new important bit" will be saved as bofreq-{{ user.person.last_name|xslugify|slice:"64" }}-a-new-important-bit-00.md.

    -

    All the items in the template MUST be filed in.

    +

    All the items in the template MUST be filed in.

    diff --git a/ietf/templates/doc/bofreq/upload_content.html b/ietf/templates/doc/bofreq/upload_content.html index 902b2900a9..74827dd485 100644 --- a/ietf/templates/doc/bofreq/upload_content.html +++ b/ietf/templates/doc/bofreq/upload_content.html @@ -7,7 +7,7 @@

    Upload New Revision
    - {{ doc.name }} + {{ doc.name }}

    Change responsible AD
    - {{ titletext }} + {{ titletext }} {% csrf_token %} {% bootstrap_form form %} + href="{% url "ietf.doc.views_doc.document_main" name=doc.name %}"> Back
    diff --git a/ietf/templates/doc/change_shepherd.html b/ietf/templates/doc/change_shepherd.html index a102ce24ee..755bd5d316 100644 --- a/ietf/templates/doc/change_shepherd.html +++ b/ietf/templates/doc/change_shepherd.html @@ -10,7 +10,7 @@

    Change document shepherd
    - {{ doc.name }}-{{ doc.rev }} + {{ doc.name }}-{{ doc.rev }}

    The shepherd needs to have a Datatracker account. A new account can be diff --git a/ietf/templates/doc/change_shepherd_email.html b/ietf/templates/doc/change_shepherd_email.html index 36a8b44d15..b608c2d09a 100644 --- a/ietf/templates/doc/change_shepherd_email.html +++ b/ietf/templates/doc/change_shepherd_email.html @@ -8,7 +8,7 @@

    Change document shepherd email
    - {{ doc.name }}-{{ doc.rev }} + {{ doc.name }}-{{ doc.rev }}

    {% csrf_token %} diff --git a/ietf/templates/doc/change_state.html b/ietf/templates/doc/change_state.html index d405e059b3..9d083e2f3c 100644 --- a/ietf/templates/doc/change_state.html +++ b/ietf/templates/doc/change_state.html @@ -8,7 +8,7 @@

    Change state
    - {{ doc.title }} + {{ doc.title }}

    Help on states diff --git a/ietf/templates/doc/change_title.html b/ietf/templates/doc/change_title.html index 4f419e1dd0..14d7956cfe 100644 --- a/ietf/templates/doc/change_title.html +++ b/ietf/templates/doc/change_title.html @@ -8,14 +8,14 @@

    Change title
    - {{ titletext }} + {{ titletext }}

    {% csrf_token %} {% bootstrap_form form %} + href="{% url "ietf.doc.views_doc.document_main" name=doc.name %}"> Back
    diff --git a/ietf/templates/doc/charter/action_announcement_text.html b/ietf/templates/doc/charter/action_announcement_text.html index 5722b342a1..e087b175b4 100644 --- a/ietf/templates/doc/charter/action_announcement_text.html +++ b/ietf/templates/doc/charter/action_announcement_text.html @@ -11,7 +11,7 @@

    {{ charter.chartered_group.type.name }} action announcement writeup
    - {{ charter.chartered_group.acronym }} + {{ charter.chartered_group.acronym }}

    {% csrf_token %} @@ -21,7 +21,7 @@

    {% if user|has_role:"Secretariat" %} + href="{% url 'ietf.doc.views_charter.approve' name=charter.name %}"> Charter approval page {% endif %} diff --git a/ietf/templates/doc/charter/action_text.txt b/ietf/templates/doc/charter/action_text.txt index fbe9ed3fe1..9a1e715222 100644 --- a/ietf/templates/doc/charter/action_text.txt +++ b/ietf/templates/doc/charter/action_text.txt @@ -1,4 +1,4 @@ -{% load ietf_filters %}{% autoescape off %}From: The IESG +{% load ietf_filters %}{% autoescape off %}From: {% if group.type_id == "rg" %}The IRTF {% else %}The IESG {% endif %} To: {{ to }}{% if cc %} Cc: {{ cc }} {% endif %} Subject: {{ group.type.name }} Action: {{ action_type }} {{ group.name }} ({{ group.acronym }}) diff --git a/ietf/templates/doc/charter/approve.html b/ietf/templates/doc/charter/approve.html index f109da6872..2a8654482e 100644 --- a/ietf/templates/doc/charter/approve.html +++ b/ietf/templates/doc/charter/approve.html @@ -2,16 +2,16 @@ {# Copyright The IETF Trust 2015, All Rights Reserved #} {% load origin %} {% load django_bootstrap5 %} -{% block title %}Approve {{ charter.canonical_name }}{% endblock %} +{% block title %}Approve {{ charter.name }}{% endblock %} {% block content %} {% origin %} -

    Approve {{ charter.canonical_name }}-{{ charter.rev }}

    +

    Approve {{ charter.name }}-{{ charter.rev }}

    {% csrf_token %}
    {{ announcement }}
    + href="{% url "ietf.doc.views_charter.action_announcement_text" name=charter.name %}?next=approve"> Edit/regenerate announcement Ballot issued
    - {{ doc.name }} + {{ doc.name }}

    Ballot has been sent out. diff --git a/ietf/templates/doc/charter/ballot_writeup.txt b/ietf/templates/doc/charter/ballot_writeup.txt index 552fe65e8d..bf753bdfaf 100644 --- a/ietf/templates/doc/charter/ballot_writeup.txt +++ b/ietf/templates/doc/charter/ballot_writeup.txt @@ -36,14 +36,6 @@ RFC Editor Note (Insert RFC Editor Note here or remove section) -IRTF Note - - (Insert IRTF Note here or remove section) - -IESG Note - - (Insert IESG Note here or remove section) - IANA Note (Insert IANA Note here or remove section) diff --git a/ietf/templates/doc/charter/ballot_writeupnotes.html b/ietf/templates/doc/charter/ballot_writeupnotes.html index fa4c12b617..a202d48b76 100644 --- a/ietf/templates/doc/charter/ballot_writeupnotes.html +++ b/ietf/templates/doc/charter/ballot_writeupnotes.html @@ -8,12 +8,12 @@

    Ballot writeup and notes
    - {{ charter.chartered_group }} + {{ charter.chartered_group }}

    {% csrf_token %} {% bootstrap_form ballot_writeup_form %} -
    Working group summary, personnel, IAB note, IESG note, IANA note.
    +
    Working group summary, personnel, IANA note.
    + href="{% url "ietf.doc.views_doc.document_main" name=charter.name %}"> Back
    diff --git a/ietf/templates/doc/charter/issue_ballot_mail.txt b/ietf/templates/doc/charter/issue_ballot_mail.txt index 12fc44bbbc..914935bb12 100644 --- a/ietf/templates/doc/charter/issue_ballot_mail.txt +++ b/ietf/templates/doc/charter/issue_ballot_mail.txt @@ -1,6 +1,6 @@ -{% load ietf_filters %}{% autoescape off %}To: {{ to }} {% if cc %} -Cc: {{ cc }} -{% endif %}From: IESG Secretary +{% load ietf_filters %}{% autoescape off %}To: {{ to }}{% if cc %} +Cc: {{ cc }}{% endif %} +From: IESG Secretary Reply-To: IESG Secretary Subject: Evaluation: {{ doc.name }} diff --git a/ietf/templates/doc/charter/review_announcement_text.html b/ietf/templates/doc/charter/review_announcement_text.html index 1e2542edb4..c50f5956bf 100644 --- a/ietf/templates/doc/charter/review_announcement_text.html +++ b/ietf/templates/doc/charter/review_announcement_text.html @@ -9,7 +9,7 @@

    WG Review announcement writeup
    - {{ charter.chartered_group.acronym }} + {{ charter.chartered_group.acronym }}

    {% csrf_token %} diff --git a/ietf/templates/doc/charter/submit.html b/ietf/templates/doc/charter/submit.html index 54bd68a419..d391804071 100644 --- a/ietf/templates/doc/charter/submit.html +++ b/ietf/templates/doc/charter/submit.html @@ -39,7 +39,7 @@

    Charter submission

    State {{ group.state.name }} - {% if requested_close %}
    In the process of being closed
    {% endif %} + {% if requested_close %}
    In the process of being closed
    {% endif %} diff --git a/ietf/templates/doc/conflict_review/approval_text.txt b/ietf/templates/doc/conflict_review/approval_text.txt index cf00e1003d..a52ac11a71 100644 --- a/ietf/templates/doc/conflict_review/approval_text.txt +++ b/ietf/templates/doc/conflict_review/approval_text.txt @@ -1,9 +1,9 @@ {% load ietf_filters %}{% load mail_filters %}{% autoescape off %}From: The IESG To: {{ to }} Cc: {{ cc }} -Subject: Results of IETF-conflict review for {{conflictdoc.canonical_name}}-{{conflictdoc.rev}} +Subject: Results of IETF-conflict review for {{conflictdoc.name}}-{{conflictdoc.rev}} -{% filter wordwrap:78 %}The IESG has completed a review of {{conflictdoc.canonical_name}}-{{conflictdoc.rev}} consistent with RFC5742. +{% filter wordwrap:78 %}The IESG has completed a review of {{conflictdoc.name}}-{{conflictdoc.rev}} consistent with RFC5742. {% if review.get_state_slug == 'appr-reqnopub-pend' %} The IESG recommends that '{{ conflictdoc.title }}' {{ conflictdoc.file_tag|safe }} NOT be published as {{ conflictdoc|std_level_prompt_with_article }}. @@ -18,7 +18,7 @@ The IESG would also like the {{receiver}} to review the comments in the datatrac The IESG review is documented at: {{review_url}} -A URL of the reviewed Internet Draft is: +A URL of the reviewed Internet-Draft is: {{conflictdoc_url}} The process for such documents is described {% if conflictdoc.stream_id == 'ise' %}at https://www.rfc-editor.org/indsubs.html {% else %}{% if conflictdoc.stream_id == 'irtf' %}in RFC 5743 {% endif %} {% endif %} diff --git a/ietf/templates/doc/conflict_review/approve.html b/ietf/templates/doc/conflict_review/approve.html index 5283587f07..ccbac9c4cb 100644 --- a/ietf/templates/doc/conflict_review/approve.html +++ b/ietf/templates/doc/conflict_review/approve.html @@ -2,10 +2,10 @@ {# Copyright The IETF Trust 2015, All Rights Reserved #} {% load origin %} {% load django_bootstrap5 %} -{% block title %}Approve {{ review.canonical_name }}{% endblock %} +{% block title %}Approve {{ review.name }}{% endblock %} {% block content %} {% origin %} -

    Approve {{ review.canonical_name }}

    +

    Approve {{ review.name }}

    {% csrf_token %} {% bootstrap_form form %} diff --git a/ietf/templates/doc/conflict_review/start.html b/ietf/templates/doc/conflict_review/start.html index f1b33bc6b2..d8abc2b811 100644 --- a/ietf/templates/doc/conflict_review/start.html +++ b/ietf/templates/doc/conflict_review/start.html @@ -3,13 +3,13 @@ {% load origin %} {% load django_bootstrap5 %} {% load ietf_filters %} -{% block title %}Begin IETF conflict review for {{ doc_to_review.canonical_name }}-{{ doc_to_review.rev }}{% endblock %} +{% block title %}Begin IETF conflict review for {{ doc_to_review.name }}-{{ doc_to_review.rev }}{% endblock %} {% block content %} {% origin %}

    Begin IETF conflict review
    - {{ doc_to_review.canonical_name }}-{{ doc_to_review.rev }} + {{ doc_to_review.name }}-{{ doc_to_review.rev }}

    {% if user|has_role:"Secretariat" %}

    diff --git a/ietf/templates/doc/conflict_review/submit.html b/ietf/templates/doc/conflict_review/submit.html index a95a3c5776..8259c6b12f 100644 --- a/ietf/templates/doc/conflict_review/submit.html +++ b/ietf/templates/doc/conflict_review/submit.html @@ -2,16 +2,16 @@ {# Copyright The IETF Trust 2015, All Rights Reserved #} {% load origin %} {% load django_bootstrap5 %} -{% block title %}Edit conflict review for {{ conflictdoc.canonical_name }}-{{ conflictdoc.rev }}{% endblock %} +{% block title %}Edit conflict review for {{ conflictdoc.name }}-{{ conflictdoc.rev }}{% endblock %} {% block content %} {% origin %}

    Edit conflict review
    - {{ conflictdoc.canonical_name }}-{{ conflictdoc.rev }} + {{ conflictdoc.name }}-{{ conflictdoc.rev }}

    - The text will be submitted as {{ review.canonical_name }}-{{ next_rev }} + The text will be submitted as {{ review.name }}-{{ next_rev }}

    {% csrf_token %} @@ -27,7 +27,7 @@

    Reset to template text + href="{% url "ietf.doc.views_doc.document_main" name=review.name %}"> Back diff --git a/ietf/templates/doc/disclaimer.html b/ietf/templates/doc/disclaimer.html new file mode 100644 index 0000000000..db4c42ed68 --- /dev/null +++ b/ietf/templates/doc/disclaimer.html @@ -0,0 +1,34 @@ +{# Copyright The IETF Trust 2016-2023, All Rights Reserved #} +{% load origin %} +{% load ietf_filters %} +{% origin %} +{% if doc.type_id == "rfc" %} + {% if doc.stream.slug != "ietf" and doc.stream.desc != "Legacy" and doc.std_level.slug|default:"unk" not in "bcp,ds,ps,std"|split:"," %} + + {% elif doc.stream.slug != "ietf" and doc.stream.desc == "Legacy" and doc.std_level.slug|default:"unk" not in "bcp,ds,ps,std"|split:"," %} + + {% endif %} +{% elif doc|is_in_stream %} + {% if doc.stream.slug != "ietf" and doc.std_level.slug|default:"unk" not in "bcp,ds,ps,std"|split:"," %} + + {% endif %} +{% else %} + +{% endif %} diff --git a/ietf/templates/doc/document_ballot_content.html b/ietf/templates/doc/document_ballot_content.html index 1fc4680f8b..e0feb78bc7 100644 --- a/ietf/templates/doc/document_ballot_content.html +++ b/ietf/templates/doc/document_ballot_content.html @@ -1,4 +1,4 @@ -{# Copyright The IETF Trust 2015, All Rights Reserved #} +{# Copyright The IETF Trust 2015-2022, All Rights Reserved #} {% load origin %} {% origin %} {% load ietf_filters %} @@ -11,10 +11,10 @@

    {% for p in positions %}
    - {% if p.is_old_pos %}({% endif %}{% if p.comment or p.discuss %}{% endif %}{{ p.balloter.plain_name }}{% if p.comment or p.discuss %}{% endif %}{% if p.is_old_pos %}){% endif %} + {% if p.is_old_pos %}({% endif %}{% if p.comment or p.discuss %}{% endif %}{{ p.balloter.plain_name }}{% if p.comment or p.discuss %}{% endif %}{% if p.is_old_pos %}){% endif %}
    {% empty %} - (None) + (None) {% endfor %}
    {% endfor %} @@ -53,33 +53,33 @@ Ballot question: "{{ ballot.ballot_type.question }}"

    {% endif %} - {% if editable and user|has_role:"Area Director,Secretariat,IRSG Member" %} + {% if editable and user|has_role:"Area Director,Secretariat,IRSG Member,RSAB Member" %} + href="https://mailarchive.ietf.org/arch/search/?q=subject:{{ doc.name }}+AND+subject:(discuss+OR+comment+OR+review+OR+concern)"> Search Mailarchive {% if user|can_ballot:doc %} + href="{% url "ietf.doc.views_ballot.edit_position" name=doc.name ballot_id=ballot.pk %}?ballot_edit_return_point={{ request.path|urlencode }}"> Edit position {% endif %} - {% if doc.type_id == "draft" or doc.type_id == "conflrev" or doc.type_id == "statchg" %} + {% if user|can_defer:doc %} {% if deferred %} Undefer ballot + href="{% url 'ietf.doc.views_ballot.undefer_ballot' name=doc.name %}">Undefer ballot {% else %} {% if doc.telechat_date %} Defer ballot + href="{% url 'ietf.doc.views_ballot.defer_ballot' name=doc.name %}">Defer ballot {% endif %} {% endif %} - {% if user|has_role:"Area Director,Secretariat" and ballot.ballot_type.slug != "irsg-approve" %} + {% endif %} + {% if user|can_clear_ballot:doc %} + href="{% url 'ietf.doc.views_ballot.clear_ballot' name=doc.name ballot_type_slug=ballot.ballot_type.slug %}"> Clear ballot - {% endif %} {% endif %} {% endif %} {% for n, positions in position_groups %} @@ -91,7 +91,7 @@ {{ p.balloter.plain_name }}
    - {% if p.old_positions %}(was {{ p.old_positions|join:", " }}){% endif %} + {% if p.old_positions %}(was {{ p.old_positions|join:", " }}){% endif %} {{ p.pos }} {% if user|has_role:"Secretariat" %} {% if p.pos.blocking and p.discuss %}
    -
    +
    {{ p.pos.name }} ({{ p.discuss_time|date:"Y-m-d" }}{% if not p.for_current_revision and p.get_dochistory.rev %}{% if p.discuss_time %} {% endif %}for -{{ p.get_dochistory.rev }}{% endif %}) + {# TODO: This logic is in place for discusses/ballots/comments, consider centralizing somewhere #} {% if p.send_email %} - + + Sent + {% elif p.any_email_sent == True %} - + + Sent for earlier + {% elif p.any_email_sent == False %} - + + Not sent + {% else %} - + + Unknown + {% endif %}
    @@ -135,17 +144,25 @@ Comment ({{ p.comment_time|date:"Y-m-d" }}{% if not p.for_current_revision and p.get_dochistory.rev %}{% if p.comment_time %} {% endif %}for -{{ p.get_dochistory.rev }}{% endif %}) {% if p.send_email %} - + + Sent + {% elif p.any_email_sent == True %} - + + Sent for earlier + {% elif p.any_email_sent == False %} - + + Not sent + {% else %} - + + Unknown + {% endif %}
    @@ -162,11 +179,11 @@ {% if p.is_old_pos %}
    - diff --git a/ietf/templates/doc/document_bibtex.bib b/ietf/templates/doc/document_bibtex.bib index 5dda4649eb..5e52ec3c58 100644 --- a/ietf/templates/doc/document_bibtex.bib +++ b/ietf/templates/doc/document_bibtex.bib @@ -3,7 +3,7 @@ {% load ietf_filters %} {% load textfilters %} -{% if doc.get_state_slug == "rfc" %} +{% if doc.type_id == "rfc" %} {% if doc.stream|slugify == "legacy" %} % Datatracker information for RFCs on the Legacy Stream is unfortunately often % incorrect. Please correct the bibtex below based on the information in the @@ -16,7 +16,7 @@ @misc{ publisher = {RFC Editor}, doi = {% templatetag openbrace %}{{ doi }}{% templatetag closebrace %}, url = {% templatetag openbrace %}{{ doc.rfc_number|rfceditor_info_url }}{% templatetag closebrace %},{% else %} -{% if published %}%% You should probably cite rfc{{ latest_revision.doc.rfc_number }} instead of this I-D.{% else %}{% if replaced_by %}%% You should probably cite {{replaced_by|join:" or "}} instead of this I-D.{% else %} +{% if published_as %}%% You should probably cite rfc{{ published_as.rfc_number }} instead of this I-D.{% else %}{% if replaced_by %}%% You should probably cite {{replaced_by|join:" or "}} instead of this I-D.{% else %} {% if doc.rev != latest_revision.rev %}%% You should probably cite {{latest_revision.doc.name}}-{{latest_revision.rev}} instead of this revision.{%endif%}{% endif %}{% endif %} @techreport{% templatetag openbrace %}{{doc.name|slice:"6:"}}-{{doc.rev}}, number = {% templatetag openbrace %}{{doc.name}}-{{doc.rev}}{% templatetag closebrace %}, @@ -25,11 +25,11 @@ @techreport{ publisher = {% templatetag openbrace %}Internet Engineering Task Force{% templatetag closebrace %}, note = {% templatetag openbrace %}Work in Progress{% templatetag closebrace %}, url = {% templatetag openbrace %}{{ settings.IDTRACKER_BASE_URL }}{% url 'ietf.doc.views_doc.document_main' name=doc.name rev=doc.rev %}{% templatetag closebrace %},{% endif %} - author = {% templatetag openbrace %}{% for author in doc.documentauthor_set.all %}{{ author.person.name|texescape}}{% if not forloop.last %} and {% endif %}{% endfor %}{% templatetag closebrace %}, + author = {% templatetag openbrace %}{% for author in doc.documentauthor_set.all %}{{ author.person.name|texescape}}{% if not forloop.last %} and {% endif %}{% endfor %}{% templatetag closebrace %}, title = {% templatetag openbrace %}{% templatetag openbrace %}{{doc.title|texescape}}{% templatetag closebrace %}{% templatetag closebrace %}, pagetotal = {{ doc.pages }}, year = {{ doc.pub_date.year }}, - month = {{ doc.pub_date|date:"b" }},{% if not doc.rfc_number or doc.pub_date.day == 1 and doc.pub_date.month == 4 %} + month = {{ doc.pub_date|date:"b" }},{% if not doc.type_id == "rfc" or doc.pub_date.day == 1 and doc.pub_date.month == 4 %} day = {{ doc.pub_date.day }},{% endif %} abstract = {% templatetag openbrace %}{{ doc.abstract|clean_whitespace|texescape }}{% templatetag closebrace %}, {% templatetag closebrace %} diff --git a/ietf/templates/doc/document_bofreq.html b/ietf/templates/doc/document_bofreq.html index 6cb5df88b4..cbf64c1484 100644 --- a/ietf/templates/doc/document_bofreq.html +++ b/ietf/templates/doc/document_bofreq.html @@ -9,7 +9,7 @@ {% origin %} {{ top|safe }} {% include "doc/revisions_list.html" %} -
    +
    {% if doc.rev != latest_rev %}
    The information below is for an older version of this BOF request.
    {% endif %} @@ -21,7 +21,7 @@ {{ doc.get_state.slug|capfirst }} BOF request - {% if snapshot %}Snapshot{% endif %} + {% if snapshot %}Snapshot{% endif %} @@ -135,7 +135,7 @@ {% endif %} - {{ doc.notify|default:'(None)' }} + {{ doc.notify|default:'(None)' }} @@ -154,7 +154,7 @@
    {{ doc.name }}-{{ doc.rev }}
    -
    +
    {{ content }}
    diff --git a/ietf/templates/doc/document_charter.html b/ietf/templates/doc/document_charter.html index a811f26aba..7564e1d213 100644 --- a/ietf/templates/doc/document_charter.html +++ b/ietf/templates/doc/document_charter.html @@ -14,7 +14,7 @@ {% origin %} {{ top|safe }} {% include "doc/revisions_list.html" %} -
    +
    {% if doc.rev|charter_major_rev != latest_rev|charter_major_rev %}
    The information below is for an older @@ -59,7 +59,7 @@ {{ group.name }} {{ group.type.name }} ({{ group.acronym }}) - {% if snapshot %}Snapshot{% endif %} + {% if snapshot %}Snapshot{% endif %} @@ -94,8 +94,8 @@ {% else %} No document state {% endif %} - {% if chartering == "initial" %}Initial chartering{% endif %} - {% if chartering == "rechartering" %}Rechartering{% endif %} + {% if chartering == "initial" %}Initial chartering{% endif %} + {% if chartering == "rechartering" %}Rechartering{% endif %} @@ -152,7 +152,7 @@ {% if not telechat %} - (None) + (None) {% else %} On agenda of {{ telechat.telechat_date|date:"Y-m-d" }} IESG telechat {% endif %} @@ -175,7 +175,7 @@ {% endif %} - {{ doc.notify|default:'(None)' }} + {{ doc.notify|default:'(None)' }} @@ -227,10 +227,10 @@ {% if doc.rev != "" %}
    - {{ doc.canonical_name }}-{{ doc.rev }} + {{ doc.name }}-{{ doc.rev }}
    -
    {{ content|maybewordwrap|urlize_ietf_docs|linkify }}
    + {{ content }}
    {% endif %} diff --git a/ietf/templates/doc/document_chatlog.html b/ietf/templates/doc/document_chatlog.html index 1e88810fea..7c97471e69 100644 --- a/ietf/templates/doc/document_chatlog.html +++ b/ietf/templates/doc/document_chatlog.html @@ -8,7 +8,7 @@ {% origin %} {{ top|safe }} {% include "doc/revisions_list.html" %} -
    +
    {% if doc.rev != latest_rev %}
    The information below is for an old version of the document.
    {% endif %} @@ -26,13 +26,13 @@ ({{ doc.group.acronym }}) {{ doc.group.type.name }} {% endif %} - {% if snapshot %}Snapshot{% endif %} + {% if snapshot %}Snapshot{% endif %} Title - {{ doc.title|default:'(None)' }} + {{ doc.title|default:'(None)' }} Session diff --git a/ietf/templates/doc/document_conflict_review.html b/ietf/templates/doc/document_conflict_review.html index 9e9f78b685..8a2361832b 100644 --- a/ietf/templates/doc/document_conflict_review.html +++ b/ietf/templates/doc/document_conflict_review.html @@ -10,7 +10,7 @@ {% origin %} {{ top|safe }} {% include "doc/revisions_list.html" %} -
    +
    {% if doc.rev != latest_rev %}
    The information below is for an old version of the document.
    {% endif %} @@ -27,10 +27,10 @@ - - {% if conflictdoc.get_state_slug == 'rfc' %}{{ conflictdoc.canonical_name|prettystdname }}{% else %}{{ conflictdoc.canonical_name }}-{{ conflictdoc.rev }}{% endif %} - {{ conflictdoc.stream }} stream - {% if snapshot %}Snapshot{% endif %} + + {% if conflictdoc.type_id == 'rfc' %}{{ conflictdoc.name|prettystdname }}{% else %}{{ conflictdoc.name }}-{{ conflictdoc.rev }}{% endif %} + {{ conflictdoc.stream }} stream + {% if snapshot %}Snapshot{% endif %} @@ -54,11 +54,7 @@ {% endif %} - {% if "no-problem" in doc.get_state.name|slugify %} - {{ doc.get_state.name}} - {% else %} - {{ doc.get_state.name }} - {% endif %} + {{ doc.get_state.name|badgeify }} @@ -90,7 +86,7 @@ {% if not telechat %} - (None) + (None) {% else %} On agenda of {{ telechat.telechat_date|date:"Y-m-d" }} IESG telechat {% if doc.returning_item %}(returning item){% endif %} @@ -149,4 +145,4 @@ -{% endblock %} +{% endblock %} \ No newline at end of file diff --git a/ietf/templates/doc/document_draft.html b/ietf/templates/doc/document_draft.html index f763159fca..eab1d779fb 100644 --- a/ietf/templates/doc/document_draft.html +++ b/ietf/templates/doc/document_draft.html @@ -1,5 +1,5 @@ {% extends "base.html" %} -{# Copyright The IETF Trust 2016-2020, All Rights Reserved #} +{# Copyright The IETF Trust 2016-2024, All Rights Reserved #} {% load origin %} {% load static %} {% load ietf_filters %} @@ -13,11 +13,11 @@ title="Document changes" href="/feed/document-changes/{{ name }}/"> + content="{{ doc.title }} {% if doc.type_id == 'rfc' and not snapshot %}(RFC {{ rfc_number }}{% if published %}, {{ doc.pub_date|date:'F Y' }}{% endif %}{% if obsoleted_by %}; obsoleted by {% for rel in obsoleted_by %}{{ rel.source.name|prettystdname}}{% if not forloop.last%}, {% endif %}{% endfor %}{% endif %}){% endif %}"> {% endblock %} {% block morecss %}.inline { display: inline; }{% endblock %} {% block title %} - {% if doc.get_state_slug == "rfc" and not snapshot %} + {% if doc.type_id == "rfc" and not snapshot %} RFC {{ rfc_number }} - {{ doc.title }} {% else %} {{ name }}-{{ doc.rev }} - {{ doc.title }} @@ -27,11 +27,12 @@ {% origin %} {{ top|safe }} {% include "doc/revisions_list.html" with document_html=document_html %} -
    + {% include "doc/disclaimer.html" with document_html=document_html %} +
    {% if doc.rev != latest_rev %}
    The information below is for an old version of the document.
    {% else %} - {% if doc.get_state_slug == "rfc" and snapshot %} + {% if doc.became_rfc %}
    The information below is for an old version of the document that is already published as an RFC.
    @@ -47,7 +48,7 @@ {% if doc.stream %} {% if doc.stream.slug != "legacy" %} - + {% if doc.stream_id == 'ietf' %} WG {% else %} @@ -62,7 +63,12 @@ {% if doc.stream and can_edit_stream_info and doc.stream.slug != "legacy" and not snapshot %} + {% if doc|is_doc_ietf_adoptable or doc|can_issue_ietf_wg_lc or doc|can_submit_to_iesg %} + href="{% url 'ietf.doc.views_draft.offer_wg_action_helpers' name=doc.name %}" + {% else %} + href="{% url 'ietf.doc.views_draft.change_stream_state' name=doc.name state_type=stream_state_type_slug %}" + {% endif %} + > Edit {% endif %} @@ -73,7 +79,7 @@ {{ stream_state }} {% else %} - + (None) {% endif %} @@ -85,7 +91,7 @@
    {% endif %} {% if due_date %} - (Due date {{ due_date }}) + (Due date {{ due_date }}) {% endif %} {% else %} @@ -95,7 +101,7 @@ - (No stream defined) + (No stream defined) {% endif %} @@ -166,18 +172,18 @@ {% if pres.rev and pres.rev != doc.rev %}(version -{{ pres.rev }}){% endif %}{% if not forloop.last %},{% endif %} {% endfor %} {% else %} - + (None) {% endif %} {% endif %} - {% if doc.stream_id == 'ietf' or doc.stream_id == 'ise' or doc.stream_id == 'irtf' or doc.stream_id == 'editorial' %} + {% if doc.stream_id == 'ietf' or doc.stream_id == 'ise' or doc.stream_id == 'irtf' or doc.stream_id == 'editorial' or doc.stream_id == 'iab' %} - Document shepherd + {% if doc.stream_id == 'iab' %}IAB shepherd{% else %}Document shepherd{% endif %} {% if can_edit_stream_info and not snapshot %} @@ -196,7 +202,7 @@ {% if doc.shepherd %} {% person_link doc.shepherd.person %} {% else %} - + (None) {% endif %} @@ -224,9 +230,9 @@ href="{% url 'ietf.doc.views_doc.document_shepherd_writeup' name=doc.name %}"> Show - Last changed {{ shepherd_writeup.time|date:"Y-m-d" }} + Last changed {{ shepherd_writeup.time|date:"Y-m-d" }} {% else %} - + (None) {% endif %} @@ -251,7 +257,7 @@ {% if doc.has_rfc_editor_note %} (last changed {{ doc.has_rfc_editor_note|date:"Y-m-d" }}) {% else %} - + (None) {% endif %} @@ -267,19 +273,19 @@ {% endif %} - {% if not doc.stream_id == 'iab' %} + {% if doc.stream_id != 'iab' and doc.stream_id != 'editorial' %} IESG - + IESG state - {% if iesg_state.slug != 'idexists' and can_edit %} + {% if iesg_state.slug != 'idexists' and iesg_state.slug != 'dead' and can_edit or user|has_role:"Secretariat" %} Edit @@ -303,7 +309,7 @@ Action Holder{{ doc.documentactionholder_set.all|pluralize }} - {% if can_edit %} + {% if can_edit_action_holders %} Edit @@ -318,7 +324,7 @@ {% person_link action_holder.person title=action_holder.role_for_doc %} {{ action_holder|action_holder_badge }}
    {% endfor %} - {% if can_edit %} + {% if can_edit_action_holders %} @@ -327,7 +333,7 @@ {% endif %} {% else %} - + (None) {% endif %} @@ -375,7 +381,7 @@ On agenda of {{ telechat.telechat_date }} IESG telechat {% if telechat.returning_item %}(returning item){% endif %} {% else %} - + (None) {% endif %} @@ -402,38 +408,25 @@ {% if doc.ad %} {% person_link doc.ad %} {% else %} - + (None) {% endif %} - {% if iesg_state.slug != 'idexists' %} - {% if doc.note or can_edit %} - - - - IESG note - - - {% if can_edit and not snapshot %} - - Edit - - {% endif %} - - - {% if doc.note %} - {{ doc.note|linebreaksbr }} - {% else %} - - (None) - - {% endif %} - - - {% endif %} + {% if iesg_state.slug != 'idexists' and doc.note %} + + + + IESG note + + + {# IESG Notes are historic and read-only now #} + + + {{ doc.notedoc.note|urlize_ietf_docs|linkify|linebreaksbr }} + + {% endif %} @@ -452,7 +445,7 @@ {% if doc.notify %} {{ doc.notify|linkify }} {% else %} - + (None) {% endif %} @@ -460,125 +453,127 @@ {% endif %} - {% if can_edit_iana_state or iana_review_state or iana_experts_state or iana_experts_comment %} - - {% if iana_review_state or can_edit_iana_state %} - - - IANA - - - - IANA review state - - - - {% if can_edit_iana_state and not snapshot %} - - Edit - - {% endif %} - - - {% if not iana_review_state %} - - (None) - - {% else %} - {{ iana_review_state }} - {% endif %} - - - {% endif %} - {% if iana_action_state or can_edit_iana_state %} - - - {% if not can_edit_iana_state and not iana_review_state %}IANA{% endif %} - - - - IANA action state - - - - {% if can_edit_iana_state and not snapshot %} - - Edit + {% if doc.stream_id != 'editorial' %} + {% if can_edit_iana_state or iana_review_state or iana_experts_state or iana_experts_comment %} + + {% if iana_review_state or can_edit_iana_state %} + + + IANA + + + + IANA review state - {% endif %} - - - {% if not iana_action_state %} - - (None) - - {% else %} - {{ iana_action_state }} - {% endif %} - - - {% endif %} - {% if iana_experts_state or can_edit_iana_state %} - - - {% if not can_edit_iana_state and not iana_review_state and not iana_action_state %}IANA{% endif %} - - - - IANA expert review state - - - - {% if can_edit_iana_state and not snapshot %} - - Edit + + + {% if can_edit_iana_state and not snapshot %} + + Edit + + {% endif %} + + + {% if not iana_review_state %} + + (None) + + {% else %} + {{ iana_review_state }} + {% endif %} + + + {% endif %} + {% if iana_action_state or can_edit_iana_state %} + + + {% if not can_edit_iana_state and not iana_review_state %}IANA{% endif %} + + + + IANA action state - {% endif %} - - - {% if not iana_experts_state %} - - (None) - - {% else %} - {{ iana_experts_state }} - {% endif %} - - - {% endif %} - {% if iana_experts_comment or can_edit_iana_state %} - - - {% if not can_edit_iana_state and not iana_review_state and not iana_action_state and not iana_experts_state %} - IANA - {% endif %} - - - IANA expert review comments - - - {% if can_edit_iana_state and not snapshot %} - - Edit + + + {% if can_edit_iana_state and not snapshot %} + + Edit + + {% endif %} + + + {% if not iana_action_state %} + + (None) + + {% else %} + {{ iana_action_state }} + {% endif %} + + + {% endif %} + {% if iana_experts_state or can_edit_iana_state %} + + + {% if not can_edit_iana_state and not iana_review_state and not iana_action_state %}IANA{% endif %} + + + + IANA expert review state - {% endif %} - - - {% if not iana_experts_comment %} - - (None) - - {% else %} - {{ iana_experts_comment }} - {% endif %} - - - {% endif %} - + + + {% if can_edit_iana_state and not snapshot %} + + Edit + + {% endif %} + + + {% if not iana_experts_state %} + + (None) + + {% else %} + {{ iana_experts_state }} + {% endif %} + + + {% endif %} + {% if iana_experts_comment or can_edit_iana_state %} + + + {% if not can_edit_iana_state and not iana_review_state and not iana_action_state and not iana_experts_state %} + IANA + {% endif %} + + + IANA expert review comments + + + {% if can_edit_iana_state and not snapshot %} + + Edit + + {% endif %} + + + {% if not iana_experts_comment %} + + (None) + + {% else %} + {{ iana_experts_comment }} + {% endif %} + + + {% endif %} + + {% endif %} {% endif %} {% if rfc_editor_state %} @@ -587,7 +582,7 @@ RFC Editor - + RFC Editor state @@ -644,20 +639,20 @@ IPR {% if doc.related_ipr %} - + {{ doc.related_ipr|length }} {% endif %} References @@ -671,70 +666,50 @@ Nits - + {% if user|has_role:"Area Director" %} + {# IDNITS3 is an experimental service, so only show it to Area Directors #} + + + + Nits-v3 (Experimental) + + {% endif %} + + + + Search email archive + {% if user.is_authenticated %} - + Untrack - + Track {% endif %} - {% if user.review_teams %} - Remove review wishes - @@ -764,10 +739,10 @@ {% endfor %} {% endif %}
    - {% if doc.get_state_slug == "active" or doc.get_state_slug == "rfc" %} + {% if doc.get_state_slug == "active" or doc.type_id == "rfc" or doc.became_rfc %}
    - {% if doc.get_state_slug == "rfc" and not snapshot %} + {% if doc.type_id == "rfc" and not snapshot %} RFC {{ rfc_number }} {% else %} {{ name }}-{{ doc.rev }} @@ -786,7 +761,7 @@ {% endif %} {% else %}
    - {% else %} - (not online) + (not online) {% endif %} \ No newline at end of file diff --git a/ietf/templates/doc/document_history.html b/ietf/templates/doc/document_history.html index 67e34d0cd9..78471e08d6 100644 --- a/ietf/templates/doc/document_history.html +++ b/ietf/templates/doc/document_history.html @@ -11,16 +11,42 @@ - + {% endblock %} {% block content %} {% origin %} {{ top|safe }} {% if diff_revisions and diff_revisions|length > 1 or doc.name|rfcbis %}

    Revision differences

    - {% include "doc/document_history_form.html" with doc=doc diff_revisions=diff_revisions action=rfcdiff_base_url only %} + {% include "doc/document_history_form.html" with doc=doc diff_revisions=diff_revisions action=rfcdiff_base_url snapshot=snapshot only %} + {% endif %} + {% if ballot_doc_rev %} +

    + Your most recent ballot position was provided when the document was at version {{ ballot_doc_rev }}. +

    {% endif %}

    Document history

    + {% if doc.came_from_draft %} +
    + {% endif %} + {% if doc.became_rfc %} + + {% endif %} + {% if can_add_comment %}

    Date - Rev. + {% if doc.type_id not in "rfc,bcp,std,fyi" %}Rev.{% endif %} By Action @@ -45,7 +71,7 @@

    Document history

    {{ e.time|date:"Y-m-d" }}
    - {{ e.rev }} + {% if doc.type_id not in "rfc,bcp,std,fyi" %}{{ e.rev }}{% endif %} {{ e.by|escape }} {{ e.desc|format_history_text }} diff --git a/ietf/templates/doc/document_history_form.html b/ietf/templates/doc/document_history_form.html index 0045a70d9c..646da0038b 100644 --- a/ietf/templates/doc/document_history_form.html +++ b/ietf/templates/doc/document_history_form.html @@ -13,20 +13,19 @@ {% endif %} - {% for name, rev, time, url in diff_revisions %} - diff --git a/ietf/templates/doc/document_html.html b/ietf/templates/doc/document_html.html index 6a7e86eb7b..a85b1f5310 100644 --- a/ietf/templates/doc/document_html.html +++ b/ietf/templates/doc/document_html.html @@ -4,20 +4,24 @@ {% load origin %} {% load static %} {% load ietf_filters textfilters %} +{% load document_type_badge %} +{% load django_vite %} {% origin %} - + {% analytical_head_top %} - {% if not snapshot and doc.get_state_slug == "rfc" %} + {% if doc.type_id == "rfc" %} RFC {{ doc.rfc_number }} - {{ doc.title }} {% else %} {{ doc.name }}-{{ doc.rev }} {% endif %} + + {% if request.COOKIES.pagedeps == 'inline' %} @@ -26,44 +30,85 @@ {% if html %} {% endif %} + {% vite_asset 'client/embedded.js' %} + {% endif %} - + {% include "base/icons.html" %} {% include "doc/opengraph.html" %} - {% analytical_head_bottom %} + {% analytical_head_bottom %} + {% analytical_body_top %} - -