Refactoring (#1)
This PR is very far from perfect, and there are some things that do not work yet, but I think it's good enough to merge Co-authored-by: Shyam Saraswati <shyam.saraswati@gmail.com> Co-authored-by: Nikolai Rodionov <nikolai.rodionov@grandcentrix.net> Reviewed-on: https://git.badhouseplants.net/allanger/mkdocs-with-confluence/pulls/1
This commit is contained in:
parent
04600fbbe4
commit
8efe5ceda6
14
.coveragerc
14
.coveragerc
@ -1,14 +0,0 @@
|
|||||||
[run]
|
|
||||||
branch = True
|
|
||||||
source = mkdocs_with_confluence
|
|
||||||
include = */mkdocs_with_confluence/*
|
|
||||||
|
|
||||||
[report]
|
|
||||||
exclude_lines =
|
|
||||||
if self.debug:
|
|
||||||
pragma: no cover
|
|
||||||
raise NotImplementedError
|
|
||||||
if __name__ == .__main__.:
|
|
||||||
ignore_errors = True
|
|
||||||
omit =
|
|
||||||
tests/*
|
|
1
.dockerignore
Normal file
1
.dockerignore
Normal file
@ -0,0 +1 @@
|
|||||||
|
Dockerfile
|
60
.github/workflows/release_version.yaml
vendored
Normal file
60
.github/workflows/release_version.yaml
vendored
Normal file
@ -0,0 +1,60 @@
|
|||||||
|
---
|
||||||
|
name: "Version build"
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
tags:
|
||||||
|
- "v*.*.*"
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
build:
|
||||||
|
runs-on: ${{ matrix.os }}
|
||||||
|
strategy:
|
||||||
|
fail-fast: false
|
||||||
|
# Currently, only for one arch, because I've got no idea about cross-platform and python
|
||||||
|
matrix:
|
||||||
|
include:
|
||||||
|
- os: ubuntu-latest
|
||||||
|
steps:
|
||||||
|
- name: Checkout
|
||||||
|
uses: actions/checkout@v3
|
||||||
|
|
||||||
|
- uses: actions/setup-python@v4
|
||||||
|
with:
|
||||||
|
python-version: 3.9
|
||||||
|
|
||||||
|
- name: Setup poetry
|
||||||
|
uses: abatilo/actions-poetry@v2
|
||||||
|
with:
|
||||||
|
poetry-version: 1.3.2
|
||||||
|
|
||||||
|
- name: Build
|
||||||
|
run: poetry build
|
||||||
|
|
||||||
|
- name: Get archive name
|
||||||
|
run: echo "ARCHIVE_NAME=$(find $GITHUB_WORKSPACE/dist -maxdepth 1 -mindepth 1 -name '*tar.gz' -execdir basename '{}' ';')" >> $GITHUB_ENV
|
||||||
|
|
||||||
|
- name: Archive build artifacts
|
||||||
|
uses: actions/upload-artifact@v3
|
||||||
|
with:
|
||||||
|
name: ${{ env.ARCHIVE_NAME }}
|
||||||
|
path: ${{ github.workspace }}/dist/${{ env.ARCHIVE_NAME }}
|
||||||
|
|
||||||
|
release:
|
||||||
|
needs: build
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
steps:
|
||||||
|
- name: Checkout
|
||||||
|
uses: actions/checkout@v3
|
||||||
|
|
||||||
|
- name: Download artifact
|
||||||
|
uses: actions/download-artifact@v3
|
||||||
|
with:
|
||||||
|
path: ${{ github.workspace }}/release/
|
||||||
|
- name: check
|
||||||
|
run: ls -R
|
||||||
|
|
||||||
|
- name: Release
|
||||||
|
uses: softprops/action-gh-release@v1
|
||||||
|
with:
|
||||||
|
files: ${{ github.workspace }}/release/*/*
|
1
.gitignore
vendored
1
.gitignore
vendored
@ -1,2 +1,3 @@
|
|||||||
mkdocs_with_confluence/__pycache__/*
|
mkdocs_with_confluence/__pycache__/*
|
||||||
mkdocs_with_confluence.egg-info/*
|
mkdocs_with_confluence.egg-info/*
|
||||||
|
dist
|
||||||
|
@ -1,35 +0,0 @@
|
|||||||
repos:
|
|
||||||
- repo: local
|
|
||||||
hooks:
|
|
||||||
- id: black
|
|
||||||
name: black
|
|
||||||
entry: black
|
|
||||||
language: python
|
|
||||||
types: [python]
|
|
||||||
language_version: python3.7
|
|
||||||
args: [--line-length=120]
|
|
||||||
- repo: local
|
|
||||||
hooks:
|
|
||||||
- id: mypy
|
|
||||||
name: mypy
|
|
||||||
entry: mypy
|
|
||||||
language: system
|
|
||||||
types: [python]
|
|
||||||
args: [--ignore-missing-imports, --namespace-packages, --show-error-codes, --pretty]
|
|
||||||
- repo: local
|
|
||||||
hooks:
|
|
||||||
- id: flake8
|
|
||||||
name: flake8
|
|
||||||
entry: flake8
|
|
||||||
language: system
|
|
||||||
types: [python]
|
|
||||||
args: [--max-line-length=120, "--ignore=D101,D104,D212,D200,E203,W293,D412,W503"]
|
|
||||||
# D100 requires all Python files (modules) to have a "public" docstring even if all functions within have a docstring.
|
|
||||||
# D104 requires __init__ files to have a docstring
|
|
||||||
# D212
|
|
||||||
# D200
|
|
||||||
# D412 No blank lines allowed between a section header and its content
|
|
||||||
# E203
|
|
||||||
# W293 blank line contains whitespace
|
|
||||||
# W503 line break before binary operator (for compatibility with black)
|
|
||||||
|
|
6
.sops.yaml
Normal file
6
.sops.yaml
Normal file
@ -0,0 +1,6 @@
|
|||||||
|
creation_rules:
|
||||||
|
- path_regex: .*
|
||||||
|
key_groups:
|
||||||
|
- age:
|
||||||
|
- age1yxxn8xscjdue05apn2knfzjy99gwk4x9djfqzgzg83h0hqnwra7qny775y
|
||||||
|
|
@ -1,9 +0,0 @@
|
|||||||
language: python
|
|
||||||
script:
|
|
||||||
- pip install -r requirements_dev.txt
|
|
||||||
- python setup.py install
|
|
||||||
- flake8 --max-line-length=120 --ignore=D101,D104,D212,D200,E203,W293,D412,W503 mkdocs_with_confluence/
|
|
||||||
- black --check --line-length=120 mkdocs_with_confluence/
|
|
||||||
- nosetests --with-coverage
|
|
||||||
after_success:
|
|
||||||
- bash <(curl -s https://codecov.io/bash)
|
|
16
Makefile
Normal file
16
Makefile
Normal file
@ -0,0 +1,16 @@
|
|||||||
|
|
||||||
|
build:
|
||||||
|
poetry build
|
||||||
|
|
||||||
|
venv:
|
||||||
|
python3 -m venv /tmp/venv
|
||||||
|
|
||||||
|
run_example:
|
||||||
|
|
||||||
|
@docker build -t mkdocs-example -f ./example/Dockerfile --build-arg JIRA_PASSWORD=$(shell sops --decrypt ./example/secret.yaml | yq '.JIRA_PASSWORD' ) .
|
||||||
|
@docker run -p 8000:8000 mkdocs-example
|
||||||
|
|
||||||
|
lint:
|
||||||
|
@docker run --rm -v ${PWD}:/data cytopia/pylint ./mkdocs_with_confluence/plugin.py
|
||||||
|
|
||||||
|
|
20
README.md
20
README.md
@ -1,15 +1,18 @@
|
|||||||

|
|
||||||
[](https://travis-ci.com/pawelsikora/mkdocs-with-confluence)
|
|
||||||
[](https://codecov.io/gh/pawelsikora/mkdocs-with-confluence)
|
|
||||||

|
|
||||||

|
|
||||||

|
|
||||||

|
|
||||||
# mkdocs-with-confluence
|
# mkdocs-with-confluence
|
||||||
|
|
||||||
MkDocs plugin that converts markdown pages into confluence markup
|
MkDocs plugin that converts markdown pages into confluence markup
|
||||||
and export it to the Confluence page
|
and export it to the Confluence page
|
||||||
|
|
||||||
|
# How to use
|
||||||
|
To enable plugin, you need to set the `MKDOCS_TO_CONFLUENCE` environment variable.
|
||||||
|
```BASH
|
||||||
|
export MKDOCS_TO_CONFLUENCE=1
|
||||||
|
```
|
||||||
|
By default the dry-run mode is turned off. If you wan't to enable it, you can use the config file, ot the `MKDOCS_TO_CONFLUENCE_DRY_RUN` environment variable
|
||||||
|
```BASH
|
||||||
|
export MKDOCS_TO_CONFLUENCE_DRY_RUN=1
|
||||||
|
```
|
||||||
|
|
||||||
## Setup
|
## Setup
|
||||||
Install the plugin using pip:
|
Install the plugin using pip:
|
||||||
|
|
||||||
@ -36,9 +39,6 @@ Use following config and adjust it according to your needs:
|
|||||||
parent_page_name: <YOUR_ROOT_PARENT_PAGE>
|
parent_page_name: <YOUR_ROOT_PARENT_PAGE>
|
||||||
username: <YOUR_USERNAME_TO_CONFLUENCE>
|
username: <YOUR_USERNAME_TO_CONFLUENCE>
|
||||||
password: <YOUR_PASSWORD_TO_CONFLUENCE>
|
password: <YOUR_PASSWORD_TO_CONFLUENCE>
|
||||||
enabled_if_env: MKDOCS_TO_CONFLUENCE
|
|
||||||
#verbose: true
|
|
||||||
#debug: true
|
|
||||||
dryrun: true
|
dryrun: true
|
||||||
```
|
```
|
||||||
|
|
||||||
|
28
codecov.yml
28
codecov.yml
@ -1,28 +0,0 @@
|
|||||||
comment: false
|
|
||||||
|
|
||||||
coverage:
|
|
||||||
status:
|
|
||||||
project:
|
|
||||||
default:
|
|
||||||
# Commits pushed to master should not make the overall
|
|
||||||
# project coverage decrease by more than x%:
|
|
||||||
target: auto
|
|
||||||
threshold: 20%
|
|
||||||
patch:
|
|
||||||
default:
|
|
||||||
# Be tolerant on slight code coverage diff on PRs to limit
|
|
||||||
# noisy red coverage status on github PRs.
|
|
||||||
# Note: The coverage stats are still uploaded
|
|
||||||
# to codecov so that PR reviewers can see uncovered lines
|
|
||||||
target: auto
|
|
||||||
threshold: 20%
|
|
||||||
|
|
||||||
codecov:
|
|
||||||
notify:
|
|
||||||
# Prevent coverage status to upload multiple times for parallel and long
|
|
||||||
# running CI pipelines. This configuration is particularly useful on PRs
|
|
||||||
# to avoid confusion. Note that this value is set to the number of Azure
|
|
||||||
# Pipeline jobs uploading coverage reports.
|
|
||||||
after_n_builds: 6
|
|
||||||
token: 21ccebfb-1239-4ec6-a01d-039067b9ab30
|
|
||||||
|
|
22
example/Dockerfile
Normal file
22
example/Dockerfile
Normal file
@ -0,0 +1,22 @@
|
|||||||
|
FROM python:latest as BUILDER
|
||||||
|
|
||||||
|
FROM BUILDER as plugin_builder
|
||||||
|
WORKDIR /src
|
||||||
|
COPY . /src
|
||||||
|
RUN pip install poetry
|
||||||
|
RUN poetry build
|
||||||
|
RUN mkdir /out
|
||||||
|
RUN mv $(find /src/dist -maxdepth 1 -mindepth 1 -name '*tar.gz') /out/mkdocs_with_confluence.tar.gz
|
||||||
|
|
||||||
|
FROM BUILDER as common_builder
|
||||||
|
ENV MKDOCS_TO_CONFLUENCE=true
|
||||||
|
ARG JIRA_PASSWORD
|
||||||
|
ENV JIRA_PASSWORD=$JIRA_PASSWORD
|
||||||
|
RUN pip install mkdocs mkdocs-material
|
||||||
|
WORKDIR /src
|
||||||
|
COPY ./example /src
|
||||||
|
COPY --from=plugin_builder /out/mkdocs_with_confluence.tar.gz /tmp/
|
||||||
|
RUN pip install /tmp/mkdocs_with_confluence.tar.gz
|
||||||
|
ENTRYPOINT ["mkdocs"]
|
||||||
|
CMD ["serve", "-a", "0.0.0.0:8000"]
|
||||||
|
|
BIN
example/docs/assets/argocd.png
Normal file
BIN
example/docs/assets/argocd.png
Normal file
Binary file not shown.
After Width: | Height: | Size: 5.7 KiB |
17
example/docs/index.md
Normal file
17
example/docs/index.md
Normal file
@ -0,0 +1,17 @@
|
|||||||
|
# Welcome to MkDocs
|
||||||
|
|
||||||
|
For full documentation visit [mkdocs.org](https://www.mkdocs.org).
|
||||||
|
|
||||||
|
## Commands
|
||||||
|
|
||||||
|
* `mkdocs new [dir-name]` - Create a new project.
|
||||||
|
* `mkdocs serve` - Start the live-reloading docs server.
|
||||||
|
* `mkdocs build` - Build the documentation site.
|
||||||
|
* `mkdocs -h` - Print help message and exit.
|
||||||
|
|
||||||
|
## Project layout
|
||||||
|
|
||||||
|
mkdocs.yml # The configuration file.
|
||||||
|
docs/
|
||||||
|
index.md # The documentation homepage.
|
||||||
|
... # Other markdown pages, images and other files.
|
5
example/docs/section1/page1.md
Normal file
5
example/docs/section1/page1.md
Normal file
@ -0,0 +1,5 @@
|
|||||||
|
# Page1
|
||||||
|
|
||||||
|
Hello there
|
||||||
|
|
||||||
|

|
6
example/docs/section1/page2.md
Normal file
6
example/docs/section1/page2.md
Normal file
@ -0,0 +1,6 @@
|
|||||||
|
# Page2
|
||||||
|
|
||||||
|
Hello there!
|
||||||
|
<img src="assets/argocd.png"
|
||||||
|
alt="Markdown Monster icon"
|
||||||
|
style="float: left; margin-right: 10px;" />
|
3
example/docs/section1/sub_section/page1.md
Normal file
3
example/docs/section1/sub_section/page1.md
Normal file
@ -0,0 +1,3 @@
|
|||||||
|
# Page1
|
||||||
|
|
||||||
|
Hello there
|
3
example/docs/section2/page1.md
Normal file
3
example/docs/section2/page1.md
Normal file
@ -0,0 +1,3 @@
|
|||||||
|
# Page1
|
||||||
|
|
||||||
|
Hello there, it's section two
|
3
example/docs/section2/page2.md
Normal file
3
example/docs/section2/page2.md
Normal file
@ -0,0 +1,3 @@
|
|||||||
|
# Page2
|
||||||
|
|
||||||
|
Hello there! Section two, can't you see?
|
24
example/mkdocs.yml
Normal file
24
example/mkdocs.yml
Normal file
@ -0,0 +1,24 @@
|
|||||||
|
---
|
||||||
|
site_name: My Docs
|
||||||
|
repo_url: https://github.com/allanger/mkdocs_with_confluence
|
||||||
|
nav:
|
||||||
|
# - index: 'index.md'
|
||||||
|
- Section1:
|
||||||
|
- page-1: './section1/page1.md'
|
||||||
|
- page-2: './section1/page2.md'
|
||||||
|
- Sub Section:
|
||||||
|
- page-1: './section1/sub_section/page1.md'
|
||||||
|
- Section2:
|
||||||
|
- page-1: './section2/page1.md'
|
||||||
|
- page-2: './section2/page2.md'
|
||||||
|
theme:
|
||||||
|
name: material
|
||||||
|
plugins:
|
||||||
|
- search
|
||||||
|
- mkdocs-with-confluence:
|
||||||
|
host_url: https://mkdocs-to-conflunce.atlassian.net/wiki/rest/api/content
|
||||||
|
space: MKDOC
|
||||||
|
parent_page_name: MKdocs
|
||||||
|
username: allanguor@gmail.com
|
||||||
|
dryrun: false
|
||||||
|
|
21
example/secret.yaml
Normal file
21
example/secret.yaml
Normal file
@ -0,0 +1,21 @@
|
|||||||
|
JIRA_PASSWORD: ENC[AES256_GCM,data:wBghxjYOmFBVJ/lwTXXiMy4sizH2BnSdmQLOoGqmh3opExNTcmKyUgRk8pVXxBumLkxz2PwHnxAlSNoma5hDwUh/z8UrZdpv3zf1U9JkMLpyp41a/D29sjWjizfL2YURgRmrjmaEk4cQzgF2lXh+1LYcH4s8rhvHGAtMIgezTX1J2DG+yyF+P3bp3IM/eq3Ch1xg08MgFht5nl4qnMsTUnZk8qWpSMiRc6nZiHNFIHlEUsDasfG7AMOZMbjo9tWS,iv:UhYmW6eTtgjId+yZE/acOxayNjXTax2dq1Xs1m3xNpY=,tag:yb/ce2ehAUXFbw5m1v1bzA==,type:str]
|
||||||
|
sops:
|
||||||
|
kms: []
|
||||||
|
gcp_kms: []
|
||||||
|
azure_kv: []
|
||||||
|
hc_vault: []
|
||||||
|
age:
|
||||||
|
- recipient: age1yxxn8xscjdue05apn2knfzjy99gwk4x9djfqzgzg83h0hqnwra7qny775y
|
||||||
|
enc: |
|
||||||
|
-----BEGIN AGE ENCRYPTED FILE-----
|
||||||
|
YWdlLWVuY3J5cHRpb24ub3JnL3YxCi0+IFgyNTUxOSB1TkRNcXZJQ1c1U1M0Sngx
|
||||||
|
YittS1g2Q1VIaC84YWwzNzh0Z0VtRnY2WGxVClBWcXBYM3kvQ1BmWHZ0NFJqSlBi
|
||||||
|
Mjltc0pMRjNnNmpRd1dabmQ4Y205dDQKLS0tIEZMK1JnZlZHaER3SkNNY0pOVXpN
|
||||||
|
a0tPYUI5TXdYUVdudVdDR0svOWd2Z00KNeCnkrIB1Ik0uazO3+aQKY0EEsqFR7Ay
|
||||||
|
FOGlI4GlJu2OXBpQU22eZOxIhwAQILbn29u8yD0ZqT1u4jplW7IdEA==
|
||||||
|
-----END AGE ENCRYPTED FILE-----
|
||||||
|
lastmodified: "2023-02-20T19:09:10Z"
|
||||||
|
mac: ENC[AES256_GCM,data:K2ZPW7fLYf+o9jBiVqfxPJIgasWBqWm02DBFHpwo+ZiQy2AQlEfzBECm/rudHTBV7hyw46UkA3cZf1bypsIcgkUTI7ztrGN/PNMmq1+Qr29l6jgphwBtaAUK3Kxc9PI5lxYjkzVjybCFuWLP/HTwIrLz2qGRRLb5BeGTrGT+MT0=,iv:/t7RIwP0AeT6ifSuVd/S4+EFfvHzPdeU6HFSKySyjlo=,tag:QrI/8agEDwaLCeN4Ccwl1A==,type:str]
|
||||||
|
pgp: []
|
||||||
|
unencrypted_suffix: _unencrypted
|
||||||
|
version: 3.7.3
|
@ -1,7 +1,7 @@
|
|||||||
import time
|
|
||||||
import os
|
import os
|
||||||
import hashlib
|
import hashlib
|
||||||
import sys
|
import sys
|
||||||
|
import glob
|
||||||
import re
|
import re
|
||||||
import tempfile
|
import tempfile
|
||||||
import shutil
|
import shutil
|
||||||
@ -9,16 +9,24 @@ import requests
|
|||||||
import mimetypes
|
import mimetypes
|
||||||
import mistune
|
import mistune
|
||||||
import contextlib
|
import contextlib
|
||||||
|
import time
|
||||||
|
|
||||||
from time import sleep
|
from time import sleep
|
||||||
from mkdocs.config import config_options
|
from mkdocs.config import config_options
|
||||||
from mkdocs.plugins import BasePlugin
|
from mkdocs.plugins import BasePlugin
|
||||||
from md2cf.confluence_renderer import ConfluenceRenderer
|
from md2cf.confluence_renderer import ConfluenceRenderer
|
||||||
from os import environ
|
from os import environ
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
|
from loguru import logger
|
||||||
|
|
||||||
|
ENABLE_ENV_VAR = "MKDOCS_TO_CONFLUENCE"
|
||||||
|
DRY_RUN_ENV_VAR = "MKDOCS_TO_CONFLUENCE_DRY_RUN"
|
||||||
TEMPLATE_BODY = "<p> TEMPLATE </p>"
|
TEMPLATE_BODY = "<p> TEMPLATE </p>"
|
||||||
|
HEADER_MESSAGE = "‼️ This page is created automatically, all you changes will be overwritten during the next MKDocs deployment. Do not edit a page here ‼️"
|
||||||
|
|
||||||
|
SECTION_PAGE_CONTENT = "<p> It's juat a Section Page </p>"
|
||||||
|
|
||||||
|
# -- I don't know why it's here
|
||||||
@contextlib.contextmanager
|
@contextlib.contextmanager
|
||||||
def nostdout():
|
def nostdout():
|
||||||
save_stdout = sys.stdout
|
save_stdout = sys.stdout
|
||||||
@ -26,341 +34,140 @@ def nostdout():
|
|||||||
yield
|
yield
|
||||||
sys.stdout = save_stdout
|
sys.stdout = save_stdout
|
||||||
|
|
||||||
|
# -- I don't know why it's here
|
||||||
class DummyFile(object):
|
class DummyFile(object):
|
||||||
def write(self, x):
|
def write(self, x):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
class MkdocsWithConfluence(BasePlugin):
|
class MkdocsWithConfluence(BasePlugin):
|
||||||
_id = 0
|
|
||||||
config_scheme = (
|
config_scheme = (
|
||||||
("host_url", config_options.Type(str, default=None)),
|
("host_url", config_options.Type(str, default=None)),
|
||||||
("space", config_options.Type(str, default=None)),
|
("space", config_options.Type(str, default=None)),
|
||||||
("parent_page_name", config_options.Type(str, default=None)),
|
("parent_page_name", config_options.Type(str, default=None)),
|
||||||
("username", config_options.Type(str, default=environ.get("JIRA_USERNAME", None))),
|
("username", config_options.Type(str, default=environ.get("JIRA_USERNAME", None))),
|
||||||
("password", config_options.Type(str, default=environ.get("JIRA_PASSWORD", None))),
|
("password", config_options.Type(str, default=environ.get("JIRA_PASSWORD", None))),
|
||||||
("enabled_if_env", config_options.Type(str, default=None)),
|
|
||||||
("verbose", config_options.Type(bool, default=False)),
|
|
||||||
("debug", config_options.Type(bool, default=False)),
|
|
||||||
("dryrun", config_options.Type(bool, default=False)),
|
("dryrun", config_options.Type(bool, default=False)),
|
||||||
|
("header_message", config_options.Type(str, default=HEADER_MESSAGE)),
|
||||||
)
|
)
|
||||||
|
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
self.enabled = True
|
self.enabled = False
|
||||||
self.confluence_renderer = ConfluenceRenderer(use_xhtml=True)
|
self.confluence_renderer = ConfluenceRenderer(use_xhtml=True)
|
||||||
self.confluence_mistune = mistune.Markdown(renderer=self.confluence_renderer)
|
self.confluence_mistune = mistune.Markdown(renderer=self.confluence_renderer)
|
||||||
self.simple_log = False
|
self.simple_log = False
|
||||||
self.flen = 1
|
self.flen = 1
|
||||||
self.session = requests.Session()
|
self.session = requests.Session()
|
||||||
self.page_attachments = {}
|
self.page_attachments = {}
|
||||||
|
self.repo_url = None
|
||||||
|
|
||||||
def on_nav(self, nav, config, files):
|
def on_config(self, config):
|
||||||
MkdocsWithConfluence.tab_nav = []
|
logger.info(config)
|
||||||
navigation_items = nav.__repr__()
|
# ------------------------------------------------------
|
||||||
|
# -- Enable the plugin by setting environment variable
|
||||||
for n in navigation_items.split("\n"):
|
# ------------------------------------------------------
|
||||||
leading_spaces = len(n) - len(n.lstrip(" "))
|
if os.environ.get(ENABLE_ENV_VAR):
|
||||||
spaces = leading_spaces * " "
|
logger.info("MKDocs with Confluence is enabled")
|
||||||
if "Page" in n:
|
self.enabled = True
|
||||||
try:
|
else:
|
||||||
self.page_title = self.__get_page_title(n)
|
logger.info(
|
||||||
if self.page_title is None:
|
f"MKDocs with Confluence is disabled, set the {ENABLE_ENV_VAR} to enable the plugin"
|
||||||
raise AttributeError
|
)
|
||||||
except AttributeError:
|
# ------------------------------------------------------
|
||||||
self.page_local_path = self.__get_page_url(n)
|
# -- Set the dry-run mode
|
||||||
print(
|
# ------------------------------------------------------
|
||||||
f"WARN - Page from path {self.page_local_path} has no"
|
if self.config["dryrun"] or os.environ.get(DRY_RUN_ENV_VAR):
|
||||||
f" entity in the mkdocs.yml nav section. It will be uploaded"
|
logger.info("dry-run mode is turned on, your changes won't be synced with Confluence")
|
||||||
f" to the Confluence, but you may not see it on the web server!"
|
self.dryrun = True
|
||||||
)
|
else:
|
||||||
self.page_local_name = self.__get_page_name(n)
|
logger.info("dry-run mode is turned off, your changes will be synced with Confluence")
|
||||||
self.page_title = self.page_local_name
|
self.dryrun = False
|
||||||
|
# ------------------------------------------------------
|
||||||
p = spaces + self.page_title
|
# -- Set git url to add to a confluence page
|
||||||
MkdocsWithConfluence.tab_nav.append(p)
|
# ------------------------------------------------------
|
||||||
if "Section" in n:
|
if config["repo_url"]:
|
||||||
try:
|
self.repo_url = config["repo_url"]
|
||||||
self.section_title = self.__get_section_title(n)
|
logger.info(f"git url is set to {self.repo_url}")
|
||||||
if self.section_title is None:
|
|
||||||
raise AttributeError
|
|
||||||
except AttributeError:
|
|
||||||
self.section_local_path = self.__get_page_url(n)
|
|
||||||
print(
|
|
||||||
f"WARN - Section from path {self.section_local_path} has no"
|
|
||||||
f" entity in the mkdocs.yml nav section. It will be uploaded"
|
|
||||||
f" to the Confluence, but you may not see it on the web server!"
|
|
||||||
)
|
|
||||||
self.section_local_name = self.__get_section_title(n)
|
|
||||||
self.section_title = self.section_local_name
|
|
||||||
s = spaces + self.section_title
|
|
||||||
MkdocsWithConfluence.tab_nav.append(s)
|
|
||||||
|
|
||||||
def on_files(self, files, config):
|
def on_files(self, files, config):
|
||||||
pages = files.documentation_pages()
|
pages = files.documentation_pages()
|
||||||
try:
|
try:
|
||||||
self.flen = len(pages)
|
self.flen = len(pages)
|
||||||
print(f"Number of Files in directory tree: {self.flen}")
|
logger.debug(f"number of Files in directory tree: {self.flen}")
|
||||||
except 0:
|
except 0:
|
||||||
print("ERR: You have no documentation pages" "in the directory tree, please add at least one!")
|
logger.error("no files found to be synced")
|
||||||
|
|
||||||
def on_post_template(self, output_content, template_name, config):
|
|
||||||
if self.config["verbose"] is False and self.config["debug"] is False:
|
|
||||||
self.simple_log = True
|
|
||||||
print("INFO - Mkdocs With Confluence: Start exporting markdown pages... (simple logging)")
|
|
||||||
else:
|
|
||||||
self.simple_log = False
|
|
||||||
|
|
||||||
def on_config(self, config):
|
|
||||||
if "enabled_if_env" in self.config:
|
|
||||||
env_name = self.config["enabled_if_env"]
|
|
||||||
if env_name:
|
|
||||||
self.enabled = os.environ.get(env_name) == "1"
|
|
||||||
if not self.enabled:
|
|
||||||
print(
|
|
||||||
"WARNING - Mkdocs With Confluence: Exporting MKDOCS pages to Confluence turned OFF: "
|
|
||||||
f"(set environment variable {env_name} to 1 to enable)"
|
|
||||||
)
|
|
||||||
return
|
|
||||||
else:
|
|
||||||
print(
|
|
||||||
"INFO - Mkdocs With Confluence: Exporting MKDOCS pages to Confluence "
|
|
||||||
f"turned ON by var {env_name}==1!"
|
|
||||||
)
|
|
||||||
self.enabled = True
|
|
||||||
else:
|
|
||||||
print(
|
|
||||||
"WARNING - Mkdocs With Confluence: Exporting MKDOCS pages to Confluence turned OFF: "
|
|
||||||
f"(set environment variable {env_name} to 1 to enable)"
|
|
||||||
)
|
|
||||||
return
|
|
||||||
else:
|
|
||||||
print("INFO - Mkdocs With Confluence: Exporting MKDOCS pages to Confluence turned ON by default!")
|
|
||||||
self.enabled = True
|
|
||||||
|
|
||||||
if self.config["dryrun"]:
|
|
||||||
print("WARNING - Mkdocs With Confluence - DRYRUN MODE turned ON")
|
|
||||||
self.dryrun = True
|
|
||||||
else:
|
|
||||||
self.dryrun = False
|
|
||||||
|
|
||||||
def on_page_markdown(self, markdown, page, config, files):
|
def on_page_markdown(self, markdown, page, config, files):
|
||||||
MkdocsWithConfluence._id += 1
|
# TODO: Modify pages here
|
||||||
|
logger.warning("TODO: page should be modified in this block")
|
||||||
self.session.auth = (self.config["username"], self.config["password"])
|
self.session.auth = (self.config["username"], self.config["password"])
|
||||||
|
confluencePageName = page.url[0:-1]
|
||||||
|
#.replace("/", "-")
|
||||||
|
if self.config["parent_page_name"] is not None:
|
||||||
|
parent_page = self.config["parent_page_name"]
|
||||||
|
else:
|
||||||
|
parent_page = self.config["space"]
|
||||||
|
page_name = ""
|
||||||
|
|
||||||
if self.enabled:
|
# TODO: Refactor
|
||||||
if self.simple_log is True:
|
if confluencePageName.rsplit('/',1)[0]:
|
||||||
print("INFO - Mkdocs With Confluence: Page export progress: [", end="", flush=True)
|
confluencePageName = (f"{confluencePageName.rsplit('/',1)[0]}+{page.title.replace(' ', ' ')}")
|
||||||
for i in range(MkdocsWithConfluence._id):
|
else:
|
||||||
print("#", end="", flush=True)
|
confluencePageName = (f"{page.title.replace(' ', ' ')}")
|
||||||
for j in range(self.flen - MkdocsWithConfluence._id):
|
# Create empty pages for sections only
|
||||||
print("-", end="", flush=True)
|
logger.info("preparing emtpy pages for sections")
|
||||||
print(f"] ({MkdocsWithConfluence._id} / {self.flen})", end="\r", flush=True)
|
for path in page.url.rsplit("/", 2)[0].split("/"):
|
||||||
|
logger.debug(f"path is {path}")
|
||||||
if self.config["debug"]:
|
parent_id = self.find_page_id(parent_page)
|
||||||
print(f"\nDEBUG - Handling Page '{page.title}' (And Parent Nav Pages if necessary):\n")
|
if path:
|
||||||
if not all(self.config_scheme):
|
if page_name:
|
||||||
print("DEBUG - ERR: YOU HAVE EMPTY VALUES IN YOUR CONFIG. ABORTING")
|
page_name = page_name + " " + path
|
||||||
return markdown
|
|
||||||
|
|
||||||
try:
|
|
||||||
if self.config["debug"]:
|
|
||||||
print("DEBUG - Get section first parent title...: ")
|
|
||||||
try:
|
|
||||||
|
|
||||||
parent = self.__get_section_title(page.ancestors[0].__repr__())
|
|
||||||
except IndexError as e:
|
|
||||||
if self.config["debug"]:
|
|
||||||
print(
|
|
||||||
f"DEBUG - WRN({e}): No first parent! Assuming "
|
|
||||||
f"DEBUG - {self.config['parent_page_name']}..."
|
|
||||||
)
|
|
||||||
parent = None
|
|
||||||
if self.config["debug"]:
|
|
||||||
print(f"DEBUG - {parent}")
|
|
||||||
if not parent:
|
|
||||||
parent = self.config["parent_page_name"]
|
|
||||||
|
|
||||||
if self.config["parent_page_name"] is not None:
|
|
||||||
main_parent = self.config["parent_page_name"]
|
|
||||||
else:
|
else:
|
||||||
main_parent = self.config["space"]
|
page_name = path
|
||||||
|
logger.info(f"Will create a page {page_name} under the {parent_page}")
|
||||||
if self.config["debug"]:
|
self.add_page(page_name, parent_id, "<p> I want to make sections pages better: https://git.badhouseplants.net/allanger/mkdocs-with-confluence/issues/2 </p>")
|
||||||
print("DEBUG - Get section second parent title...: ")
|
parent_page = page_name
|
||||||
try:
|
parent_id = self.find_page_id(parent_page)
|
||||||
parent1 = self.__get_section_title(page.ancestors[1].__repr__())
|
confluencePageName = parent_page + " " + page.title
|
||||||
except IndexError as e:
|
new_markdown = markdown
|
||||||
if self.config["debug"]:
|
if self.repo_url:
|
||||||
print(
|
new_markdown = f">You can edit documentation here: {self.repo_url}\n" + new_markdown
|
||||||
f"DEBUG - ERR({e}) No second parent! Assuming "
|
new_markdown = f">{self.config['header_message']}\n\n" + new_markdown
|
||||||
f"second parent is main parent: {main_parent}..."
|
# -------------------------------------------------
|
||||||
)
|
# -- Sync attachments
|
||||||
parent1 = None
|
# -------------------------------------------------
|
||||||
if self.config["debug"]:
|
attachments = []
|
||||||
print(f"{parent}")
|
# -- TODO: support named picture
|
||||||
|
md_image_reg = "(?:[!]\[(?P<caption>.*?)\])\((?P<image>.*?)\)(?P<options>\{.*\})?"
|
||||||
if not parent1:
|
try:
|
||||||
parent1 = main_parent
|
for match in re.finditer(md_image_reg, markdown):
|
||||||
if self.config["debug"]:
|
# -- TODO: I'm sure it can be done better
|
||||||
print(
|
attachment_path = "./docs" + match.group(2)
|
||||||
f"DEBUG - ONLY ONE PARENT FOUND. ASSUMING AS A "
|
logger.info(f"found image: ./docs{match.group(2)}")
|
||||||
f"FIRST NODE after main parent config {main_parent}"
|
images = re.search(md_image_reg, new_markdown)
|
||||||
)
|
# -- TODO: Options maybe the reason why page is invalid, but I'm not sure about it yet
|
||||||
|
# new_markdown = new_markdown.replace(images.group("options"), "")
|
||||||
if self.config["debug"]:
|
new_markdown = re.sub(md_image_reg, f"<p><ac:image><ri:attachment ri:filename=\"{os.path.basename(attachment_path)}\"/></ac:image></p>", new_markdown)
|
||||||
print(f"DEBUG - PARENT0: {parent}, PARENT1: {parent1}, MAIN PARENT: {main_parent}")
|
attachments.append(attachment_path)
|
||||||
|
except AttributeError as e:
|
||||||
tf = tempfile.NamedTemporaryFile(delete=False)
|
logger.warning(e)
|
||||||
f = open(tf.name, "w")
|
logger.debug(f"attachments: {attachments}")
|
||||||
|
confluence_body = self.confluence_mistune(new_markdown)
|
||||||
attachments = []
|
self.add_page(confluencePageName, parent_id, confluence_body)
|
||||||
try:
|
if attachments:
|
||||||
for match in re.finditer(r'img src="file://(.*)" s', markdown):
|
logger.debug(f"UPLOADING ATTACHMENTS TO CONFLUENCE FOR {page.title}, DETAILS:")
|
||||||
if self.config["debug"]:
|
logger.debug(f"FILES: {attachments}")
|
||||||
print(f"DEBUG - FOUND IMAGE: {match.group(1)}")
|
for attachment in attachments:
|
||||||
attachments.append(match.group(1))
|
logger.debug(f"trying to upload {attachment} to {confluencePageName}")
|
||||||
for match in re.finditer(r"!\[[\w\. -]*\]\((?!http|file)([^\s,]*).*\)", markdown):
|
if self.enabled:
|
||||||
file_path = match.group(1).lstrip("./\\")
|
try:
|
||||||
attachments.append(file_path)
|
self.add_or_update_attachment(confluencePageName, attachment)
|
||||||
|
except Exception as Argument:
|
||||||
if self.config["debug"]:
|
logger.warning(Argument)
|
||||||
print(f"DEBUG - FOUND IMAGE: {file_path}")
|
|
||||||
attachments.append("docs/" + file_path.replace("../", ""))
|
|
||||||
|
|
||||||
except AttributeError as e:
|
|
||||||
if self.config["debug"]:
|
|
||||||
print(f"DEBUG - WARN(({e}): No images found in markdown. Proceed..")
|
|
||||||
new_markdown = re.sub(
|
|
||||||
r'<img src="file:///tmp/', '<p><ac:image ac:height="350"><ri:attachment ri:filename="', markdown
|
|
||||||
)
|
|
||||||
new_markdown = re.sub(r'" style="page-break-inside: avoid;">', '"/></ac:image></p>', new_markdown)
|
|
||||||
confluence_body = self.confluence_mistune(new_markdown)
|
|
||||||
f.write(confluence_body)
|
|
||||||
if self.config["debug"]:
|
|
||||||
print(confluence_body)
|
|
||||||
page_name = page.title
|
|
||||||
new_name = "confluence_page_" + page_name.replace(" ", "_") + ".html"
|
|
||||||
shutil.copy(f.name, new_name)
|
|
||||||
f.close()
|
|
||||||
|
|
||||||
if self.config["debug"]:
|
|
||||||
print(
|
|
||||||
f"\nDEBUG - UPDATING PAGE TO CONFLUENCE, DETAILS:\n"
|
|
||||||
f"DEBUG - HOST: {self.config['host_url']}\n"
|
|
||||||
f"DEBUG - SPACE: {self.config['space']}\n"
|
|
||||||
f"DEBUG - TITLE: {page.title}\n"
|
|
||||||
f"DEBUG - PARENT: {parent}\n"
|
|
||||||
f"DEBUG - BODY: {confluence_body}\n"
|
|
||||||
)
|
|
||||||
|
|
||||||
page_id = self.find_page_id(page.title)
|
|
||||||
if page_id is not None:
|
|
||||||
if self.config["debug"]:
|
|
||||||
print(
|
|
||||||
f"DEBUG - JUST ONE STEP FROM UPDATE OF PAGE '{page.title}' \n"
|
|
||||||
f"DEBUG - CHECKING IF PARENT PAGE ON CONFLUENCE IS THE SAME AS HERE"
|
|
||||||
)
|
|
||||||
|
|
||||||
parent_name = self.find_parent_name_of_page(page.title)
|
|
||||||
|
|
||||||
if parent_name == parent:
|
|
||||||
if self.config["debug"]:
|
|
||||||
print("DEBUG - Parents match. Continue...")
|
|
||||||
else:
|
|
||||||
if self.config["debug"]:
|
|
||||||
print(f"DEBUG - ERR, Parents does not match: '{parent}' =/= '{parent_name}' Aborting...")
|
|
||||||
return markdown
|
|
||||||
self.update_page(page.title, confluence_body)
|
|
||||||
for i in MkdocsWithConfluence.tab_nav:
|
|
||||||
if page.title in i:
|
|
||||||
print(f"INFO - Mkdocs With Confluence: {i} *UPDATE*")
|
|
||||||
else:
|
|
||||||
if self.config["debug"]:
|
|
||||||
print(
|
|
||||||
f"DEBUG - PAGE: {page.title}, PARENT0: {parent}, "
|
|
||||||
f"PARENT1: {parent1}, MAIN PARENT: {main_parent}"
|
|
||||||
)
|
|
||||||
parent_id = self.find_page_id(parent)
|
|
||||||
self.wait_until(parent_id, 1, 20)
|
|
||||||
second_parent_id = self.find_page_id(parent1)
|
|
||||||
self.wait_until(second_parent_id, 1, 20)
|
|
||||||
main_parent_id = self.find_page_id(main_parent)
|
|
||||||
if not parent_id:
|
|
||||||
if not second_parent_id:
|
|
||||||
main_parent_id = self.find_page_id(main_parent)
|
|
||||||
if not main_parent_id:
|
|
||||||
print("ERR: MAIN PARENT UNKNOWN. ABORTING!")
|
|
||||||
return markdown
|
|
||||||
|
|
||||||
if self.config["debug"]:
|
|
||||||
print(
|
|
||||||
f"DEBUG - Trying to ADD page '{parent1}' to "
|
|
||||||
f"main parent({main_parent}) ID: {main_parent_id}"
|
|
||||||
)
|
|
||||||
body = TEMPLATE_BODY.replace("TEMPLATE", parent1)
|
|
||||||
self.add_page(parent1, main_parent_id, body)
|
|
||||||
for i in MkdocsWithConfluence.tab_nav:
|
|
||||||
if parent1 in i:
|
|
||||||
print(f"INFO - Mkdocs With Confluence: {i} *NEW PAGE*")
|
|
||||||
time.sleep(1)
|
|
||||||
|
|
||||||
if self.config["debug"]:
|
|
||||||
print(
|
|
||||||
f"DEBUG - Trying to ADD page '{parent}' "
|
|
||||||
f"to parent1({parent1}) ID: {second_parent_id}"
|
|
||||||
)
|
|
||||||
body = TEMPLATE_BODY.replace("TEMPLATE", parent)
|
|
||||||
self.add_page(parent, second_parent_id, body)
|
|
||||||
for i in MkdocsWithConfluence.tab_nav:
|
|
||||||
if parent in i:
|
|
||||||
print(f"INFO - Mkdocs With Confluence: {i} *NEW PAGE*")
|
|
||||||
time.sleep(1)
|
|
||||||
|
|
||||||
if parent_id is None:
|
|
||||||
for i in range(11):
|
|
||||||
while parent_id is None:
|
|
||||||
try:
|
|
||||||
self.add_page(page.title, parent_id, confluence_body)
|
|
||||||
except requests.exceptions.HTTPError:
|
|
||||||
print(
|
|
||||||
f"ERR - HTTP error on adding page. It probably occured due to "
|
|
||||||
f"parent ID('{parent_id}') page is not YET synced on server. Retry nb {i}/10..."
|
|
||||||
)
|
|
||||||
sleep(5)
|
|
||||||
parent_id = self.find_page_id(parent)
|
|
||||||
break
|
|
||||||
|
|
||||||
self.add_page(page.title, parent_id, confluence_body)
|
|
||||||
|
|
||||||
print(f"Trying to ADD page '{page.title}' to parent0({parent}) ID: {parent_id}")
|
|
||||||
for i in MkdocsWithConfluence.tab_nav:
|
|
||||||
if page.title in i:
|
|
||||||
print(f"INFO - Mkdocs With Confluence: {i} *NEW PAGE*")
|
|
||||||
|
|
||||||
if attachments:
|
|
||||||
self.page_attachments[page.title] = attachments
|
|
||||||
|
|
||||||
except IndexError as e:
|
|
||||||
if self.config["debug"]:
|
|
||||||
print(f"DEBUG - ERR({e}): Exception error!")
|
|
||||||
return markdown
|
|
||||||
|
|
||||||
return markdown
|
return markdown
|
||||||
|
|
||||||
def on_post_page(self, output, page, config):
|
def on_post_page(self, output, page, config):
|
||||||
site_dir = config.get("site_dir")
|
logger.info("The author was uploading images here, maybe there was a reason for that")
|
||||||
attachments = self.page_attachments.get(page.title, [])
|
|
||||||
|
|
||||||
if self.config["debug"]:
|
|
||||||
print(f"\nDEBUG - UPLOADING ATTACHMENTS TO CONFLUENCE FOR {page.title}, DETAILS:")
|
|
||||||
print(f"FILES: {attachments} \n")
|
|
||||||
for attachment in attachments:
|
|
||||||
if self.config["debug"]:
|
|
||||||
print(f"DEBUG - looking for {attachment} in {site_dir}")
|
|
||||||
for p in Path(site_dir).rglob(f"*{attachment}"):
|
|
||||||
self.add_or_update_attachment(page.title, p)
|
|
||||||
return output
|
|
||||||
|
|
||||||
def on_page_content(self, html, page, config, files):
|
def on_page_content(self, html, page, config, files):
|
||||||
return html
|
return html
|
||||||
@ -372,19 +179,17 @@ class MkdocsWithConfluence(BasePlugin):
|
|||||||
return os.path.basename(re.search("url='(.*)'\\)", section).group(1)[:-1])
|
return os.path.basename(re.search("url='(.*)'\\)", section).group(1)[:-1])
|
||||||
|
|
||||||
def __get_section_name(self, section):
|
def __get_section_name(self, section):
|
||||||
if self.config["debug"]:
|
logger.debug(f"SECTION name: {section}")
|
||||||
print(f"DEBUG - SECTION name: {section}")
|
|
||||||
return os.path.basename(re.search("url='(.*)'\\/", section).group(1)[:-1])
|
return os.path.basename(re.search("url='(.*)'\\/", section).group(1)[:-1])
|
||||||
|
|
||||||
def __get_section_title(self, section):
|
def __get_section_title(self, section):
|
||||||
if self.config["debug"]:
|
logger.debug(f"SECTION title: {section}")
|
||||||
print(f"DEBUG - SECTION title: {section}")
|
|
||||||
try:
|
try:
|
||||||
r = re.search("Section\\(title='(.*)'\\)", section)
|
r = re.search("Section\\(title='(.*)'\\)", section)
|
||||||
return r.group(1)
|
return r.group(1)
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
name = self.__get_section_name(section)
|
name = self.__get_section_name(section)
|
||||||
print(f"WRN - Section '{name}' doesn't exist in the mkdocs.yml nav section!")
|
logger.warning(f"Section '{name}' doesn't exist in the mkdocs.yml nav section!")
|
||||||
return name
|
return name
|
||||||
|
|
||||||
def __get_page_title(self, section):
|
def __get_page_title(self, section):
|
||||||
@ -393,7 +198,7 @@ class MkdocsWithConfluence(BasePlugin):
|
|||||||
return r.group(1)
|
return r.group(1)
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
name = self.__get_page_url(section)
|
name = self.__get_page_url(section)
|
||||||
print(f"WRN - Page '{name}' doesn't exist in the mkdocs.yml nav section!")
|
logger.warning(f"Page '{name}' doesn't exist in the mkdocs.yml nav section!")
|
||||||
return name
|
return name
|
||||||
|
|
||||||
# Adapted from https://stackoverflow.com/a/3431838
|
# Adapted from https://stackoverflow.com/a/3431838
|
||||||
@ -405,9 +210,8 @@ class MkdocsWithConfluence(BasePlugin):
|
|||||||
return hash_sha1.hexdigest()
|
return hash_sha1.hexdigest()
|
||||||
|
|
||||||
def add_or_update_attachment(self, page_name, filepath):
|
def add_or_update_attachment(self, page_name, filepath):
|
||||||
print(f"INFO - Mkdocs With Confluence * {page_name} *ADD/Update ATTACHMENT if required* {filepath}")
|
logger.warning(f"Mkdocs With Confluence * {page_name} *ADD/Update ATTACHMENT if required* {filepath}")
|
||||||
if self.config["debug"]:
|
logger.debug(f"Mkdocs With Confluence: Add Attachment: PAGE NAME: {page_name}, FILE: {filepath}")
|
||||||
print(f" * Mkdocs With Confluence: Add Attachment: PAGE NAME: {page_name}, FILE: {filepath}")
|
|
||||||
page_id = self.find_page_id(page_name)
|
page_id = self.find_page_id(page_name)
|
||||||
if page_id:
|
if page_id:
|
||||||
file_hash = self.get_file_sha1(filepath)
|
file_hash = self.get_file_sha1(filepath)
|
||||||
@ -417,25 +221,21 @@ class MkdocsWithConfluence(BasePlugin):
|
|||||||
file_hash_regex = re.compile(r"\[v([a-f0-9]{40})]$")
|
file_hash_regex = re.compile(r"\[v([a-f0-9]{40})]$")
|
||||||
existing_match = file_hash_regex.search(existing_attachment["version"]["message"])
|
existing_match = file_hash_regex.search(existing_attachment["version"]["message"])
|
||||||
if existing_match is not None and existing_match.group(1) == file_hash:
|
if existing_match is not None and existing_match.group(1) == file_hash:
|
||||||
if self.config["debug"]:
|
logger.debug(f" * Mkdocs With Confluence * {page_name} * Existing attachment skipping * {filepath}")
|
||||||
print(f" * Mkdocs With Confluence * {page_name} * Existing attachment skipping * {filepath}")
|
|
||||||
else:
|
else:
|
||||||
self.update_attachment(page_id, filepath, existing_attachment, attachment_message)
|
self.update_attachment(page_id, filepath, existing_attachment, attachment_message)
|
||||||
else:
|
else:
|
||||||
self.create_attachment(page_id, filepath, attachment_message)
|
self.create_attachment(page_id, filepath, attachment_message)
|
||||||
else:
|
else:
|
||||||
if self.config["debug"]:
|
logger.debug("PAGE DOES NOT EXISTS")
|
||||||
print("PAGE DOES NOT EXISTS")
|
|
||||||
|
|
||||||
def get_attachment(self, page_id, filepath):
|
def get_attachment(self, page_id, filepath):
|
||||||
name = os.path.basename(filepath)
|
name = os.path.basename(filepath)
|
||||||
if self.config["debug"]:
|
logger.debug(f" * Mkdocs With Confluence: Get Attachment: PAGE ID: {page_id}, FILE: {filepath}")
|
||||||
print(f" * Mkdocs With Confluence: Get Attachment: PAGE ID: {page_id}, FILE: {filepath}")
|
|
||||||
|
|
||||||
url = self.config["host_url"] + "/" + page_id + "/child/attachment"
|
url = self.config["host_url"] + "/" + page_id + "/child/attachment"
|
||||||
headers = {"X-Atlassian-Token": "no-check"} # no content-type here!
|
headers = {"X-Atlassian-Token": "no-check"} # no content-type here!
|
||||||
if self.config["debug"]:
|
logger.debug(f"URL: {url}")
|
||||||
print(f"URL: {url}")
|
|
||||||
|
|
||||||
r = self.session.get(url, headers=headers, params={"filename": name, "expand": "version"})
|
r = self.session.get(url, headers=headers, params={"filename": name, "expand": "version"})
|
||||||
r.raise_for_status()
|
r.raise_for_status()
|
||||||
@ -445,15 +245,11 @@ class MkdocsWithConfluence(BasePlugin):
|
|||||||
return response_json["results"][0]
|
return response_json["results"][0]
|
||||||
|
|
||||||
def update_attachment(self, page_id, filepath, existing_attachment, message):
|
def update_attachment(self, page_id, filepath, existing_attachment, message):
|
||||||
if self.config["debug"]:
|
logger.debug(f" * Mkdocs With Confluence: Update Attachment: PAGE ID: {page_id}, FILE: {filepath}")
|
||||||
print(f" * Mkdocs With Confluence: Update Attachment: PAGE ID: {page_id}, FILE: {filepath}")
|
|
||||||
|
|
||||||
url = self.config["host_url"] + "/" + page_id + "/child/attachment/" + existing_attachment["id"] + "/data"
|
url = self.config["host_url"] + "/" + page_id + "/child/attachment/" + existing_attachment["id"] + "/data"
|
||||||
headers = {"X-Atlassian-Token": "no-check"} # no content-type here!
|
headers = {"X-Atlassian-Token": "no-check"} # no content-type here!
|
||||||
|
logger.debug(f"URL: {url}")
|
||||||
if self.config["debug"]:
|
|
||||||
print(f"URL: {url}")
|
|
||||||
|
|
||||||
filename = os.path.basename(filepath)
|
filename = os.path.basename(filepath)
|
||||||
|
|
||||||
# determine content-type
|
# determine content-type
|
||||||
@ -465,21 +261,18 @@ class MkdocsWithConfluence(BasePlugin):
|
|||||||
if not self.dryrun:
|
if not self.dryrun:
|
||||||
r = self.session.post(url, headers=headers, files=files)
|
r = self.session.post(url, headers=headers, files=files)
|
||||||
r.raise_for_status()
|
r.raise_for_status()
|
||||||
print(r.json())
|
logger.debug(r.json())
|
||||||
if r.status_code == 200:
|
if r.status_code == 200:
|
||||||
print("OK!")
|
logger.info("OK!")
|
||||||
else:
|
else:
|
||||||
print("ERR!")
|
print("ERR!")
|
||||||
|
|
||||||
def create_attachment(self, page_id, filepath, message):
|
def create_attachment(self, page_id, filepath, message):
|
||||||
if self.config["debug"]:
|
logger.debug(f" * Mkdocs With Confluence: Create Attachment: PAGE ID: {page_id}, FILE: {filepath}")
|
||||||
print(f" * Mkdocs With Confluence: Create Attachment: PAGE ID: {page_id}, FILE: {filepath}")
|
|
||||||
|
|
||||||
url = self.config["host_url"] + "/" + page_id + "/child/attachment"
|
url = self.config["host_url"] + "/" + page_id + "/child/attachment"
|
||||||
headers = {"X-Atlassian-Token": "no-check"} # no content-type here!
|
headers = {"X-Atlassian-Token": "no-check"} # no content-type here!
|
||||||
|
logger.debug(f"URL: {url}")
|
||||||
if self.config["debug"]:
|
|
||||||
print(f"URL: {url}")
|
|
||||||
|
|
||||||
filename = os.path.basename(filepath)
|
filename = os.path.basename(filepath)
|
||||||
|
|
||||||
@ -490,73 +283,67 @@ class MkdocsWithConfluence(BasePlugin):
|
|||||||
files = {"file": (filename, open(filepath, "rb"), content_type), "comment": message}
|
files = {"file": (filename, open(filepath, "rb"), content_type), "comment": message}
|
||||||
if not self.dryrun:
|
if not self.dryrun:
|
||||||
r = self.session.post(url, headers=headers, files=files)
|
r = self.session.post(url, headers=headers, files=files)
|
||||||
print(r.json())
|
logger.debug(r.json())
|
||||||
r.raise_for_status()
|
r.raise_for_status()
|
||||||
if r.status_code == 200:
|
if r.status_code == 200:
|
||||||
print("OK!")
|
logger.debug("OK!")
|
||||||
else:
|
else:
|
||||||
print("ERR!")
|
logger.debug("ERR!")
|
||||||
|
|
||||||
def find_page_id(self, page_name):
|
def find_page_id(self, page_name):
|
||||||
if self.config["debug"]:
|
logger.info(f"looking for a page id of the page: {page_name}")
|
||||||
print(f"INFO - * Mkdocs With Confluence: Find Page ID: PAGE NAME: {page_name}")
|
|
||||||
name_confl = page_name.replace(" ", "+")
|
name_confl = page_name.replace(" ", "+")
|
||||||
url = self.config["host_url"] + "?title=" + name_confl + "&spaceKey=" + self.config["space"] + "&expand=history"
|
url = self.config["host_url"] + "?title=" + name_confl + "&spaceKey=" + self.config["space"] + "&expand=history"
|
||||||
if self.config["debug"]:
|
logger.debug(f"URL: {url}")
|
||||||
print(f"URL: {url}")
|
|
||||||
r = self.session.get(url)
|
r = self.session.get(url)
|
||||||
r.raise_for_status()
|
r.raise_for_status()
|
||||||
with nostdout():
|
with nostdout():
|
||||||
response_json = r.json()
|
response_json = r.json()
|
||||||
if response_json["results"]:
|
if response_json["results"]:
|
||||||
if self.config["debug"]:
|
logger.debug(f"response: {response_json}")
|
||||||
print(f"ID: {response_json['results'][0]['id']}")
|
|
||||||
return response_json["results"][0]["id"]
|
return response_json["results"][0]["id"]
|
||||||
else:
|
else:
|
||||||
if self.config["debug"]:
|
logger.debug(f"page {page_name} doens't exist")
|
||||||
print("PAGE DOES NOT EXIST")
|
|
||||||
return None
|
return None
|
||||||
|
|
||||||
def add_page(self, page_name, parent_page_id, page_content_in_storage_format):
|
def add_page(self, page_name, parent_page_id, page_content_in_storage_format):
|
||||||
print(f"INFO - * Mkdocs With Confluence: {page_name} - *NEW PAGE*")
|
logger.info(f"Creating a new page: {page_name} under page with ID: {parent_page_id}")
|
||||||
|
if self.enabled:
|
||||||
if self.config["debug"]:
|
if self.find_page_id(page_name):
|
||||||
print(f" * Mkdocs With Confluence: Adding Page: PAGE NAME: {page_name}, parent ID: {parent_page_id}")
|
self.update_page(page_name, page_content_in_storage_format)
|
||||||
url = self.config["host_url"] + "/"
|
|
||||||
if self.config["debug"]:
|
|
||||||
print(f"URL: {url}")
|
|
||||||
headers = {"Content-Type": "application/json"}
|
|
||||||
space = self.config["space"]
|
|
||||||
data = {
|
|
||||||
"type": "page",
|
|
||||||
"title": page_name,
|
|
||||||
"space": {"key": space},
|
|
||||||
"ancestors": [{"id": parent_page_id}],
|
|
||||||
"body": {"storage": {"value": page_content_in_storage_format, "representation": "storage"}},
|
|
||||||
}
|
|
||||||
if self.config["debug"]:
|
|
||||||
print(f"DATA: {data}")
|
|
||||||
if not self.dryrun:
|
|
||||||
r = self.session.post(url, json=data, headers=headers)
|
|
||||||
r.raise_for_status()
|
|
||||||
if r.status_code == 200:
|
|
||||||
if self.config["debug"]:
|
|
||||||
print("OK!")
|
|
||||||
else:
|
else:
|
||||||
if self.config["debug"]:
|
logger.info(f"Creating a new page: {page_name} under page with ID: {parent_page_id}")
|
||||||
print("ERR!")
|
url = self.config["host_url"] + "/"
|
||||||
|
logger.debug(f"URL: {url}")
|
||||||
|
headers = {"Content-Type": "application/json"}
|
||||||
|
space = self.config["space"]
|
||||||
|
data = {
|
||||||
|
"type": "page",
|
||||||
|
"title": page_name,
|
||||||
|
"space": {"key": space},
|
||||||
|
"ancestors": [{"id": parent_page_id}],
|
||||||
|
"body": {"storage": {"value": page_content_in_storage_format, "representation": "storage"}},
|
||||||
|
}
|
||||||
|
logger.debug(f"DATA: {data}")
|
||||||
|
if not self.dryrun:
|
||||||
|
try:
|
||||||
|
r = self.session.post(url, json=data, headers=headers)
|
||||||
|
r.raise_for_status()
|
||||||
|
except Exception as exp:
|
||||||
|
logger.error(exp)
|
||||||
|
if r.status_code == 200:
|
||||||
|
logger.info(f"page created: {page_name}")
|
||||||
|
else:
|
||||||
|
logger.error(f"page can't be created: {page_name}")
|
||||||
|
|
||||||
def update_page(self, page_name, page_content_in_storage_format):
|
def update_page(self, page_name, page_content_in_storage_format):
|
||||||
page_id = self.find_page_id(page_name)
|
page_id = self.find_page_id(page_name)
|
||||||
print(f"INFO - * Mkdocs With Confluence: {page_name} - *UPDATE*")
|
logger.debug(f"updating page {page_name}")
|
||||||
if self.config["debug"]:
|
|
||||||
print(f" * Mkdocs With Confluence: Update PAGE ID: {page_id}, PAGE NAME: {page_name}")
|
|
||||||
if page_id:
|
if page_id:
|
||||||
page_version = self.find_page_version(page_name)
|
page_version = self.find_page_version(page_name)
|
||||||
page_version = page_version + 1
|
page_version = page_version + 1
|
||||||
url = self.config["host_url"] + "/" + page_id
|
url = self.config["host_url"] + "/" + page_id
|
||||||
if self.config["debug"]:
|
|
||||||
print(f"URL: {url}")
|
|
||||||
headers = {"Content-Type": "application/json"}
|
headers = {"Content-Type": "application/json"}
|
||||||
space = self.config["space"]
|
space = self.config["space"]
|
||||||
data = {
|
data = {
|
||||||
@ -569,21 +356,20 @@ class MkdocsWithConfluence(BasePlugin):
|
|||||||
}
|
}
|
||||||
|
|
||||||
if not self.dryrun:
|
if not self.dryrun:
|
||||||
r = self.session.put(url, json=data, headers=headers)
|
try:
|
||||||
r.raise_for_status()
|
r = self.session.put(url, json=data, headers=headers)
|
||||||
|
r.raise_for_status()
|
||||||
|
except Exception as exp:
|
||||||
|
logger.error(exp)
|
||||||
if r.status_code == 200:
|
if r.status_code == 200:
|
||||||
if self.config["debug"]:
|
logger.info(f"page created: {page_name}")
|
||||||
print("OK!")
|
|
||||||
else:
|
else:
|
||||||
if self.config["debug"]:
|
logger.error(f"page can't be created: {page_name}")
|
||||||
print("ERR!")
|
|
||||||
else:
|
else:
|
||||||
if self.config["debug"]:
|
logger.warning("page {page_name} doesn't exist")
|
||||||
print("PAGE DOES NOT EXIST YET!")
|
|
||||||
|
|
||||||
def find_page_version(self, page_name):
|
def find_page_version(self, page_name):
|
||||||
if self.config["debug"]:
|
logger.debug(f"INFO - * Mkdocs With Confluence: Find PAGE VERSION, PAGE NAME: {page_name}")
|
||||||
print(f"INFO - * Mkdocs With Confluence: Find PAGE VERSION, PAGE NAME: {page_name}")
|
|
||||||
name_confl = page_name.replace(" ", "+")
|
name_confl = page_name.replace(" ", "+")
|
||||||
url = self.config["host_url"] + "?title=" + name_confl + "&spaceKey=" + self.config["space"] + "&expand=version"
|
url = self.config["host_url"] + "?title=" + name_confl + "&spaceKey=" + self.config["space"] + "&expand=version"
|
||||||
r = self.session.get(url)
|
r = self.session.get(url)
|
||||||
@ -591,17 +377,14 @@ class MkdocsWithConfluence(BasePlugin):
|
|||||||
with nostdout():
|
with nostdout():
|
||||||
response_json = r.json()
|
response_json = r.json()
|
||||||
if response_json["results"] is not None:
|
if response_json["results"] is not None:
|
||||||
if self.config["debug"]:
|
logger.debug(f"VERSION: {response_json['results'][0]['version']['number']}")
|
||||||
print(f"VERSION: {response_json['results'][0]['version']['number']}")
|
|
||||||
return response_json["results"][0]["version"]["number"]
|
return response_json["results"][0]["version"]["number"]
|
||||||
else:
|
else:
|
||||||
if self.config["debug"]:
|
logger.debug("PAGE DOES NOT EXISTS")
|
||||||
print("PAGE DOES NOT EXISTS")
|
|
||||||
return None
|
return None
|
||||||
|
|
||||||
def find_parent_name_of_page(self, name):
|
def find_parent_name_of_page(self, name):
|
||||||
if self.config["debug"]:
|
logger.debug(f"INFO - * Mkdocs With Confluence: Find PARENT OF PAGE, PAGE NAME: {name}")
|
||||||
print(f"INFO - * Mkdocs With Confluence: Find PARENT OF PAGE, PAGE NAME: {name}")
|
|
||||||
idp = self.find_page_id(name)
|
idp = self.find_page_id(name)
|
||||||
url = self.config["host_url"] + "/" + idp + "?expand=ancestors"
|
url = self.config["host_url"] + "/" + idp + "?expand=ancestors"
|
||||||
|
|
||||||
@ -610,15 +393,14 @@ class MkdocsWithConfluence(BasePlugin):
|
|||||||
with nostdout():
|
with nostdout():
|
||||||
response_json = r.json()
|
response_json = r.json()
|
||||||
if response_json:
|
if response_json:
|
||||||
if self.config["debug"]:
|
logger.debug(f"PARENT NAME: {response_json['ancestors'][-1]['title']}")
|
||||||
print(f"PARENT NAME: {response_json['ancestors'][-1]['title']}")
|
|
||||||
return response_json["ancestors"][-1]["title"]
|
return response_json["ancestors"][-1]["title"]
|
||||||
else:
|
else:
|
||||||
if self.config["debug"]:
|
logger.debug("PAGE DOES NOT HAVE PARENT")
|
||||||
print("PAGE DOES NOT HAVE PARENT")
|
|
||||||
return None
|
return None
|
||||||
|
|
||||||
def wait_until(self, condition, interval=0.1, timeout=1):
|
def wait_until(self, condition, interval=0.1, timeout=1):
|
||||||
start = time.time()
|
start = time.time()
|
||||||
while not condition and time.time() - start < timeout:
|
while not condition and time.time() - start < timeout:
|
||||||
time.sleep(interval)
|
time.sleep(interval)
|
||||||
|
|
||||||
|
1417
poetry.lock
generated
Normal file
1417
poetry.lock
generated
Normal file
File diff suppressed because it is too large
Load Diff
22
pyproject.toml
Normal file
22
pyproject.toml
Normal file
@ -0,0 +1,22 @@
|
|||||||
|
[tool.poetry]
|
||||||
|
name = "mkdocs-with-confluence"
|
||||||
|
version = "0.3.0"
|
||||||
|
description = "MkDocs plugin for uploading markdown documentation to Confluence via Confluence REST API"
|
||||||
|
authors = ["Nikolai Rodionov <allanger@zohomail.com>"]
|
||||||
|
readme = "README.md"
|
||||||
|
packages = [{include = "mkdocs_with_confluence/plugin.py"}]
|
||||||
|
|
||||||
|
[tool.poetry.dependencies]
|
||||||
|
python = "^3.9"
|
||||||
|
poetry = "^1.3.2"
|
||||||
|
poetry2setup = "^1.1.0"
|
||||||
|
md2cf = "^2.0.1"
|
||||||
|
loguru = "^0.6.0"
|
||||||
|
|
||||||
|
|
||||||
|
[build-system]
|
||||||
|
requires = ["poetry-core"]
|
||||||
|
build-backend = "poetry.core.masonry.api"
|
||||||
|
|
||||||
|
[tool.poetry.plugins."mkdocs.plugins"]
|
||||||
|
mkdocs-with-confluence = 'mkdocs_with_confluence.plugin:MkdocsWithConfluence'
|
@ -1,4 +0,0 @@
|
|||||||
pre-commit
|
|
||||||
mime
|
|
||||||
mistune
|
|
||||||
md2cf
|
|
@ -1,4 +0,0 @@
|
|||||||
black
|
|
||||||
flake8
|
|
||||||
nose
|
|
||||||
pytest-cov
|
|
16
setup.py
16
setup.py
@ -1,16 +0,0 @@
|
|||||||
from setuptools import setup, find_packages
|
|
||||||
|
|
||||||
setup(
|
|
||||||
name="mkdocs-with-confluence",
|
|
||||||
version="0.2.7",
|
|
||||||
description="MkDocs plugin for uploading markdown documentation to Confluence via Confluence REST API",
|
|
||||||
keywords="mkdocs markdown confluence documentation rest python",
|
|
||||||
url="https://github.com/pawelsikora/mkdocs-with-confluence/",
|
|
||||||
author="Pawel Sikora",
|
|
||||||
author_email="sikor6@gmail.com",
|
|
||||||
license="MIT",
|
|
||||||
python_requires=">=3.6",
|
|
||||||
install_requires=["mkdocs>=1.1", "jinja2", "mistune", "md2cf", "requests"],
|
|
||||||
packages=find_packages(),
|
|
||||||
entry_points={"mkdocs.plugins": ["mkdocs-with-confluence = mkdocs_with_confluence.plugin:MkdocsWithConfluence"]},
|
|
||||||
)
|
|
137
tests/test.py
137
tests/test.py
@ -1,137 +0,0 @@
|
|||||||
import requests
|
|
||||||
import mimetypes
|
|
||||||
|
|
||||||
# -----------------------------------------------------------------------------
|
|
||||||
# Globals
|
|
||||||
|
|
||||||
BASE_URL = "<YOUR_DOMAIN>/rest/api/content"
|
|
||||||
SPACE_NAME = "<YOUR_SPACE_NAME>"
|
|
||||||
USERNAME = "<YOUR_USERNAME>"
|
|
||||||
PASSWORD = "<YOUR_PASSWORD>"
|
|
||||||
|
|
||||||
|
|
||||||
def upload_attachment(page_id, filepath):
|
|
||||||
url = BASE_URL + "/" + page_id + "/child/attachment/"
|
|
||||||
headers = {"X-Atlassian-Token": "no-check"} # no content-type here!
|
|
||||||
print(f"URL: {url}")
|
|
||||||
filename = filepath
|
|
||||||
|
|
||||||
# determine content-type
|
|
||||||
content_type, encoding = mimetypes.guess_type(filename)
|
|
||||||
if content_type is None:
|
|
||||||
content_type = "multipart/form-data"
|
|
||||||
|
|
||||||
# provide content-type explicitly
|
|
||||||
files = {"file": (filename, open(filename, "rb"), content_type)}
|
|
||||||
print(f"FILES: {files}")
|
|
||||||
|
|
||||||
auth = (USERNAME, PASSWORD)
|
|
||||||
r = requests.post(url, headers=headers, files=files, auth=auth)
|
|
||||||
r.raise_for_status()
|
|
||||||
|
|
||||||
|
|
||||||
def find_parent_name_of_page(name):
|
|
||||||
idp = find_page_id(name)
|
|
||||||
url = BASE_URL + "/" + idp + "?expand=ancestors"
|
|
||||||
print(f"URL: {url}")
|
|
||||||
|
|
||||||
auth = (USERNAME, PASSWORD)
|
|
||||||
r = requests.get(url, auth=auth)
|
|
||||||
r.raise_for_status()
|
|
||||||
response_json = r.json()
|
|
||||||
if response_json:
|
|
||||||
print(f"ID: {response_json['ancestors'][0]['title']}")
|
|
||||||
return response_json
|
|
||||||
else:
|
|
||||||
print("PAGE DOES NOT EXIST")
|
|
||||||
return None
|
|
||||||
|
|
||||||
|
|
||||||
def find_page_id(name):
|
|
||||||
name_confl = name.replace(" ", "+")
|
|
||||||
url = BASE_URL + "?title=" + name_confl + "&spaceKey=" + SPACE_NAME + "&expand=history"
|
|
||||||
print(f"URL: {url}")
|
|
||||||
|
|
||||||
auth = (USERNAME, PASSWORD)
|
|
||||||
r = requests.get(url, auth=auth)
|
|
||||||
r.raise_for_status()
|
|
||||||
response_json = r.json()
|
|
||||||
if response_json["results"]:
|
|
||||||
print(f"ID: {response_json['results']}")
|
|
||||||
return response_json["results"]
|
|
||||||
else:
|
|
||||||
print("PAGE DOES NOT EXIST")
|
|
||||||
return None
|
|
||||||
|
|
||||||
|
|
||||||
def add_page(page_name, parent_page_id):
|
|
||||||
url = BASE_URL + "/"
|
|
||||||
print(f"URL: {url}")
|
|
||||||
headers = {"Content-Type": "application/json"}
|
|
||||||
auth = (USERNAME, PASSWORD)
|
|
||||||
data = {
|
|
||||||
"type": "page",
|
|
||||||
"title": page_name,
|
|
||||||
"space": {"key": SPACE_NAME},
|
|
||||||
"ancestors": [{"id": parent_page_id}],
|
|
||||||
"body": {"storage": {"value": "<p>This is a new page</p>", "representation": "storage"}},
|
|
||||||
}
|
|
||||||
|
|
||||||
r = requests.post(url, json=data, headers=headers, auth=auth)
|
|
||||||
r.raise_for_status()
|
|
||||||
print(r.json())
|
|
||||||
|
|
||||||
|
|
||||||
def update_page(page_name):
|
|
||||||
page_id = find_page_id(page_name)
|
|
||||||
if page_id:
|
|
||||||
page_version = find_page_version(page_name)
|
|
||||||
page_version = page_version + 1
|
|
||||||
print(f"PAGE ID: {page_id}, PAGE NAME: {page_name}")
|
|
||||||
url = BASE_URL + "/" + page_id
|
|
||||||
print(f"URL: {url}")
|
|
||||||
headers = {"Content-Type": "application/json"}
|
|
||||||
auth = (USERNAME, PASSWORD)
|
|
||||||
data = {
|
|
||||||
"type": "page",
|
|
||||||
"space": {"key": SPACE_NAME},
|
|
||||||
"body": {"storage": {"value": "<p>Let the dragons out!</p>", "representation": "storage"}},
|
|
||||||
"version": {"number": page_version},
|
|
||||||
}
|
|
||||||
|
|
||||||
data["id"] = page_id
|
|
||||||
data["title"] = page_name
|
|
||||||
print(data)
|
|
||||||
|
|
||||||
r = requests.put(url, json=data, headers=headers, auth=auth)
|
|
||||||
r.raise_for_status()
|
|
||||||
print(r.json())
|
|
||||||
else:
|
|
||||||
print("PAGE DOES NOT EXIST. CREATING WITH DEFAULT BODY")
|
|
||||||
add_page(page_name)
|
|
||||||
|
|
||||||
|
|
||||||
def find_page_version(name):
|
|
||||||
name_confl = name.replace(" ", "+")
|
|
||||||
url = BASE_URL + "?title=" + name_confl + "&spaceKey=" + SPACE_NAME + "&expand=version"
|
|
||||||
|
|
||||||
print(f"URL: {url}")
|
|
||||||
|
|
||||||
auth = (USERNAME, PASSWORD)
|
|
||||||
r = requests.get(url, auth=auth)
|
|
||||||
r.raise_for_status()
|
|
||||||
response_json = r.json()
|
|
||||||
if response_json["results"]:
|
|
||||||
print(f"VERSION: {response_json['results'][0]['version']['number']}")
|
|
||||||
return response_json["results"][0]["version"]["number"]
|
|
||||||
else:
|
|
||||||
print("PAGE DOES NOT EXISTS")
|
|
||||||
return None
|
|
||||||
|
|
||||||
|
|
||||||
# add_page()
|
|
||||||
# update_page("Test Page")
|
|
||||||
# find_page_version("Test Page")
|
|
||||||
# find_parent_name_of_page("Test Parent Page")
|
|
||||||
# find_page_id("Test Page")
|
|
||||||
# upload_attachment()
|
|
Loading…
Reference in New Issue
Block a user