Merge branch 'main' of https://github.com/PaintYourDragon/circuitpython
This commit is contained in:
commit
fa55669da5
55
.github/actions/deps/external/action.yml
vendored
55
.github/actions/deps/external/action.yml
vendored
@ -9,54 +9,31 @@ inputs:
|
||||
- cache
|
||||
- restore
|
||||
|
||||
platform:
|
||||
port:
|
||||
required: false
|
||||
default: none
|
||||
type: choice
|
||||
options:
|
||||
- arm
|
||||
- aarch
|
||||
- esp
|
||||
- riscv
|
||||
- none
|
||||
type: string
|
||||
|
||||
runs:
|
||||
using: composite
|
||||
steps:
|
||||
# aarch
|
||||
- name: Get aarch toolchain
|
||||
if: inputs.platform == 'aarch'
|
||||
run: |
|
||||
wget --no-verbose https://adafruit-circuit-python.s3.amazonaws.com/gcc-arm-10.3-2021.07-x86_64-aarch64-none-elf.tar.xz
|
||||
sudo tar -C /usr --strip-components=1 -xaf gcc-arm-10.3-2021.07-x86_64-aarch64-none-elf.tar.xz
|
||||
sudo apt-get install -y mtools
|
||||
shell: bash
|
||||
- name: Install mkfs.fat
|
||||
if: inputs.platform == 'aarch'
|
||||
run: |
|
||||
wget https://github.com/dosfstools/dosfstools/releases/download/v4.2/dosfstools-4.2.tar.gz
|
||||
tar -xaf dosfstools-4.2.tar.gz
|
||||
cd dosfstools-4.2
|
||||
./configure
|
||||
make -j 2
|
||||
cd src
|
||||
echo >> $GITHUB_PATH $(pwd)
|
||||
shell: bash
|
||||
|
||||
# arm
|
||||
- name: Get arm toolchain
|
||||
if: inputs.platform == 'aarch' || inputs.platform == 'arm'
|
||||
if: >-
|
||||
inputs.port != 'none' &&
|
||||
inputs.port != 'litex' &&
|
||||
inputs.port != 'espressif'
|
||||
uses: carlosperate/arm-none-eabi-gcc-action@v1
|
||||
with:
|
||||
release: '10-2020-q4'
|
||||
|
||||
# esp
|
||||
- name: Get esp toolchain
|
||||
if: inputs.platform == 'esp'
|
||||
# espressif
|
||||
- name: Get espressif toolchain
|
||||
if: inputs.port == 'espressif'
|
||||
run: sudo apt-get install -y ninja-build
|
||||
shell: bash
|
||||
- name: Install IDF tools
|
||||
if: inputs.platform == 'esp'
|
||||
if: inputs.port == 'espressif'
|
||||
run: |
|
||||
echo "Installing ESP-IDF tools"
|
||||
$IDF_PATH/tools/idf_tools.py --non-interactive install required
|
||||
@ -66,24 +43,16 @@ runs:
|
||||
rm -rf $IDF_TOOLS_PATH/dist
|
||||
shell: bash
|
||||
- name: Set environment
|
||||
if: inputs.platform == 'esp'
|
||||
if: inputs.port == 'espressif'
|
||||
run: |
|
||||
source $IDF_PATH/export.sh
|
||||
echo >> $GITHUB_ENV "IDF_PYTHON_ENV_PATH=$IDF_PYTHON_ENV_PATH"
|
||||
echo >> $GITHUB_PATH "$PATH"
|
||||
shell: bash
|
||||
|
||||
# riscv
|
||||
- name: Get riscv toolchain
|
||||
if: inputs.platform == 'riscv'
|
||||
run: |
|
||||
wget https://static.dev.sifive.com/dev-tools/riscv64-unknown-elf-gcc-8.3.0-2019.08.0-x86_64-linux-centos6.tar.gz
|
||||
sudo tar -C /usr --strip-components=1 -xaf riscv64-unknown-elf-gcc-8.3.0-2019.08.0-x86_64-linux-centos6.tar.gz
|
||||
shell: bash
|
||||
|
||||
# common
|
||||
- name: Cache python dependencies
|
||||
if: inputs.platform != 'esp'
|
||||
if: inputs.port != 'espressif'
|
||||
uses: ./.github/actions/deps/python
|
||||
with:
|
||||
action: ${{ inputs.action }}
|
||||
|
12
.github/actions/deps/ports/action.yml
vendored
12
.github/actions/deps/ports/action.yml
vendored
@ -19,10 +19,18 @@ runs:
|
||||
if [ -z $PORT ]; then (exit 1); else echo >> $GITHUB_OUTPUT "port=$PORT"; fi
|
||||
shell: bash
|
||||
|
||||
- name: Set up espressif port
|
||||
- name: Set up broadcom
|
||||
if: steps.board-to-port.outputs.port == 'broadcom'
|
||||
uses: ./.github/actions/deps/ports/broadcom
|
||||
|
||||
- name: Set up espressif
|
||||
if: steps.board-to-port.outputs.port == 'espressif'
|
||||
uses: ./.github/actions/deps/ports/espressif
|
||||
|
||||
- name: Set up nrf port
|
||||
- name: Set up litex
|
||||
if: steps.board-to-port.outputs.port == 'litex'
|
||||
uses: ./.github/actions/deps/ports/litex
|
||||
|
||||
- name: Set up nrf
|
||||
if: steps.board-to-port.outputs.port == 'nrf'
|
||||
uses: ./.github/actions/deps/ports/nrf
|
||||
|
21
.github/actions/deps/ports/broadcom/action.yml
vendored
Normal file
21
.github/actions/deps/ports/broadcom/action.yml
vendored
Normal file
@ -0,0 +1,21 @@
|
||||
name: Fetch broadcom port deps
|
||||
|
||||
runs:
|
||||
using: composite
|
||||
steps:
|
||||
- name: Get broadcom toolchain
|
||||
run: |
|
||||
wget --no-verbose https://adafruit-circuit-python.s3.amazonaws.com/gcc-arm-10.3-2021.07-x86_64-aarch64-none-elf.tar.xz
|
||||
sudo tar -C /usr --strip-components=1 -xaf gcc-arm-10.3-2021.07-x86_64-aarch64-none-elf.tar.xz
|
||||
sudo apt-get install -y mtools
|
||||
shell: bash
|
||||
- name: Install mkfs.fat
|
||||
run: |
|
||||
wget https://github.com/dosfstools/dosfstools/releases/download/v4.2/dosfstools-4.2.tar.gz
|
||||
tar -xaf dosfstools-4.2.tar.gz
|
||||
cd dosfstools-4.2
|
||||
./configure
|
||||
make -j 2
|
||||
cd src
|
||||
echo >> $GITHUB_PATH $(pwd)
|
||||
shell: bash
|
10
.github/actions/deps/ports/litex/action.yml
vendored
Normal file
10
.github/actions/deps/ports/litex/action.yml
vendored
Normal file
@ -0,0 +1,10 @@
|
||||
name: Fetch litex port deps
|
||||
|
||||
runs:
|
||||
using: composite
|
||||
steps:
|
||||
- name: Get litex toolchain
|
||||
run: |
|
||||
wget https://static.dev.sifive.com/dev-tools/riscv64-unknown-elf-gcc-8.3.0-2019.08.0-x86_64-linux-centos6.tar.gz
|
||||
sudo tar -C /usr --strip-components=1 -xaf riscv64-unknown-elf-gcc-8.3.0-2019.08.0-x86_64-linux-centos6.tar.gz
|
||||
shell: bash
|
8
.github/workflows/build-boards.yml
vendored
8
.github/workflows/build-boards.yml
vendored
@ -3,9 +3,6 @@ name: Build boards
|
||||
on:
|
||||
workflow_call:
|
||||
inputs:
|
||||
platform:
|
||||
required: true
|
||||
type: string
|
||||
boards:
|
||||
required: true
|
||||
type: string
|
||||
@ -19,7 +16,7 @@ on:
|
||||
required: false
|
||||
|
||||
jobs:
|
||||
build:
|
||||
board:
|
||||
runs-on: ubuntu-22.04
|
||||
env:
|
||||
CP_VERSION: ${{ inputs.cp-version }}
|
||||
@ -38,6 +35,7 @@ jobs:
|
||||
with:
|
||||
python-version: 3.x
|
||||
- name: Set up port
|
||||
id: set-up-port
|
||||
uses: ./.github/actions/deps/ports
|
||||
with:
|
||||
board: ${{ matrix.board }}
|
||||
@ -47,7 +45,7 @@ jobs:
|
||||
- name: Set up external
|
||||
uses: ./.github/actions/deps/external
|
||||
with:
|
||||
platform: ${{ inputs.platform }}
|
||||
port: ${{ steps.set-up-port.outputs.port }}
|
||||
- name: Set up mpy-cross
|
||||
if: steps.set-up-submodules.outputs.frozen == 'True'
|
||||
uses: ./.github/actions/mpy_cross
|
||||
|
76
.github/workflows/build.yml
vendored
76
.github/workflows/build.yml
vendored
@ -18,16 +18,11 @@ concurrency:
|
||||
|
||||
jobs:
|
||||
scheduler:
|
||||
runs-on: ubuntu-20.04
|
||||
runs-on: ubuntu-22.04
|
||||
outputs:
|
||||
build-doc: ${{ steps.set-matrix.outputs.build-doc }}
|
||||
build-boards: ${{ steps.set-matrix.outputs.build-boards }}
|
||||
build-windows: ${{ steps.set-matrix.outputs.build-windows }}
|
||||
boards-aarch: ${{ steps.set-matrix.outputs.boards-aarch }}
|
||||
boards-arm: ${{ steps.set-matrix.outputs.boards-arm }}
|
||||
boards-atmel: ${{ steps.set-matrix.outputs.boards-atmel }}
|
||||
boards-esp: ${{ steps.set-matrix.outputs.boards-esp }}
|
||||
boards-riscv: ${{ steps.set-matrix.outputs.boards-riscv }}
|
||||
docs: ${{ steps.set-matrix.outputs.docs }}
|
||||
ports: ${{ steps.set-matrix.outputs.ports }}
|
||||
windows: ${{ steps.set-matrix.outputs.windows }}
|
||||
cp-version: ${{ steps.set-up-submodules.outputs.version }}
|
||||
steps:
|
||||
- name: Dump GitHub context
|
||||
@ -99,7 +94,7 @@ jobs:
|
||||
|
||||
mpy-cross:
|
||||
needs: scheduler
|
||||
if: needs.scheduler.outputs.build-boards == 'True'
|
||||
if: needs.scheduler.outputs.ports != '{}'
|
||||
uses: ./.github/workflows/build-mpy-cross.yml
|
||||
secrets: inherit
|
||||
with:
|
||||
@ -108,7 +103,7 @@ jobs:
|
||||
mpy-cross-mac:
|
||||
runs-on: macos-11
|
||||
needs: scheduler
|
||||
if: needs.scheduler.outputs.build-boards == 'True'
|
||||
if: needs.scheduler.outputs.ports != '{}'
|
||||
env:
|
||||
CP_VERSION: ${{ needs.scheduler.outputs.cp-version }}
|
||||
steps:
|
||||
@ -160,10 +155,10 @@ jobs:
|
||||
AWS_ACCESS_KEY_ID: ${{ secrets.AWS_ACCESS_KEY_ID }}
|
||||
AWS_SECRET_ACCESS_KEY: ${{ secrets.AWS_SECRET_ACCESS_KEY }}
|
||||
|
||||
build-doc:
|
||||
docs:
|
||||
runs-on: ubuntu-22.04
|
||||
needs: scheduler
|
||||
if: needs.scheduler.outputs.build-doc == 'True'
|
||||
if: needs.scheduler.outputs.docs == 'True'
|
||||
env:
|
||||
CP_VERSION: ${{ needs.scheduler.outputs.cp-version }}
|
||||
steps:
|
||||
@ -218,10 +213,10 @@ jobs:
|
||||
[ -z "$TWINE_USERNAME" ] || echo "Uploading dev release to PyPi"
|
||||
[ -z "$TWINE_USERNAME" ] || twine upload circuitpython-stubs/dist/*
|
||||
|
||||
build-windows:
|
||||
windows:
|
||||
runs-on: windows-2022
|
||||
needs: scheduler
|
||||
if: needs.scheduler.outputs.build-windows == 'True'
|
||||
if: needs.scheduler.outputs.windows == 'True'
|
||||
env:
|
||||
CP_VERSION: ${{ needs.scheduler.outputs.cp-version }}
|
||||
defaults:
|
||||
@ -286,52 +281,15 @@ jobs:
|
||||
# ERROR: Platform MINGW64_NT-10.0-17763-x86_64 appears to be unsupported
|
||||
# https://github.com/espressif/esp-idf/issues/7062
|
||||
|
||||
aarch:
|
||||
ports:
|
||||
needs: [scheduler, mpy-cross, tests]
|
||||
if: ${{ needs.scheduler.outputs.boards-aarch != '[]' }}
|
||||
if: needs.scheduler.outputs.ports != '{}'
|
||||
uses: ./.github/workflows/build-boards.yml
|
||||
secrets: inherit
|
||||
strategy:
|
||||
fail-fast: false
|
||||
matrix:
|
||||
port: ${{ fromJSON(needs.scheduler.outputs.ports).ports }}
|
||||
with:
|
||||
platform: aarch
|
||||
boards: ${{ needs.scheduler.outputs.boards-aarch }}
|
||||
cp-version: ${{ needs.scheduler.outputs.cp-version }}
|
||||
|
||||
arm:
|
||||
needs: [scheduler, mpy-cross, tests]
|
||||
if: ${{ needs.scheduler.outputs.boards-arm != '[]' }}
|
||||
uses: ./.github/workflows/build-boards.yml
|
||||
secrets: inherit
|
||||
with:
|
||||
platform: arm
|
||||
boards: ${{ needs.scheduler.outputs.boards-arm }}
|
||||
cp-version: ${{ needs.scheduler.outputs.cp-version }}
|
||||
|
||||
atmel:
|
||||
needs: [scheduler, mpy-cross, tests]
|
||||
if: ${{ needs.scheduler.outputs.boards-atmel != '[]' }}
|
||||
uses: ./.github/workflows/build-boards.yml
|
||||
secrets: inherit
|
||||
with:
|
||||
platform: arm
|
||||
boards: ${{ needs.scheduler.outputs.boards-atmel }}
|
||||
cp-version: ${{ needs.scheduler.outputs.cp-version }}
|
||||
|
||||
esp:
|
||||
needs: [scheduler, mpy-cross, tests]
|
||||
if: ${{ needs.scheduler.outputs.boards-esp != '[]' }}
|
||||
uses: ./.github/workflows/build-boards.yml
|
||||
secrets: inherit
|
||||
with:
|
||||
platform: esp
|
||||
boards: ${{ needs.scheduler.outputs.boards-esp }}
|
||||
cp-version: ${{ needs.scheduler.outputs.cp-version }}
|
||||
|
||||
riscv:
|
||||
needs: [scheduler, mpy-cross, tests]
|
||||
if: ${{ needs.scheduler.outputs.boards-riscv != '[]' }}
|
||||
uses: ./.github/workflows/build-boards.yml
|
||||
secrets: inherit
|
||||
with:
|
||||
platform: riscv
|
||||
boards: ${{ needs.scheduler.outputs.boards-riscv }}
|
||||
boards: ${{ toJSON(fromJSON(needs.scheduler.outputs.ports)[matrix.port]) }}
|
||||
cp-version: ${{ needs.scheduler.outputs.cp-version }}
|
||||
|
@ -32,16 +32,6 @@ run-name: ${{ inputs.board }}-${{ inputs.language }}-${{ inputs.version }}${{ in
|
||||
jobs:
|
||||
build:
|
||||
runs-on: ubuntu-22.04
|
||||
env:
|
||||
PLATFORM_atmel-samd: arm
|
||||
PLATFORM_broadcom: aarch
|
||||
PLATFORM_cxd56: arm
|
||||
PLATFORM_espressif: esp
|
||||
PLATFORM_litex: riscv
|
||||
PLATFORM_mimxrt10xx: arm
|
||||
PLATFORM_nrf: arm
|
||||
PLATFORM_raspberrypi: arm
|
||||
PLATFORM_stm: arm
|
||||
steps:
|
||||
- name: Set up repository
|
||||
run: |
|
||||
@ -70,7 +60,7 @@ jobs:
|
||||
uses: ./.github/actions/deps/external
|
||||
with:
|
||||
action: cache
|
||||
platform: ${{ env[format('PLATFORM_{0}', steps.set-up-port.outputs.port)] }}
|
||||
port: ${{ steps.set-up-port.outputs.port }}
|
||||
- name: Set up mpy-cross
|
||||
if: steps.set-up-submodules.outputs.frozen == 'True'
|
||||
uses: ./.github/actions/mpy_cross
|
4
.github/workflows/run-tests.yml
vendored
4
.github/workflows/run-tests.yml
vendored
@ -9,7 +9,7 @@ on:
|
||||
|
||||
jobs:
|
||||
run:
|
||||
runs-on: ubuntu-20.04
|
||||
runs-on: ubuntu-22.04
|
||||
strategy:
|
||||
fail-fast: false
|
||||
matrix:
|
||||
@ -31,7 +31,7 @@ jobs:
|
||||
- name: Set up python
|
||||
uses: actions/setup-python@v4
|
||||
with:
|
||||
python-version: 3.x
|
||||
python-version: 3.8
|
||||
- name: Set up submodules
|
||||
uses: ./.github/actions/deps/submodules
|
||||
with:
|
||||
|
@ -6,7 +6,7 @@ msgstr ""
|
||||
"Project-Id-Version: PACKAGE VERSION\n"
|
||||
"Report-Msgid-Bugs-To: \n"
|
||||
"POT-Creation-Date: 2021-01-04 12:55-0600\n"
|
||||
"PO-Revision-Date: 2023-02-26 06:37+0000\n"
|
||||
"PO-Revision-Date: 2023-03-08 07:10+0000\n"
|
||||
"Last-Translator: Wellington Terumi Uemura <wellingtonuemura@gmail.com>\n"
|
||||
"Language-Team: \n"
|
||||
"Language: pt_BR\n"
|
||||
@ -14,7 +14,7 @@ msgstr ""
|
||||
"Content-Type: text/plain; charset=UTF-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"Plural-Forms: nplurals=2; plural=n > 1;\n"
|
||||
"X-Generator: Weblate 4.16-dev\n"
|
||||
"X-Generator: Weblate 4.16.2-dev\n"
|
||||
|
||||
#: main.c
|
||||
msgid ""
|
||||
@ -847,11 +847,11 @@ msgstr ""
|
||||
|
||||
#: shared-bindings/bitmaptools/__init__.c
|
||||
msgid "Coordinate arrays have different lengths"
|
||||
msgstr ""
|
||||
msgstr "As coordenadas das matrizes possuem comprimentos diferentes"
|
||||
|
||||
#: shared-bindings/bitmaptools/__init__.c
|
||||
msgid "Coordinate arrays types have different sizes"
|
||||
msgstr ""
|
||||
msgstr "Os tipos das coordenadas das matrizes possuem tamanhos diferentes"
|
||||
|
||||
#: py/persistentcode.c
|
||||
msgid "Corrupt .mpy file"
|
||||
|
@ -6,7 +6,7 @@ msgstr ""
|
||||
"Project-Id-Version: PACKAGE VERSION\n"
|
||||
"Report-Msgid-Bugs-To: \n"
|
||||
"POT-Creation-Date: 2021-01-04 12:55-0600\n"
|
||||
"PO-Revision-Date: 2023-02-26 06:37+0000\n"
|
||||
"PO-Revision-Date: 2023-03-08 07:10+0000\n"
|
||||
"Last-Translator: Jonny Bergdahl <jonny@bergdahl.it>\n"
|
||||
"Language-Team: LANGUAGE <LL@li.org>\n"
|
||||
"Language: sv\n"
|
||||
@ -14,7 +14,7 @@ msgstr ""
|
||||
"Content-Type: text/plain; charset=UTF-8\n"
|
||||
"Content-Transfer-Encoding: 8bit\n"
|
||||
"Plural-Forms: nplurals=2; plural=n != 1;\n"
|
||||
"X-Generator: Weblate 4.16-dev\n"
|
||||
"X-Generator: Weblate 4.16.2-dev\n"
|
||||
|
||||
#: main.c
|
||||
msgid ""
|
||||
@ -839,11 +839,11 @@ msgstr ""
|
||||
|
||||
#: shared-bindings/bitmaptools/__init__.c
|
||||
msgid "Coordinate arrays have different lengths"
|
||||
msgstr ""
|
||||
msgstr "Arrayer för koordinater har olika längd"
|
||||
|
||||
#: shared-bindings/bitmaptools/__init__.c
|
||||
msgid "Coordinate arrays types have different sizes"
|
||||
msgstr ""
|
||||
msgstr "Arrayer för koordinater har olika storlek"
|
||||
|
||||
#: py/persistentcode.c
|
||||
msgid "Corrupt .mpy file"
|
||||
|
@ -367,7 +367,7 @@ $(BUILD)/firmware.elf: $(OBJ) $(GENERATED_LD_FILE)
|
||||
$(STEPECHO) "LINK $@"
|
||||
$(Q)echo $(OBJ) > $(BUILD)/firmware.objs
|
||||
$(Q)$(CC) -o $@ $(LDFLAGS) @$(BUILD)/firmware.objs -Wl,--start-group $(LIBS) -Wl,--end-group
|
||||
$(Q)$(SIZE) $@ | $(PYTHON) $(TOP)/tools/build_memory_info.py $(GENERATED_LD_FILE)
|
||||
$(Q)$(SIZE) $@ | $(PYTHON) $(TOP)/tools/build_memory_info.py $(GENERATED_LD_FILE) $(BUILD)
|
||||
|
||||
$(BUILD)/firmware.bin: $(BUILD)/firmware.elf
|
||||
$(STEPECHO) "Create $@"
|
||||
|
@ -434,7 +434,7 @@ $(BUILD)/firmware.elf: $(OBJ) | esp-idf-stamp
|
||||
$(BUILD)/circuitpython-firmware.bin: $(BUILD)/firmware.elf | tools/build_memory_info.py
|
||||
$(STEPECHO) "Create $@"
|
||||
$(Q)esptool.py --chip $(IDF_TARGET) elf2image $(FLASH_FLAGS) --elf-sha256-offset 0xb0 -o $@ $^
|
||||
$(Q)$(PYTHON) tools/build_memory_info.py $< $(BUILD)/esp-idf/sdkconfig $@
|
||||
$(Q)$(PYTHON) tools/build_memory_info.py $< $(BUILD)/esp-idf/sdkconfig $@ $(BUILD)
|
||||
|
||||
$(BUILD)/firmware.bin: $(BUILD)/circuitpython-firmware.bin | esp-idf-stamp
|
||||
$(Q)$(PYTHON) ../../tools/join_bins.py $@ $(BOOTLOADER_OFFSET) $(BUILD)/esp-idf/bootloader/bootloader.bin $(PARTITION_TABLE_OFFSET) $(BUILD)/esp-idf/partition_table/partition-table.bin $(FIRMWARE_OFFSET) $(BUILD)/circuitpython-firmware.bin
|
||||
|
@ -6,13 +6,12 @@
|
||||
# SPDX-License-Identifier: MIT
|
||||
|
||||
import csv
|
||||
import json
|
||||
import os
|
||||
import re
|
||||
import sys
|
||||
|
||||
from elftools.elf.elffile import ELFFile
|
||||
|
||||
print()
|
||||
|
||||
internal_memory = {
|
||||
"esp32": [
|
||||
@ -117,8 +116,12 @@ with open(sys.argv[1], "rb") as stream:
|
||||
|
||||
# This file is the bin
|
||||
used_flash = os.stat(sys.argv[3]).st_size
|
||||
|
||||
free_flash = firmware_region - used_flash
|
||||
|
||||
with open(f"{sys.argv[4]}/firmware.size.json", "w") as f:
|
||||
json.dump({"used_flash": used_flash, "firmware_region": firmware_region}, f)
|
||||
|
||||
print()
|
||||
print(
|
||||
"{:7} bytes used, {:7} bytes free in flash firmware space out of {} bytes ({}kB).".format(
|
||||
used_flash, free_flash, firmware_region, firmware_region / 1024
|
||||
|
@ -136,7 +136,7 @@ all: $(BUILD)/firmware.bin $(BUILD)/firmware.dfu
|
||||
$(BUILD)/firmware.elf: $(OBJ)
|
||||
$(STEPECHO) "LINK $@"
|
||||
$(Q)$(CC) -o $@ $(LDFLAGS) $^ -Wl,--start-group $(LIBS) -Wl,--end-group
|
||||
$(Q)$(SIZE) $@ | $(PYTHON) $(TOP)/tools/build_memory_info.py $(LD_FILE)
|
||||
$(Q)$(SIZE) $@ | $(PYTHON) $(TOP)/tools/build_memory_info.py $(LD_FILE) $(BUILD)
|
||||
|
||||
$(BUILD)/firmware.bin: $(BUILD)/firmware.elf
|
||||
$(STEPECHO) "Create $@"
|
||||
|
@ -212,7 +212,7 @@ $(BUILD)/firmware.elf: $(OBJ) $(GENERATED_LD_FILE)
|
||||
$(STEPECHO) "LINK $@"
|
||||
$(Q)echo $(OBJ) > $(BUILD)/firmware.objs
|
||||
$(Q)$(CC) -o $@ $(LDFLAGS) @$(BUILD)/firmware.objs -Wl,--start-group $(LIBS) -Wl,--end-group
|
||||
$(Q)$(SIZE) $@ | $(PYTHON) $(TOP)/tools/build_memory_info.py $(GENERATED_LD_FILE)
|
||||
$(Q)$(SIZE) $@ | $(PYTHON) $(TOP)/tools/build_memory_info.py $(GENERATED_LD_FILE) $(BUILD)
|
||||
|
||||
$(BUILD)/firmware.bin: $(BUILD)/firmware.elf
|
||||
$(STEPECHO) "Create $@"
|
||||
|
@ -417,7 +417,7 @@ $(BUILD)/firmware.elf: $(OBJ) $(LINK_LD)
|
||||
$(Q)echo $(OBJ) > $(BUILD)/firmware.objs
|
||||
$(Q)echo $(PICO_LDFLAGS) > $(BUILD)/firmware.ldflags
|
||||
$(Q)$(CC) -o $@ $(CFLAGS) @$(BUILD)/firmware.ldflags -Wl,-T,$(LINK_LD) -Wl,-Map=$@.map -Wl,-cref -Wl,--gc-sections @$(BUILD)/firmware.objs -Wl,-lc
|
||||
$(Q)$(SIZE) $@ | $(PYTHON) $(TOP)/tools/build_memory_info.py $(LINK_LD)
|
||||
$(Q)$(SIZE) $@ | $(PYTHON) $(TOP)/tools/build_memory_info.py $(LINK_LD) $(BUILD)
|
||||
|
||||
$(BUILD)/firmware.bin: $(BUILD)/firmware.elf
|
||||
$(STEPECHO) "Create $@"
|
||||
|
@ -268,7 +268,7 @@ $(BUILD)/firmware.elf: $(OBJ)
|
||||
$(STEPECHO) "LINK $@"
|
||||
$(Q)echo $^ > $(BUILD)/firmware.objs
|
||||
$(Q)$(CC) -o $@ $(LDFLAGS) @$(BUILD)/firmware.objs -Wl,--start-group $(LIBS) -Wl,--end-group
|
||||
$(Q)$(SIZE) $@ | $(PYTHON) $(TOP)/tools/build_memory_info.py $(LD_FILE)
|
||||
$(Q)$(SIZE) $@ | $(PYTHON) $(TOP)/tools/build_memory_info.py $(LD_FILE) $(BUILD)
|
||||
|
||||
$(BUILD)/firmware.bin: $(BUILD)/firmware.elf
|
||||
$(STEPECHO) "Create $@"
|
||||
|
@ -20,7 +20,7 @@
|
||||
//| self,
|
||||
//| key: ReadableBuffer,
|
||||
//| mode: int = 0,
|
||||
//| iv: Optional[ReadableBuffer] = None,
|
||||
//| IV: Optional[ReadableBuffer] = None,
|
||||
//| segment_size: int = 8,
|
||||
//| ) -> None:
|
||||
//| """Create a new AES state with the given key.
|
||||
@ -28,7 +28,7 @@
|
||||
//| :param ~circuitpython_typing.ReadableBuffer key: A 16-, 24-, or 32-byte key
|
||||
//| :param int mode: AES mode to use. One of: `MODE_ECB`, `MODE_CBC`, or
|
||||
//| `MODE_CTR`
|
||||
//| :param ~circuitpython_typing.ReadableBuffer iv: Initialization vector to use for CBC or CTR mode
|
||||
//| :param ~circuitpython_typing.ReadableBuffer IV: Initialization vector to use for CBC or CTR mode
|
||||
//|
|
||||
//| Additional arguments are supported for legacy reasons.
|
||||
//|
|
||||
@ -98,33 +98,49 @@ STATIC mp_obj_t aesio_aes_make_new(const mp_obj_type_t *type, size_t n_args,
|
||||
return MP_OBJ_FROM_PTR(self);
|
||||
}
|
||||
|
||||
STATIC mp_obj_t aesio_aes_rekey(size_t n_args, const mp_obj_t *pos_args) {
|
||||
//| def rekey(
|
||||
//| self,
|
||||
//| key: ReadableBuffer,
|
||||
//| IV: Optional[ReadableBuffer] = None,
|
||||
//| ) -> None:
|
||||
//| """Update the AES state with the given key.
|
||||
//|
|
||||
//| :param ~circuitpython_typing.ReadableBuffer key: A 16-, 24-, or 32-byte key
|
||||
//| :param ~circuitpython_typing.ReadableBuffer IV: Initialization vector to use
|
||||
//| for CBC or CTR mode"""
|
||||
//| ...
|
||||
STATIC mp_obj_t aesio_aes_rekey(size_t n_args, const mp_obj_t *pos_args, mp_map_t *kw_args) {
|
||||
aesio_aes_obj_t *self = MP_OBJ_TO_PTR(pos_args[0]);
|
||||
enum { ARG_key, ARG_IV };
|
||||
static const mp_arg_t allowed_args[] = {
|
||||
{MP_QSTR_key, MP_ARG_OBJ | MP_ARG_REQUIRED, {.u_obj = MP_OBJ_NULL} },
|
||||
{MP_QSTR_IV, MP_ARG_OBJ, {.u_obj = MP_OBJ_NULL} },
|
||||
};
|
||||
mp_arg_val_t args[MP_ARRAY_SIZE(allowed_args)];
|
||||
mp_arg_parse_all(n_args - 1, pos_args + 1, kw_args, MP_ARRAY_SIZE(allowed_args), allowed_args, args);
|
||||
|
||||
mp_buffer_info_t bufinfo;
|
||||
mp_get_buffer_raise(pos_args[1], &bufinfo, MP_BUFFER_READ);
|
||||
|
||||
mp_get_buffer_raise(args[ARG_key].u_obj, &bufinfo, MP_BUFFER_READ);
|
||||
const uint8_t *key = bufinfo.buf;
|
||||
size_t key_length = bufinfo.len;
|
||||
if (key == NULL) {
|
||||
mp_raise_ValueError(translate("No key was specified"));
|
||||
}
|
||||
|
||||
if ((key_length != 16) && (key_length != 24) && (key_length != 32)) {
|
||||
mp_raise_ValueError(translate("Key must be 16, 24, or 32 bytes long"));
|
||||
}
|
||||
|
||||
const uint8_t *iv = NULL;
|
||||
if (n_args > 2) {
|
||||
mp_get_buffer_raise(pos_args[2], &bufinfo, MP_BUFFER_READ);
|
||||
size_t iv_length = bufinfo.len;
|
||||
iv = (const uint8_t *)bufinfo.buf;
|
||||
(void)mp_arg_validate_length(iv_length, AES_BLOCKLEN, MP_QSTR_IV);
|
||||
if (args[ARG_IV].u_obj != NULL &&
|
||||
mp_get_buffer(args[ARG_IV].u_obj, &bufinfo, MP_BUFFER_READ)) {
|
||||
(void)mp_arg_validate_length(bufinfo.len, AES_BLOCKLEN, MP_QSTR_IV);
|
||||
|
||||
iv = bufinfo.buf;
|
||||
}
|
||||
|
||||
common_hal_aesio_aes_rekey(self, key, key_length, iv);
|
||||
return mp_const_none;
|
||||
}
|
||||
|
||||
MP_DEFINE_CONST_FUN_OBJ_VAR(aesio_aes_rekey_obj, 2, aesio_aes_rekey);
|
||||
MP_DEFINE_CONST_FUN_OBJ_KW(aesio_aes_rekey_obj, 1, aesio_aes_rekey);
|
||||
|
||||
STATIC void validate_length(aesio_aes_obj_t *self, size_t src_length,
|
||||
size_t dest_length) {
|
||||
|
@ -7,6 +7,8 @@
|
||||
|
||||
import re
|
||||
import sys
|
||||
import json
|
||||
|
||||
|
||||
# Handle size constants with K or M suffixes (allowed in .ld but not in Python).
|
||||
K_PATTERN = re.compile(r"([0-9]+)[kK]")
|
||||
@ -15,11 +17,10 @@ K_REPLACE = r"(\1*1024)"
|
||||
M_PATTERN = re.compile(r"([0-9]+)[mM]")
|
||||
M_REPLACE = r"(\1*1024*1024)"
|
||||
|
||||
print()
|
||||
|
||||
text = 0
|
||||
data = 0
|
||||
bss = 0
|
||||
|
||||
# stdin is the linker output.
|
||||
for line in sys.stdin:
|
||||
# Uncomment to see linker output.
|
||||
@ -29,6 +30,7 @@ for line in sys.stdin:
|
||||
text, data, bss = map(int, line.split()[:3])
|
||||
|
||||
regions = {}
|
||||
|
||||
# This file is the linker script.
|
||||
with open(sys.argv[1], "r") as f:
|
||||
for line in f:
|
||||
@ -51,6 +53,11 @@ used_flash = data + text
|
||||
free_flash = firmware_region - used_flash
|
||||
used_ram = data + bss
|
||||
free_ram = ram_region - used_ram
|
||||
|
||||
with open(f"{sys.argv[2]}/firmware.size.json", "w") as f:
|
||||
json.dump({"used_flash": used_flash, "firmware_region": firmware_region}, f)
|
||||
|
||||
print()
|
||||
print(
|
||||
"{} bytes used, {} bytes free in flash firmware space out of {} bytes ({}kB).".format(
|
||||
used_flash, free_flash, firmware_region, firmware_region / 1024
|
||||
|
@ -11,6 +11,7 @@ import subprocess
|
||||
import shutil
|
||||
import build_board_info as build_info
|
||||
import time
|
||||
import json
|
||||
|
||||
sys.path.append("../docs")
|
||||
from shared_bindings_matrix import get_settings_from_makefile
|
||||
@ -18,7 +19,7 @@ from shared_bindings_matrix import get_settings_from_makefile
|
||||
for port in build_info.SUPPORTED_PORTS:
|
||||
result = subprocess.run("rm -rf ../ports/{port}/build*".format(port=port), shell=True)
|
||||
|
||||
PARALLEL = "-j 5"
|
||||
PARALLEL = "-j 4"
|
||||
if "GITHUB_ACTION" in os.environ:
|
||||
PARALLEL = "-j 2"
|
||||
|
||||
@ -29,6 +30,11 @@ if "BOARDS" in os.environ:
|
||||
|
||||
sha, version = build_info.get_version_info()
|
||||
|
||||
build_all = os.environ.get("GITHUB_EVENT_NAME") != "pull_request"
|
||||
|
||||
LANGUAGE_FIRST = "en_US"
|
||||
LANGUAGE_THRESHOLD = 10 * 1024
|
||||
|
||||
languages = build_info.get_languages()
|
||||
|
||||
all_languages = build_info.get_languages(list_all=True)
|
||||
@ -44,6 +50,9 @@ for board in build_boards:
|
||||
board_info = all_boards[board]
|
||||
board_settings = get_settings_from_makefile("../ports/" + board_info["port"], board)
|
||||
|
||||
languages.remove(LANGUAGE_FIRST)
|
||||
languages.insert(0, LANGUAGE_FIRST)
|
||||
|
||||
for language in languages:
|
||||
bin_directory = "../bin/{board}/{language}".format(board=board, language=language)
|
||||
os.makedirs(bin_directory, exist_ok=True)
|
||||
@ -130,4 +139,16 @@ for board in build_boards:
|
||||
# Flush so we will see something before 10 minutes has passed.
|
||||
print(flush=True)
|
||||
|
||||
if (not build_all) and (language is LANGUAGE_FIRST) and (exit_status is 0):
|
||||
try:
|
||||
with open(
|
||||
f"../ports/{board_info['port']}/{build_dir}/firmware.size.json", "r"
|
||||
) as f:
|
||||
firmware = json.load(f)
|
||||
if firmware["used_flash"] + LANGUAGE_THRESHOLD < firmware["firmware_region"]:
|
||||
print("Skipping languages")
|
||||
break
|
||||
except FileNotFoundError:
|
||||
pass
|
||||
|
||||
sys.exit(exit_status)
|
||||
|
@ -167,8 +167,8 @@ def get_commit_depth_and_check_suite(query_commits):
|
||||
|
||||
|
||||
def get_bad_check_runs(query_check_runs):
|
||||
bad_runs = {}
|
||||
more_pages = True
|
||||
bad_runs_by_matrix = {}
|
||||
|
||||
run_types = ["failed", "incomplete"]
|
||||
|
||||
@ -184,13 +184,12 @@ def get_bad_check_runs(query_check_runs):
|
||||
|
||||
for check_run in check_runs[run_type]["nodes"]:
|
||||
name = check_run["name"]
|
||||
res_matrix = regex_matrix.search(name)
|
||||
if res_matrix:
|
||||
matrix = name.split(" /", 1)[0]
|
||||
matrix_job = name.split(" (", 1)[1][:-1]
|
||||
bad_runs_by_matrix.setdefault(matrix, []).append(matrix_job)
|
||||
if name.startswith("ports") or regex_matrix.search(name):
|
||||
matrix = name.split(" ", 1)[0]
|
||||
matrix_job = name.rsplit(" (", 1)[1][:-1]
|
||||
bad_runs.setdefault(matrix, []).append(matrix_job)
|
||||
elif name != "scheduler":
|
||||
bad_runs_by_matrix[name] = True
|
||||
bad_runs[name] = True
|
||||
else:
|
||||
return {}
|
||||
|
||||
@ -200,7 +199,7 @@ def get_bad_check_runs(query_check_runs):
|
||||
query_check_runs.variables["include" + run_type_camel] = True
|
||||
more_pages = True
|
||||
|
||||
return bad_runs_by_matrix
|
||||
return bad_runs
|
||||
|
||||
|
||||
def set_commit(commit):
|
||||
|
@ -78,13 +78,13 @@ def main():
|
||||
submodules = ["extmod/ulab", "lib/", "tools/"]
|
||||
elif TARGET == "tests":
|
||||
submodules = ["extmod/ulab", "lib/", "tools/"]
|
||||
elif TARGET == "build-doc":
|
||||
elif TARGET == "docs":
|
||||
# used in .readthedocs.yml to generate RTD
|
||||
submodules = ["extmod/ulab"]
|
||||
submodules_tags = ["frozen/"]
|
||||
elif TARGET == "mpy-cross" or TARGET == "mpy-cross-mac":
|
||||
submodules = ["tools/"] # for huffman
|
||||
elif TARGET == "build-windows":
|
||||
elif TARGET == "windows":
|
||||
# This builds one board from a number of ports so fill out a bunch of submodules
|
||||
submodules = ["extmod/ulab", "lib/", "tools/", "ports/", "data/nvm.toml"]
|
||||
elif TARGET == "website":
|
||||
|
@ -42,18 +42,6 @@ from shared_bindings_matrix import (
|
||||
all_ports_all_boards,
|
||||
)
|
||||
|
||||
PORT_TO_JOB = {
|
||||
"atmel-samd": "atmel",
|
||||
"broadcom": "aarch",
|
||||
"cxd56": "arm",
|
||||
"espressif": "esp",
|
||||
"litex": "riscv",
|
||||
"mimxrt10xx": "arm",
|
||||
"nrf": "arm",
|
||||
"raspberrypi": "arm",
|
||||
"stm": "arm",
|
||||
}
|
||||
|
||||
IGNORE = [
|
||||
"tools/ci_set_matrix.py",
|
||||
"tools/ci_check_duplicate_usb_vid_pid.py",
|
||||
@ -109,7 +97,7 @@ def set_output(name: str, value):
|
||||
print(f"Would set GitHub actions output {name} to '{value}'")
|
||||
|
||||
|
||||
def set_boards_to_build(build_all: bool):
|
||||
def set_boards(build_all: bool):
|
||||
if last_failed_jobs.get("mpy-cross") or last_failed_jobs.get("tests"):
|
||||
build_all = True
|
||||
|
||||
@ -224,18 +212,13 @@ def set_boards_to_build(build_all: bool):
|
||||
boards_to_build = all_board_ids
|
||||
break
|
||||
|
||||
# Split boards by job
|
||||
job_to_boards = {"aarch": [], "arm": [], "atmel": [], "esp": [], "riscv": []}
|
||||
|
||||
# Append previously failed boards
|
||||
for job in job_to_boards:
|
||||
if job in last_failed_jobs:
|
||||
for board in last_failed_jobs[job]:
|
||||
boards_to_build.add(board)
|
||||
boards_to_build.update(last_failed_jobs.get("ports") or [])
|
||||
|
||||
build_boards = bool(boards_to_build)
|
||||
print("Building boards:", build_boards)
|
||||
set_output("build-boards", build_boards)
|
||||
print("Building boards:", bool(boards_to_build))
|
||||
|
||||
# Split boards by port
|
||||
port_to_boards_to_build = {}
|
||||
|
||||
# Append boards according to job
|
||||
for board in sorted(boards_to_build):
|
||||
@ -244,17 +227,19 @@ def set_boards_to_build(build_all: bool):
|
||||
# if this happens it's not in `board_to_port`.
|
||||
if not port:
|
||||
continue
|
||||
job_to_boards[PORT_TO_JOB[port]].append(board)
|
||||
port_to_boards_to_build.setdefault(port, []).append(board)
|
||||
print(" ", board)
|
||||
|
||||
# Set the step outputs for each job
|
||||
for job in job_to_boards:
|
||||
set_output(f"boards-{job}", json.dumps(job_to_boards[job]))
|
||||
if port_to_boards_to_build:
|
||||
port_to_boards_to_build["ports"] = sorted(list(port_to_boards_to_build.keys()))
|
||||
|
||||
# Set the step outputs
|
||||
set_output("ports", json.dumps(port_to_boards_to_build))
|
||||
|
||||
|
||||
def set_docs_to_build(build_doc: bool):
|
||||
def set_docs(build_doc: bool):
|
||||
if not build_doc:
|
||||
if last_failed_jobs.get("build-doc"):
|
||||
if last_failed_jobs.get("docs"):
|
||||
build_doc = True
|
||||
else:
|
||||
doc_pattern = re.compile(PATTERN_DOCS)
|
||||
@ -277,12 +262,12 @@ def set_docs_to_build(build_doc: bool):
|
||||
|
||||
# Set the step outputs
|
||||
print("Building docs:", build_doc)
|
||||
set_output("build-doc", build_doc)
|
||||
set_output("docs", build_doc)
|
||||
|
||||
|
||||
def set_windows_to_build(build_windows):
|
||||
def set_windows(build_windows: bool):
|
||||
if not build_windows:
|
||||
if last_failed_jobs.get("build-windows"):
|
||||
if last_failed_jobs.get("windows"):
|
||||
build_windows = True
|
||||
else:
|
||||
for file in changed_files:
|
||||
@ -296,23 +281,22 @@ def set_windows_to_build(build_windows):
|
||||
|
||||
# Set the step outputs
|
||||
print("Building windows:", build_windows)
|
||||
set_output("build-windows", build_windows)
|
||||
|
||||
|
||||
def check_changed_files():
|
||||
if not changed_files:
|
||||
print("Building all docs/boards")
|
||||
return True
|
||||
else:
|
||||
print("Adding docs/boards to build based on changed files")
|
||||
return False
|
||||
set_output("windows", build_windows)
|
||||
|
||||
|
||||
def main():
|
||||
build_all = check_changed_files()
|
||||
set_docs_to_build(build_all)
|
||||
set_windows_to_build(build_all)
|
||||
set_boards_to_build(build_all)
|
||||
# Build all if no changed files
|
||||
build_all = not changed_files
|
||||
print(
|
||||
"Building all docs/boards"
|
||||
if build_all
|
||||
else "Adding docs/boards to build based on changed files"
|
||||
)
|
||||
|
||||
# Set jobs
|
||||
set_docs(build_all)
|
||||
set_windows(build_all)
|
||||
set_boards(build_all)
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
|
Loading…
x
Reference in New Issue
Block a user