[[!tag standards]] Obligatory Dilbert: Links: * * walkthrough video (19jun2022) * PDF version of this DRAFT specification **SV is in DRAFT STATUS**. SV has not yet been submitted to the OpenPOWER Foundation ISA WG for review. === # Scalable Vectors for the Power ISA SV is designed as a strict RISC-paradigm Scalable Vector ISA for Hybrid 3D CPU GPU VPU workloads. As such it brings features normally only found in Cray Supercomputers (Cray-1, NEC SX-Aurora) and in GPUs, but keeps strictly to a *Simple* RISC principle of leveraging a *Scalar* ISA, exclusively using "Prefixing". **Not one single actual explicit Vector opcode exists in SV, at all**. It is suitable for low-power Embedded and DSP Workloads as much as it is for power-efficient Supercomputing. Fundamental design principles: * Taking the simplicity of the RISC paradigm and applying it strictly and uniformly to create a Scalable Vector ISA. * Effectively a hardware for-loop, pausing PC, issuing multiple scalar operations * Preserving the underlying scalar execution dependencies as if the for-loop had been expanded as actual scalar instructions (termed "preserving Program Order") * Specifically designed to be Precise-Interruptible at all times (many Vector ISAs have operations which, due to higher internal accuracy or other complexity, must be effectively atomic only for the full Vector operation's duration, adversely affecting interrupt response latency, or be abandoned and started again) * Augments ("tags") existing instructions, providing Vectorisation "context" rather than adding new instructions. * Strictly does not interfere with or alter the non-Scalable Power ISA in any way * In the Prefix space, does not modify or deviate from the underlying scalar Power ISA unless it provides significant performance or other advantage to do so in the Vector space (dropping the "sticky" characteristics of XER.SO and CR0.SO for example) * Designed for Supercomputing: avoids creating significant sequential dependency hazards, allowing standard high performance superscalar multi-issue micro-architectures to be leveraged. * Divided into Compliancy Levels to reduce cost of implementation for specific needs. Advantages of these design principles: * Simplicity of introduction and implementation on top of the existing Power ISA without disruption. * It is therefore easy to create a first (and sometimes only) implementation as literally a for-loop in hardware, simulators, and compilers. * Hardware Architects may understand and implement SV as being an extra pipeline stage, inserted between decode and issue, that is a simple for-loop issuing element-level sub-instructions. * More complex HDL can be done by repeating existing scalar ALUs and pipelines as blocks and leveraging existing Multi-Issue Infrastructure * As (mostly) a high-level "context" that does not (significantly) deviate from scalar Power ISA and, in its purest form being "a for loop around scalar instructions", it is minimally-disruptive and consequently stands a reasonable chance of broad community adoption and acceptance * Completely wipes not just SIMD opcode proliferation off the map (SIMD is O(N^6) opcode proliferation) but off of Vectorisation ISAs as well. No more separate Vector instructions. Comparative instruction count: * ARM NEON SIMD: around 2,000 instructions, prerequisite: ARM Scalar. * ARM SVE: around 4,000 instructions, prerequisite: NEON and ARM Scalar * ARM SVE2: around 1,000 instructions, prerequisite: SVE, NEON, and ARM Scalar for a grand total of well over 7,000 instructions. * Intel AVX-512: around 4,000 instructions, prerequisite AVX, AVX2, AVX-128 and AVX-256 which in turn critically rely on the rest of x86, for a grand total of well over 10,000 instructions. * RISV-V RVV: 192 instructions, prerequisite 96 Scalar RV64GC instructions * SVP64: **six** instructions, two of which are in the same space (svshape, svshape2), with 24-bit prefixing of prerequisite SFS (150) or SFFS (214) Compliancy Subsets. **There are no dedicated Vector instructions, only Scalar-prefixed**. Comparative Basic Design Principle: * ARM NEON and VSX: PackedSIMD. No instruction-overloaded meaning (every instruction is unique for a given register bitwidth, guaranteeing binary interoperability) * Intel AVX-512 (and below): Hybrid Packed-Predicated SIMD with no instruction-overloading, guaranteeing binary interoperability but at the same time penalising the ISA with runaway opcode proliferation. * ARM SVE/SVE2: Hybrid Packed-Predicated SIMD with instruction-overloading that destroys binary interoperability. This is hidden behind the misuse of the word "Scalable" and is **permitted under License** by "Silicon Partners". * RISC-V RVV: Cray-style Scalable Vector but with instruction-overloading **permitted by the specification** that destroys binary interoperability. * SVP64: Cray-style Scalable Vector with no instruction-overloaded meanings. The regfile numbers and bitwidths shall **not** change in a future revision (for the same instruction encoding): "Silicon Partner" Scaling is prohibited, in order to guarantee binary interoperability. Future revisions of SVP64 may extend VSX instructions to achieve larger regfiles, and non-interoperability on the same will likewise be prohibited. SV comprises several [[sv/compliancy_levels]] suited to Embedded, Energy efficient High-Performance Compute, Distributed Computing and Advanced Computational Supercomputing. The Compliancy Levels are arranged such that even at the bare minimum Level, full Soft-Emulation of all optional and future features is possible. # Sub-pages Pages being developed and examples * [[sv/executive_summary]] * [[sv/overview]] explaining the basics. * [[sv/compliancy_levels]] for minimum subsets through to Advanced Supercomputing. * [[sv/implementation]] implementation planning and coordination * [[sv/po9_encoding]] a new DRAFT 64-bit space similar to EXT1xx, introducing new areas EXT232-63 and EXT300-363 * [[sv/svp64]] contains the packet-format *only*, the [[svp64/appendix]] contains explanations and further details * [[sv/svp64-single]] still under development * [[sv/svp64_quirks]] things in SVP64 that slightly break the rules or are not immediately apparent despite the RISC paradigm * [[opcode_regs_deduped]] autogenerated table of SVP64 decoder augmentation * [[sv/sprs]] SPRs * [[sv/rfc]] RFCs to the [OPF ISA WG](https://openpower.foundation/isarfc/) SVP64 "Modes": * For condition register operations see [[sv/cr_ops]] - SVP64 Condition Register ops: Guidelines on Vectorisation of any v3.0B base operations which return or modify a Condition Register bit or field. * For LD/ST Modes, see [[sv/ldst]]. * For Branch modes, see [[sv/branches]] - SVP64 Conditional Branch behaviour: All/Some Vector CRs * For arithmetic and logical, see [[sv/normal]] * [[sv/mv.vec]] pack/unpack move to and from vec2/3/4, actually an RM.EXTRA Mode and a [[sv/remap]] mode Core SVP64 instructions: * [[sv/setvl]] the Cray-style "Vector Length" instruction * svremap, svindex and svshape: part of [[sv/remap]] "Remapping" for Matrix Multiply, DCT/FFT and RGB-style "Structure Packing" as well as general-purpose Indexing. Also describes associated SPRs. * [[sv/svstep]] Key stepping instruction, primarily for Vertical-First Mode and also providing traditional "Vector Iota" capability. *Please note: there are only six instructions in the whole of SV. Beyond this point are additional **Scalar** instructions related to specific workloads that have nothing to do with the SV Specification* # Stability Guarantees in Simple-V Providing long-term stability in an ISA is extremely challenging but critically important. It requires certain guarantees to be provided. * Firstly: that instructions will never be ambiguously-defined. * Secondly, that no instruction shall change meaning to produce different results on different hardware (present or future). * Thirdly, that Scalar "defined words" (32 bit instruction encodings) if Vectorised will also always be implemented as identical Scalar instructions (the sole semi-exception being Vectorised Branch-Conditional) * Fourthly, that implementors are not permitted to either add arbitrary features nor implement features in an incompatible way. *(Performance may differ, but differing results are not permitted)*. * Fifthly, that any part of Simple-V not implemented by a lower Compliancy Level is *required* to raise an illegal instruction trap (allowing soft-emulation), including if Simple-V is not implemented at all. * Sixthly, that any `UNDEFINED` behaviour for practical implementation reasons is clearly documented for both programmers and hardware implementors. In particular, given the strong recent emphasis and interest in "Scalable Vector" ISAs, it is most unfortunate that both ARM SVE and RISC-V RVV permit the exact same instruction to produce different results on different hardware depending on a "Silicon Partner" hardware choice. This choice catastrophically and irrevocably causes binary non-interoperability *despite being a "feature"*. Explained in it is the exact same binary-incompatibility issue faced by Power ISA on its 32- to 64-bit transition: 32-bit hardware was **unable** to trap-and-emulate 64-bit binaries because the opcodes were (are) the same. It is therefore *guaranteed* that extensions to the register file width and quantity in Simple-V shall only be made in future by explicit means, ensuring binary compatibility. # Optional Scalar instructions **Additional Instructions for specific purposes (not SVP64)** All of these instructions below have nothing to do with SV. They are all entirely designed as Scalar instructions that, as Scalar instructions, stand on their own merit. Considerable lengths have been made to provide justifications for each of these *Scalar* instructions in a *Scalar* context, completely independently of SVP64. Some of these Scalar instructions happen also designed to make Scalable Vector binaries more efficient, such as the crweird group. Others are to bring the Scalar Power ISA up-to-date within specific workloads, such as a JavaScript Rounding instruction (which saves 32 scalar instructions including seven branch instructions). None of them are strictly necessary but performance and power consumption may be (or, is already) compromised in certain workloads and use-cases without them. Vector-related but still Scalar: * [[sv/mv.swizzle]] vec2/3/4 Swizzles (RGBA, XYZW) for 3D and CUDA. designed as a Scalar instruction. * [[sv/vector_ops]] scalar operations needed for supporting vectors * [[sv/cr_int_predication]] scalar instructions needed for effective predication Stand-alone Scalar Instructions: * [[sv/bitmanip]] * [[sv/fcvt]] FP Conversion (due to OpenPOWER Scalar FP32) * [[sv/fclass]] detect class of FP numbers * [[sv/int_fp_mv]] Move and convert GPR <-> FPR, needed for !VSX * [[sv/av_opcodes]] scalar opcodes for Audio/Video * [[prefix_codes]] Decode/encode prefix-codes, used by JPEG, DEFLATE, etc. * TODO: OpenPOWER adaptation [[openpower/transcendentals]] Twin targetted instructions (two registers out, one implicit, just like Load-with-Update). * [[isa/svfixedarith]] * [[isa/svfparith]] * [[sv/biginteger]] Operations that help with big arithmetic Explanation of the rules for twin register targets (implicit RS, FRS) explained in SVP64 [[svp64/appendix]] # Architectural Note This section is primarily for the ISA Working Group and for IBM in their capacity and responsibility for allocating "Architectural Resources" (opcodes), but it is also useful for general understanding of Simple-V. Simple-V is effectively a type of "Zero-Overhead Loop Control" to which an entire 24 bits are exclusively dedicated in a fully RISC-abstracted manner. Within those 24-bits there are no Scalar instructions, and no Vector instructions: there is *only* "Loop Control". This is why there are no actual Vector operations in Simple-V: *all* suitable Scalar Operations are Vectorised or not at all. This has some extremely important implications when considering adding new instructions, and especially when allocating the Opcode Space for them. To protect SVP64 from damage, a "Hard Rule" has to be set: Scalar Instructions must be simultaneously added in the corresponding SVP64 opcode space with the exact same 32-bit "Defined Word" or they must not be added at all. Likewise, instructions planned for addition in what is considered (wrongly) to be the exclusive "Vector" domain must correspondingly be added in the Scalar space with the exact same 32-bit "Defined Word", or they must not be added at all. Some explanation of the above is needed. Firstly, "Defined Word" is a term used in Section 1.6.3 of the Power ISA v3 1 Book I: it means, in short, "a 32 bit instruction", which can then be Prefixed by EXT001 to extend it to 64-bit (named EXT100-163). Prefixed-Prefixed (96-bit Variable-Length) encodings are prohibited in v3.1 and they are just as prohibited in Simple-V: it's too complex in hardware. This means that **only** 32-bit "Defined Words" may be Vectorised, and in particular it means that no 64-bit instruction (EXT100-163) may **ever** be Vectorised. Secondly, the term "Vectoriseable" was used. This refers to "instructions which if SVP64-Prefixed are actually meaningful". `sc` is meaningless to Vectorise, for example, as is `sync` and `mtmsr` (there is only ever going to be one MSR). The problem comes if the rationale is applied, "if unused, Unvectoriseable opcodes can therefore be allocated to alternative instructions mixing inside the SVP64 Opcode space", which unfortunately results in huge inadviseable complexity in HDL at the Decode Phase, attempting to discern between the two types. Worse than that, if the alternate 64-bit instruction is Vectoriseable but the 32-bit Scalar "Defined Word" is already allocated, how can there ever be a Scalar version of the alternate instruction? It would have to be added as a **completely different** 32-bit "Defined Word", and things go rapidly downhill in the Decoder as well as the ISA from there. Therefore to avoid risk and long-term damage to the Power ISA: * *even Unvectoriseable* "Defined Words" (`mtmsr`) must have the corresponding SVP64 Prefixed Space `RESERVED`, permanently requiring Illegal Instruction to be raised (the 64-bit encoding corresponding to an illegal `sv.mtmsr` if ever incorrectly attempted must be **defined** to raise an Exception) * *Even instructions that may not be Scalar* (although for various practical reasons this is extremely rare if not impossible, if not just generally "strongly discouraged") which have no meaning or use as a 32-bit Scalar "Defined Word", **must** still have the Scalar "Defined Word" `RESERVED` in the scalar opcode space, as an Illegal Instruction. A good example of the former is `mtmsr` because there is only one MSR register (`sv.mtmsr` is meaningless, as is `sv.sc`), and a good example of the latter is [[sv/mv.x]] which is so deeply problematic to add to any Scalar ISA that it was rejected outright and an alternative route taken (Indexed REMAP). Another good example would be Cross Product which has no meaning at all in a Scalar ISA (Cross Product as a concept only applies to Mathematical Vectors). If any such Vector operation were ever added, it would be **critically** important to reserve the exact same *Scalar* opcode with the exact same "Defined Word" in the *Scalar* Power ISA opcode space, as an Illegal Instruction. There are good reasons why Cross Product has not been proposed, but it serves to illustrate the point as far as Architectural Resource Allocation is concerned. Bottom line is that whilst this seems wasteful the alternatives are a destabilisation of the Power ISA and impractically-complex Hardware Decoders. With the Scalar Power ISA (v3.0, v3.1) already being comprehensive in the number of instructions, keeping further Decode complexity down is a high priority. # Other Scalable Vector ISAs These Scalable Vector ISAs are listed to aid in understanding and context of what is involved. * Original Cray ISA * NEC SX Aurora (still in production, inspired by Cray) * RISC-V RVV (inspired by Cray) * MRISC32 ISA Manual (under active development) * Mitch Alsup's MyISA 66000 Vector Processor ISA Manual is available from Mitch on request. A comprehensive list of 3D GPU, Packed SIMD, Predicated-SIMD and true Scalable Vector ISAs may be found at the [[sv/vector_isa_comparison]] page. Note: AVX-512 and SVE2 are *not Vector ISAs*, they are Predicated-SIMD. *Public discussions have taken place at Conferences attended by both Intel and ARM on adding a `setvl` instruction which would easily make both AVX-512 and SVE2 truly "Scalable".* [[sv/comparison_table]] in tabular form. # Major opcodes summary Simple-V itself only requires six instructions with 6-bit Minor XO (bits 26-31), and the SVP64 Prefix Encoding requires 25% space of the EXT001 Major Opcode. There are **no** Vector Instructions and consequently **no further opcode space is required**. Even though they are currently placed in the EXT022 Sandbox, the "Management" instructions (setvl, svstep, svremap, svshape, svindex) are designed to fit cleanly into EXT019 (exactly like `addpcis`) or other 5/6-bit Minor XO area (bits 25-31) that has space for Rc=1. That said: for the target workloads for which Scalable Vectors are typically used, the Scalar ISA on which those workloads critically rely is somewhat anaemic. The Libre-SOC Team has therefore been addressing that by developing a number of Scalar instructions in specialist areas (Big Integer, Cryptography, 3D, Audio/Video, DSP) and it is these which require considerable Scalar opcode space. Please be advised that even though SV is entirely DRAFT status, there is considerable concern that because there is not yet any two-way day-to-day communication established with the OPF ISA WG, we have no idea if any of these are conflicting with future plans by any OPF Members. **The External ISA WG RFC Process has now been ratified but Libre-SOC may not join the OPF as an entity because it does not exist except in name. Even if it existed it would be a conflict of interest to join the OPF, due to our funding remit from NLnet**. We therefore proceed on the basis of making public the intention to submit RFCs once the External ISA WG RFC Process is in place and, in a wholly unsatisfactory manner have to *hope and trust* that OPF ISA WG Members are reading this and take it into consideration. **Scalar Summary** As in above sections, it is emphasised strongly that Simple-V in no way critically depends on the 100 or so *Scalar* instructions also being developed by Libre-SOC. **None of these Draft opcodes are intended for private custom secret proprietary usage. They are all intended for entirely public, upstream, high-profile mass-volume day-to-day usage at the same level as add, popcnt and fld** * bitmanip requires two major opcodes (due to 16+ bit immediates) those are currently EXT022 and EXT05. * brownfield encoding in one of those two major opcodes still requires multiple VA-Form operations (in greater numbers than EXT04 has spare) * space in EXT019 next to addpcis and crops is recommended (or any other 5-6 bit Minor XO areas) * many X-Form opcodes currently in EXT022 have no preference for a location at all, and may be moved to EXT059, EXT019, EXT031 or other much more suitable location. * even if ratified and even if the majority (mostly X-Form) is moved to other locations, the large immediate sizes of the remaining bitmanip instructions means it would be highly likely these remaining instructions would need two major opcodes. Fortuitously the v3.1 Spec states that both EXT005 and EXT009 are available. **Additional observations** Note that there is no Sandbox allocation in the published ISA Spec for v3.1 EXT01 usage, and because SVP64 is already 64-bit Prefixed, Prefixed-Prefixed-instructions (SVP64 Prefixed v3.1 Prefixed) would become a whopping 96-bit long instruction. Avoiding this situation is a high priority which in turn by necessity puts pressure on the 32-bit Major Opcode space. Note also that EXT022, the Official Architectural Sandbox area available for "Custom non-approved purposes" according to the Power ISA Spec, is under severe design pressure as it is insufficient to hold the full extent of the instruction additions required to create a Hybrid 3D CPU-VPU-GPU. Although the wording of the Power ISA Specification leaves open the *possibility* of not needing to propose ISA Extensions to the ISA WG, it is clear that EXT022 is an inappropriate location for a large high-profile Extension intended for mass-volume product deployment. Every in-good-faith effort will therefore be made to work with the OPF ISA WG to submit SVP64 via the External RFC Process. **Whilst SVP64 is only 6 instructions the heavy focus on VSX for the past 12 years has left the SFFS Level anaemic and out-of-date compared to ARM and x86.** This is very much a blessing, as the Scalar ISA has remained clean, making it highly suited to RISC-paradigm Scalable Vector Prefixing. Approximately 100 additional (optional) Scalar Instructions are up for proposal to bring SFFS up-to-date. None of them require or depend on PackedSIMD VSX (or VMX). # Other Examples experiments future ideas discussion: * [Scalar register access](https://bugs.libre-soc.org/show_bug.cgi?id=905) above r31 and CR7. * [[sv/propagation]] Context propagation including svp64, swizzle and remap * [[sv/masked_vector_chaining]] * [[sv/discussion]] * [[sv/example_dep_matrices]] * [[sv/major_opcode_allocation]] * [[sv/byteswap]] * [[sv/16_bit_compressed]] experimental * [[sv/toc_data_pointer]] experimental * [[sv/predication]] discussion on predication concepts * [[sv/register_type_tags]] * [[sv/mv.x]] deprecated in favour of Indexed REMAP Additional links: * * [[sv/vector_isa_comparison]] - a list of Packed SIMD, GPU, and other Scalable Vector ISAs * [[sv/comparison_table]] - a one-off (experimental) table comparing ISAs * [[simple_v_extension]] old (deprecated) version * [[openpower/sv/llvm]]