armv7a ,cortex a : add L1, L2 cache support, va to pa support
authorMichel Jaouen <michel.jaouen@stericsson.com>
Thu, 29 Sep 2011 15:17:27 +0000 (17:17 +0200)
committerØyvind Harboe <oyvind.harboe@zylin.com>
Fri, 30 Sep 2011 07:45:26 +0000 (09:45 +0200)
src/target/arm_dpm.c
src/target/arm_dpm.h
src/target/armv7a.c
src/target/armv7a.h
src/target/cortex_a.c
src/target/cortex_a.h

index 012316bf8ad78e65be79874d95fe79ecb4239ef8..dfd3f275baec243816bfa70e385d67dde72199c9 100644 (file)
@@ -109,7 +109,7 @@ static int dpm_mcr(struct target *target, int cpnum,
 /* Toggles between recorded core mode (USR, SVC, etc) and a temporary one.
  * Routines *must* restore the original mode before returning!!
  */
-static int dpm_modeswitch(struct arm_dpm *dpm, enum arm_mode mode)
+int dpm_modeswitch(struct arm_dpm *dpm, enum arm_mode mode)
 {
        int retval;
        uint32_t cpsr;
index b20184c8d5475c2957e489314af6240be37e9580..ba2e155c24481f2c88259cd41e25fc552e32d794 100644 (file)
@@ -133,6 +133,9 @@ int arm_dpm_setup(struct arm_dpm *dpm);
 int arm_dpm_initialize(struct arm_dpm *dpm);
 
 int arm_dpm_read_current_registers(struct arm_dpm *);
+int dpm_modeswitch(struct arm_dpm *dpm, enum arm_mode mode);
+
+
 int arm_dpm_write_dirty_registers(struct arm_dpm *, bool bpwp);
 
 void arm_dpm_report_wfar(struct arm_dpm *, uint32_t wfar);
index 151deb41534ac19afdecd49c2beec78501002b89..e0d08827f62a730eed52eef462efac54fcb5292e 100644 (file)
@@ -1,6 +1,8 @@
 /***************************************************************************
  *    Copyright (C) 2009 by David Brownell                                 *
  *                                                                         *
+ *    Copyright (C) ST-Ericsson SA 2011 michel.jaouen@stericsson.com       *
+ *                                                                         *
  *   This program is free software; you can redistribute it and/or modify  *
  *   it under the terms of the GNU General Public License as published by  *
  *   the Free Software Foundation; either version 2 of the License, or     *
@@ -34,7 +36,8 @@
 #include <unistd.h>
 
 #include "arm_opcodes.h"
-
+#include "target.h"
+#include "target_type.h"
 
 static void armv7a_show_fault_registers(struct target *target)
 {
@@ -84,6 +87,661 @@ done:
        /* (void) */ dpm->finish(dpm);
 }
 
+int armv7a_read_ttbcr(struct target *target)
+{
+       struct armv7a_common *armv7a = target_to_armv7a(target);
+       struct arm_dpm *dpm = armv7a->armv4_5_common.dpm;
+       uint32_t ttbcr;
+       int retval = dpm->prepare(dpm);
+       if (retval!=ERROR_OK) goto done;
+       /*  MRC p15,0,<Rt>,c2,c0,2 ; Read CP15 Translation Table Base Control Register*/
+       retval = dpm->instr_read_data_r0(dpm,
+                       ARMV4_5_MRC(15, 0, 0, 2, 0, 2),
+                       &ttbcr);
+       if (retval!=ERROR_OK) goto done;
+    armv7a->armv7a_mmu.ttbr1_used = ((ttbcr & 0x7)!=0)? 1: 0;
+    armv7a->armv7a_mmu.ttbr0_mask  = 7 << (32 -((ttbcr & 0x7)));
+#if 0
+    LOG_INFO("ttb1 %s ,ttb0_mask %x",
+       armv7a->armv7a_mmu.ttbr1_used ? "used":"not used",
+       armv7a->armv7a_mmu.ttbr0_mask);
+#endif
+       if (armv7a->armv7a_mmu.ttbr1_used == 1)
+       {
+               LOG_INFO("SVC access above %x",
+                               (0xffffffff & armv7a->armv7a_mmu.ttbr0_mask));
+               armv7a->armv7a_mmu.os_border = 0xffffffff & armv7a->armv7a_mmu.ttbr0_mask;
+       }
+       else
+       {
+               /*  fix me , default is hard coded LINUX border  */
+        armv7a->armv7a_mmu.os_border = 0xc0000000;
+       }
+done:
+       dpm->finish(dpm);
+       return retval;
+}
+
+
+/*  method adapted to cortex A : reused arm v4 v5 method*/
+int armv7a_mmu_translate_va(struct target *target,  uint32_t va, uint32_t *val)
+{
+       uint32_t first_lvl_descriptor = 0x0;
+       uint32_t second_lvl_descriptor = 0x0;
+       int retval;
+       uint32_t cb;
+       struct armv7a_common *armv7a = target_to_armv7a(target);
+       struct arm_dpm *dpm = armv7a->armv4_5_common.dpm;
+       uint32_t ttb = 0; /*  default ttb0 */
+       if (armv7a->armv7a_mmu.ttbr1_used == -1) armv7a_read_ttbcr(target);
+       if ((armv7a->armv7a_mmu.ttbr1_used) &&
+                       (va > (0xffffffff & armv7a->armv7a_mmu.ttbr0_mask)))
+       {
+               /*  select ttb 1 */
+               ttb = 1;
+       }
+       retval = dpm->prepare(dpm);
+       if (retval != ERROR_OK)
+               goto done;
+
+       /*  MRC p15,0,<Rt>,c2,c0,ttb */
+       retval = dpm->instr_read_data_r0(dpm,
+                       ARMV4_5_MRC(15, 0, 0, 2, 0, ttb),
+                       &ttb);
+       retval = armv7a->armv7a_mmu.read_physical_memory(target,
+                       (ttb & 0xffffc000) | ((va & 0xfff00000) >> 18),
+                       4, 1, (uint8_t*)&first_lvl_descriptor);
+       if (retval != ERROR_OK)
+               return retval;
+       first_lvl_descriptor = target_buffer_get_u32(target, (uint8_t*)
+                       &first_lvl_descriptor);
+       /*  reuse armv4_5 piece of code, specific armv7a changes may come later */
+       LOG_DEBUG("1st lvl desc: %8.8" PRIx32 "", first_lvl_descriptor);
+
+       if ((first_lvl_descriptor & 0x3) == 0)
+       {
+               LOG_ERROR("Address translation failure");
+               return ERROR_TARGET_TRANSLATION_FAULT;
+       }
+
+
+       if ((first_lvl_descriptor & 0x3) == 2)
+       {
+               /* section descriptor */
+               cb = (first_lvl_descriptor & 0xc) >> 2;
+               *val = (first_lvl_descriptor & 0xfff00000) | (va & 0x000fffff);
+               return ERROR_OK;
+       }
+
+       if ((first_lvl_descriptor & 0x3) == 1)
+       {
+               /* coarse page table */
+               retval = armv7a->armv7a_mmu.read_physical_memory(target,
+                               (first_lvl_descriptor & 0xfffffc00) | ((va & 0x000ff000) >> 10),
+                               4, 1, (uint8_t*)&second_lvl_descriptor);
+               if (retval != ERROR_OK)
+                       return retval;
+       }
+       else if ((first_lvl_descriptor & 0x3) == 3)
+       {
+               /* fine page table */
+               retval = armv7a->armv7a_mmu.read_physical_memory(target,
+                               (first_lvl_descriptor & 0xfffff000) | ((va & 0x000ffc00) >> 8),
+                               4, 1, (uint8_t*)&second_lvl_descriptor);
+               if (retval != ERROR_OK)
+                       return retval;
+       }
+
+       second_lvl_descriptor = target_buffer_get_u32(target, (uint8_t*)
+                       &second_lvl_descriptor);
+
+       LOG_DEBUG("2nd lvl desc: %8.8" PRIx32 "", second_lvl_descriptor);
+
+       if ((second_lvl_descriptor & 0x3) == 0)
+       {
+               LOG_ERROR("Address translation failure");
+               return ERROR_TARGET_TRANSLATION_FAULT;
+       }
+
+       /* cacheable/bufferable is always specified in bits 3-2 */
+       cb = (second_lvl_descriptor & 0xc) >> 2;
+
+       if ((second_lvl_descriptor & 0x3) == 1)
+       {
+               /* large page descriptor */
+               *val = (second_lvl_descriptor & 0xffff0000) | (va & 0x0000ffff);
+               return ERROR_OK;
+       }
+
+       if ((second_lvl_descriptor & 0x3) == 2)
+       {
+               /* small page descriptor */
+               *val = (second_lvl_descriptor & 0xfffff000) | (va & 0x00000fff);
+               return ERROR_OK;
+       }
+
+       if ((second_lvl_descriptor & 0x3) == 3)
+       {
+               *val = (second_lvl_descriptor & 0xfffffc00) | (va & 0x000003ff);
+               return ERROR_OK;
+       }
+
+       /* should not happen */
+       LOG_ERROR("Address translation failure");
+       return ERROR_TARGET_TRANSLATION_FAULT;
+
+done:
+       return retval;
+}
+
+
+/*  V7 method VA TO PA  */
+int armv7a_mmu_translate_va_pa(struct target *target, uint32_t va,
+               uint32_t *val, int meminfo)
+{
+       int retval = ERROR_FAIL;
+       struct armv7a_common *armv7a = target_to_armv7a(target);
+       struct arm_dpm *dpm = armv7a->armv4_5_common.dpm;
+       uint32_t virt = va & ~0xfff;
+       uint32_t NOS,NS,SH,INNER,OUTER;
+       *val = 0xdeadbeef;
+       retval = dpm->prepare(dpm);
+       if (retval != ERROR_OK)
+               goto done;
+       /*  mmu must be enable in order to get a correct translation */
+       /*  use VA to PA CP15 register for conversion */
+       retval = dpm->instr_write_data_r0(dpm,
+                       ARMV4_5_MCR(15, 0, 0, 7, 8, 0),
+                       virt);
+       if (retval!=ERROR_OK) goto done;
+       retval = dpm->instr_read_data_r0(dpm,
+                       ARMV4_5_MRC(15, 0, 0, 7, 4, 0),
+                       val);
+       /* decode memory attribute */
+       NOS = (*val >> 10) & 1; /*  Not Outer shareable */
+       NS = (*val >> 9) & 1; /* Non secure */
+       SH = (*val >> 7 )& 1; /*  shareable */
+       INNER = (*val >> 4) &  0x7;
+       OUTER = (*val >> 2) & 0x3;
+
+       if (retval!=ERROR_OK) goto done;
+       *val = (*val & ~0xfff)  +  (va & 0xfff);
+       if (*val == va)
+               LOG_WARNING("virt = phys  : MMU disable !!");
+       if (meminfo)
+       {
+               LOG_INFO("%x : %x %s outer shareable %s secured",
+                               va, *val,
+                               NOS == 1 ? "not" : " ",
+                               NS == 1 ? "not" :"");
+               switch (OUTER) {
+               case 0 : LOG_INFO("outer: Non-Cacheable");
+                                break;
+               case 1 : LOG_INFO("outer: Write-Back, Write-Allocate");
+                                break;
+               case 2 : LOG_INFO("outer: Write-Through, No Write-Allocate");
+                                break;
+               case 3 : LOG_INFO("outer: Write-Back, no Write-Allocate");
+                                break;
+               }
+               switch (INNER) {
+               case 0 : LOG_INFO("inner: Non-Cacheable");
+                                break;
+               case 1 : LOG_INFO("inner: Strongly-ordered");
+                                break;
+               case 3 : LOG_INFO("inner: Device");
+                                break;
+               case 5 : LOG_INFO("inner: Write-Back, Write-Allocate");
+                                break;
+               case 6 : LOG_INFO("inner:  Write-Through");
+                                break;
+               case 7 : LOG_INFO("inner: Write-Back, no Write-Allocate");
+
+               default: LOG_INFO("inner: %x ???",INNER);
+               }
+       }
+
+done:
+       dpm->finish(dpm);
+
+       return retval;
+}
+
+static int armv7a_handle_inner_cache_info_command(struct command_context *cmd_ctx,
+               struct armv7a_cache_common *armv7a_cache)
+{
+       if (armv7a_cache->ctype == -1)
+       {
+               command_print(cmd_ctx, "cache not yet identified");
+               return ERROR_OK;
+       }
+
+       command_print(cmd_ctx,
+                       "D-Cache: linelen %i, associativity %i, nsets %i, cachesize %d KBytes",
+                       armv7a_cache->d_u_size.linelen,
+                       armv7a_cache->d_u_size.associativity,
+                       armv7a_cache->d_u_size.nsets,
+                       armv7a_cache->d_u_size.cachesize);
+
+       command_print(cmd_ctx,
+                       "I-Cache: linelen %i, associativity %i, nsets %i, cachesize %d KBytes",
+                       armv7a_cache->i_size.linelen,
+                       armv7a_cache->i_size.associativity,
+                       armv7a_cache->i_size.nsets,
+                       armv7a_cache->i_size.cachesize);
+
+       return ERROR_OK;
+}
+
+static int _armv7a_flush_all_data(struct target *target)
+{
+       struct armv7a_common *armv7a = target_to_armv7a(target);
+       struct arm_dpm *dpm = armv7a->armv4_5_common.dpm;
+       struct armv7a_cachesize *d_u_size =
+               &(armv7a->armv7a_mmu.armv7a_cache.d_u_size);
+       int32_t c_way, c_index = d_u_size->index;
+       int retval;
+       /*  check that cache data is on at target halt */
+       if (!armv7a->armv7a_mmu.armv7a_cache.d_u_cache_enabled)
+       {
+               LOG_INFO("flushed not performed :cache not on at target halt");
+               return ERROR_OK;
+       }
+       retval = dpm->prepare(dpm);
+       if (retval != ERROR_OK) goto done;
+       do {
+               c_way = d_u_size->way;
+               do {
+                       uint32_t value = (c_index << d_u_size->index_shift)
+                               | (c_way << d_u_size->way_shift);
+                       /*  DCCISW */
+                       //LOG_INFO ("%d %d %x",c_way,c_index,value);
+                       retval = dpm->instr_write_data_r0(dpm,
+                                       ARMV4_5_MCR(15, 0, 0, 7, 14, 2),
+                                       value);
+                       if (retval!= ERROR_OK) goto done;
+                       c_way -= 1;
+               } while (c_way >=0);
+               c_index -= 1;
+       } while (c_index >=0);
+       return retval;
+done:
+       LOG_ERROR("flushed failed");
+       dpm->finish(dpm);
+       return retval;
+}
+
+static int  armv7a_flush_all_data( struct target * target)
+{
+       int retval = ERROR_FAIL;
+       /*  check that armv7a_cache is correctly identify */
+       struct armv7a_common *armv7a = target_to_armv7a(target);
+       if (armv7a->armv7a_mmu.armv7a_cache.ctype == -1)
+       {
+               LOG_ERROR("trying to flush un-identified cache");
+               return retval;
+       }
+
+       if (target->smp)
+       {
+               /*  look if all the other target have been flushed in order to flush level
+                *  2 */
+               struct target_list *head;
+               struct target *curr;
+               head = target->head;
+               while(head != (struct target_list*)NULL)
+               {
+                       curr = head->target;
+                       if ((curr->state == TARGET_HALTED))
+                       {   LOG_INFO("Wait flushing data l1 on core %d",curr->coreid);
+                               retval = _armv7a_flush_all_data(curr);
+                       }
+                       head = head->next;
+               }
+       }
+       else retval = _armv7a_flush_all_data(target);
+       return retval;
+}
+
+
+/* L2 is not specific to armv7a  a specific file is needed */
+static int armv7a_l2x_flush_all_data(struct target * target)
+{
+
+#define L2X0_CLEAN_INV_WAY             0x7FC
+       int retval = ERROR_FAIL;
+       struct armv7a_common *armv7a = target_to_armv7a(target);
+       struct armv7a_l2x_cache *l2x_cache = (struct armv7a_l2x_cache*)
+               (armv7a->armv7a_mmu.armv7a_cache.l2_cache);
+       uint32_t base = l2x_cache->base;
+       uint32_t l2_way = l2x_cache->way;
+       uint32_t l2_way_val = (1<<l2_way) -1;
+       retval = armv7a_flush_all_data(target);
+       if (retval!=ERROR_OK) return retval;
+       retval = target->type->write_phys_memory(target,
+                       (uint32_t)(base+(uint32_t)L2X0_CLEAN_INV_WAY),
+                       (uint32_t)4,
+                       (uint32_t)1,
+                       (uint8_t*)&l2_way_val);
+       return retval;
+}
+
+static int armv7a_handle_l2x_cache_info_command(struct command_context *cmd_ctx,
+               struct armv7a_cache_common *armv7a_cache)
+{
+
+       struct armv7a_l2x_cache *l2x_cache = (struct armv7a_l2x_cache*)
+                       (armv7a_cache->l2_cache);
+
+       if (armv7a_cache->ctype == -1)
+       {
+               command_print(cmd_ctx, "cache not yet identified");
+               return ERROR_OK;
+       }
+
+       command_print(cmd_ctx,
+                       "L1 D-Cache: linelen %i, associativity %i, nsets %i, cachesize %d KBytes",
+               armv7a_cache->d_u_size.linelen,
+               armv7a_cache->d_u_size.associativity,
+               armv7a_cache->d_u_size.nsets,
+               armv7a_cache->d_u_size.cachesize);
+
+       command_print(cmd_ctx,
+                       "L1 I-Cache: linelen %i, associativity %i, nsets %i, cachesize %d KBytes",
+               armv7a_cache->i_size.linelen,
+               armv7a_cache->i_size.associativity,
+               armv7a_cache->i_size.nsets,
+               armv7a_cache->i_size.cachesize);
+    command_print(cmd_ctx, "L2 unified cache Base Address 0x%x, %d ways",
+                       l2x_cache->base, l2x_cache->way);
+
+
+       return ERROR_OK;
+}
+
+
+int armv7a_l2x_cache_init(struct target *target, uint32_t base, uint32_t way)
+{
+       struct armv7a_l2x_cache *l2x_cache;
+       struct target_list *head = target->head;
+       struct target *curr;
+
+       struct armv7a_common *armv7a = target_to_armv7a(target);
+       if (armv7a == NULL)
+               LOG_ERROR("not an armv7a target");
+       l2x_cache = calloc(1, sizeof(struct armv7a_l2x_cache));
+       l2x_cache->base = base;
+       l2x_cache->way = way;
+       /*LOG_INFO("cache l2 initialized base %x  way %d",
+       l2x_cache->base,l2x_cache->way);*/
+       if (armv7a->armv7a_mmu.armv7a_cache.l2_cache)
+       {
+               LOG_INFO("cache l2 already initialized\n");
+       }
+       armv7a->armv7a_mmu.armv7a_cache.l2_cache = (void*) l2x_cache;
+       /*  initialize l1 / l2x cache function  */
+       armv7a->armv7a_mmu.armv7a_cache.flush_all_data_cache
+               = armv7a_l2x_flush_all_data;
+       armv7a->armv7a_mmu.armv7a_cache.display_cache_info =
+               armv7a_handle_l2x_cache_info_command;
+       /*  initialize all target in this cluster (smp target)*/
+       /*  l2 cache must be configured after smp declaration */
+       while(head != (struct target_list*)NULL)
+       {
+               curr = head->target;
+               if (curr != target)
+               {
+                       armv7a = target_to_armv7a(curr);
+                       if (armv7a->armv7a_mmu.armv7a_cache.l2_cache)
+                       {
+                               LOG_ERROR("smp target : cache l2 already initialized\n");
+                       }
+                       armv7a->armv7a_mmu.armv7a_cache.l2_cache = (void*) l2x_cache;
+                       armv7a->armv7a_mmu.armv7a_cache.flush_all_data_cache =
+                               armv7a_l2x_flush_all_data;
+                       armv7a->armv7a_mmu.armv7a_cache.display_cache_info =
+                               armv7a_handle_l2x_cache_info_command;
+               }
+               head = head -> next;
+       }
+       return JIM_OK;
+}
+
+COMMAND_HANDLER(handle_cache_l2x)
+{
+       struct target *target = get_current_target(CMD_CTX);
+    uint32_t base, way;
+switch (CMD_ARGC) {
+       case 0:
+               return ERROR_COMMAND_SYNTAX_ERROR;
+               break;
+       case 2:
+               //command_print(CMD_CTX, "%s %s", CMD_ARGV[0], CMD_ARGV[1]);
+
+
+               COMMAND_PARSE_NUMBER(u32, CMD_ARGV[0], base);
+               COMMAND_PARSE_NUMBER(u32, CMD_ARGV[1], way);
+
+               /* AP address is in bits 31:24 of DP_SELECT */
+                armv7a_l2x_cache_init(target, base, way);
+               break;
+       default:
+               return ERROR_COMMAND_SYNTAX_ERROR;
+       }
+return ERROR_OK;
+}
+
+
+int armv7a_handle_cache_info_command(struct command_context *cmd_ctx,
+               struct armv7a_cache_common *armv7a_cache)
+{
+       if (armv7a_cache->ctype == -1)
+       {
+               command_print(cmd_ctx, "cache not yet identified");
+               return ERROR_OK;
+       }
+
+       if (armv7a_cache->display_cache_info)
+               armv7a_cache->display_cache_info(cmd_ctx, armv7a_cache);
+       return ERROR_OK;
+}
+
+
+/*  retrieve core id cluster id  */
+int arnv7a_read_mpidr(struct target *target)
+{
+    int retval = ERROR_FAIL;
+       struct armv7a_common *armv7a = target_to_armv7a(target);
+       struct arm_dpm *dpm = armv7a->armv4_5_common.dpm;
+    uint32_t mpidr;
+       retval = dpm->prepare(dpm);
+       if (retval!=ERROR_OK) goto done;
+       /* MRC p15,0,<Rd>,c0,c0,5; read Multiprocessor ID register*/
+
+       retval = dpm->instr_read_data_r0(dpm,
+                       ARMV4_5_MRC(15, 0, 0, 0, 0, 5),
+                       &mpidr);
+       if (retval!=ERROR_OK) goto done;
+    if (mpidr & 1<<31)
+       {
+               armv7a->multi_processor_system = (mpidr >> 30) & 1;
+               armv7a->cluster_id = (mpidr >> 8) & 0xf;
+               armv7a->cpu_id = mpidr & 0x3;
+               LOG_INFO("%s cluster %x core %x %s", target->cmd_name,
+                               armv7a->cluster_id,
+                               armv7a->cpu_id,
+                               armv7a->multi_processor_system == 0 ? "multi core": "mono core");
+
+       }
+       else
+       LOG_ERROR("mpdir not in multiprocessor format");
+
+done:
+       dpm->finish(dpm);
+       return retval;
+
+
+}
+
+
+int armv7a_identify_cache(struct target *target)
+{
+       /*  read cache descriptor */
+       int retval = ERROR_FAIL;
+       struct armv7a_common *armv7a = target_to_armv7a(target);
+       struct arm_dpm *dpm = armv7a->armv4_5_common.dpm;
+       uint32_t cache_selected,clidr;
+       uint32_t cache_i_reg, cache_d_reg;
+       struct armv7a_cache_common *cache = &(armv7a->armv7a_mmu.armv7a_cache);
+       armv7a_read_ttbcr(target);
+       retval = dpm->prepare(dpm);
+
+       if (retval!=ERROR_OK) goto done;
+       /*  retrieve CLIDR */
+       /*  mrc p15, 1, r0, c0, c0, 1           @ read clidr */
+       retval = dpm->instr_read_data_r0(dpm,
+                       ARMV4_5_MRC(15, 1, 0, 0, 0, 1),
+                       &clidr);
+       if (retval!=ERROR_OK) goto done;
+       clidr = (clidr & 0x7000000) >> 23;
+       LOG_INFO("number of cache level %d",clidr /2 );
+       if ((clidr /2) > 1)
+       {
+               // FIXME not supported present in cortex A8 and later
+               //  in cortex A7, A15
+               LOG_ERROR("cache l2 present :not supported");
+       }
+       /*  retrieve selected cache  */
+       /*  MRC p15, 2,<Rd>, c0, c0, 0; Read CSSELR */
+       retval = dpm->instr_read_data_r0(dpm,
+                       ARMV4_5_MRC(15, 2, 0, 0, 0, 0),
+                       &cache_selected);
+       if (retval!=ERROR_OK) goto done;
+
+       retval = armv7a->armv4_5_common.mrc(target, 15,
+                       2, 0,   /* op1, op2 */
+                       0, 0,   /* CRn, CRm */
+                       &cache_selected);
+       /* select instruction cache*/
+       /*  MCR p15, 2,<Rd>, c0, c0, 0; Write CSSELR */
+       /*  [0]  : 1 instruction cache selection , 0 data cache selection */
+       retval = dpm->instr_write_data_r0(dpm,
+                       ARMV4_5_MRC(15, 2, 0, 0, 0, 0),
+                       1);
+       if (retval!=ERROR_OK) goto done;
+
+       /* read CCSIDR*/
+       /* MRC P15,1,<RT>,C0, C0,0 ;on cortex A9 read CCSIDR */
+       /* [2:0] line size  001 eight word per line */
+       /* [27:13] NumSet 0x7f 16KB, 0xff 32Kbytes, 0x1ff 64Kbytes */
+       retval = dpm->instr_read_data_r0(dpm,
+                       ARMV4_5_MRC(15, 1, 0, 0, 0, 0),
+                       &cache_i_reg);
+       if (retval!=ERROR_OK) goto done;
+
+       /*  select data cache*/
+       retval = dpm->instr_write_data_r0(dpm,
+                       ARMV4_5_MRC(15, 2, 0, 0, 0, 0),
+                       0);
+       if (retval!=ERROR_OK) goto done;
+
+       retval = dpm->instr_read_data_r0(dpm,
+                       ARMV4_5_MRC(15, 1, 0, 0, 0, 0),
+                       &cache_d_reg);
+       if (retval!=ERROR_OK) goto done;
+
+       /*  restore selected cache  */
+       dpm->instr_write_data_r0(dpm,
+                       ARMV4_5_MRC(15, 2, 0, 0, 0, 0),
+                       cache_selected);
+
+       if (retval != ERROR_OK) goto done;
+       dpm->finish(dpm);
+
+       // put fake type
+       cache->d_u_size.linelen = 16 << (cache_d_reg & 0x7);
+       cache->d_u_size.cachesize = (((cache_d_reg >> 13) & 0x7fff)+1)/8;
+       cache->d_u_size.nsets = (cache_d_reg >> 13) & 0x7fff;
+       cache->d_u_size.associativity = ((cache_d_reg >> 3) & 0x3ff) +1;
+       /*  compute info for set way operation on cache */
+       cache->d_u_size.index_shift = (cache_d_reg & 0x7) + 4;
+    cache->d_u_size.index = (cache_d_reg >> 13) & 0x7fff;
+    cache->d_u_size.way = ((cache_d_reg >> 3) & 0x3ff);
+       cache->d_u_size.way_shift = cache->d_u_size.way+1;
+       {
+               int i=0;
+               while(((cache->d_u_size.way_shift >> i) & 1)!=1) i++;
+               cache->d_u_size.way_shift = 32-i;
+       }
+    /*LOG_INFO("data cache index %d << %d, way %d << %d",
+                       cache->d_u_size.index, cache->d_u_size.index_shift,
+                cache->d_u_size.way,   cache->d_u_size.way_shift);
+
+       LOG_INFO("data cache %d bytes %d KBytes asso %d ways",
+                       cache->d_u_size.linelen,
+                       cache->d_u_size.cachesize,
+                       cache->d_u_size.associativity
+                       );*/
+       cache->i_size.linelen = 16 << (cache_i_reg & 0x7);
+       cache->i_size.associativity = ((cache_i_reg >> 3) & 0x3ff) +1;
+       cache->i_size.nsets = (cache_i_reg >> 13) & 0x7fff;
+       cache->i_size.cachesize = (((cache_i_reg >> 13) & 0x7fff)+1)/8;
+       /*  compute info for set way operation on cache */
+       cache->i_size.index_shift = (cache_i_reg & 0x7) + 4;
+    cache->i_size.index = (cache_i_reg >> 13) & 0x7fff;
+    cache->i_size.way = ((cache_i_reg >> 3) & 0x3ff);
+       cache->i_size.way_shift = cache->i_size.way+1;
+       {
+               int i=0;
+               while(((cache->i_size.way_shift >> i) & 1)!=1) i++;
+               cache->i_size.way_shift = 32-i;
+       }
+    /*LOG_INFO("instruction cache index %d << %d, way %d << %d",
+                       cache->i_size.index, cache->i_size.index_shift,
+                cache->i_size.way, cache->i_size.way_shift);
+
+       LOG_INFO("instruction cache %d bytes %d KBytes asso %d ways",
+                       cache->i_size.linelen,
+                       cache->i_size.cachesize,
+                       cache->i_size.associativity
+                       );*/
+       /*  if no l2 cache initialize l1 data cache flush function function */
+       if (armv7a->armv7a_mmu.armv7a_cache.flush_all_data_cache == NULL)
+       {
+               armv7a->armv7a_mmu.armv7a_cache.display_cache_info =
+                       armv7a_handle_inner_cache_info_command;
+               armv7a->armv7a_mmu.armv7a_cache.flush_all_data_cache =
+                       armv7a_flush_all_data;
+       }
+               armv7a->armv7a_mmu.armv7a_cache.ctype = 0;
+
+done:
+       dpm->finish(dpm);
+        arnv7a_read_mpidr(target);
+       return retval;
+
+}
+
+
+
+int armv7a_init_arch_info(struct target *target, struct armv7a_common *armv7a)
+{
+       struct armv7a_common *again;
+       struct arm *armv4_5 = &armv7a->armv4_5_common;
+    armv4_5->arch_info = armv7a;
+       target->arch_info = &armv7a->armv4_5_common;
+       /*  target is useful in all function arm v4 5 compatible */
+       armv7a->armv4_5_common.target = target;
+       armv7a->armv4_5_common.common_magic =  ARM_COMMON_MAGIC;
+       armv7a->common_magic = ARMV7_COMMON_MAGIC;
+       armv7a->armv7a_mmu.armv7a_cache.l2_cache = NULL;
+       armv7a->armv7a_mmu.armv7a_cache.ctype = -1;
+       armv7a->armv7a_mmu.armv7a_cache.flush_all_data_cache = NULL;
+    armv7a->armv7a_mmu.armv7a_cache.display_cache_info = NULL;
+    again =target_to_armv7a(target);
+       return ERROR_OK;
+}
+
 int armv7a_arch_state(struct target *target)
 {
        static const char *state[] =
@@ -103,9 +761,9 @@ int armv7a_arch_state(struct target *target)
        arm_arch_state(target);
 
        LOG_USER("MMU: %s, D-Cache: %s, I-Cache: %s",
-                state[armv7a->armv4_5_mmu.mmu_enabled],
-                state[armv7a->armv4_5_mmu.armv4_5_cache.d_u_cache_enabled],
-                state[armv7a->armv4_5_mmu.armv4_5_cache.i_cache_enabled]);
+                state[armv7a->armv7a_mmu.mmu_enabled],
+                state[armv7a->armv7a_mmu.armv7a_cache.d_u_cache_enabled],
+                state[armv7a->armv7a_mmu.armv7a_cache.i_cache_enabled]);
 
        if (armv4_5->core_mode == ARM_MODE_ABT)
                armv7a_show_fault_registers(target);
@@ -116,11 +774,37 @@ int armv7a_arch_state(struct target *target)
        return ERROR_OK;
 }
 
+static const struct command_registration l2_cache_commands[] = {
+       {
+               .name = "l2x",
+               .handler = handle_cache_l2x,
+               .mode = COMMAND_EXEC,
+               .help = "configure l2x cache "
+                       "",
+               .usage = "[base_addr] [number_of_way]",
+       },
+               COMMAND_REGISTRATION_DONE
+
+};
+
+const struct command_registration l2x_cache_command_handlers[] = {
+       {
+               .name = "cache_config",
+               .mode = COMMAND_EXEC,
+               .help = "cache configuation for a target",
+               .chain = l2_cache_commands,
+       },
+       COMMAND_REGISTRATION_DONE
+};
+
 
 const struct command_registration armv7a_command_handlers[] = {
        {
                .chain = dap_command_handlers,
        },
+       {
+               .chain = l2x_cache_command_handlers,
+       },
        COMMAND_REGISTRATION_DONE
 };
 
index f93245634adb3bd147f9f076b4c1cfa1f779af06..dde1f23e0f97bb5cd48bb2ac65a1ef602cecf7a5 100644 (file)
@@ -43,6 +43,56 @@ enum
 #define V2POWPW  5
 #define V2POWUR  6
 #define V2POWUW  7
+/*   L210/L220 cache controller support */
+struct armv7a_l2x_cache {
+       uint32_t base;
+       uint32_t way;
+};
+
+struct armv7a_cachesize
+{
+       uint32_t level_num;
+       /*  cache dimensionning */
+       uint32_t linelen;
+       uint32_t associativity;
+       uint32_t nsets;
+       uint32_t cachesize;
+       /* info for set way operation on cache */
+       uint32_t index;
+       uint32_t index_shift;
+       uint32_t way;
+       uint32_t way_shift;
+};
+
+
+struct armv7a_cache_common
+{
+       int ctype;
+       struct armv7a_cachesize d_u_size;       /* data cache */
+       struct armv7a_cachesize i_size;     /* instruction cache */
+       int i_cache_enabled;
+       int d_u_cache_enabled;
+       /* l2 external unified cache if some */
+       void *l2_cache;
+       int  (*flush_all_data_cache)(struct target *target);
+       int  (*display_cache_info)(struct command_context *cmd_ctx,
+                       struct armv7a_cache_common *armv7a_cache);
+};
+
+
+struct armv7a_mmu_common
+{
+       /*  following field mmu working way */
+       int32_t ttbr1_used; /*  -1 not initialized, 0 no ttbr1 1 ttbr1 used and  */
+       uint32_t ttbr0_mask;/*  masked to be used  */
+       uint32_t os_border;
+
+       int (*read_physical_memory)(struct target *target, uint32_t address, uint32_t size, uint32_t count, uint8_t *buffer);
+       struct armv7a_cache_common armv7a_cache;
+       uint32_t mmu_enabled;
+};
+
+
 
 struct armv7a_common
 {
@@ -57,9 +107,13 @@ struct armv7a_common
        uint32_t debug_base;
        uint8_t debug_ap;
        uint8_t memory_ap;
+       /* mdir */
+       uint8_t multi_processor_system;
+       uint8_t cluster_id;
+       uint8_t cpu_id;
 
-       /* Cache and Memory Management Unit */
-       struct armv4_5_mmu_common armv4_5_mmu;
+       /* cache specific to V7 Memory Management Unit compatible with v4_5*/
+       struct armv7a_mmu_common armv7a_mmu;
 
        int (*examine_debug_reason)(struct target *target);
        int (*post_debug_entry)(struct target *target);
@@ -112,9 +166,16 @@ target_to_armv7a(struct target *target)
 #define CPUDBG_AUTHSTATUS      0xFB8
 
 int armv7a_arch_state(struct target *target);
+int armv7a_identify_cache(struct target *target);
 struct reg_cache *armv7a_build_reg_cache(struct target *target,
                struct armv7a_common *armv7a_common);
 int armv7a_init_arch_info(struct target *target, struct armv7a_common *armv7a);
+int armv7a_mmu_translate_va_pa(struct target *target, uint32_t va,
+               uint32_t *val,int meminfo);
+int armv7a_mmu_translate_va(struct target *target,  uint32_t va, uint32_t *val);
+
+int armv7a_handle_cache_info_command(struct command_context *cmd_ctx,
+               struct armv7a_cache_common *armv7a_cache);
 
 extern const struct command_registration armv7a_command_handlers[];
 
index 39f1b9e1ae541eb2004faceff31010e268389ff3..7547f17f6328b97fe2e357864d5a7c0d89b9502e 100755 (executable)
@@ -66,12 +66,6 @@ static int cortex_a8_dap_write_coreregister_u32(struct target *target,
 static int cortex_a8_mmu(struct target *target, int *enabled);
 static int cortex_a8_virt2phys(struct target *target,
                 uint32_t virt, uint32_t *phys);
-static int cortex_a8_disable_mmu_caches(struct target *target, int mmu,
-                int d_u_cache, int i_cache);
-static int cortex_a8_enable_mmu_caches(struct target *target, int mmu,
-                int d_u_cache, int i_cache);
-static int cortex_a8_get_ttb(struct target *target, uint32_t *result);
-
 
 /*
  * FIXME do topology discovery using the ROM; don't
@@ -82,6 +76,99 @@ static int cortex_a8_get_ttb(struct target *target, uint32_t *result);
 #define swjdp_memoryap 0
 #define swjdp_debugap 1
 
+/*  restore cp15_control_reg at resume */
+static int cortex_a8_restore_cp15_control_reg(struct target* target)
+{
+       int retval = ERROR_OK;
+       struct cortex_a8_common *cortex_a8 = target_to_cortex_a8(target);
+    struct armv7a_common *armv7a = target_to_armv7a(target);
+
+       if (cortex_a8->cp15_control_reg !=cortex_a8->cp15_control_reg_curr)
+       {
+               cortex_a8->cp15_control_reg_curr = cortex_a8->cp15_control_reg;
+               //LOG_INFO("cp15_control_reg: %8.8" PRIx32, cortex_a8->cp15_control_reg);
+               retval = armv7a->armv4_5_common.mcr(target, 15,
+                               0, 0,   /* op1, op2 */
+                               1, 0,   /* CRn, CRm */
+                               cortex_a8->cp15_control_reg);
+       }
+       return ERROR_OK;
+}
+
+/*  check address before cortex_a8_apb read write access with mmu on
+ *  remove apb predictible data abort */
+static int cortex_a8_check_address(struct target *target, uint32_t address)
+{
+       struct armv7a_common *armv7a = target_to_armv7a(target);
+       struct cortex_a8_common *cortex_a8 = target_to_cortex_a8(target);
+       uint32_t os_border = armv7a->armv7a_mmu.os_border;
+       if ((address < os_border) &&
+                       (armv7a->armv4_5_common.core_mode == ARM_MODE_SVC)){
+               LOG_ERROR("%x access in userspace and target in supervisor",address);
+               return ERROR_FAIL;
+       }
+       if ((address >= os_border) &&
+                       ( cortex_a8->curr_mode != ARM_MODE_SVC)){
+        dpm_modeswitch(&armv7a->dpm, ARM_MODE_SVC);
+        cortex_a8->curr_mode = ARM_MODE_SVC;
+               LOG_INFO("%x access in kernel space and target not in supervisor",
+                               address);
+               return ERROR_OK;
+       }
+       if ((address < os_border) &&
+                       (cortex_a8->curr_mode == ARM_MODE_SVC)){
+        dpm_modeswitch(&armv7a->dpm, ARM_MODE_ANY);
+     cortex_a8->curr_mode = ARM_MODE_ANY;
+       }
+       return ERROR_OK;
+}
+/*  modify cp15_control_reg in order to enable or disable mmu for :
+ *  - virt2phys address conversion
+ *  - read or write memory in phys or virt address */
+static int cortex_a8_mmu_modify(struct target *target, int enable)
+{
+       struct cortex_a8_common *cortex_a8 = target_to_cortex_a8(target);
+       struct armv7a_common *armv7a = target_to_armv7a(target);
+       int retval = ERROR_OK;
+       if (enable)
+       {
+               /*  if mmu enabled at target stop and mmu not enable */
+               if (!(cortex_a8->cp15_control_reg & 0x1U))
+               {
+                       LOG_ERROR("trying to enable mmu on target stopped with mmu disable");
+                       return ERROR_FAIL;
+               }
+               if (!(cortex_a8->cp15_control_reg_curr & 0x1U))
+               {
+                       cortex_a8->cp15_control_reg_curr |= 0x1U;
+                       retval = armv7a->armv4_5_common.mcr(target, 15,
+                                       0, 0,   /* op1, op2 */
+                                       1, 0,   /* CRn, CRm */
+                                       cortex_a8->cp15_control_reg_curr);
+               }
+       }
+       else
+       {
+               if (cortex_a8->cp15_control_reg_curr & 0x4U)
+               {
+                   /*  data cache is active */
+                       cortex_a8->cp15_control_reg_curr &= ~0x4U;
+                       /* flush data cache armv7 function to be called */
+                       if (armv7a->armv7a_mmu.armv7a_cache.flush_all_data_cache)
+                       armv7a->armv7a_mmu.armv7a_cache.flush_all_data_cache(target);
+               }
+               if ( (cortex_a8->cp15_control_reg_curr & 0x1U))
+               {
+                       cortex_a8->cp15_control_reg_curr &= ~0x1U;
+                       retval = armv7a->armv4_5_common.mcr(target, 15,
+                                       0, 0,   /* op1, op2 */
+                                       1, 0,   /* CRn, CRm */
+                                       cortex_a8->cp15_control_reg_curr);
+               }
+       }
+       return retval;
+}
+
 /*
  * Cortex-A8 Basic debug access, very low level assumes state is saved
  */
@@ -929,7 +1016,11 @@ static int cortex_a8_internal_restore(struct target *target, int current,
        buf_set_u32(armv4_5->pc->value, 0, 32, resume_pc);
        armv4_5->pc->dirty = 1;
        armv4_5->pc->valid = 1;
-
+       /* restore dpm_mode at system halt */
+    dpm_modeswitch(&armv7a->dpm, ARM_MODE_ANY);
+    /* called it now before restoring context because it uses cpu
+        * register r0 for restoring cp15 control register */
+       retval = cortex_a8_restore_cp15_control_reg(target);
        retval = cortex_a8_restore_context(target, handle_breakpoints);
        if (retval != ERROR_OK)
                return retval;
@@ -1147,6 +1238,7 @@ static int cortex_a8_debug_entry(struct target *target)
 
                /* read Current PSR */
                retval = cortex_a8_dap_read_coreregister_u32(target, &cpsr, 16);
+               /*  store current cpsr */
                if (retval != ERROR_OK)
                        return retval;
 
@@ -1220,32 +1312,21 @@ static int cortex_a8_post_debug_entry(struct target *target)
        if (retval != ERROR_OK)
                return retval;
        LOG_DEBUG("cp15_control_reg: %8.8" PRIx32, cortex_a8->cp15_control_reg);
+    cortex_a8->cp15_control_reg_curr = cortex_a8->cp15_control_reg;
 
-       if (armv7a->armv4_5_mmu.armv4_5_cache.ctype == -1)
+       if (armv7a->armv7a_mmu.armv7a_cache.ctype == -1)
        {
-               uint32_t cache_type_reg;
-
-               /* MRC p15,0,<Rt>,c0,c0,1 ; Read CP15 Cache Type Register */
-               retval = armv7a->armv4_5_common.mrc(target, 15,
-                               0, 1,   /* op1, op2 */
-                               0, 0,   /* CRn, CRm */
-                               &cache_type_reg);
-               if (retval != ERROR_OK)
-                       return retval;
-               LOG_DEBUG("cp15 cache type: %8.8x", (unsigned) cache_type_reg);
-
-               /* FIXME the armv4_4 cache info DOES NOT APPLY to Cortex-A8 */
-               armv4_5_identify_cache(cache_type_reg,
-                               &armv7a->armv4_5_mmu.armv4_5_cache);
+               armv7a_identify_cache(target);
        }
 
-       armv7a->armv4_5_mmu.mmu_enabled =
+       armv7a->armv7a_mmu.mmu_enabled =
                        (cortex_a8->cp15_control_reg & 0x1U) ? 1 : 0;
-       armv7a->armv4_5_mmu.armv4_5_cache.d_u_cache_enabled =
+       armv7a->armv7a_mmu.armv7a_cache.d_u_cache_enabled =
                        (cortex_a8->cp15_control_reg & 0x4U) ? 1 : 0;
-       armv7a->armv4_5_mmu.armv4_5_cache.i_cache_enabled =
+       armv7a->armv7a_mmu.armv7a_cache.i_cache_enabled =
                        (cortex_a8->cp15_control_reg & 0x1000U) ? 1 : 0;
-
+    cortex_a8->curr_mode = armv7a->armv4_5_common.core_mode;
+    
        return ERROR_OK;
 }
 
@@ -1990,18 +2071,9 @@ static int cortex_a8_read_phys_memory(struct target *target,
                } else {
 
                        /* read memory through APB-AP */
-                       int enabled = 0;
-
-                       retval = cortex_a8_mmu(target, &enabled);
-                       if (retval != ERROR_OK)
-                               return retval;
-
-                       if (enabled)
-                       {
-                               LOG_WARNING("Reading physical memory through \
-                                               APB with MMU enabled is not yet implemented");
-                               return ERROR_TARGET_FAILURE;
-                       }
+                       /*  disable mmu */
+                       retval = cortex_a8_mmu_modify(target, 0);
+            if (retval != ERROR_OK) return retval;
                        retval =  cortex_a8_read_apb_ab_memory(target, address, size, count, buffer);
                }
        }
@@ -2040,6 +2112,11 @@ static int cortex_a8_read_memory(struct target *target, uint32_t address,
                }
                retval = cortex_a8_read_phys_memory(target, address, size, count, buffer);
        } else {
+               retval = cortex_a8_check_address(target, address);
+        if (retval != ERROR_OK) return retval;
+               /*  enable mmu */
+               retval = cortex_a8_mmu_modify(target, 1);
+           if (retval != ERROR_OK) return retval;
                retval = cortex_a8_read_apb_ab_memory(target, address, size, count, buffer);
        }
        return retval;
@@ -2081,19 +2158,10 @@ static int cortex_a8_write_phys_memory(struct target *target,
                } else {
 
                        /* write memory through APB-AP */
-                       int enabled = 0;
-
-                       retval = cortex_a8_mmu(target, &enabled);
+                       retval = cortex_a8_mmu_modify(target, 0);
                        if (retval != ERROR_OK)
                                return retval;
-
-                       if (enabled)
-                       {
-                               LOG_WARNING("Writing physical memory through APB with MMU" \
-                                               "enabled is not yet implemented");
-                               return ERROR_TARGET_FAILURE;
-                       }
-                       return cortex_a8_write_apb_ab_memory(target, address, size, count, buffer);
+                       return  cortex_a8_write_apb_ab_memory(target, address, size, count, buffer);
                }
        }
 
@@ -2117,7 +2185,7 @@ static int cortex_a8_write_phys_memory(struct target *target,
                 */
 
                /* invalidate I-Cache */
-               if (armv7a->armv4_5_mmu.armv4_5_cache.i_cache_enabled)
+               if (armv7a->armv7a_mmu.armv7a_cache.i_cache_enabled)
                {
                        /* ICIMVAU - Invalidate Cache single entry
                         * with MVA to PoU
@@ -2135,7 +2203,7 @@ static int cortex_a8_write_phys_memory(struct target *target,
                }
 
                /* invalidate D-Cache */
-               if (armv7a->armv4_5_mmu.armv4_5_cache.d_u_cache_enabled)
+               if (armv7a->armv7a_mmu.armv7a_cache.d_u_cache_enabled)
                {
                        /* DCIMVAC - Invalidate data Cache line
                         * with MVA to PoC
@@ -2191,6 +2259,11 @@ static int cortex_a8_write_memory(struct target *target, uint32_t address,
                                count, buffer);
        }
        else {
+               retval = cortex_a8_check_address(target, address);
+               if (retval != ERROR_OK) return retval;
+               /*  enable mmu  */
+               retval = cortex_a8_mmu_modify(target, 1);
+               if (retval != ERROR_OK) return retval;
                retval = cortex_a8_write_apb_ab_memory(target, address, size, count, buffer);
        }
     return retval;
@@ -2375,7 +2448,6 @@ static int cortex_a8_init_arch_info(struct target *target,
                struct cortex_a8_common *cortex_a8, struct jtag_tap *tap)
 {
        struct armv7a_common *armv7a = &cortex_a8->armv7a_common;
-       struct arm *armv4_5 = &armv7a->armv4_5_common;
        struct adiv5_dap *dap = &armv7a->dap;
 
        armv7a->armv4_5_common.dap = dap;
@@ -2387,7 +2459,6 @@ static int cortex_a8_init_arch_info(struct target *target,
        {
        armv7a->armv4_5_common.dap = dap;
        /* Setup struct cortex_a8_common */
-       armv4_5->arch_info = armv7a;
 
        /* prepare JTAG information for the new target */
        cortex_a8->jtag_info.tap = tap;
@@ -2406,31 +2477,20 @@ static int cortex_a8_init_arch_info(struct target *target,
 
        cortex_a8->fast_reg_read = 0;
 
-       /* Set default value */
-       cortex_a8->current_address_mode = ARM_MODE_ANY;
-
        /* register arch-specific functions */
        armv7a->examine_debug_reason = NULL;
 
        armv7a->post_debug_entry = cortex_a8_post_debug_entry;
 
        armv7a->pre_restore_context = NULL;
-       armv7a->armv4_5_mmu.armv4_5_cache.ctype = -1;
-       armv7a->armv4_5_mmu.get_ttb = cortex_a8_get_ttb;
-       armv7a->armv4_5_mmu.read_memory = cortex_a8_read_phys_memory;
-       armv7a->armv4_5_mmu.write_memory = cortex_a8_write_phys_memory;
-       armv7a->armv4_5_mmu.disable_mmu_caches = cortex_a8_disable_mmu_caches;
-       armv7a->armv4_5_mmu.enable_mmu_caches = cortex_a8_enable_mmu_caches;
-       armv7a->armv4_5_mmu.has_tiny_pages = 1;
-       armv7a->armv4_5_mmu.mmu_enabled = 0;
 
+       armv7a->armv7a_mmu.read_physical_memory = cortex_a8_read_phys_memory;
 
+       
 //     arm7_9->handle_target_request = cortex_a8_handle_target_request;
 
        /* REVISIT v7a setup should be in a v7a-specific routine */
-       arm_init_arch_info(target, armv4_5);
-       armv7a->common_magic = ARMV7_COMMON_MAGIC;
-
+       armv7a_init_arch_info(target, armv7a);
        target_register_timer_callback(cortex_a8_handle_target_request, 1, 1, target);
 
        return ERROR_OK;
@@ -2443,133 +2503,6 @@ static int cortex_a8_target_create(struct target *target, Jim_Interp *interp)
        return cortex_a8_init_arch_info(target, cortex_a8, target->tap);
 }
 
-static int cortex_a8_get_ttb(struct target *target, uint32_t *result)
-{
-       struct cortex_a8_common *cortex_a8 = target_to_cortex_a8(target);
-    struct armv7a_common *armv7a = &cortex_a8->armv7a_common;
-    uint32_t ttb = 0, retval = ERROR_OK;
-
-    /* current_address_mode is set inside cortex_a8_virt2phys()
-       where we can determine if address belongs to user or kernel */
-    if(cortex_a8->current_address_mode == ARM_MODE_SVC)
-    {
-        /* MRC p15,0,<Rt>,c1,c0,0 ; Read CP15 System Control Register */
-        retval = armv7a->armv4_5_common.mrc(target, 15,
-                    0, 1,   /* op1, op2 */
-                    2, 0,   /* CRn, CRm */
-                    &ttb);
-               if (retval != ERROR_OK)
-                       return retval;
-    }
-    else if(cortex_a8->current_address_mode == ARM_MODE_USR)
-    {
-        /* MRC p15,0,<Rt>,c1,c0,0 ; Read CP15 System Control Register */
-        retval = armv7a->armv4_5_common.mrc(target, 15,
-                    0, 0,   /* op1, op2 */
-                    2, 0,   /* CRn, CRm */
-                    &ttb);
-               if (retval != ERROR_OK)
-                       return retval;
-    }
-    /* we don't know whose address is: user or kernel
-       we assume that if we are in kernel mode then
-       address belongs to kernel else if in user mode
-       - to user */
-    else if(armv7a->armv4_5_common.core_mode == ARM_MODE_SVC)
-    {
-        /* MRC p15,0,<Rt>,c1,c0,0 ; Read CP15 System Control Register */
-        retval = armv7a->armv4_5_common.mrc(target, 15,
-                    0, 1,   /* op1, op2 */
-                    2, 0,   /* CRn, CRm */
-                    &ttb);
-               if (retval != ERROR_OK)
-                       return retval;
-    }
-    else if(armv7a->armv4_5_common.core_mode == ARM_MODE_USR)
-    {
-        /* MRC p15,0,<Rt>,c1,c0,0 ; Read CP15 System Control Register */
-        retval = armv7a->armv4_5_common.mrc(target, 15,
-                    0, 0,   /* op1, op2 */
-                    2, 0,   /* CRn, CRm */
-                    &ttb);
-               if (retval != ERROR_OK)
-                       return retval;
-    }
-    /* finally we don't know whose ttb to use: user or kernel */
-    else
-        LOG_ERROR("Don't know how to get ttb for current mode!!!");
-
-    ttb &= 0xffffc000;
-
-    *result = ttb;
-
-    return ERROR_OK;
-}
-
-static int cortex_a8_disable_mmu_caches(struct target *target, int mmu,
-                int d_u_cache, int i_cache)
-{
-    struct cortex_a8_common *cortex_a8 = target_to_cortex_a8(target);
-    struct armv7a_common *armv7a = &cortex_a8->armv7a_common;
-    uint32_t cp15_control;
-    int retval;
-
-    /* read cp15 control register */
-    retval = armv7a->armv4_5_common.mrc(target, 15,
-                    0, 0,   /* op1, op2 */
-                    1, 0,   /* CRn, CRm */
-                    &cp15_control);
-    if (retval != ERROR_OK)
-       return retval;
-
-
-    if (mmu)
-            cp15_control &= ~0x1U;
-
-    if (d_u_cache)
-            cp15_control &= ~0x4U;
-
-    if (i_cache)
-            cp15_control &= ~0x1000U;
-
-    retval = armv7a->armv4_5_common.mcr(target, 15,
-                    0, 0,   /* op1, op2 */
-                    1, 0,   /* CRn, CRm */
-                    cp15_control);
-       return retval;
-}
-
-static int cortex_a8_enable_mmu_caches(struct target *target, int mmu,
-                int d_u_cache, int i_cache)
-{
-    struct cortex_a8_common *cortex_a8 = target_to_cortex_a8(target);
-    struct armv7a_common *armv7a = &cortex_a8->armv7a_common;
-    uint32_t cp15_control;
-    int retval;
-
-    /* read cp15 control register */
-    retval = armv7a->armv4_5_common.mrc(target, 15,
-                    0, 0,   /* op1, op2 */
-                    1, 0,   /* CRn, CRm */
-                    &cp15_control);
-    if (retval != ERROR_OK)
-       return retval;
-
-    if (mmu)
-            cp15_control |= 0x1U;
-
-    if (d_u_cache)
-            cp15_control |= 0x4U;
-
-    if (i_cache)
-            cp15_control |= 0x1000U;
-
-    retval = armv7a->armv4_5_common.mcr(target, 15,
-                    0, 0,   /* op1, op2 */
-                    1, 0,   /* CRn, CRm */
-                    cp15_control);
-       return retval;
-}
 
 
 static int cortex_a8_mmu(struct target *target, int *enabled)
@@ -2579,36 +2512,35 @@ static int cortex_a8_mmu(struct target *target, int *enabled)
                return ERROR_TARGET_INVALID;
        }
 
-       *enabled = target_to_cortex_a8(target)->armv7a_common.armv4_5_mmu.mmu_enabled;
+       *enabled = target_to_cortex_a8(target)->armv7a_common.armv7a_mmu.mmu_enabled;
        return ERROR_OK;
 }
 
 static int cortex_a8_virt2phys(struct target *target,
                uint32_t virt, uint32_t *phys)
 {
-       uint32_t cb;
-       struct cortex_a8_common *cortex_a8 = target_to_cortex_a8(target);
-       // struct armv7a_common *armv7a = &cortex_a8->armv7a_common;
+       int retval = ERROR_FAIL;
        struct armv7a_common *armv7a = target_to_armv7a(target);
-
-    /* We assume that virtual address is separated
-       between user and kernel in Linux style:
-       0x00000000-0xbfffffff - User space
-       0xc0000000-0xffffffff - Kernel space */
-    if( virt < 0xc0000000 ) /* Linux user space */
-        cortex_a8->current_address_mode = ARM_MODE_USR;
-    else /* Linux kernel */
-        cortex_a8->current_address_mode = ARM_MODE_SVC;
-       uint32_t ret;
-       int retval = armv4_5_mmu_translate_va(target,
-                       &armv7a->armv4_5_mmu, virt, &cb, &ret);
-       if (retval != ERROR_OK)
-               return retval;
-    /* Reset the flag. We don't want someone else to use it by error */
-    cortex_a8->current_address_mode = ARM_MODE_ANY;
-
-       *phys = ret;
-       return ERROR_OK;
+       struct adiv5_dap *swjdp = armv7a->armv4_5_common.dap;
+       uint8_t apsel = swjdp->apsel;
+       if (apsel == swjdp_memoryap)
+       {
+               uint32_t ret;
+               retval = armv7a_mmu_translate_va(target,
+                                virt, &ret);
+               if (retval != ERROR_OK)
+                       goto done;
+               *phys = ret;
+       } 
+       else
+       { /*  use this method if swjdp_memoryap not selected */
+               /*  mmu must be enable in order to get a correct translation */
+               retval = cortex_a8_mmu_modify(target, 1);
+               if (retval != ERROR_OK) goto done;
+               retval = armv7a_mmu_translate_va_pa(target, virt,  phys, 1);
+       }
+done:
+       return retval;
 }
 
 COMMAND_HANDLER(cortex_a8_handle_cache_info_command)
@@ -2616,8 +2548,8 @@ COMMAND_HANDLER(cortex_a8_handle_cache_info_command)
        struct target *target = get_current_target(CMD_CTX);
        struct armv7a_common *armv7a = target_to_armv7a(target);
 
-       return armv4_5_handle_cache_info_command(CMD_CTX,
-                       &armv7a->armv4_5_mmu.armv4_5_cache);
+       return armv7a_handle_cache_info_command(CMD_CTX,
+                       &armv7a->armv7a_mmu.armv7a_cache);
 }
 
 
@@ -2789,5 +2721,4 @@ struct target_type cortexa8_target = {
        .write_phys_memory = cortex_a8_write_phys_memory,
        .mmu = cortex_a8_mmu,
        .virt2phys = cortex_a8_virt2phys,
-
 };
index b49e670f21eddcae67793f2e5842c9f62c2d2cb9..17e44e210578de9c764d3f9e929c268ee1bed304 100644 (file)
@@ -63,6 +63,10 @@ struct cortex_a8_common
 
        /* Saved cp15 registers */
        uint32_t cp15_control_reg;
+       /* latest cp15 register value written and cpsr processor mode */
+       uint32_t cp15_control_reg_curr;
+    enum arm_mode curr_mode;
+
 
        /* Breakpoint register pairs */
        int brp_num_context;
@@ -73,10 +77,8 @@ struct cortex_a8_common
        /* Use cortex_a8_read_regs_through_mem for fast register reads */
        int fast_reg_read;
 
-       /* Flag that helps to resolve what ttb to use: user or kernel */
-       int current_address_mode;
-
        struct armv7a_common armv7a_common;
+
 };
 
 static inline struct cortex_a8_common *

Linking to existing account procedure

If you already have an account and want to add another login method you MUST first sign in with your existing account and then change URL to read https://review.openocd.org/login/?link to get to this page again but this time it'll work for linking. Thank you.

SSH host keys fingerprints

1024 SHA256:YKx8b7u5ZWdcbp7/4AeXNaqElP49m6QrwfXaqQGJAOk gerrit-code-review@openocd.zylin.com (DSA)
384 SHA256:jHIbSQa4REvwCFG4cq5LBlBLxmxSqelQPem/EXIrxjk gerrit-code-review@openocd.org (ECDSA)
521 SHA256:UAOPYkU9Fjtcao0Ul/Rrlnj/OsQvt+pgdYSZ4jOYdgs gerrit-code-review@openocd.org (ECDSA)
256 SHA256:A13M5QlnozFOvTllybRZH6vm7iSt0XLxbA48yfc2yfY gerrit-code-review@openocd.org (ECDSA)
256 SHA256:spYMBqEYoAOtK7yZBrcwE8ZpYt6b68Cfh9yEVetvbXg gerrit-code-review@openocd.org (ED25519)
+--[ED25519 256]--+
|=..              |
|+o..   .         |
|*.o   . .        |
|+B . . .         |
|Bo. = o S        |
|Oo.+ + =         |
|oB=.* = . o      |
| =+=.+   + E     |
|. .=o   . o      |
+----[SHA256]-----+
2048 SHA256:0Onrb7/PHjpo6iVZ7xQX2riKN83FJ3KGU0TvI0TaFG4 gerrit-code-review@openocd.zylin.com (RSA)