~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/block/blk-mq-cpumap.c

Version: ~ [ linux-6.11.5 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.58 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.114 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.169 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.228 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.284 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.322 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.9 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

Diff markup

Differences between /block/blk-mq-cpumap.c (Version linux-6.11.5) and /block/blk-mq-cpumap.c (Version linux-5.6.19)


  1 // SPDX-License-Identifier: GPL-2.0                 1 // SPDX-License-Identifier: GPL-2.0
  2 /*                                                  2 /*
  3  * CPU <-> hardware queue mapping helpers           3  * CPU <-> hardware queue mapping helpers
  4  *                                                  4  *
  5  * Copyright (C) 2013-2014 Jens Axboe               5  * Copyright (C) 2013-2014 Jens Axboe
  6  */                                                 6  */
  7 #include <linux/kernel.h>                           7 #include <linux/kernel.h>
  8 #include <linux/threads.h>                          8 #include <linux/threads.h>
  9 #include <linux/module.h>                           9 #include <linux/module.h>
 10 #include <linux/mm.h>                              10 #include <linux/mm.h>
 11 #include <linux/smp.h>                             11 #include <linux/smp.h>
 12 #include <linux/cpu.h>                             12 #include <linux/cpu.h>
 13 #include <linux/group_cpus.h>                  << 
 14                                                    13 
                                                   >>  14 #include <linux/blk-mq.h>
 15 #include "blk.h"                                   15 #include "blk.h"
 16 #include "blk-mq.h"                                16 #include "blk-mq.h"
 17                                                    17 
 18 void blk_mq_map_queues(struct blk_mq_queue_map !!  18 static int queue_index(struct blk_mq_queue_map *qmap,
                                                   >>  19                        unsigned int nr_queues, const int q)
 19 {                                                  20 {
 20         const struct cpumask *masks;           !!  21         return qmap->queue_offset + (q % nr_queues);
 21         unsigned int queue, cpu;               !!  22 }
                                                   >>  23 
                                                   >>  24 static int get_first_sibling(unsigned int cpu)
                                                   >>  25 {
                                                   >>  26         unsigned int ret;
                                                   >>  27 
                                                   >>  28         ret = cpumask_first(topology_sibling_cpumask(cpu));
                                                   >>  29         if (ret < nr_cpu_ids)
                                                   >>  30                 return ret;
 22                                                    31 
 23         masks = group_cpus_evenly(qmap->nr_que !!  32         return cpu;
 24         if (!masks) {                          !!  33 }
 25                 for_each_possible_cpu(cpu)     !!  34 
 26                         qmap->mq_map[cpu] = qm !!  35 int blk_mq_map_queues(struct blk_mq_queue_map *qmap)
 27                 return;                        !!  36 {
                                                   >>  37         unsigned int *map = qmap->mq_map;
                                                   >>  38         unsigned int nr_queues = qmap->nr_queues;
                                                   >>  39         unsigned int cpu, first_sibling, q = 0;
                                                   >>  40 
                                                   >>  41         for_each_possible_cpu(cpu)
                                                   >>  42                 map[cpu] = -1;
                                                   >>  43 
                                                   >>  44         /*
                                                   >>  45          * Spread queues among present CPUs first for minimizing
                                                   >>  46          * count of dead queues which are mapped by all un-present CPUs
                                                   >>  47          */
                                                   >>  48         for_each_present_cpu(cpu) {
                                                   >>  49                 if (q >= nr_queues)
                                                   >>  50                         break;
                                                   >>  51                 map[cpu] = queue_index(qmap, nr_queues, q++);
 28         }                                          52         }
 29                                                    53 
 30         for (queue = 0; queue < qmap->nr_queue !!  54         for_each_possible_cpu(cpu) {
 31                 for_each_cpu(cpu, &masks[queue !!  55                 if (map[cpu] != -1)
 32                         qmap->mq_map[cpu] = qm !!  56                         continue;
                                                   >>  57                 /*
                                                   >>  58                  * First do sequential mapping between CPUs and queues.
                                                   >>  59                  * In case we still have CPUs to map, and we have some number of
                                                   >>  60                  * threads per cores then map sibling threads to the same queue
                                                   >>  61                  * for performance optimizations.
                                                   >>  62                  */
                                                   >>  63                 if (q < nr_queues) {
                                                   >>  64                         map[cpu] = queue_index(qmap, nr_queues, q++);
                                                   >>  65                 } else {
                                                   >>  66                         first_sibling = get_first_sibling(cpu);
                                                   >>  67                         if (first_sibling == cpu)
                                                   >>  68                                 map[cpu] = queue_index(qmap, nr_queues, q++);
                                                   >>  69                         else
                                                   >>  70                                 map[cpu] = map[first_sibling];
                                                   >>  71                 }
 33         }                                          72         }
 34         kfree(masks);                          !!  73 
                                                   >>  74         return 0;
 35 }                                                  75 }
 36 EXPORT_SYMBOL_GPL(blk_mq_map_queues);              76 EXPORT_SYMBOL_GPL(blk_mq_map_queues);
 37                                                    77 
 38 /**                                                78 /**
 39  * blk_mq_hw_queue_to_node - Look up the memor     79  * blk_mq_hw_queue_to_node - Look up the memory node for a hardware queue index
 40  * @qmap: CPU to hardware queue map.               80  * @qmap: CPU to hardware queue map.
 41  * @index: hardware queue index.                   81  * @index: hardware queue index.
 42  *                                                 82  *
 43  * We have no quick way of doing reverse looku     83  * We have no quick way of doing reverse lookups. This is only used at
 44  * queue init time, so runtime isn't important     84  * queue init time, so runtime isn't important.
 45  */                                                85  */
 46 int blk_mq_hw_queue_to_node(struct blk_mq_queu     86 int blk_mq_hw_queue_to_node(struct blk_mq_queue_map *qmap, unsigned int index)
 47 {                                                  87 {
 48         int i;                                     88         int i;
 49                                                    89 
 50         for_each_possible_cpu(i) {                 90         for_each_possible_cpu(i) {
 51                 if (index == qmap->mq_map[i])      91                 if (index == qmap->mq_map[i])
 52                         return cpu_to_node(i); !!  92                         return local_memory_node(cpu_to_node(i));
 53         }                                          93         }
 54                                                    94 
 55         return NUMA_NO_NODE;                       95         return NUMA_NO_NODE;
 56 }                                                  96 }
 57                                                    97 

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php