~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/arch/x86/kernel/cpu/mtrr/mtrr.c

Version: ~ [ linux-6.11-rc3 ] ~ [ linux-6.10.4 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.45 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.104 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.164 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.223 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.281 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.319 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.9 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

  1 /*  Generic MTRR (Memory Type Range Register) driver.
  2 
  3     Copyright (C) 1997-2000  Richard Gooch
  4     Copyright (c) 2002       Patrick Mochel
  5 
  6     This library is free software; you can redistribute it and/or
  7     modify it under the terms of the GNU Library General Public
  8     License as published by the Free Software Foundation; either
  9     version 2 of the License, or (at your option) any later version.
 10 
 11     This library is distributed in the hope that it will be useful,
 12     but WITHOUT ANY WARRANTY; without even the implied warranty of
 13     MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 14     Library General Public License for more details.
 15 
 16     You should have received a copy of the GNU Library General Public
 17     License along with this library; if not, write to the Free
 18     Software Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
 19 
 20     Richard Gooch may be reached by email at  rgooch@atnf.csiro.au
 21     The postal address is:
 22       Richard Gooch, c/o ATNF, P. O. Box 76, Epping, N.S.W., 2121, Australia.
 23 
 24     Source: "Pentium Pro Family Developer's Manual, Volume 3:
 25     Operating System Writer's Guide" (Intel document number 242692),
 26     section 11.11.7
 27 
 28     This was cleaned and made readable by Patrick Mochel <mochel@osdl.org>
 29     on 6-7 March 2002.
 30     Source: Intel Architecture Software Developers Manual, Volume 3:
 31     System Programming Guide; Section 9.11. (1997 edition - PPro).
 32 */
 33 
 34 #include <linux/types.h> /* FIXME: kvm_para.h needs this */
 35 
 36 #include <linux/stop_machine.h>
 37 #include <linux/kvm_para.h>
 38 #include <linux/uaccess.h>
 39 #include <linux/export.h>
 40 #include <linux/mutex.h>
 41 #include <linux/init.h>
 42 #include <linux/sort.h>
 43 #include <linux/cpu.h>
 44 #include <linux/pci.h>
 45 #include <linux/smp.h>
 46 #include <linux/syscore_ops.h>
 47 #include <linux/rcupdate.h>
 48 
 49 #include <asm/cacheinfo.h>
 50 #include <asm/cpufeature.h>
 51 #include <asm/e820/api.h>
 52 #include <asm/mtrr.h>
 53 #include <asm/msr.h>
 54 #include <asm/memtype.h>
 55 
 56 #include "mtrr.h"
 57 
 58 /* arch_phys_wc_add returns an MTRR register index plus this offset. */
 59 #define MTRR_TO_PHYS_WC_OFFSET 1000
 60 
 61 u32 num_var_ranges;
 62 
 63 unsigned int mtrr_usage_table[MTRR_MAX_VAR_RANGES];
 64 DEFINE_MUTEX(mtrr_mutex);
 65 
 66 const struct mtrr_ops *mtrr_if;
 67 
 68 /*  Returns non-zero if we have the write-combining memory type  */
 69 static int have_wrcomb(void)
 70 {
 71         struct pci_dev *dev;
 72 
 73         dev = pci_get_class(PCI_CLASS_BRIDGE_HOST << 8, NULL);
 74         if (dev != NULL) {
 75                 /*
 76                  * ServerWorks LE chipsets < rev 6 have problems with
 77                  * write-combining. Don't allow it and leave room for other
 78                  * chipsets to be tagged
 79                  */
 80                 if (dev->vendor == PCI_VENDOR_ID_SERVERWORKS &&
 81                     dev->device == PCI_DEVICE_ID_SERVERWORKS_LE &&
 82                     dev->revision <= 5) {
 83                         pr_info("Serverworks LE rev < 6 detected. Write-combining disabled.\n");
 84                         pci_dev_put(dev);
 85                         return 0;
 86                 }
 87                 /*
 88                  * Intel 450NX errata # 23. Non ascending cacheline evictions to
 89                  * write combining memory may resulting in data corruption
 90                  */
 91                 if (dev->vendor == PCI_VENDOR_ID_INTEL &&
 92                     dev->device == PCI_DEVICE_ID_INTEL_82451NX) {
 93                         pr_info("Intel 450NX MMC detected. Write-combining disabled.\n");
 94                         pci_dev_put(dev);
 95                         return 0;
 96                 }
 97                 pci_dev_put(dev);
 98         }
 99         return mtrr_if->have_wrcomb ? mtrr_if->have_wrcomb() : 0;
100 }
101 
102 static void __init init_table(void)
103 {
104         int i, max;
105 
106         max = num_var_ranges;
107         for (i = 0; i < max; i++)
108                 mtrr_usage_table[i] = 1;
109 }
110 
111 struct set_mtrr_data {
112         unsigned long   smp_base;
113         unsigned long   smp_size;
114         unsigned int    smp_reg;
115         mtrr_type       smp_type;
116 };
117 
118 /**
119  * mtrr_rendezvous_handler - Work done in the synchronization handler. Executed
120  * by all the CPUs.
121  * @info: pointer to mtrr configuration data
122  *
123  * Returns nothing.
124  */
125 static int mtrr_rendezvous_handler(void *info)
126 {
127         struct set_mtrr_data *data = info;
128 
129         mtrr_if->set(data->smp_reg, data->smp_base,
130                      data->smp_size, data->smp_type);
131         return 0;
132 }
133 
134 static inline int types_compatible(mtrr_type type1, mtrr_type type2)
135 {
136         return type1 == MTRR_TYPE_UNCACHABLE ||
137                type2 == MTRR_TYPE_UNCACHABLE ||
138                (type1 == MTRR_TYPE_WRTHROUGH && type2 == MTRR_TYPE_WRBACK) ||
139                (type1 == MTRR_TYPE_WRBACK && type2 == MTRR_TYPE_WRTHROUGH);
140 }
141 
142 /**
143  * set_mtrr - update mtrrs on all processors
144  * @reg:        mtrr in question
145  * @base:       mtrr base
146  * @size:       mtrr size
147  * @type:       mtrr type
148  *
149  * This is kinda tricky, but fortunately, Intel spelled it out for us cleanly:
150  *
151  * 1. Queue work to do the following on all processors:
152  * 2. Disable Interrupts
153  * 3. Wait for all procs to do so
154  * 4. Enter no-fill cache mode
155  * 5. Flush caches
156  * 6. Clear PGE bit
157  * 7. Flush all TLBs
158  * 8. Disable all range registers
159  * 9. Update the MTRRs
160  * 10. Enable all range registers
161  * 11. Flush all TLBs and caches again
162  * 12. Enter normal cache mode and reenable caching
163  * 13. Set PGE
164  * 14. Wait for buddies to catch up
165  * 15. Enable interrupts.
166  *
167  * What does that mean for us? Well, stop_machine() will ensure that
168  * the rendezvous handler is started on each CPU. And in lockstep they
169  * do the state transition of disabling interrupts, updating MTRR's
170  * (the CPU vendors may each do it differently, so we call mtrr_if->set()
171  * callback and let them take care of it.) and enabling interrupts.
172  *
173  * Note that the mechanism is the same for UP systems, too; all the SMP stuff
174  * becomes nops.
175  */
176 static void set_mtrr(unsigned int reg, unsigned long base, unsigned long size,
177                      mtrr_type type)
178 {
179         struct set_mtrr_data data = { .smp_reg = reg,
180                                       .smp_base = base,
181                                       .smp_size = size,
182                                       .smp_type = type
183                                     };
184 
185         stop_machine_cpuslocked(mtrr_rendezvous_handler, &data, cpu_online_mask);
186 
187         generic_rebuild_map();
188 }
189 
190 /**
191  * mtrr_add_page - Add a memory type region
192  * @base: Physical base address of region in pages (in units of 4 kB!)
193  * @size: Physical size of region in pages (4 kB)
194  * @type: Type of MTRR desired
195  * @increment: If this is true do usage counting on the region
196  *
197  * Memory type region registers control the caching on newer Intel and
198  * non Intel processors. This function allows drivers to request an
199  * MTRR is added. The details and hardware specifics of each processor's
200  * implementation are hidden from the caller, but nevertheless the
201  * caller should expect to need to provide a power of two size on an
202  * equivalent power of two boundary.
203  *
204  * If the region cannot be added either because all regions are in use
205  * or the CPU cannot support it a negative value is returned. On success
206  * the register number for this entry is returned, but should be treated
207  * as a cookie only.
208  *
209  * On a multiprocessor machine the changes are made to all processors.
210  * This is required on x86 by the Intel processors.
211  *
212  * The available types are
213  *
214  * %MTRR_TYPE_UNCACHABLE - No caching
215  *
216  * %MTRR_TYPE_WRBACK - Write data back in bursts whenever
217  *
218  * %MTRR_TYPE_WRCOMB - Write data back soon but allow bursts
219  *
220  * %MTRR_TYPE_WRTHROUGH - Cache reads but not writes
221  *
222  * BUGS: Needs a quiet flag for the cases where drivers do not mind
223  * failures and do not wish system log messages to be sent.
224  */
225 int mtrr_add_page(unsigned long base, unsigned long size,
226                   unsigned int type, bool increment)
227 {
228         unsigned long lbase, lsize;
229         int i, replace, error;
230         mtrr_type ltype;
231 
232         if (!mtrr_enabled())
233                 return -ENXIO;
234 
235         error = mtrr_if->validate_add_page(base, size, type);
236         if (error)
237                 return error;
238 
239         if (type >= MTRR_NUM_TYPES) {
240                 pr_warn("type: %u invalid\n", type);
241                 return -EINVAL;
242         }
243 
244         /* If the type is WC, check that this processor supports it */
245         if ((type == MTRR_TYPE_WRCOMB) && !have_wrcomb()) {
246                 pr_warn("your processor doesn't support write-combining\n");
247                 return -ENOSYS;
248         }
249 
250         if (!size) {
251                 pr_warn("zero sized request\n");
252                 return -EINVAL;
253         }
254 
255         if ((base | (base + size - 1)) >>
256             (boot_cpu_data.x86_phys_bits - PAGE_SHIFT)) {
257                 pr_warn("base or size exceeds the MTRR width\n");
258                 return -EINVAL;
259         }
260 
261         error = -EINVAL;
262         replace = -1;
263 
264         /* No CPU hotplug when we change MTRR entries */
265         cpus_read_lock();
266 
267         /* Search for existing MTRR  */
268         mutex_lock(&mtrr_mutex);
269         for (i = 0; i < num_var_ranges; ++i) {
270                 mtrr_if->get(i, &lbase, &lsize, &ltype);
271                 if (!lsize || base > lbase + lsize - 1 ||
272                     base + size - 1 < lbase)
273                         continue;
274                 /*
275                  * At this point we know there is some kind of
276                  * overlap/enclosure
277                  */
278                 if (base < lbase || base + size - 1 > lbase + lsize - 1) {
279                         if (base <= lbase &&
280                             base + size - 1 >= lbase + lsize - 1) {
281                                 /*  New region encloses an existing region  */
282                                 if (type == ltype) {
283                                         replace = replace == -1 ? i : -2;
284                                         continue;
285                                 } else if (types_compatible(type, ltype))
286                                         continue;
287                         }
288                         pr_warn("0x%lx000,0x%lx000 overlaps existing 0x%lx000,0x%lx000\n", base, size, lbase,
289                                 lsize);
290                         goto out;
291                 }
292                 /* New region is enclosed by an existing region */
293                 if (ltype != type) {
294                         if (types_compatible(type, ltype))
295                                 continue;
296                         pr_warn("type mismatch for %lx000,%lx000 old: %s new: %s\n",
297                                 base, size, mtrr_attrib_to_str(ltype),
298                                 mtrr_attrib_to_str(type));
299                         goto out;
300                 }
301                 if (increment)
302                         ++mtrr_usage_table[i];
303                 error = i;
304                 goto out;
305         }
306         /* Search for an empty MTRR */
307         i = mtrr_if->get_free_region(base, size, replace);
308         if (i >= 0) {
309                 set_mtrr(i, base, size, type);
310                 if (likely(replace < 0)) {
311                         mtrr_usage_table[i] = 1;
312                 } else {
313                         mtrr_usage_table[i] = mtrr_usage_table[replace];
314                         if (increment)
315                                 mtrr_usage_table[i]++;
316                         if (unlikely(replace != i)) {
317                                 set_mtrr(replace, 0, 0, 0);
318                                 mtrr_usage_table[replace] = 0;
319                         }
320                 }
321         } else {
322                 pr_info("no more MTRRs available\n");
323         }
324         error = i;
325  out:
326         mutex_unlock(&mtrr_mutex);
327         cpus_read_unlock();
328         return error;
329 }
330 
331 static int mtrr_check(unsigned long base, unsigned long size)
332 {
333         if ((base & (PAGE_SIZE - 1)) || (size & (PAGE_SIZE - 1))) {
334                 pr_warn("size and base must be multiples of 4 kiB\n");
335                 Dprintk("size: 0x%lx  base: 0x%lx\n", size, base);
336                 dump_stack();
337                 return -1;
338         }
339         return 0;
340 }
341 
342 /**
343  * mtrr_add - Add a memory type region
344  * @base: Physical base address of region
345  * @size: Physical size of region
346  * @type: Type of MTRR desired
347  * @increment: If this is true do usage counting on the region
348  *
349  * Memory type region registers control the caching on newer Intel and
350  * non Intel processors. This function allows drivers to request an
351  * MTRR is added. The details and hardware specifics of each processor's
352  * implementation are hidden from the caller, but nevertheless the
353  * caller should expect to need to provide a power of two size on an
354  * equivalent power of two boundary.
355  *
356  * If the region cannot be added either because all regions are in use
357  * or the CPU cannot support it a negative value is returned. On success
358  * the register number for this entry is returned, but should be treated
359  * as a cookie only.
360  *
361  * On a multiprocessor machine the changes are made to all processors.
362  * This is required on x86 by the Intel processors.
363  *
364  * The available types are
365  *
366  * %MTRR_TYPE_UNCACHABLE - No caching
367  *
368  * %MTRR_TYPE_WRBACK - Write data back in bursts whenever
369  *
370  * %MTRR_TYPE_WRCOMB - Write data back soon but allow bursts
371  *
372  * %MTRR_TYPE_WRTHROUGH - Cache reads but not writes
373  *
374  * BUGS: Needs a quiet flag for the cases where drivers do not mind
375  * failures and do not wish system log messages to be sent.
376  */
377 int mtrr_add(unsigned long base, unsigned long size, unsigned int type,
378              bool increment)
379 {
380         if (!mtrr_enabled())
381                 return -ENODEV;
382         if (mtrr_check(base, size))
383                 return -EINVAL;
384         return mtrr_add_page(base >> PAGE_SHIFT, size >> PAGE_SHIFT, type,
385                              increment);
386 }
387 
388 /**
389  * mtrr_del_page - delete a memory type region
390  * @reg: Register returned by mtrr_add
391  * @base: Physical base address
392  * @size: Size of region
393  *
394  * If register is supplied then base and size are ignored. This is
395  * how drivers should call it.
396  *
397  * Releases an MTRR region. If the usage count drops to zero the
398  * register is freed and the region returns to default state.
399  * On success the register is returned, on failure a negative error
400  * code.
401  */
402 int mtrr_del_page(int reg, unsigned long base, unsigned long size)
403 {
404         int i, max;
405         mtrr_type ltype;
406         unsigned long lbase, lsize;
407         int error = -EINVAL;
408 
409         if (!mtrr_enabled())
410                 return -ENODEV;
411 
412         max = num_var_ranges;
413         /* No CPU hotplug when we change MTRR entries */
414         cpus_read_lock();
415         mutex_lock(&mtrr_mutex);
416         if (reg < 0) {
417                 /*  Search for existing MTRR  */
418                 for (i = 0; i < max; ++i) {
419                         mtrr_if->get(i, &lbase, &lsize, &ltype);
420                         if (lbase == base && lsize == size) {
421                                 reg = i;
422                                 break;
423                         }
424                 }
425                 if (reg < 0) {
426                         Dprintk("no MTRR for %lx000,%lx000 found\n", base, size);
427                         goto out;
428                 }
429         }
430         if (reg >= max) {
431                 pr_warn("register: %d too big\n", reg);
432                 goto out;
433         }
434         mtrr_if->get(reg, &lbase, &lsize, &ltype);
435         if (lsize < 1) {
436                 pr_warn("MTRR %d not used\n", reg);
437                 goto out;
438         }
439         if (mtrr_usage_table[reg] < 1) {
440                 pr_warn("reg: %d has count=0\n", reg);
441                 goto out;
442         }
443         if (--mtrr_usage_table[reg] < 1)
444                 set_mtrr(reg, 0, 0, 0);
445         error = reg;
446  out:
447         mutex_unlock(&mtrr_mutex);
448         cpus_read_unlock();
449         return error;
450 }
451 
452 /**
453  * mtrr_del - delete a memory type region
454  * @reg: Register returned by mtrr_add
455  * @base: Physical base address
456  * @size: Size of region
457  *
458  * If register is supplied then base and size are ignored. This is
459  * how drivers should call it.
460  *
461  * Releases an MTRR region. If the usage count drops to zero the
462  * register is freed and the region returns to default state.
463  * On success the register is returned, on failure a negative error
464  * code.
465  */
466 int mtrr_del(int reg, unsigned long base, unsigned long size)
467 {
468         if (!mtrr_enabled())
469                 return -ENODEV;
470         if (mtrr_check(base, size))
471                 return -EINVAL;
472         return mtrr_del_page(reg, base >> PAGE_SHIFT, size >> PAGE_SHIFT);
473 }
474 
475 /**
476  * arch_phys_wc_add - add a WC MTRR and handle errors if PAT is unavailable
477  * @base: Physical base address
478  * @size: Size of region
479  *
480  * If PAT is available, this does nothing.  If PAT is unavailable, it
481  * attempts to add a WC MTRR covering size bytes starting at base and
482  * logs an error if this fails.
483  *
484  * The called should provide a power of two size on an equivalent
485  * power of two boundary.
486  *
487  * Drivers must store the return value to pass to mtrr_del_wc_if_needed,
488  * but drivers should not try to interpret that return value.
489  */
490 int arch_phys_wc_add(unsigned long base, unsigned long size)
491 {
492         int ret;
493 
494         if (pat_enabled() || !mtrr_enabled())
495                 return 0;  /* Success!  (We don't need to do anything.) */
496 
497         ret = mtrr_add(base, size, MTRR_TYPE_WRCOMB, true);
498         if (ret < 0) {
499                 pr_warn("Failed to add WC MTRR for [%p-%p]; performance may suffer.",
500                         (void *)base, (void *)(base + size - 1));
501                 return ret;
502         }
503         return ret + MTRR_TO_PHYS_WC_OFFSET;
504 }
505 EXPORT_SYMBOL(arch_phys_wc_add);
506 
507 /*
508  * arch_phys_wc_del - undoes arch_phys_wc_add
509  * @handle: Return value from arch_phys_wc_add
510  *
511  * This cleans up after mtrr_add_wc_if_needed.
512  *
513  * The API guarantees that mtrr_del_wc_if_needed(error code) and
514  * mtrr_del_wc_if_needed(0) do nothing.
515  */
516 void arch_phys_wc_del(int handle)
517 {
518         if (handle >= 1) {
519                 WARN_ON(handle < MTRR_TO_PHYS_WC_OFFSET);
520                 mtrr_del(handle - MTRR_TO_PHYS_WC_OFFSET, 0, 0);
521         }
522 }
523 EXPORT_SYMBOL(arch_phys_wc_del);
524 
525 /*
526  * arch_phys_wc_index - translates arch_phys_wc_add's return value
527  * @handle: Return value from arch_phys_wc_add
528  *
529  * This will turn the return value from arch_phys_wc_add into an mtrr
530  * index suitable for debugging.
531  *
532  * Note: There is no legitimate use for this function, except possibly
533  * in printk line.  Alas there is an illegitimate use in some ancient
534  * drm ioctls.
535  */
536 int arch_phys_wc_index(int handle)
537 {
538         if (handle < MTRR_TO_PHYS_WC_OFFSET)
539                 return -1;
540         else
541                 return handle - MTRR_TO_PHYS_WC_OFFSET;
542 }
543 EXPORT_SYMBOL_GPL(arch_phys_wc_index);
544 
545 int __initdata changed_by_mtrr_cleanup;
546 
547 /**
548  * mtrr_bp_init - initialize MTRRs on the boot CPU
549  *
550  * This needs to be called early; before any of the other CPUs are
551  * initialized (i.e. before smp_init()).
552  */
553 void __init mtrr_bp_init(void)
554 {
555         bool generic_mtrrs = cpu_feature_enabled(X86_FEATURE_MTRR);
556         const char *why = "(not available)";
557         unsigned long config, dummy;
558 
559         phys_hi_rsvd = GENMASK(31, boot_cpu_data.x86_phys_bits - 32);
560 
561         if (!generic_mtrrs && mtrr_state.enabled) {
562                 /*
563                  * Software overwrite of MTRR state, only for generic case.
564                  * Note that X86_FEATURE_MTRR has been reset in this case.
565                  */
566                 init_table();
567                 mtrr_build_map();
568                 pr_info("MTRRs set to read-only\n");
569 
570                 return;
571         }
572 
573         if (generic_mtrrs)
574                 mtrr_if = &generic_mtrr_ops;
575         else
576                 mtrr_set_if();
577 
578         if (mtrr_enabled()) {
579                 /* Get the number of variable MTRR ranges. */
580                 if (mtrr_if == &generic_mtrr_ops)
581                         rdmsr(MSR_MTRRcap, config, dummy);
582                 else
583                         config = mtrr_if->var_regs;
584                 num_var_ranges = config & MTRR_CAP_VCNT;
585 
586                 init_table();
587                 if (mtrr_if == &generic_mtrr_ops) {
588                         /* BIOS may override */
589                         if (get_mtrr_state()) {
590                                 memory_caching_control |= CACHE_MTRR;
591                                 changed_by_mtrr_cleanup = mtrr_cleanup();
592                                 mtrr_build_map();
593                         } else {
594                                 mtrr_if = NULL;
595                                 why = "by BIOS";
596                         }
597                 }
598         }
599 
600         if (!mtrr_enabled())
601                 pr_info("MTRRs disabled %s\n", why);
602 }
603 
604 /**
605  * mtrr_save_state - Save current fixed-range MTRR state of the first
606  *      cpu in cpu_online_mask.
607  */
608 void mtrr_save_state(void)
609 {
610         int first_cpu;
611 
612         if (!mtrr_enabled() || !mtrr_state.have_fixed)
613                 return;
614 
615         first_cpu = cpumask_first(cpu_online_mask);
616         smp_call_function_single(first_cpu, mtrr_save_fixed_ranges, NULL, 1);
617 }
618 
619 static int __init mtrr_init_finalize(void)
620 {
621         /*
622          * Map might exist if mtrr_overwrite_state() has been called or if
623          * mtrr_enabled() returns true.
624          */
625         mtrr_copy_map();
626 
627         if (!mtrr_enabled())
628                 return 0;
629 
630         if (memory_caching_control & CACHE_MTRR) {
631                 if (!changed_by_mtrr_cleanup)
632                         mtrr_state_warn();
633                 return 0;
634         }
635 
636         mtrr_register_syscore();
637 
638         return 0;
639 }
640 subsys_initcall(mtrr_init_finalize);
641 

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php