OpenCores
URL https://opencores.org/ocsvn/or1k_old/or1k_old/trunk

Subversion Repositories or1k_old

[/] [or1k_old/] [trunk/] [rc203soc/] [sw/] [uClinux/] [mm/] [page_alloc.c] - Diff between revs 1634 and 1765

Go to most recent revision | Only display areas with differences | Details | Blame | View Log

Rev 1634 Rev 1765
/*
/*
 *  linux/mm/page_alloc.c
 *  linux/mm/page_alloc.c
 *
 *
 *  Copyright (C) 1991, 1992, 1993, 1994  Linus Torvalds
 *  Copyright (C) 1991, 1992, 1993, 1994  Linus Torvalds
 *  Swap reorganised 29.12.95, Stephen Tweedie
 *  Swap reorganised 29.12.95, Stephen Tweedie
 */
 */
 
 
#include <linux/mm.h>
#include <linux/mm.h>
#include <linux/sched.h>
#include <linux/sched.h>
#include <linux/head.h>
#include <linux/head.h>
#include <linux/kernel.h>
#include <linux/kernel.h>
#include <linux/kernel_stat.h>
#include <linux/kernel_stat.h>
#include <linux/errno.h>
#include <linux/errno.h>
#include <linux/string.h>
#include <linux/string.h>
#include <linux/stat.h>
#include <linux/stat.h>
#include <linux/swap.h>
#include <linux/swap.h>
#include <linux/fs.h>
#include <linux/fs.h>
#include <linux/swapctl.h>
#include <linux/swapctl.h>
#include <linux/interrupt.h>
#include <linux/interrupt.h>
 
 
#include <asm/dma.h>
#include <asm/dma.h>
#include <asm/system.h> /* for cli()/sti() */
#include <asm/system.h> /* for cli()/sti() */
#include <asm/segment.h> /* for memcpy_to/fromfs */
#include <asm/segment.h> /* for memcpy_to/fromfs */
#include <asm/bitops.h>
#include <asm/bitops.h>
#include <asm/pgtable.h>
#include <asm/pgtable.h>
 
 
int nr_swap_pages = 0;
int nr_swap_pages = 0;
int nr_free_pages = 0;
int nr_free_pages = 0;
 
 
extern struct wait_queue *buffer_wait;
extern struct wait_queue *buffer_wait;
 
 
/*
/*
 * Free area management
 * Free area management
 *
 *
 * The free_area_list arrays point to the queue heads of the free areas
 * The free_area_list arrays point to the queue heads of the free areas
 * of different sizes
 * of different sizes
 */
 */
 
 
#define NR_MEM_LISTS 6
#define NR_MEM_LISTS 6
 
 
/* The start of this MUST match the start of "struct page" */
/* The start of this MUST match the start of "struct page" */
struct free_area_struct {
struct free_area_struct {
        struct page *next;
        struct page *next;
        struct page *prev;
        struct page *prev;
        unsigned int * map;
        unsigned int * map;
};
};
 
 
#define memory_head(x) ((struct page *)(x))
#define memory_head(x) ((struct page *)(x))
 
 
static struct free_area_struct free_area[NR_MEM_LISTS];
static struct free_area_struct free_area[NR_MEM_LISTS];
 
 
static inline void init_mem_queue(struct free_area_struct * head)
static inline void init_mem_queue(struct free_area_struct * head)
{
{
        head->next = memory_head(head);
        head->next = memory_head(head);
        head->prev = memory_head(head);
        head->prev = memory_head(head);
}
}
 
 
static inline void add_mem_queue(struct free_area_struct * head, struct page * entry)
static inline void add_mem_queue(struct free_area_struct * head, struct page * entry)
{
{
        struct page * next = head->next;
        struct page * next = head->next;
 
 
        entry->prev = memory_head(head);
        entry->prev = memory_head(head);
        entry->next = next;
        entry->next = next;
        next->prev = entry;
        next->prev = entry;
        head->next = entry;
        head->next = entry;
}
}
 
 
static inline void remove_mem_queue(struct page * entry)
static inline void remove_mem_queue(struct page * entry)
{
{
        struct page * next = entry->next;
        struct page * next = entry->next;
        struct page * prev = entry->prev;
        struct page * prev = entry->prev;
        next->prev = prev;
        next->prev = prev;
        prev->next = next;
        prev->next = next;
}
}
 
 
/*
/*
 * Free_page() adds the page to the free lists. This is optimized for
 * Free_page() adds the page to the free lists. This is optimized for
 * fast normal cases (no error jumps taken normally).
 * fast normal cases (no error jumps taken normally).
 *
 *
 * The way to optimize jumps for gcc-2.2.2 is to:
 * The way to optimize jumps for gcc-2.2.2 is to:
 *  - select the "normal" case and put it inside the if () { XXX }
 *  - select the "normal" case and put it inside the if () { XXX }
 *  - no else-statements if you can avoid them
 *  - no else-statements if you can avoid them
 *
 *
 * With the above two rules, you get a straight-line execution path
 * With the above two rules, you get a straight-line execution path
 * for the normal case, giving better asm-code.
 * for the normal case, giving better asm-code.
 *
 *
 * free_page() may sleep since the page being freed may be a buffer
 * free_page() may sleep since the page being freed may be a buffer
 * page or present in the swap cache. It will not sleep, however,
 * page or present in the swap cache. It will not sleep, however,
 * for a freshly allocated page (get_free_page()).
 * for a freshly allocated page (get_free_page()).
 */
 */
 
 
/*
/*
 * Buddy system. Hairy. You really aren't expected to understand this
 * Buddy system. Hairy. You really aren't expected to understand this
 *
 *
 * Hint: -mask = 1+~mask
 * Hint: -mask = 1+~mask
 */
 */
static inline void free_pages_ok(unsigned long map_nr, unsigned long order)
static inline void free_pages_ok(unsigned long map_nr, unsigned long order)
{
{
        struct free_area_struct *area = free_area + order;
        struct free_area_struct *area = free_area + order;
        unsigned long index = map_nr >> (1 + order);
        unsigned long index = map_nr >> (1 + order);
        unsigned long mask = (~0UL) << order;
        unsigned long mask = (~0UL) << order;
        unsigned long flags;
        unsigned long flags;
 
 
        save_flags(flags);
        save_flags(flags);
        cli();
        cli();
 
 
#define list(x) (mem_map+(x))
#define list(x) (mem_map+(x))
 
 
        map_nr &= mask;
        map_nr &= mask;
        nr_free_pages -= mask;
        nr_free_pages -= mask;
        while (mask + (1 << (NR_MEM_LISTS-1))) {
        while (mask + (1 << (NR_MEM_LISTS-1))) {
                if (!change_bit(index, area->map))
                if (!change_bit(index, area->map))
                        break;
                        break;
                remove_mem_queue(list(map_nr ^ -mask));
                remove_mem_queue(list(map_nr ^ -mask));
                mask <<= 1;
                mask <<= 1;
                area++;
                area++;
                index >>= 1;
                index >>= 1;
                map_nr &= mask;
                map_nr &= mask;
        }
        }
        add_mem_queue(area, list(map_nr));
        add_mem_queue(area, list(map_nr));
 
 
#undef list
#undef list
 
 
        restore_flags(flags);
        restore_flags(flags);
        if (!waitqueue_active(&buffer_wait))
        if (!waitqueue_active(&buffer_wait))
                return;
                return;
        wake_up(&buffer_wait);
        wake_up(&buffer_wait);
}
}
 
 
void __free_page(struct page *page)
void __free_page(struct page *page)
{
{
        if (!PageReserved(page) && atomic_dec_and_test(&page->count)) {
        if (!PageReserved(page) && atomic_dec_and_test(&page->count)) {
                unsigned long map_nr = page->map_nr;
                unsigned long map_nr = page->map_nr;
                delete_from_swap_cache(map_nr);
                delete_from_swap_cache(map_nr);
                free_pages_ok(map_nr, 0);
                free_pages_ok(map_nr, 0);
        }
        }
}
}
 
 
void free_pages(unsigned long addr, unsigned long order)
void free_pages(unsigned long addr, unsigned long order)
{
{
        unsigned long map_nr = MAP_NR(addr);
        unsigned long map_nr = MAP_NR(addr);
 
 
        if (map_nr < MAP_NR(high_memory)) {
        if (map_nr < MAP_NR(high_memory)) {
                mem_map_t * map = mem_map + map_nr;
                mem_map_t * map = mem_map + map_nr;
                if (PageReserved(map))
                if (PageReserved(map))
                        return;
                        return;
                if (atomic_dec_and_test(&map->count)) {
                if (atomic_dec_and_test(&map->count)) {
                        delete_from_swap_cache(map_nr);
                        delete_from_swap_cache(map_nr);
                        free_pages_ok(map_nr, order);
                        free_pages_ok(map_nr, order);
                        return;
                        return;
                }
                }
        }
        }
}
}
 
 
/*
/*
 * Some ugly macros to speed up __get_free_pages()..
 * Some ugly macros to speed up __get_free_pages()..
 */
 */
#define MARK_USED(index, order, area) \
#define MARK_USED(index, order, area) \
        change_bit((index) >> (1+(order)), (area)->map)
        change_bit((index) >> (1+(order)), (area)->map)
#define CAN_DMA(x) (PageDMA(x))
#define CAN_DMA(x) (PageDMA(x))
#define ADDRESS(x) (PAGE_OFFSET + ((x) << PAGE_SHIFT))
#define ADDRESS(x) (PAGE_OFFSET + ((x) << PAGE_SHIFT))
#define RMQUEUE(order, dma) \
#define RMQUEUE(order, dma) \
do { struct free_area_struct * area = free_area+order; \
do { struct free_area_struct * area = free_area+order; \
     unsigned long new_order = order; \
     unsigned long new_order = order; \
        do { struct page *prev = memory_head(area), *ret; \
        do { struct page *prev = memory_head(area), *ret; \
                while (memory_head(area) != (ret = prev->next)) { \
                while (memory_head(area) != (ret = prev->next)) { \
                        if (!dma || CAN_DMA(ret)) { \
                        if (!dma || CAN_DMA(ret)) { \
                                unsigned long map_nr = ret->map_nr; \
                                unsigned long map_nr = ret->map_nr; \
                                (prev->next = ret->next)->prev = prev; \
                                (prev->next = ret->next)->prev = prev; \
                                MARK_USED(map_nr, new_order, area); \
                                MARK_USED(map_nr, new_order, area); \
                                nr_free_pages -= 1 << order; \
                                nr_free_pages -= 1 << order; \
                                EXPAND(ret, map_nr, order, new_order, area); \
                                EXPAND(ret, map_nr, order, new_order, area); \
                                restore_flags(flags); \
                                restore_flags(flags); \
                                return ADDRESS(map_nr); \
                                return ADDRESS(map_nr); \
                        } \
                        } \
                        prev = ret; \
                        prev = ret; \
                } \
                } \
                new_order++; area++; \
                new_order++; area++; \
        } while (new_order < NR_MEM_LISTS); \
        } while (new_order < NR_MEM_LISTS); \
} while (0)
} while (0)
 
 
#define EXPAND(map,index,low,high,area) \
#define EXPAND(map,index,low,high,area) \
do { unsigned long size = 1 << high; \
do { unsigned long size = 1 << high; \
        while (high > low) { \
        while (high > low) { \
                area--; high--; size >>= 1; \
                area--; high--; size >>= 1; \
                add_mem_queue(area, map); \
                add_mem_queue(area, map); \
                MARK_USED(index, high, area); \
                MARK_USED(index, high, area); \
                index += size; \
                index += size; \
                map += size; \
                map += size; \
        } \
        } \
        map->count = 1; \
        map->count = 1; \
        map->age = PAGE_INITIAL_AGE; \
        map->age = PAGE_INITIAL_AGE; \
} while (0)
} while (0)
 
 
unsigned long __get_free_pages(int priority, unsigned long order, int dma)
unsigned long __get_free_pages(int priority, unsigned long order, int dma)
{
{
        unsigned long flags;
        unsigned long flags;
        int reserved_pages;
        int reserved_pages;
 
 
        if (order >= NR_MEM_LISTS)
        if (order >= NR_MEM_LISTS)
                return 0;
                return 0;
        if (intr_count && priority != GFP_ATOMIC) {
        if (intr_count && priority != GFP_ATOMIC) {
                static int count = 0;
                static int count = 0;
                if (++count < 5) {
                if (++count < 5) {
                        printk("gfp called nonatomically from interrupt %p\n",
                        printk("gfp called nonatomically from interrupt %p\n",
                                __builtin_return_address(0));
                                __builtin_return_address(0));
                        priority = GFP_ATOMIC;
                        priority = GFP_ATOMIC;
                }
                }
        }
        }
        reserved_pages = 5;
        reserved_pages = 5;
        if (priority != GFP_NFS)
        if (priority != GFP_NFS)
                reserved_pages = min_free_pages;
                reserved_pages = min_free_pages;
        if ((priority == GFP_BUFFER || priority == GFP_IO) && reserved_pages >= 48)
        if ((priority == GFP_BUFFER || priority == GFP_IO) && reserved_pages >= 48)
                reserved_pages -= (12 + (reserved_pages>>3));
                reserved_pages -= (12 + (reserved_pages>>3));
        save_flags(flags);
        save_flags(flags);
repeat:
repeat:
        cli();
        cli();
        if ((priority==GFP_ATOMIC) || nr_free_pages > reserved_pages) {
        if ((priority==GFP_ATOMIC) || nr_free_pages > reserved_pages) {
                RMQUEUE(order, dma);
                RMQUEUE(order, dma);
                restore_flags(flags);
                restore_flags(flags);
                return 0;
                return 0;
        }
        }
        restore_flags(flags);
        restore_flags(flags);
        if (priority != GFP_BUFFER && try_to_free_page(priority, dma, 1))
        if (priority != GFP_BUFFER && try_to_free_page(priority, dma, 1))
                goto repeat;
                goto repeat;
        return 0;
        return 0;
}
}
 
 
/*
/*
 * Show free area list (used inside shift_scroll-lock stuff)
 * Show free area list (used inside shift_scroll-lock stuff)
 * We also calculate the percentage fragmentation. We do this by counting the
 * We also calculate the percentage fragmentation. We do this by counting the
 * memory on each free list with the exception of the first item on the list.
 * memory on each free list with the exception of the first item on the list.
 */
 */
void show_free_areas(void)
void show_free_areas(void)
{
{
        unsigned long order, flags;
        unsigned long order, flags;
        unsigned long total = 0;
        unsigned long total = 0;
 
 
        printk("Free pages:      %6dkB\n ( ",nr_free_pages<<(PAGE_SHIFT-10));
        printk("Free pages:      %6dkB\n ( ",nr_free_pages<<(PAGE_SHIFT-10));
        save_flags(flags);
        save_flags(flags);
        cli();
        cli();
        for (order=0 ; order < NR_MEM_LISTS; order++) {
        for (order=0 ; order < NR_MEM_LISTS; order++) {
                struct page * tmp;
                struct page * tmp;
                unsigned long nr = 0;
                unsigned long nr = 0;
                for (tmp = free_area[order].next ; tmp != memory_head(free_area+order) ; tmp = tmp->next) {
                for (tmp = free_area[order].next ; tmp != memory_head(free_area+order) ; tmp = tmp->next) {
                        nr ++;
                        nr ++;
                }
                }
                total += nr * ((PAGE_SIZE>>10) << order);
                total += nr * ((PAGE_SIZE>>10) << order);
                printk("%lu*%lukB ", nr, (PAGE_SIZE>>10) << order);
                printk("%lu*%lukB ", nr, (PAGE_SIZE>>10) << order);
        }
        }
        restore_flags(flags);
        restore_flags(flags);
        printk("= %lukB)\n", total);
        printk("= %lukB)\n", total);
#ifdef SWAP_CACHE_INFO
#ifdef SWAP_CACHE_INFO
        show_swap_cache_info();
        show_swap_cache_info();
#endif  
#endif  
}
}
 
 
#define LONG_ALIGN(x) (((x)+(sizeof(long))-1)&~((sizeof(long))-1))
#define LONG_ALIGN(x) (((x)+(sizeof(long))-1)&~((sizeof(long))-1))
 
 
/*
/*
 * set up the free-area data structures:
 * set up the free-area data structures:
 *   - mark all pages reserved
 *   - mark all pages reserved
 *   - mark all memory queues empty
 *   - mark all memory queues empty
 *   - clear the memory bitmaps
 *   - clear the memory bitmaps
 */
 */
unsigned long free_area_init(unsigned long start_mem, unsigned long end_mem)
unsigned long free_area_init(unsigned long start_mem, unsigned long end_mem)
{
{
        mem_map_t * p;
        mem_map_t * p;
        unsigned long mask = PAGE_MASK;
        unsigned long mask = PAGE_MASK;
        int i;
        int i;
 
 
        /*
        /*
         * select nr of pages we try to keep free for important stuff
         * select nr of pages we try to keep free for important stuff
         * with a minimum of 48 pages. This is totally arbitrary
         * with a minimum of 48 pages. This is totally arbitrary
         */
         */
        i = (end_mem - PAGE_OFFSET) >> (PAGE_SHIFT+7);
        i = (end_mem - PAGE_OFFSET) >> (PAGE_SHIFT+7);
        if (i < 24)
        if (i < 24)
                i = 24;
                i = 24;
        i += 24;   /* The limit for buffer pages in __get_free_pages is
        i += 24;   /* The limit for buffer pages in __get_free_pages is
                    * decreased by 12+(i>>3) */
                    * decreased by 12+(i>>3) */
        min_free_pages = i;
        min_free_pages = i;
        free_pages_low = i + (i>>1);
        free_pages_low = i + (i>>1);
        free_pages_high = i + i;
        free_pages_high = i + i;
        start_mem = init_swap_cache(start_mem, end_mem);
        start_mem = init_swap_cache(start_mem, end_mem);
        mem_map = (mem_map_t *) start_mem;
        mem_map = (mem_map_t *) start_mem;
        p = mem_map + MAP_NR(end_mem);
        p = mem_map + MAP_NR(end_mem);
        start_mem = LONG_ALIGN((unsigned long) p);
        start_mem = LONG_ALIGN((unsigned long) p);
        memset(mem_map, 0, start_mem - (unsigned long) mem_map);
        memset(mem_map, 0, start_mem - (unsigned long) mem_map);
        do {
        do {
                --p;
                --p;
                p->flags = (1 << PG_DMA) | (1 << PG_reserved);
                p->flags = (1 << PG_DMA) | (1 << PG_reserved);
                p->map_nr = p - mem_map;
                p->map_nr = p - mem_map;
        } while (p > mem_map);
        } while (p > mem_map);
 
 
        for (i = 0 ; i < NR_MEM_LISTS ; i++) {
        for (i = 0 ; i < NR_MEM_LISTS ; i++) {
                unsigned long bitmap_size;
                unsigned long bitmap_size;
                init_mem_queue(free_area+i);
                init_mem_queue(free_area+i);
                mask += mask;
                mask += mask;
                end_mem = (end_mem + ~mask) & mask;
                end_mem = (end_mem + ~mask) & mask;
                bitmap_size = (end_mem - PAGE_OFFSET) >> (PAGE_SHIFT + i);
                bitmap_size = (end_mem - PAGE_OFFSET) >> (PAGE_SHIFT + i);
                bitmap_size = (bitmap_size + 7) >> 3;
                bitmap_size = (bitmap_size + 7) >> 3;
                bitmap_size = LONG_ALIGN(bitmap_size);
                bitmap_size = LONG_ALIGN(bitmap_size);
                free_area[i].map = (unsigned int *) start_mem;
                free_area[i].map = (unsigned int *) start_mem;
                memset((void *) start_mem, 0, bitmap_size);
                memset((void *) start_mem, 0, bitmap_size);
                start_mem += bitmap_size;
                start_mem += bitmap_size;
        }
        }
        return start_mem;
        return start_mem;
}
}
 
 
/*
/*
 * The tests may look silly, but it essentially makes sure that
 * The tests may look silly, but it essentially makes sure that
 * no other process did a swap-in on us just as we were waiting.
 * no other process did a swap-in on us just as we were waiting.
 *
 *
 * Also, don't bother to add to the swap cache if this page-in
 * Also, don't bother to add to the swap cache if this page-in
 * was due to a write access.
 * was due to a write access.
 */
 */
void swap_in(struct task_struct * tsk, struct vm_area_struct * vma,
void swap_in(struct task_struct * tsk, struct vm_area_struct * vma,
        pte_t * page_table, unsigned long entry, int write_access)
        pte_t * page_table, unsigned long entry, int write_access)
{
{
        unsigned long page = __get_free_page(GFP_KERNEL);
        unsigned long page = __get_free_page(GFP_KERNEL);
 
 
        if (pte_val(*page_table) != entry) {
        if (pte_val(*page_table) != entry) {
                if (page)
                if (page)
                        free_page(page);
                        free_page(page);
                return;
                return;
        }
        }
        if (!page) {
        if (!page) {
                printk("swap_in:");
                printk("swap_in:");
                set_pte(page_table, BAD_PAGE);
                set_pte(page_table, BAD_PAGE);
                swap_free(entry);
                swap_free(entry);
                oom(tsk);
                oom(tsk);
                return;
                return;
        }
        }
        read_swap_page(entry, (char *) page);
        read_swap_page(entry, (char *) page);
        if (pte_val(*page_table) != entry) {
        if (pte_val(*page_table) != entry) {
                free_page(page);
                free_page(page);
                return;
                return;
        }
        }
        vma->vm_mm->rss++;
        vma->vm_mm->rss++;
        tsk->maj_flt++;
        tsk->maj_flt++;
 
 
        /* Give the physical reallocated page a bigger start */
        /* Give the physical reallocated page a bigger start */
        if (vma->vm_mm->rss < (MAP_NR(high_memory) >> 2))
        if (vma->vm_mm->rss < (MAP_NR(high_memory) >> 2))
                mem_map[MAP_NR(page)].age = (PAGE_INITIAL_AGE + PAGE_ADVANCE);
                mem_map[MAP_NR(page)].age = (PAGE_INITIAL_AGE + PAGE_ADVANCE);
 
 
        if (!write_access && add_to_swap_cache(MAP_NR(page), entry)) {
        if (!write_access && add_to_swap_cache(MAP_NR(page), entry)) {
                /* keep swap page allocated for the moment (swap cache) */
                /* keep swap page allocated for the moment (swap cache) */
                set_pte(page_table, mk_pte(page, vma->vm_page_prot));
                set_pte(page_table, mk_pte(page, vma->vm_page_prot));
                return;
                return;
        }
        }
        set_pte(page_table, pte_mkwrite(pte_mkdirty(mk_pte(page, vma->vm_page_prot))));
        set_pte(page_table, pte_mkwrite(pte_mkdirty(mk_pte(page, vma->vm_page_prot))));
        swap_free(entry);
        swap_free(entry);
        return;
        return;
}
}
 
 
 
 

powered by: WebSVN 2.1.0

© copyright 1999-2024 OpenCores.org, equivalent to Oliscience, all rights reserved. OpenCores®, registered trademark.