Skip to content

Commit

Permalink
x86: more header fixes
Browse files Browse the repository at this point in the history
Summary: Add missing include guards for some x86 headers.

This has only had the most rudimentary testing, but is hopefully obviously
correct.

Signed-off-by: Ingo Molnar <mingo@elte.hu>
  • Loading branch information
Vegard Nossum authored and Ingo Molnar committed Jun 18, 2008
1 parent e6e07d8 commit 0db125c
Show file tree
Hide file tree
Showing 4 changed files with 16 additions and 0 deletions.
1 change: 1 addition & 0 deletions include/asm-x86/seccomp_64.h
Original file line number Diff line number Diff line change
@@ -1,4 +1,5 @@
#ifndef _ASM_SECCOMP_H
#define _ASM_SECCOMP_H

#include <linux/thread_info.h>

Expand Down
5 changes: 5 additions & 0 deletions include/asm-x86/suspend_32.h
Original file line number Diff line number Diff line change
Expand Up @@ -3,6 +3,9 @@
* Based on code
* Copyright 2001 Patrick Mochel <mochel@osdl.org>
*/
#ifndef __ASM_X86_32_SUSPEND_H
#define __ASM_X86_32_SUSPEND_H

#include <asm/desc.h>
#include <asm/i387.h>

Expand Down Expand Up @@ -44,3 +47,5 @@ static inline void acpi_save_register_state(unsigned long return_point)
/* routines for saving/restoring kernel state */
extern int acpi_save_state_mem(void);
#endif

#endif /* __ASM_X86_32_SUSPEND_H */
5 changes: 5 additions & 0 deletions include/asm-x86/xor_32.h
Original file line number Diff line number Diff line change
@@ -1,3 +1,6 @@
#ifndef ASM_X86__XOR_32_H
#define ASM_X86__XOR_32_H

/*
* Optimized RAID-5 checksumming functions for MMX and SSE.
*
Expand Down Expand Up @@ -881,3 +884,5 @@ do { \
deals with a load to a line that is being prefetched. */
#define XOR_SELECT_TEMPLATE(FASTEST) \
(cpu_has_xmm ? &xor_block_pIII_sse : FASTEST)

#endif /* ASM_X86__XOR_32_H */
5 changes: 5 additions & 0 deletions include/asm-x86/xor_64.h
Original file line number Diff line number Diff line change
@@ -1,3 +1,6 @@
#ifndef ASM_X86__XOR_64_H
#define ASM_X86__XOR_64_H

/*
* Optimized RAID-5 checksumming functions for MMX and SSE.
*
Expand Down Expand Up @@ -354,3 +357,5 @@ do { \
We may also be able to load into the L1 only depending on how the cpu
deals with a load to a line that is being prefetched. */
#define XOR_SELECT_TEMPLATE(FASTEST) (&xor_block_sse)

#endif /* ASM_X86__XOR_64_H */

0 comments on commit 0db125c

Please sign in to comment.